Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

  1. Create an AWS Account
    • Use your sagebasefirstName.lastName@sagebase.org email address for the account name
    • Enter Sage Bionetworks' physical address for the address
    • You will need to use your own credit card temporarily
  2. Send Mike Kellen an email to have your new AWS account added to the consolidated bill
    • Once this is done, you will no longer be billed on your own credit card

...

  1. Use the AWS console to create a new SSH key named SageKeyPair
  2. Download it to your home directory on the shared servers
  3. Fix the permissions on it
    Code Block
    chmod 600 ~/SageKeyPair.pem
    

Configure S3

...

  1. Use the AWS console to make a new S3 bucket named sagetestsagebio-YourUnixUsername and make these four subdirectories
    1. scripts
    2. input
    3. output
    4. results
    5. logs
    Note: Do not put any underscores in your bucket name. Only use hyphens, lowercase letters and numbers.

...


  1. Image Added
  2. Make these five subdirectories
    1. scripts
    2. input
    3. output
    4. results
    5. logs
      Image Added

Set up your config files for the AWS command line tools installed on the shared servers

...

  1. ssh to belltown
  2. Create the configuration file for s3curl command line tool
    Code Block
    ~>cat .s3curl 
    #!/bin/perl
    %awsSecretAccessKeys = (
        YourUsernameYourUnixUsername => {
            id => 'YourAccessKeyID',
            key => 'YourSecretAccessKey',
        },
    );
    
  3. Test that you can run s3curl
    Code Block
    ~>/work/platform/bin/s3curl.pl --id $USER https://s3.amazonaws.com/sagetestsagebio-$USER/ | head -c 200
    <?xml version="1.0" encoding="UTF-8"?>
    <ListBucketResult xmlns="http://s3.amazonaws.com/doc/2006-03-01/"><Name>sagetestemr</Name><Prefix></Prefix><Marker></Marker><MaxKeys>1000</MaxKeys><IsTruncated>
    

...

  1. ssh to belltown
  2. Create the configuration file for the Elastic Map Reduce command line tool
    Code Block
    ~>cat $USER-credentials.json 
    {
    "access_id": "YourAccessKeyIDYourAWSAccessKeyID",
    "private_key": "YourSecretAccessKeyYourAWSSecretAccessKey",
    "keypair": "SageKeyPair",
    "key-pair-file": "~/SageKeyPair.pem",
    "log_uri": "s3n://sagetestsagebio-YourUsernameYourUnixUsername/logs/",
    "region": "us-east-1"
    }
    
  3. Test that you can run it
    Code Block
    ~>/work/platform/bin/elastic-mapreduce-cli/elastic-mapreduce --credentials ~/$USER-credentials.json --help
    Usage: elastic-mapreduce [options]
    
      Creating Job Flows
            --create                     Create a new job flow
            --name NAME                  The name of the job flow being created
            --alive                      Create a job flow that stays running even though it has executed all its steps
            --with-termination-protection
                                         Create a job with termination protection (default is no termination protection)
            --num-instances NUM          Number of instances in the job flow
    ...
    

...