dleThe following provides instructions on how to log on to the Sage Scientific Compute workspace using your Synapse credentials, and how to use the products provided in the AWS Service Catalog to setup or modify EC2 instances and S3 buckets.
Gaining Access
Access to Sage Scientific Compute workspace is organized by communities. Community membership is defined by a Synapse Team and managed by its community manager. Each community also has a defined entry point URL, as shown below:
Community | Synapse Team | Service Catalog entry point |
---|---|---|
Sage Bionetworks | ||
AD Portal Users | ||
Accelerating Medicines Partnership - Alzheimer’s Disease (AMP-AD) Consortium | ||
Bill & Melinda Gates - Ki Team |
For Sage Bionetworks employees, access is granted during employee on boarding. For other groups, the community manager will add your Synapse account to the list of allowed users for the compute workspace.
Login
To begin, visit the appropriate entry point URL for your community, listed above, and log in with your Synapse credentials. You will be prompted to allow access to some information from your Synapse profile.
Note: Service Catalog products, discussed below, are owned by the Synapse account under which you log in. Once created, products will appear in the console only when logged in to that account, which will have sole ability to update or remove the product. Products like S3 buckets may have a life cycle beyond the project participation of any single person. To support such a case you may create and use a so-called service account in Synapse (i.e., an account meant for automation that may outlive one person’s commitment to a project). If so, then to meet regulatory requirements the credentials for the service account must be placed in a secure store whose access is limited and can be reviewed. /wiki/spaces/IT/pages/1200816129.
Service Catalog Portal
Once logged in, you will see a list of “Products” you can provision. These fall into two categories, EC2 Instances (virtual machines) and S3 Buckets used for data storage. Not all products are available to all communities.
On the left is a navigation sidebar. If you do not see, it, look for a hamburger icon in the upper left, and click on it to expand the navigation sidebar. The options in the navigation bar are “Products” and “Provisioned Products”. After you provision a product selected from “Products”, you will be able to see your products under “Provisioned Products”.
EC2
EC2 Instance Products
We currently offer three varieties of virtual machines, each preconfigured for a given purpose, described below.
Linux Docker
This product provides a basic EC2 instance with Docker installed.
Ubuntu Linux with Notebook Software
This product is an Ubuntu Linux EC2 instance with R Studio notebook software installed.
Windows
This product is a Microsoft Windows instance. This is available only to Sage employees.
Creating EC2 Products
To create an instance, select “Products List” from the navigation panel on the left. Next, select from the list one of the EC2 products described above. On the product page, click the orange “LAUNCH PRODUCT” button under the product description, then fill out the wizard as follows:
Product Version
Name: this names the product and the instance. You’ll use this to manage the product later. Please include your name in the product, e.g. if your name is Jane Doe and you are provisioning a Linux instance for your project Foo, you could name it
jdoe-linux-foo
.Version: choose a version of the product to provision.
Parameters
Use Private Network: This puts the instance on a private subnet. We strongly recommend using the default value,
true
.EC2 Instance Type: there are many instance types to choose from. To learn about their details, see https://aws.amazon.com/ec2/instance-types/. To learn about their costs, see https://aws.amazon.com/ec2/pricing/on-demand/ or use the AWS pricing calculator.
Linux Distribution: (EC2 Linux product only) the variety of Linux OS that will be installed.
Disk Size: the amount of local storage, in gigabytes. Please treat the disk as temporary storage. Long term storage of data should be in a bucket (see below).
Tag Options
Department: your department within Sage Bionetworks.
Project: the project that this instance will be used for. Please make a request to #sageit if your project is not in the list.
CostCenter: bill the product to this cost center code. If the appropriate cost center code is not in the list then select “Other / 000001” and create a custom “CostCenterOther” tag. Set the tag to a value from our official list of cost centers codes. Example:
Note: You can add additional custom tags when provisioning resources however there are 3 reserved tags that you should avoid adding: Department, Project, CostCenter and OwnerEmail. The owner email tag is automatically set to <Synapse Username>@synapse.org
Notifications
Please skip the Notifications pane. SNS notifications are not operational at this time.
Review
On the review pane, you have the opportunity to review all the choices you have made prior to clicking the “Launch” link in the lower right-hand corner.
Connecting to an Instance
A new instance takes a few minutes to be created. Once complete, it will appear in the “Provisioned Products” list, showing status Available. Select “Provisioned Product Details” from the navigation panel on the left, and click on your product. A product that has a “Succeeded” event will have outputs that include links for connecting.
ConnectionURI: if your product has a ConnectionURI link, this will open a shell prompt in a browser tab. When you are done with your session click “Terminate” in the upper right corner.
NotebookConnectionURI: Notebook products contain a NotebookConnectionURI link, which will open a notebook in the browser.
ConnectionInstructions: For Windows products, click on the ConnectionInstructions link and follow the steps provided there.
Create a Synapse personal access token
The AWS SSM allows direct access to private instances from your own computer terminal. To setup access with the AWS SSM we need to create a special Synapse personal access token (PAT) that will work with the Sage Service Catalog. This is special PAT that can only be created using this workflow, creating a PAT from the Synapse personal token manager web page will NOT work.
Request a Synapse PAT by visiting https://sc.sageit.org/personalaccesstoken , for Sage employees, or https://ad.strides.sc.sageit.org/personalaccesstoken for AMP-AD members. (You may need to login to Synapse.) If you have already created a PAT through this mechanism and are repeating the process you must first visit the token management page in Synapse and delete the existing one with the same name.
After logging into Synapse a file containing the PAT, which is a long character string (i.e. eyJ0eXAiOiJ...Z8t9Eg), is returned to you. Save the file to your local machine and note the location where you saved it to then close the browser session.
Note: At this point you can verify that the PAT for the Service Catalog was successfully created by viewing the Synapse token management page. When the PAT expires you will need to repeat these steps to create a new PAT. The PAT should look something like this
SSM access to an Instance
To setup access the AWS EC2 instances with the AWS SSM we need to install the AWS CLI and make it source credentials with an external process.
Install the AWS CLI version 2 (SSM access will not work with ver 1.x)
Install SSM session manager plugin
Create a synapse credentials script.
Linux/Mac: synapse_creds.sh with content below. Add the execute permission to thesynapse_creds.sh
file (i.e.chmod +x synapse_creds.sh
)#!/usr/bin/env bash # Inputs SC_ENDPOINT=$1 # i.e. https://sc.sageit.org SYNAPSE_PAT=$2 # The Synapse Personal Access Token # Endpoints STS_TOKEN_ENDPOINT="${SC_ENDPOINT}/ststoken" # Get Credentials AWS_STS_CREDS=$(curl --location-trusted --silent -H "Authorization:Bearer ${SYNAPSE_PAT}" ${STS_TOKEN_ENDPOINT}) echo ${AWS_STS_CREDS}
Windows: synapse_creds.bat with content below.@ECHO OFF REM Inputs REM %~1 The SC endpoint i.e. https://sc.sageit.org REM %~2 The Synapse Personal Access Token REM Use inputs to get credentials for /f %%i in ('curl --location-trusted --silent -H "Authorization:Bearer %~2" "%~1/ststoken"') do set AWS_STS_CREDS=%%i ECHO %AWS_STS_CREDS%
Open the file containing the Synapse PAT and copy the long character string (i.e. eyJ0eXAiOiJ...Z8t9Eg).
Create a ~/.aws/config file if you don’t already have one. Add the following to your
~/.aws/config
file, replacing<PERSONAL_ACCESS_TOKEN>
with the PAT you downloaded then set the/absolute/path/to/synapse_creds.sh
to the location of thesynapse_creds.sh
orsynapse_creds.bat
file.[profile service-catalog] region=us-east-1 credential_process = "/absolute/path/to/synapse_creds.sh" "https://sc.sageit.org" "<PERSONAL_ACCESS_TOKEN>"
Go to the service catalog provisioned product page. Scroll down to get your
EC2InstanceId
under the PROVISION_PRODUCT event. TheEC2InstanceId
is your--target
.Run the SSM start-session command to access the instance. Note: Windows users should do this in command prompt. In the following example, the
EC2InstanceId
from the previous step isi-0fd5c9ff0ef675ceb
.➜ aws ssm start-session --profile service-catalog \ --target i-0fd5c9ff0ef675ceb Starting session with SessionId: 3377358-0cab70190f97fcf78 sh-4.2$
Note: by default you are logged in as the ssm-user. If you prefer to start your session with a different user then you can try running SSM access with custom commands.
8. If you plan to use Docker with your instance(for example, with Rstudio or Jupyter notebooks), complete the instructions in SSM access to applications.
Debugging Access
If you encounter errors try running the AWS start-session command with the
--debug
option.If you use an invalid personal access token you will get an error similar to this
➜ aws ssm start-session --profile service-catalog \ --target i-0fd5c9ff0ef675ceb Expecting value: line 1 column 1 (char 0)
To check whether your token is valid run the following command
➜ curl -I --location-trusted \ -H Authorization:"Bearer ${SYNAPSE_PAT}" https://sc.sageit.org/ststoken
If the HTTP response status is 2xx then the PAT is valid. If the PAT is invalid the response will be 4xx
If you continue to have similar errors to
Expecting value: line X column X (char X)
then it could mean that yoursynapse_creds.sh
file is invalid. Try verifying yoursynapse_creds.sh
script independently of the AWS command by executing just the script. A successful execution should return a valid json and look something like this➜ ~/synapse_creds.sh "https://sc.sageit.org" "eyJ0eXAiO...2GLQg" {"SessionToken":"FwoGZXIvYXdzEN7//////////wEaDP2imuwAK+...13GnBrJc9SlOW6uY=","Version":1,"AccessKeyId":"XXXXXXX","SecretAccessKey":"XXXXXXXXXXXXXXXX","Expiration":"2021-07-21T22:02:17Z"}
Another problem could be that your
~/.aws/config
file is invalid. For debugging we recommend backing up your current config file and creating a new one with just oneservice-catalog
profile in it and then try re-running the start-session command.If you get a message similar to “.. AccessDeniedException when calling the TerminateSession operation..”. it could mean that the AWS SSM session plugin was not install correctly. Please verify that it was successfully installed.
SSM access with custom commands
By default the ssm start
command starts a session with the SSM-SessionManagerRunShell document which will login you in as the ssm-user
with an sh
shell. If you prefer to start your session with a ‘bash’ shell then you can use the AWS-StartInteractiveCommand document.
aws ssm start-session --profile service-catalog \ --target i-0fd5c9ff0ef675ceb \ --document-name AWS-StartInteractiveCommand \ --parameters command="bash -l"
Or if you want to start your session with the ec2-user
then you can do run this
aws ssm start-session --profile service-catalog \ --target i-0fd5c9ff0ef675ceb \ --document-name AWS-StartInteractiveCommand \ --parameters command="sudo su - ec2-user"
SSM with SSH
You can use the AWS Command Line Interface (AWS CLI) to establish Secure Shell (SSH) connections to instances using AWS Systems Manager Session Manager. Users who connect using SSH can copy files between their local machines and the EC2 instance using Secure Copy Protocol (SCP).
Setting SSH connection with ec2-user:
Use the ssm start-session command to connect to the instance
aws ssm start-session --profile service-catalog \ --target i-0fd5c9ff0ef675ceb \ --document-name AWS-StartInteractiveCommand \ --parameters command="sudo su - ec2-user"
Copy the public portion of your ssh key (on your local computer) to the instance’s
~/.ssh/authorizedkeys
file.Set the permission of the authorizedkeys file to 600 (i.e. chmod 600 ~/.ssh/authorizedkeys)
Add the following to your local machine’s
~/.ssh/config
file# SSH over Session Manager host i-* mi-* ProxyCommand sh -c "aws ssm start-session --target %h --document-name AWS-StartSSHSession --parameters 'portNumber=%p'"
From your local machine, execute the ssh command to access the instance
➜ AWS_PROFILE=service-catalog ssh -i ~/.ssh/id_rsa ec2-user@i-0fd5c9ff0ef675ceb Last login: Thu Jun 17 21:25:56 2021 __| __|_ ) _| ( / Amazon Linux 2 AMI ___|\___|___| https://aws.amazon.com/amazon-linux-2/ [ec2-user@ip-10-41-23-76 ~]$
From your local machine, execute the scp command to copy files directly to the instance
➜ AWS_PROFILE=service-catalog scp -i ~/.ssh/id_rsa README.md ec2-user@i-07eeb59282fafe244:~/. README.md 100% 814 9.2KB/s 00:00
SSM access to applications
When running apps in the instance you may want to run the apps on specific ports. The AWS SSM allows you to expose those ports to your local computer using a technique called port forwarding. Here’s an example of how to enable port forwarding to an application:
Run an application on the EC2 (i.e. docker run -p 80:80 httpd)
[ec2-user@ip-10-49-26-50 ~]$ docker run -p 80:80 httpd Unable to find image 'httpd:latest' locally latest: Pulling from library/httpd 33847f680f63: Pull complete d74938eee980: Pull complete 963cfdce5a0c: Pull complete 8d5a3cca778c: Pull complete e06a573b193b: Pull complete Digest: sha256:71a3a8e0572f18a6ce71b9bac7298d07e151e4a1b562d399779b86fef7cf580c Status: Downloaded newer image for httpd:latest AH00558: httpd: Could not reliably determine the server's fully qualified domain name, using 172.17.0.2. Set the 'ServerName' directive globally to suppress this message AH00558: httpd: Could not reliably determine the server's fully qualified domain name, using 172.17.0.2. Set the 'ServerName' directive globally to suppress this message [Thu Jul 22 23:54:12.106344 2021] [mpm_event:notice] [pid 1:tid 140706544895104] AH00489: Apache/2.4.48 (Unix) configured -- resuming normal operations [Thu Jul 22 23:54:12.107307 2021] [core:notice] [pid 1:tid 140706544895104] AH00094: Command line: 'httpd -D FOREGROUND'
To provide access to that app, an SC user can use the port forwarding feature to gain access to the app by running the AWS SSM CLI command:
aws ssm start-session --profile service-catalog \ --target i-0fd5c9ff0ef675ceb \ --document-name AWS-StartPortForwardingSession \ --parameters '{"portNumber":["80"],"localPortNumber":["9090"]}'
To provide access to that app in the Windows Command Prompt use this syntax:
aws ssm start-session --profile service-catalog \ --target i-0fd5c9ff0ef675ceb \ --document-name AWS-StartPortForwardingSession \ --parameters "{\"portNumber\":[\"80\"],\"localPortNumber\":[\"9090\"]}"
Now you should be able to access that app on your local machine at
http://localhost:9090
.
Connecting to Windows Instances
Windows will return an error when connecting via RDP, you may ignore.
“You are connecting to the RDP host "<IP Address>". The certificate couldn't be verified back to a root certificate. Your connection may not be secure. Do you want to continue?”
Connect to Windows shell
Connecting to the Windows instance’s shell is similar to accessing a linux instance’s shell. Just follow instructions in SSM access to an Instance.
Connect to Windows desktop using SSM session manager
Connecting to the Windows desktop requires a few more steps.
Connect to the Windows shell.
Create a new user and and it to the “Administrators” group
$Password = ConvertTo-SecureString "P@ssW0rD!" -AsPlainText -Force New-LocalUser "admin" -Password $Password Add-LocalGroupMember -Group "Administrators" -Member "admin"
Follow the SSM access to applications instructions to setup port forwarding to windows RDP
aws ssm start-session --profile service-catalog \ --target i-0fd5c9ff0ef675ceb \ --document-name AWS-StartPortForwardingSession \ --parameters '{"portNumber":["3389"],"localPortNumber":["3389"]}'
Install the Microsoft Remote Desktop client on your computer.
Click “+” to add a new PC. In the “PC Name” field, enter “localhost”.
Log in with username “admin” and password "P@ssW0rD!"
Connect to Windows desktop using VPN and Jumpcloud (Sage Staff Only)
Sage staff have the option to access the windows desktop using their Jumpcloud credentials. Here are the steps
Once an instance is provisioned locate its instance id (i.e. i-06531e8f977ca20ea)
Create a Jira IT issue and make a request to associate your jumpcloud user with that instance id
Once Sage IT will make the association you can login to the VPN and use remote desktop to login to the instance with your Jumpcloud credentials.
Provisioning and Using a Notebook
Provision and Connect to the Notebook
Following the instructions for provisioning a machine, provision an EC2: Linux Docker instance.
Follow the steps under SSM Access to an Instance to connect to the machine as the ec2-user
user.
Start up the Notebook Software
Log in to your Synapse account in your web browser and, under your Settings tab, create a Personal Access Token with View, Modify, and Download scope. Returning to the command line of the provisioned instance, run:
nano ~/.bash_profile
At the bottom of the file add the lines:
SYNAPSE_AUTH_TOKEN=<token here> export SYNAPSE_AUTH_TOKEN
Now save the file and run
source ~/.bash_profile
Going forward, your Synapse access token will be available as an environment variable.
We now show how to access a Jupyter Notebook. Instructions for the R Studio Notebook follow.
Run a Jupyter Notebook
Start the notebook, passing the access token:
docker run -d -p 8888:8888 -v ~/work:/home/jovyan/work -e SYNAPSE_AUTH_TOKEN=$SYNAPSE_AUTH_TOKEN \ --name jupyter jupyter/scipy-notebook
Note: Passing the access token is a convenience, allowing you to connect to Synapse in your notebook without typing your credentials.
Note: The docker image jupyter/scipy-notebook
is just an example. The flexibility of Docker is that you can use any image you wish.
Now add the Synapse Python client to your notebook.
docker exec jupyter pip install synapseclient
Connect to the notebook in your browser
Get the connection URL. Run:
docker logs jupyter
You will see a line something like:
http://127.0.0.1:8888/?token=8bd05708879ee8faa23813807874291267281cac9efe6e7b
Return to your laptop console and connect to the notebook, following the instructions SSM access to applications, taking care that “portNumber” matches the value used when starting up the notebook with Docker. The command will look something like:
aws ssm start-session --profile service-catalog --target i-01289326a8b08c72d \ --document-name AWS-StartPortForwardingSession \ --parameters '{"portNumber":["8888"],"localPortNumber":["8888"]}'
(We call this “port forwarding” because it connects a port on your local machine to a port on the remote EC2 instance.) Now, in browser visit the URL shown in the ‘docker logs’ output. You will see the Jupyter console.
Connect to Synapse from a Jupyter Notebook
In the Jupyter console, open a new notebook and run the following:
import synapseclient syn=synapseclient.Synapse() # Note, you do not need to pass credentials when logging in. syn.login() # download this sample notebook syn.get("syn25931378", downloadLocation="/home/jovyan/work")
Now return to the Files menu and you will see the downloaded notebook in the ‘work’ folder.
Clicking ‘Save’ in Jupyter will only save a copy top the provisioned EC2 machine, which you may terminate. If you want to save important results, push a copy to Synapse, like so:
myOwnProject = "syn25931363" myCopy = synapseclient.File("/home/jovyan/work/sample_notebook.ipynb", parent=myOwnProject) syn.store(myCopy)
Stopping your notebook
To stop port-forwarding, in the terminal where you started port forwarding, hit Control-C.
To stop the notebook, return to the terminal in which you have a connection to the EC2 machine or start a new connection as earlier:
aws ssm start-session --profile service-catalog --target i-0fd5c9ff0ef675ceb --parameters command="sudo su - ec2-user"
where i-0fd5c9ff0ef675ceb
is the instance ID of your EC2 instance. At the EC2 command line you can stop and remove the notebook container:
docker stop jupyter docker rm jupyter
Run an RStudio Notebook
On the EC2, run the following:
git clone https://github.com/Sage-Bionetworks/docker-rstudio.git cd docker-rstudio/ cp .env.example .env
Now edit the file .env and change the line starting SYNAPSE_TOKEN
to read:
SYNAPSE_TOKEN=$SYNAPSE_AUTH_TOKEN
Make any other customizations you desire. Note the username and password. Save the file and run:
docker-compose up
which connects an R-Studio notebook to port 8787. Start a port forwarding session forwarding port 8787
:
aws ssm start-session --profile service-catalog --target i-0fd5c9ff0ef675ceb \ --document-name AWS-StartPortForwardingSession \ --parameters '{"portNumber":["8787"],"localPortNumber":["8787"]}'
As before the value for ‘target’ is just an example.
Now turn your browser to http://localhost:8787 and you will see your R Studio instance, using the credentials from the .env
file to log in.
Managing an Instance
On the product page for each provisioned product, there is an orange “ACTIONS” button in the upper right that allows you to manage the instance. You can also visit the EC2 Console to check on the status of your instance.
Update
Using the update action allows you to change parameters or update to a new version of the product. WARNING: changes to configuration parameters usually result in a recreation (“replacement”) of the instance, any data saved on the instance will be lost, and the nature of the update by Amazon is difficult to predict. We recommend that you save any important data to S3, provision a new instance and terminate the original.
Terminate
The terminate action deletes the instance permanently.
Stop
Stops the instance. We recommend using this when the instance is not in use.
Start
Starts the instance.
Restart
Performs a restart.
Update Plan
The service catalog allow you to create a plan (AKA a changeset) when updating your product. A plan will provide you with some information on what AWS will do before you execute the planned update. After a plan is created you can either delete, modify, or execute the plan. The update does not happen until you execute the plan.
Change Owner
Please ask #sageit for help transferring ownership if that is deemed to be necessary rather than using this action. Otherwise, you may find that you cannot connect to your instance.
Cloud Storage (S3)
Note: S3 storage products currently are available only to Sage employees.
Cloud Storage Products
To understand the cost of S3 buckets see https://aws.amazon.com/s3/pricing/ or use the AWS pricing calculator. Note that while data egress can be a substantial cost, our Service Catalog provisions buckets and EC2 instances in the same AWS region. Since AWS does not charge for egress to a location within a bucket’s region, accessing data from an instance in provisioned by our Service Catalog will avoid such costs.
S3 Private Encrypted Bucket
This product builds an encrypted AWS S3 bucket with private access. To access the bucket, see the section Using an S3 Bucket, below. Note that only the user who provisioned the bucket can access it. To share a bucket, use the Synapse Bucket instead.
S3 Synapse Bucket
This product builds an AWS S3 bucket with private access for Synapse. The additional configuration parameter is the name of the Synapse user who will be allowed to link the bucket to a Synapse project. Once linked to a project, Synapse can be used to share the bucket with other users.
Creating S3 Products
Product Version and Parameters
When provisioning, you are prompted for two names. On the “Product Version” screen of the wizard, you must name your product. This is the name you will see listed in Service Catalog under “Provisioned Products” later. Please include your name in the product, e.g. if your name is Jane Doe and you are provisioning a bucket for your project Foo, you could name it jdoe-bucket-foo
. On the “Parameters” screen, you have the option of naming the bucket itself (otherwise a name will be assigned). That is the name to use when accessing the bucket through the Amazon S3 client or via the Amazon S3 console.
The S3 Synapse Bucket product requires the “SynapseUserName” field.
There is an additional field in S3 Synapse Bucket, “SynapseProdARN”, that should be left as its default value.
Tag options, Notifications, and Review
The final three screens of the wizard are the same as in Creating an EC2 Instance above.
Using an S3 Bucket
As with EC2 products, once provisioning is complete your S3 product will appear in the “Provisioned Products” list, showing status Available. Select “Provisioned Product Details” from the navigation panel on the left, and click on your product. A product that has a “Succeeded” event will have outputs that include a “BucketUrl” link.
Using a bucket from the AWS Console
Clicking on the “BucketUrl” link from the provisioned product takes you to the S3 console where you can upload and download files.
Using a bucket with the S3 client
To authenticate the S3 client for bucket access, follow the set-up steps under SSM access to an Instance, above. You can then access the bucket by including “--profile service-catalog” in the command, e.g. to download a file the syntax is:
aws --profile service-catalog s3 cp s3://<your-bucket-name>/file.txt ./file.txt
Using a bucket with the Synapse client
You can access a provisioned Synapse bucket using one of the synapse clients to push and pull data. In order to complete the setup of the Synapse Bucket, you must manually complete the final step of setting the S3 bucket as a Synapse upload location and linking it to a project. The project’s sharing settings can then be used to control what Synapse users can access the bucket.
Managing an S3 Bucket
On the product page for each provisioned product, there is an orange “ACTIONS” button in the upper right that allows you to manage the instance.
Update
Using the update action allows you to change parameters or update to a new version of the product.
Terminate
The terminate action removes the bucket product from the Service Catalog however the bucket is not immediately deleted. The bucket and data in it will be placed in an archived state where no users will have access to the bucket. Upon expiration of the archive period (30 days) the bucket and data will get automatically purged from AWS. The owner of the bucket may request Sage IT to restore access to the bucket before it is purged
Change Owner
Please ask #sageit for help transferring ownership if that is deemed to be necessary rather than using this action.
Scheduled Jobs
Note: Scheduled Jobs products currently are available only to Sage employees.
Scheduled Job Products
Scheduled jobs are essentially cron jobs that we’ve setup to run in AWS batch. This product allows you to run an arbitrary task on the cloud to process your workload. The task must be run using a docker image. The task can be manually triggered or it can be setup to run on a schedule. A scheduled job is limited to 10 minutes of run time and to the memory/CPU chosen when launching the product.
Creating Scheduled Job Products
To create a scheduled job, select the “Products List” from the navigation panel on the left. Next, select “Scheduled Jobs” from the list. On the product page, click the orange “LAUNCH PRODUCT” button under the product description, then fill out the wizard. Most of the parameters contain helpful information describing the valid inputs.
Notes:
The “Environment” parameters are required fields. You can replace the default values, however please do not leave these fields empty. Also pay special attention to the formatting that’s required for the values. The deployment will fail if the formatting isn’t correct.
There is an AWS bug that prevents disabling the scheduled job after it has been enabled. The workaround is to either (1) Terminate the job and create a new one or (2) Set the rate to some distant time in the future (i.e. 3650 days).
Tag options, Notifications, and Review
The final three screens of the wizard are the same as in Creating an EC2 Instance above.
Manually Run Scheduled Jobs
Once provisioning is complete your Scheduled Jobs product will appear in the “Provisioned Products” list, showing status Available. Select “Provisioned Product Details” from the navigation panel on the left, and click on your product. A product that has a “Succeeded” event will have outputs that include a “SubmitJobApi” link. The link is a URL to trigger the job, which can be entered in your web browser or a cUrl command. Entered into your web browser the result will look like:
NOTE: Since invoking the URL requires no authentication, anyone obtaining the URL can trigger the job. So the URL created for the provisioned product must be treated as a secret, just like any password or key. If its secrecy is compromised the provisioned product must be terminated.
View Scheduled Job Status
Click on the “Jobs” link to view the batch job status. Once triggered the job should transition from STARTING → RUNNING → SUCCEEDED.
Access Scheduled Job Logs
The job will send logs to AWS cloudwatch. To access the logs click on the “Logs” link in PROVISION_PRODUCT outputs. Below is an example of a log for a job that ran the command “printenv”.
Scheduled Job Secrets
Secrets are stored in the AWS secrets manager and exposed to the job as environment variables. The logs above print out the environment variables from the job. Take note of the “SCHEDULED_JOB_SECRETS” parameter in the logs. The secrets that are passed into this product are exposed as environment variables in the logs by the “printenv” command. Please make sure to never expose secrets in this way. DO NOT PRINT ENVIRONMENT VARIABLES.
Accessing Scheduled Job Secrets
Job secrets can be access a number of different ways. The first way is simply to get it from the docker container environment variable SCHEDULED_JOB_SECRETS.
Environment variable example:
printenv SCHEDULED_JOB_SECRETS|jq .SECRET1 "Shh1"
Passing Synapse access token as a secret using environment variables:
You can use a scheduled job to access data in Synapse. To do so:
Create a personal access token (PAT) as explained here.
In the Secrets parameter of your scheduled job, include the PAT, e.g.
"PAT":"eyJ0eXAiOiJKV1QiLCJraWQiOiJXN05OOldMSlQ..."
You may include other secrets if needed, each name/value pair separated by a comma.
Add code in your containerized script to parse the “
SCHEDULED_JOB_SECRETS
" JSON object, extract the PAT and use it to authenticate. Here is a Python example, synapse_login.py:import json, os, synapseclient secrets=json.loads(os.getenv("SCHEDULED_JOB_SECRETS")) auth_token = secrets["PAT"] syn=synapseclient.Synapse() syn.login(authToken=auth_token)
Create a container image which includes the Synapse client, this script, and the command to invoke the script. Here is a Dockerfile to create such a container image:
FROM sagebionetworks/synapsepythonclient COPY synapse_login.py synapse_login.py CMD python3 synapse_login.py
Now you can build the container image, push to a public Docker registry and use in your Scheduled Job.
Retrieving job secrets using the AWS client
To access the job secrets using either the AWS CLI or one of the AWS SDKs you must copy the “JobSecretArn” from the service catalog PROVISION_PRODUCT outputs. Then provide the JobSecretArn to either the AWS secrets manager CLI or one of the AWS SDKs to retrieve the secret.
AWS secrets manager CLI Example :
aws secretsmanager --output json get-secret-value --secret-id arn:aws:secretsmanager:us-east-1:465877038949:secret:JobSecrets-rEx1eKL9pokj-h7hCGO { "ARN": "arn:aws:secretsmanager:us-east-1:465877038949:secret:JobSecrets-rEx1eKL9pokj-h7hCGO", "Name": "JobSecrets-rEx1eKL9pokj", "VersionId": "09904c83-f4ea-4664-a773-eded857ab5a0", "SecretString": "{ \"SECRET1\":\"Shh1\" }", "VersionStages": [ "AWSCURRENT" ], "CreatedDate": "2021-12-18T09:01:23.690000-08:00" }
# Use this code snippet in your app. # If you need more information about configurations or implementing the sample code, visit the AWS docs: # https://aws.amazon.com/developers/getting-started/python/ import boto3 import base64 from botocore.exceptions import ClientError def get_secret(): secret_name = "arn:aws:secretsmanager:us-east-1:465877038949:secret:JobSecrets-rEx1eKL9pokj-h7hCGO" region_name = "us-east-1" # Create a Secrets Manager client session = boto3.session.Session() client = session.client( service_name='secretsmanager', region_name=region_name ) # In this sample we only handle the specific exceptions for the 'GetSecretValue' API. # See https://docs.aws.amazon.com/secretsmanager/latest/apireference/API_GetSecretValue.html # We rethrow the exception by default. try: get_secret_value_response = client.get_secret_value( SecretId=secret_name ) except ClientError as e: if e.response['Error']['Code'] == 'DecryptionFailureException': # Secrets Manager can't decrypt the protected secret text using the provided KMS key. # Deal with the exception here, and/or rethrow at your discretion. raise e elif e.response['Error']['Code'] == 'InternalServiceErrorException': # An error occurred on the server side. # Deal with the exception here, and/or rethrow at your discretion. raise e elif e.response['Error']['Code'] == 'InvalidParameterException': # You provided an invalid value for a parameter. # Deal with the exception here, and/or rethrow at your discretion. raise e elif e.response['Error']['Code'] == 'InvalidRequestException': # You provided a parameter value that is not valid for the current state of the resource. # Deal with the exception here, and/or rethrow at your discretion. raise e elif e.response['Error']['Code'] == 'ResourceNotFoundException': # We can't find the resource that you asked for. # Deal with the exception here, and/or rethrow at your discretion. raise e else: # Decrypts secret using the associated KMS key. # Depending on whether the secret is a string or binary, one of these fields will be populated. if 'SecretString' in get_secret_value_response: secret = get_secret_value_response['SecretString'] else: decoded_binary_secret = base64.b64decode(get_secret_value_response['SecretBinary']) # Your code goes here.
Best Practices
When to Use the Service Catalog
The benefits of the Service Catalog are that it is self-service, meant to fulfill the most common needs for compute and storage, and that it creates resources in a PHI-safe environment. We encourage you to use it preferentially. However, it will not fulfill all needs. For custom development in a PHI-safe environment, the “scicomp” account remains the preference for Sage employees. For custom development that does not concern PHI, the “sandbox” account can be used by Sage employees. For more information, see the Sage Bionetworks intranet article on computing. If you have any questions about which environment is most suitable, questions are welcome in the #sageit Slack channel!
Ephemeral Instances, Persistent Data
We encourage you to treat your instances as ephemeral. It is very easy with the Service Catalog to create new instances, and since updating the parameters of an instance frequently results in their recreation, it is best not to get too attached to any one. This makes getting in the habit of storing your data in Synapse or S3 highly desirable. If you leave the organization you should again store any data your colleagues will need in Synapse or S3, then terminate your instances. Any instances left running will be terminated upon your departure.
EC2 SSM Session Timeouts
There is a short idle timeout on the SSM sessions started in the browser by clicking on the EC2 ConnectionURI link as described above. This will happen if you don't actively use the session for twelve minutes or more. The session in a browser is designed to allow quick access to EC2 instances. If you prefer longer sessions then we recommend that you setup command line SSM access to an instance.
Frequently Asked Questions
Why didn’t I receive an email about my new product?
Emails are not enabled in the current system. You can get status updates on the provisioned products pages, including connection instructions when the provisioning succeeds.
Note: you may need to expand the product events to view connection information.
Which OS versions are currently available?
Amazon Linux 2 AMI 2.0.20190823.1 x86_64 HVM gp2
Ubuntu 18.04 (Bionic)
Windows Server 2019
Why can’t I set my own AMI ID?
The current product templates are meant to cover the most common use cases. You may request a new product if your use case is not covered. Sage employees can do this by filing a Jira issue. Others should contact their community manager.
Why does the status show “available” even after I stop my EC2?
The status on the service catalog UI is either “available” or “under change”. That status only pertains to the service catalog product, which can be any provisioned AWS resource or bundle of resources (i.e. DB, ALB, S3, ECS, etc..). That status does not pertain specifically to an EC2 instance’s activity status. Unfortunately the service catalog UI does not provide EC2 status. To view your EC2’s status you must click on the LinuxInstance
link and view the EC2 status on the AWS EC2 console UI.
What is a User ARN?
ARN stands for Amazon Resource Name, a unique identifier for a particular resource in AWS.
Why can’t I access my windows instance?
Windows instances take much longer than linux instances to initialize. It can take between 5-15 minutes to complete the initialization process. If you provision a windows instance you may need to wait up to 15 minutes after the Service Catalog completes before you can RDP into it.
How can I ask a question or otherwise get help?
Sage employees should use the #sageit Slack channel. Members of other communities can contact their community manager or reach out via the Synapse help forum.