Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

This is the minimum information needed to bootstrap our system with the information needed to load a configuration via a .properties file.  Here, the actual .properties file should be loaded in .S3

where <<encryption key>> is found on our unix system in

work/platform/PasswordsAndCredentials/platformPropertyEncryptionKey.txtS3 as described below

Setting up a Properties file in S3

For each stack, we have created a unique IAM User, encryption key, and configuration file.  These values are passed into the container of the environments as described above.  AWS access key ids, secret keys, encryption keys, and the url for an environment can be found on sodo at /work/platform/PasswordsAndCredentials/StackCredentials/IAMUsers in the appropriate .csv file.  All stack environments run under this IAM User, and have permission to access their configuration file from S3.  Configuration files can be loaded / updated in S3 under the elasticbeanstalk-us-east-1-325565585839 bucket (this is the same place the .war files are deployed).  This will give URLs of the form https://s3.amazonaws.com/elasticbeanstalk-us-east-1-325565585839/beanstalk-<stack-name>-stack.properties  If you are creating a new stack, you will have to create the IAM user and grant that user access to access the configuration file using the IAM tab of the AWS console.  In most cases you should be able to keep the configuration the file the same, or replace it with a new file of the same name.

How to run the Data Loader

On Once environments are running, you can populate the system with a set of starting data.  On one of the local servers, goto /work/platform/DatasetMetadataLoader and execute the following:

Code Block
# Make sure you have the latest version
svn up
# Execute the loader
# Replace <repo_instance> and <auth_instance> by the repository and authentication instances.
# Either make sure that <platform_admin_email> is a Synapse administrator on crowd, or replace it by a Synapse administrator account
python datasetCsvLoader.py -d ./AllDatasets.csv -l ./AllDatasetLayerLocations.csv -e http://<repo_instance>/repo/v1 
-a http://<auth_instance>/auth/v1 -m /work/platform/DatasetMetadataLoader/platform.md5sums.csv -u <platform_admin_email>  -p <platform_admin_pw>

This will create a publicly-accessible project called Sage BioCuration, and populate it with curated data from Sage's repository data team.

Verify Deployment

To verify deployment, run top-level queries against the repository instances from an authenticated account.

...