...
- Round Robin type data store - Basically any kind of storage that is fixed in size, and does semi-automatic data aggregation. The basic idea is that for a period X you have the full details of whatever data you log. Then after X has elapsed, the data is aggregated in one, or several ways (average, minimum, maximum). This aggregate data is then stored for another period Y. Repeat until the data is no longer relevant and can be dropped from the store.
- Data Collection - Since Synapse is on Amazon's Elastic Beanstalk, there is the possibility that data usage must be aggregated from several different Synapse instances. In addition, certain data (like user activity) is most easily collected from other sources than the services (like the Crowd servers). Thus some kind of data collection mechanism is needed.
- Data Interpretation - Since both metrics so far proposed (User and project activity levels) are somewhat expensive to compute (if it's even possible), ideally the front-end GUI will not request that this data be recomputed ever. Some background process - whether it is hosted in the metrics web server, or run independently - is needed to do any kind of pre-processing to the data before it is entered into the data store.
Proposed Solutions
Amazon Cloudwatch
Since it turns out Cloudwatch does no data aggregation/consolidation and holds onto Cloudwatch stores data for exactly two weeks, this is not suitable for the usage metrics use case. To extend the life span, the data must be retrieved via CloudWatch API and then stored to S3, DynamoDB, or Redshift.
Custom EC2 Instance
Pros:
- No storage limitations except cost
- No data expectations to workaround
- This EC2 instance would be able to act as both the data collection mechanism and the data store, allowing it to keep it's back end storage mechanism in a consistent state.
Cons:
- Another custom application/library to build and maintain
- It's not clear what the best way to actually implement the collecting of the data would be.
Google BigQuery
Pros:
- No storage limitations except cost
- Can store the data in full detail
- Fast search times, with no pre-calculated search parameters
- Thus exposing a new metric is as simple as thinking of it, and then implementing fetching the data from BigQuery, and then revealing it in the UI
- The infrastructure is built for us. Amazon provides all the pieces to make a system like BigQuery, or at least that solves the same problems, but they're pieces, not a product.
Cons:
- It's a Google technology, not Amazon thus doubling the number of accounts, maintenance etc.
- I feel like there may be others, but I can't think of them right now
Data Requirements
User Activity Data
Given data set: A time-ordered list (or set of lists) of user auth-events from the crowd servers
...
- Current activity status
- Creation date
- Last login date
- Number of total logins
Project Activity Data
Basically the same source data as for User activity is available (or could be), and the method for calculating it is essentially the same. So it's essentially the same metric, just for projects, not users. There may be the additional information on what type of usage it was (data access, data modification, data addition etc.), but otherwise the same.
...