SageRx is a medication ontology and medication-related data aggregator created from many different public sources of data.
Including data from DailyMed, FDA, RxNorm, Orange Book, and more!
SageRx uses Airflow to schedule jobs to extract, load, and transform (using dbt) open drug data.
Data ends up in a PostgreSQL database and can be queried using pgAdmin (included with SageRx) or via any SQL editor of your choice.
We will be moving documentation over to Github, but additional documentation exists on the SageRx Website.
Subscribe to our newsletter to keep ontop of updates.
We would love to see you contribute to SageRx. Join our Slack channel to get involved.
Style Guide: How we think about the structure and naming conventions of SageRx.
- Install Docker.
- Windows users will also install WSL 2 (a Linux subsystem that runs on Windows) as part of this process.
- Clone the repo.
- Add a
.env
file at the root of the repo. - Add ENV vars to the
.env
file.AIRFLOW_UID=<uid>
- UID can be found by runningid -u
on linux systems, typically the first user on the system is1000
or1001
.- Windows users following the Docker Desktop install guide should have WSL 2 installed. You can open up command line, type
wsl
and then within WSL 2, you can enterid -u
to see your UID.
- Windows users following the Docker Desktop install guide should have WSL 2 installed. You can open up command line, type
UMLS_API=<umls_api_key>
- if you want to use RxNorm, you need an API key from UMLS.
- Make sure Docker is installed
- Run
docker-compose up airflow-init
. - Run
docker-compose up
.
NOTE: if you have an M1 Mac
export DOCKER_DEFAULT_PLATFORM=linux/amd64
, and re-build your images
NOTE 2: if you're running WSL1/2 you may need to use
docker compose
rather thandocker-compose
per this
- Airflow UI is hosted on
localhost:8001
or0.0.0.0:8001
- Username/password =
airflow
/airflow
- Username/password =
- PgAdmin is hosted on
localhost:8002
or0.0.0.0:8002
- Username/password =
sagerx
/sagerx
- Username/password =
On docker-compose up
a dbt container will be created to be used for cli commands. To enter commands run docker exec -it dbt /bin/bash
. This will place you into a bash session in the dbt container. Then you can run dbt commands as you normally would.
To serve dbt documentation locally, enter the commands in the dbt container dbt docs generate
then dbt docs serve --port 8081
. They should generate on http://localhost:8081
The export_marts DAG is implemented to allow users to push .csv versions of the marts-layer tables to an AWS S3 bucket of their choosing. The DAG is currently configured to export 2 tables from the sagerx_dev schema: all_ndc_descriptions and atc_codes_to_rxnorm_products. Future iterations may allow for more schemas/tables as demand dictates. If a user wishes to get .csv copies of those tables pushed to an AWS S3 bucket, they will need to add an additional 3 variables to their .env file (continuing from the Installation instructions):
- ACCESS_KEY = 'your-aws-access-key-id'
- SECRET_ACCESS_KEY = 'your-aws-secret-access-key-id'
- DEST_BUCKET = 'the-name-of-your-aws-bucket'
The access and secret-access keys can be found in 2 ways:
- If the user has the AWS CLI tools installed, simply type from Mac/Linux command prompt OR Windows Powershell:
- cat ~/.aws/credentials
- If the user does not have AWS CLI tool installed or is unfamiliar with such, it is possible you would have saved your credentials in a location on your local machine - the file will be named after the IAM User in your AWS account, something like 'username_accessKeys.csv'. If you can't find that file, simply create a new set by following this procedure:
- Log in to the AWS console
- Search 'IAM' in the search bar and select the IAM Service --> the IAM Dashboard is displayed
- In the 'IAM resources' box, click on the number beneath 'Users' --> all IAM users are displayed
- choose the user with permissions in accordance with your needs (this will typically be the user with administrator access, but the only IAM User permission that is required is Read/Write permissions on your S3 bucket)
- Within the user page, click on 'Security Credentials', and scroll down to the box titled 'Access Keys'
- NOTE: A single user can only have 2 valid sets of login credentials at any one time. If you already have 2, you will need to either delete one or create a new user for this application
- NOTE: Once you have created a set of credentials, you only have 1 opportunity to view/save those credentials from the AWS UI. you should therefore be sure to save the .csv file in safe place.
- Click 'Create access key', select the 'Command Line Interface' option, click Next, fill in a name for the keys, click 'Create access keys', then download the .csv file and save it in safe place
- Open the .csv file, then paste the access key and secret access key into the .env file as described above
Currently we are utilizing 2 GCP products: Google Cloud Storage (GCS) and BigQuery (BQ).
The current workflow has all of the dbt tables are created locally with only the final products being pushed to GCP. This reduces computational expenses especially as we test out new data sources and need to run dbt more frequently.
To accomplish this yourself, you need to follow these steps:
- Set up a GCP account (including billing)
- Made a new storage bucket under GCS
- Enable IAM API
- Create a new service account with permissions to add and delete content for both GCS and BQ, dbt example
- Created new key for service account and download it as JSON, this gets added to the root directory of the project as gcp.json
- Add the necessary environment variables to .env
- GCS_BUCKET
- GCP_PROJECT
- GCP_DATASET
- Rebuild docker containers
If you get issues on folder permissions:
sudo chmod -R 777 postgres,data,extracts,logs,plugins
If you get trouble from the postgres container with errors such as password authentication failed for user "airflow"
or role "airflow" does not exist
, these are all from the same issue that postgres is not setting itself up correctly. This is because of a file permission issue solved by running chmod +x ./postgres/0_pg_stat_statement.sh
. You might need to remove any existing database configuration with rm -rf airflow/data
and docker-compose down --volumes
.