An ftrack s3 accessor updated to work with ftrack-python-api and boto3.
Install using pip:
pip install ftrack-s3-accessor
If you want to use with the transfer components action found here you'll need to additionally:
pip install ftrack-action-handler
Configure a new location within ftrack with the name 'studio.remote'. This will be used as the location for s3.
Create a storage bucket in s3 and set the bucket name using the FTRACK_S3_ACCESSOR_BUCKET environment variable (default: ftrack). Ensure your bucket name is globally unique and meets aws s3 naming restrictions.
Set all other ftrack environment variables for your ftrack instance.
Running the scripts from within your environment requires you to additionally set your sources root to the accessor directory.
PYTHONPATH=./ftrack_s3_accessor
Ensure you have an working aws configuration under your ~/.aws folder. You can check this by running:
import boto3
If this fails, your aws configuration isn't setup properly. Refer to the boto3 documentation on how to set it up. You should only need a ~/.aws/config and ~/.aws/credentials file.
The main plugin can be found in the plugins folder. This folder may be registered using the FTRACK_EVENT_PLUGIN_PATH ftrack environment variable so that it is picked up when ftrack is started.
Examples of how to use the plugin can be found in the scripts folder. The simplest way to launch ftrack with the accessor is scripts/start_ftrack_with_s3.py.
It is possible to use the transfer components action to move components between local and remote storage. Ensure it is on the FTRACK_EVENT_PLUGIN_PATH (or add it to the plugins folder) and it should become available under ftracks actions menu. You will need to ensure your local storage is also correctly configured within a script when running the accessor outside of ftrack-connect, as the connect location configured by the desktop client will not be available as an option.