This directory contains sample utility scripts to help you work with AWS Deadline Cloud. These scripts provide command-line tools for common tasks like managing job attachments, working with queues, and automating workflows.
The upload_to_job_attachments script uploads files and directories from your local workstation or server to AWS Deadline Cloud job attachments storage. It uploads files to the job attachments S3 bucket in content-addressable storage format, allowing subsequent Deadline Cloud jobs to use the data without re-uploading. This is useful for pre-populating job attachments with large datasets that multiple jobs will use.
Key features:
- Upload individual files or entire directories recursively
- Multi-threaded uploads for better performance
- Automatic deduplication (skips files already in S3)
- Generates JSON manifest of uploaded files
- Two configuration modes: direct S3 specification or queue lookup
Example usage:
# Upload using direct S3 specification
python upload_to_job_attachments.py \
--s3-bucket my-bucket \
--s3-prefix job-attachments \
--paths /path/to/files /path/to/directory
# Upload using queue lookup
python upload_to_job_attachments.py \
--farm-id farm-1234567890abcdef \
--queue-id queue-1234567890abcdef \
--paths /path/to/files /path/to/directory