Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 9 Next »

This is a small Heroku project in charge of weekly exports of our EAD files from the ArchivesSpace API to S3.

There is ruby code [stored in github] to export from ArchiveSpace API to EAD files on an S3 bucket. This code runs as a heroku app, scheduled to execute regularly (currently nightly).  There is terraform configuration [also stored in the same github repo] to create and manage the S3 and other AWS resources involved. 

More detailed documentation can also be found in the README of the GitHub repo. Here are some links and overview:

  • GitHub project sciencehistory/export_archivesspace_xml

  • Heroku project export-archivesspace-xml

    • Note: we currently have 3 Heroku add-ons: Proximo for proxying ($5 / month); Papertrail for logging (free); and the Heroku Scheduler to actually spin up the task (free).

    • Note there are heroku configuration variables needed to identify and provide access to archivespace and S3 resources, see github repo README.

  • S3 bucket

    • Configured for “static website hosting”

    • Unfortunately, currently accessed only over http. We’d need to set up cloudfront for https access to an s3 bucket, which we have not currently done.

  • CNAME ead.sciencehistory.org, which points to the s3 bucket

    • Managed using ordinary sciencehistory DNS, our external partners use this hostname so it must be kept running.

  • IAM user that can access the bucket

  • IAM policy granting that user access to the bucket

  • No labels