Skip to content
Discussion options

You must be logged in to vote

Hi @fghorow,

We use dascore to manage fairly large datasets, I think the largest was around 60 TB, and it looks like the spool indexer could handle much more. These, however, were on a local NAS. We don't (yet) have a great way to work with data in an S3 bucket directly. However, if you can mount the S3 bucket to behave like a file system in a cloud environment (eg with Amazon EFS or FSx ), the you should be able to use DASCore's spool method as shown in the docs. You can then index (spool.update) and select/iterate through the contents as you need.

However, if you only need a few files, you can browse the s3 bucket and download what you need locally. This will require a lot less setup.

B…

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@fghorow
Comment options

Answer selected by fghorow
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants