Replies: 1 comment 4 replies
-
Hi @roywei, thanks for your interest! The shared FS example in the docs uses an object store (S3/GCS/R2) as the backing persistent store. We provide access to these object stores by creating a FUSE mount using either goofys (for S3 and R2) or gcsfuse (for gcs). Performance varies depending on your chosen VM type, the backing store and the workload. For S3, we ran some benchmarks and the detailed results can be found here. To summarize, on AWS m5.8xlarge with default EBS volume attached with 4 parallel processes reading/writing to files, we observed: Disk Bandwidth (MB/s)
Disk IOPS
Please note that performance is highly dependent on the workload. The discussion section in the results file provides cases when shared storage performance might be low (e.g., many small files). Maximum file size, though I haven't tested it, should be similar to the limits put in by the backing object store. Here's S3's limits. If you have a particular use case in mind, I am happy to discuss it! |
Beta Was this translation helpful? Give feedback.
-
Hi,
I love this project and great work! One question I have is what's the file storage size limit and throughput limit for a shared file system that all nodes can access (read/write)? It's not very clear from the storage doc, and what's the underlying implementation?
Thanks and can't wait to try this out!
Beta Was this translation helpful? Give feedback.
All reactions