Mounting shared storage
Mount persistent storage volumes to your GPU instances across different providers.
Overview
Persistent storage volumes provide shared, network-attached storage that can be attached to your GPU instances. These volumes persist independently of your instances, so data is not lost when an instance terminates.
Key features:- Volumes persist independently of instances; data is not lost when an instance terminates
- A single volume can be attached to multiple instances simultaneously (provider-dependent)
- Up to 10 volumes can be attached to a single instance on Spheron AI and Verda; Voltage Park and Sesterce are limited to one volume per instance
- Provider-specific mounting protocols (NFS or block device)
- Volumes can be attached and detached from running instances without data loss
- Billed by capacity provisioned, not by usage
Prerequisites
Before mounting a volume, ensure you have:
- Created a persistent storage volume via the Spheron dashboard (Create Volume page)
- Attached the volume to your GPU instance
- Retrieved the volume's connection details (virtual IP, mount path, protocol)
- SSH access to your instance
Provider-specific mounting guides
Select your provider below for detailed mounting instructions:
Voltage Park
Protocol: Network File System (NFS) v3 Status: Available
Mount high-performance NFS volumes on Voltage Park instances with parallel connection support.
View Voltage Park mounting guide
Key features:- NFS v3 protocol with high-performance
nconnectoption - Virtual IP-based network mounting
- 1 volume per instance (a volume can only be attached to one instance at a time)
- Automatic fstab configuration for boot persistence
Verda
Protocol: Network File System (NFS) Status: Available
Mount shared filesystems on Verda instances using datacenter-region NFS endpoints and pseudopaths.
Key features:- NFS protocol with high-performance
nconnectoption - Datacenter-specific NFS endpoints (e.g.,
nfs.fin-01.datacrunch.io) - Pseudopath-based filesystem identification
- Up to 10 volumes can be attached to a single instance
- Supports concurrent multi-instance access within the same datacenter
- Cluster-wide mounting via
pdsh
Sesterce
Protocol: Block device (ext4) Status: Available
Mount persistent block storage volumes on Sesterce instances by formatting and mounting raw block devices.
Key features:- Direct-attached block storage with local disk throughput
- UUID-based fstab configuration for stable boot persistence
- ext4 filesystem for broad compatibility
Spheron AI
Protocol: Block device (ext4, xfs, or ntfs) Status: Available
Mount persistent block storage volumes on Spheron AI instances by formatting and mounting raw block devices. Choose your preferred filesystem.
View Spheron AI mounting guide
Key features:- Up to 10 volumes can be attached to a single instance
- Direct-attached block storage with local disk throughput
- Flexible filesystem choice: ext4, xfs, or ntfs
- UUID-based fstab configuration with
nofailfor safe boot persistence
Other providers
Status: Coming soon
Support for additional providers is in development. Each provider may use different protocols and mounting procedures optimized for their infrastructure.
Planned providers:- Massed Compute
Check back for updates or contact support if you need volume mounting support for a specific provider.
General concepts
Volume lifecycle
Create Volume → Attach to Instance → Mount → Use → Unmount → Detach → Delete
↓ ↓ ↓ ↓ ↓ ↓ ↓
Available Available Attached Attached Attached Available DeletedCommon operations
Creating volumes: Create persistent storage volumes from the Create Volume page in the Spheron dashboard. Specify the provider, name, and size in GB.
Attaching volumes: Attach a volume to a running instance from the volume details page in the dashboard. The volume and instance must use the same provider.
Mounting: Follow the provider-specific mounting instructions in the guides linked above.
Detaching volumes: Unmount the volume from the instance first, then detach it from the dashboard. Volume data is preserved after detachment.
What's next
- Volume API Reference: Create and manage volumes programmatically
- External Storage Access: Configure local NVMe drives on Voltage Park
- SSH Connection Setup: Secure instance access
- Security Best Practices: Secure your storage and data
- Getting Started: Complete deployment guide
For questions, join Discord or contact support via the dashboard.