Storage
Introduction
Block Storage operates at the raw block level using high-performance protocols like Fibre Channel (FC) at 8/16/32 Gbps or iSCSI over Ethernet, providing direct-attached storage or SAN connectivity with sub-millisecond latency. It presents logical unit numbers (LUNs) as raw disk volumes to the operating system, making it optimal for databases, virtual machine disk images, and transactional workloads requiring consistent IOPS performance and low-level disk control.
File Storage operates at the file system layer using network protocols like NFS (supporting NFSv3/v4 with features like client-side caching and Kerberos authentication) or SMB/CIFS (with opportunistic locking and distributed file system capabilities), providing POSIX-compliant file semantics with hierarchical directory structures, metadata management, and concurrent access controls ideal for content repositories and shared application data.
Object Storage uses RESTful HTTP/HTTPS APIs over TCP/IP, storing data as objects with unique identifiers in flat namespaces within buckets or containers, implementing eventual consistency models and offering features like versioning, lifecycle policies, cross-region replication, and virtually unlimited horizontal scaling through distributed hash tables, making it perfect for cloud-native applications, content distribution, backup repositories, and big data analytics requiring petabyte-scale storage with global accessibility.
MinIO is an open-source object storage solution that's compatible with Amazon S3's API. It's particularly popular for private cloud deployments and can be run on-premises or in any cloud environment. MinIO excels at high-performance workloads and is often used in conjunction with Kubernetes for scalable container deployments.
Amazon S3 (Simple Storage Service) is the industry standard for cloud object storage, offering virtually unlimited scalability, 99.999999999% durability, and extensive integration with AWS services. It provides different storage tiers (like Standard, Infrequent Access, and Glacier) to optimize costs based on access patterns.
Hitachi Content Platform (HCP) is an enterprise-grade object storage system that focuses on data governance, compliance, and security. It offers advanced features like data classification, retention policies, and WORM (Write Once, Read Many) capabilities. HCP can be deployed on-premises or in hybrid cloud configurations and supports multiple protocols including S3 compatibility.

Workshops
The Virtual File System (VFS) is an abstraction layer that provides a unified interface for accessing different types of file systems and file storage. It creates a consistent programming interface that hides the specific details of the underlying storage mechanisms.
VFS allows applications to access files across various storage types—local disks, network locations, cloud storage, archives, FTP servers, SFTP sites, HTTP resources, and more—using a single, consistent API. This eliminates the need to implement separate code for each storage type.
Key benefits include location transparency (uniform access regardless of physical location), protocol independence (same operations across different protocols), and enhanced functionality (metadata access, caching, security controls). VFS implementations are common in operating systems (Linux VFS), programming frameworks (Apache Commons VFS), and data processing tools (Pentaho Data Integration's VFS support).
In tools like Pentaho, VFS enables seamless reading and writing of data across diverse storage systems using a standardized URI-based path notation, significantly simplifying data integration workflows involving multiple storage technologies.
Last updated
Was this helpful?