Alluxio Confidential
AI 3.7 Launch
Universal S3 + POSIX cache for end-to-end AI workloads.
AI at Full Throughput
and Ultra Low Latency
Jingwen Ouyang
Senior Product Manager
Alluxio Confidential
AI Data Life Cycle
Data
Collection
Data
Preprocessing
Model
Training
Model
Verification
Model
Loading
Inference
Data
Archiving
Data is everywhere in every stage of the journey.
Data needs to be accessed fast, friction free, and low cost.
Alluxio Confidential
Alluxio makes it easy to share and
manage data from
any storage
to any compute engine
in any environment
with high performance and low cost.
Alluxio Confidential
Alluxio as the Universal Cache
for S3 and POSIX Workloads
AI 3.7 Highlight
Alluxio Confidential
Alluxio as the Universal Cache for S3 and POSIX Workloads
THROUGHPUT LATENCY
Simplified example: application throughput =100 MB/s, T_setup=200 ms
● Large dataset: 1 GB → ~10.2 s (throughput dominates)
● Small dataset: 128 KB → ~0.201 s (latency dominates).
Training & rollout love throughput; inference loves latency!
T_total ≈ T_setup (≈ TTFB) + (data_size / application throughput)
Alluxio Confidential
THROUGHPUT
Alluxio has always been a leader in high throughput.
Enables customers to rapidly load massive quantities
of data into GPU memory for AI training and model
deployment/cold starts.
NEW in AI 3.7… Alluxio also delivers
Ultra Low Latency Caching for data
stored on cloud storage (e.g. AWS S3).
LATENCY
Alluxio as the Universal Cache for S3 and POSIX Workloads
Alluxio Confidential
Alluxio as the Universal Cache for S3 and POSIX Workloads
● A Single Alluxio Worker achieved a high throughput comparing to HPC Storage Solutions:
○ Up to 81.6 Gbps (or 9.5 GiB/s) w/ 100 Gbps network - 2.5 GiB/s(@1 thread) to 9.5 GiB/s(@32 threads)
○ Up to 352.2 Gbps (or 41 GiB/s) w/ 400 Gbps network
Setup
● Alluxio:
1 Alluxio worker (i3en.metal)
● FIO Benchmark:
Sequential Read
bs = 256KB
Note: an Alluxio fuse client (c5n.metal) co-located
with training servers is responsible for POSIX API
access to Alluxio Workers which actually cache the
data
Throughput Microbenchmark: Reads from A Single Worker
Alluxio Confidential
Alluxio is the industry-leading
sub-ms time to first byte (TTFB) solution on S3-class storage
How much better is Alluxio? (Details next slide)
➔ 45x Lower Latency than S3 Standard
➔ 5x Lower Latency than S3 Express One Zone
➔ Unlimited, linear scalability
Alluxio as the Universal Cache for S3 and POSIX Workloads
Alluxio Confidential
Test environment references
Alluxio EE
● Version/Spec: Alluxio Enterprise AI 3.6 (50TB
cache)
● Test env: 1 FUSE (C5n.metal, 100Gbps
network) and 1 Worker (i3en.metal)
AWS S3
● Version/Spec: AWS S3 bucket (Standard Class)
● Test env: 1 FUSE (C5n.metal, 100Gbps
network)
AWS S3 Express One Zone
● Version/Spec: AWS bucket (S3 Express One
Zone Class)
● Test env: 1 FUSE (C5n.metal, 100Gbps
network)
Alluxio as the Universal Cache for S3 and POSIX Workloads
➔ 45x Lower Latency than S3 Standard
➔ 5x Lower Latency than S3 Express One Zone
Alluxio Confidential
But that’s not all…
➔ Alluxio is 100% transparent to AI workloads
➔ S3 API & POSIX API
◆ Broad application support PyTorch, Python, AWS SDK, Boto3 …
◆ NO code changes
◆ NO workflow changes
➔ NO data imports or migrations
Drop in Alluxio as the transparent S3 caching layer
for faster and more scalable AI!
Alluxio as the Universal Cache for S3 and POSIX Workloads
Alluxio Confidential
What the Public Results Show — MLPerf Training v2.0
Alluxio Achieves Exceptional GPU Utilization
Alluxio
DDN
Nutanix
Hammerspace
HPE
Source: MLCommons MLPerf Training v2.0 (retrieved 08/12/2025).
MLPerf®/MLCommons® are trademarks of MLCommons.
Alluxio Confidential
What the Public Results Show — MLPerf Training v2.0
Linear scale-out: clients↑ workers↑ throughput↑
Source: MLCommons MLPerf Training v2.0 (retrieved 08/12/2025).
MLPerf®/MLCommons® are trademarks of MLCommons.
Alluxio Confidential
AI 3.7 Key Features: Operations & Management
Deploy & Configure Alluxio in WebUI
After installing the Alluxio K8s Operator, the intuitive WebUI can be used to configure cluster
parameters, allocate resources, and customize deployments, making Alluxio deployment faster,
simpler, and more accurate than ever before.
Alluxio Confidential
AI 3.7 Key Features: Operations & Management
FUSE Non-Disruptive Upgrade: Rolling, K8s-native FUSE upgrades; mounts stay live.
Alluxio's innovative FUSE Online Upgrades feature allows admins to upgrade FUSE services
without interrupting active AI workloads
Alluxio's FUSE Online Upgrade capability maintains data accessibility throughout the upgrade
process by:
● Preserving active file handles and connections
● Queuing operations during the brief transition period
● Automatically resuming operations within tens of seconds
In this release, read operations (read, stat) are fully retained, while write operations will be
supported in future updates. This ensures your critical AI workloads keep running even during
necessary system maintenance.
Alluxio Confidential
AI 3.7 Key Features: Security & Compliance
Role Based Access Controls for S3 Data
● Define who can access which data
● Define what they can do with that data
● Integrations with Authentication &
Authorization providers
○ OIDC based Providers
(Okta, Cognito, Microsoft AD)
○ Apache Ranger, Open Policy Agent (OPA)
Audit & Analyze User Data Access & Operations
Now, every interaction is automatically recorded
with:
● User identities and authentication details
● Operations performed (read, write, delete, etc.)
● Precise timestamps
● Accessed resources and paths
Enables security teams to detect anomalies,
investigate incidents, and demonstrate compliance
with regulatory requirements, which are essential
for enterprise AI deployments handling sensitive
data.
Alluxio Confidential
Demo
Accelerate AI.
Alluxio Confidential
Demo: Transparent S3 Cache
Link 4:23
This demo showcases how to seamlessly accelerate a PyTorch data loading script using the Alluxio S3 API without any code modifications.
The presenter first runs a Python script that uses the s3torchconnectorlibrary to read data directly from an S3 bucket, establishing a
baseline performance of about 25 seconds. Then, by simply setting the S3_ENDPOINT_URLenvironment variable to point to the Alluxio S3
API endpoint, the exact same script is run again. This time, Alluxio serves the data from its cache, reducing the read time to about 15
seconds. The performance improvement is visually confirmed using a Grafana dashboard, which shows the data is now cached in Alluxio
and the cache hit rate is 100%.
● Objective (0:01): Accelerate PyTorch data pipelines that use S3 by leveraging the Alluxio S3 API, without requiring any changes to
the existing application code.
● The Mechanism (0:26): The integration works by redirecting S3 traffic to the Alluxio cluster. This is achieved by setting the
S3_ENDPOINT_URLenvironment variable to the Alluxio S3 endpoint address.
● Baseline Performance - Cold Read (1:03): The first run of the script with the endpoint URL unset reads data directly from AWS S3,
establishing a "cold read" benchmark of approximately 25 seconds.
● Accelerated Performance - Hot Read (1:52): For the second run, the S3_ENDPOINT_URLis set to the Alluxio endpoint. The exact
same script now reads the data from the Alluxio cache in about 15 seconds, demonstrating a significant performance improvement.
● Verification with Metrics (3:25): The performance gain is visually confirmed using a Grafana dashboard, which shows that 3 GiB of
data is now cached in Alluxio and the cache hit rate for the second run is 100%.
Alluxio Confidential
Demo: WebUI - Cluster Deployment
link 4:17
This demo introduces the new features of the Alluxio Management Console in version 3.7. The presenter walks through how to access
the console and use its new graphical user interface to deploy and manage Alluxio clusters. He demonstrates the two methods for
cluster creation: a manual, form-based configuration and a streamlined YAML upload option. The demo highlights how the console
provides a detailed, real-time status view of cluster resources, including workloads and individual pods, and shows how to easily
access pod logs for quick troubleshooting.
● Accessing the Console (0:11): The Alluxio Operator now deploys a dedicated alluxio-consolepod and service, which can
be accessed locally by port-forwarding its service to your machine.
● Cluster Creation (0:55): The console offers two ways to deploy a new Alluxio cluster: a step-by-step "Manual Configuration"
form or by simply uploading an existing YAML configuration file.
● Live Status Monitoring (2:56): The "View Status" page provides a real-time dashboard showing the creation progress and
health of all cluster workloads and their associated pods.
● Simplified Troubleshooting (3:29): Users can now view the logs of any pod directly from the console's UI with a single click,
making it easy to diagnose issues like a pod in an error state.
● Integrated Cluster Management (3:52): For existing clusters, the "View Console" button provides a direct link to that specific
cluster's traditional Alluxio UI, allowing for more detailed management operations like preloading data and managing storage.
Alluxio Confidential
Thanks.
Schedule Demo at alluxio.io/demo

More Related Content

PDF
AI/ML Infra Meetup | Beyond S3's Basics: Architecting for AI-Native Data Access
PDF
Alluxio Webinar | What’s New in Alluxio AI: 3X Faster Checkpoint File Creatio...
PDF
Alluxio Webinar | Optimize, Don't Overspend: Data Caching Strategy for AI Wor...
PDF
Meet You in the Middle: 1000x Performance for Parquet Queries on PB-Scale Dat...
PDF
Deploying Alluxio in the Cloud for Machine Learning
PDF
What’s new in Alluxio 2: from seamless operations to structured data management
PDF
AI/ML Infra Meetup | Maximizing GPU Efficiency : Optimizing Model Training wi...
PDF
StorageQuery: federated querying on object stores, powered by Alluxio and Presto
AI/ML Infra Meetup | Beyond S3's Basics: Architecting for AI-Native Data Access
Alluxio Webinar | What’s New in Alluxio AI: 3X Faster Checkpoint File Creatio...
Alluxio Webinar | Optimize, Don't Overspend: Data Caching Strategy for AI Wor...
Meet You in the Middle: 1000x Performance for Parquet Queries on PB-Scale Dat...
Deploying Alluxio in the Cloud for Machine Learning
What’s new in Alluxio 2: from seamless operations to structured data management
AI/ML Infra Meetup | Maximizing GPU Efficiency : Optimizing Model Training wi...
StorageQuery: federated querying on object stores, powered by Alluxio and Presto

Similar to Product Update: Alluxio AI 3.7 Now with Sub-Millisecond Latency (20)

PDF
AI Infra Day | Hands-on Lab: CV Model Training with PyTorch & Alluxio on Kube...
PDF
Accelerating Spark with Kubernetes
PDF
Building Cloud Native Analytical Pipelines on AWS
PDF
Alluxio Webinar | Accelerate AI: Alluxio 101
PDF
Accelerating Cloud Training With Alluxio
PPTX
Ceph Day San Jose - Enable Fast Big Data Analytics on Ceph with Alluxio
PDF
Enable Fast Big Data Analytics on Ceph with Alluxio at Ceph Days 2017
PDF
Alluxio 2.0 & Near Real-time Big Data Platform w/ Spark & Alluxio
PDF
Iceberg + Alluxio for Fast Data Analytics
PDF
Deep Learning and Gene Computing Acceleration with Alluxio in Kubernetes
PDF
How to get started with Oracle Cloud Infrastructure
PPTX
OS for AI: Elastic Microservices & the Next Gen of ML
PPTX
ZStack architecture overview
PDF
Alluxio Innovations for Structured Data
PDF
TIBCO BW6 and MuleSoft Anypoint Platform
PDF
Speeding up I/O for Machine Learning ft Apple Case Study using TensorFlow, N...
PDF
Introduction to Python Asyncio
PDF
Alluxio Webinar | Model Training Across Regions and Clouds – Challenges, Solu...
PDF
Alluxio 2.9 Release Overview
PDF
Building Fast SQL Analytics on Anything with Presto, Alluxio
AI Infra Day | Hands-on Lab: CV Model Training with PyTorch & Alluxio on Kube...
Accelerating Spark with Kubernetes
Building Cloud Native Analytical Pipelines on AWS
Alluxio Webinar | Accelerate AI: Alluxio 101
Accelerating Cloud Training With Alluxio
Ceph Day San Jose - Enable Fast Big Data Analytics on Ceph with Alluxio
Enable Fast Big Data Analytics on Ceph with Alluxio at Ceph Days 2017
Alluxio 2.0 & Near Real-time Big Data Platform w/ Spark & Alluxio
Iceberg + Alluxio for Fast Data Analytics
Deep Learning and Gene Computing Acceleration with Alluxio in Kubernetes
How to get started with Oracle Cloud Infrastructure
OS for AI: Elastic Microservices & the Next Gen of ML
ZStack architecture overview
Alluxio Innovations for Structured Data
TIBCO BW6 and MuleSoft Anypoint Platform
Speeding up I/O for Machine Learning ft Apple Case Study using TensorFlow, N...
Introduction to Python Asyncio
Alluxio Webinar | Model Training Across Regions and Clouds – Challenges, Solu...
Alluxio 2.9 Release Overview
Building Fast SQL Analytics on Anything with Presto, Alluxio
Ad

More from Alluxio, Inc. (20)

PDF
AI/ML Infra Meetup | LLM Agents and Implementation Challenges
PDF
Introduction to Apache Iceberg™ & Tableflow
PDF
Optimizing Tiered Storage for Low-Latency Real-Time Analytics at AI Scale
PDF
Meet in the Middle: Solving the Low-Latency Challenge for Agentic AI
PDF
From Data Preparation to Inference: How Alluxio Speeds Up AI
PDF
Best Practice for LLM Serving in the Cloud
PDF
How Coupang Leverages Distributed Cache to Accelerate ML Model Training
PDF
Alluxio Webinar | Inside Deepseek 3FS: A Deep Dive into AI-Optimized Distribu...
PDF
AI/ML Infra Meetup | Building Production Platform for Large-Scale Recommendat...
PDF
AI/ML Infra Meetup | How Uber Optimizes LLM Training and Finetune
PDF
AI/ML Infra Meetup | Optimizing ML Data Access with Alluxio: Preprocessing, ...
PDF
AI/ML Infra Meetup | Deployment, Discovery and Serving of LLMs at Uber Scale
PDF
AI/ML Infra Meetup | A Faster and More Cost Efficient LLM Inference Stack
PDF
AI/ML Infra Meetup | Balancing Cost, Performance, and Scale - Running GPU/CPU...
PDF
AI/ML Infra Meetup | RAYvolution - The Last Mile: Mastering AI Deployment wit...
PDF
AI/ML Infra Meetup | The power of Ray in the era of LLM and multi-modality AI
PDF
AI/ML Infra Meetup | Exploring Distributed Caching for Faster GPU Training wi...
PDF
AI/ML Infra Meetup | Big Data and AI, Zoom Developers
PDF
AI/ML Infra Meetup | TorchTitan, One-stop PyTorch native solution for product...
PDF
AI/ML Infra Meetup | Scaling Experimentation Platform in Digital Marketplaces...
AI/ML Infra Meetup | LLM Agents and Implementation Challenges
Introduction to Apache Iceberg™ & Tableflow
Optimizing Tiered Storage for Low-Latency Real-Time Analytics at AI Scale
Meet in the Middle: Solving the Low-Latency Challenge for Agentic AI
From Data Preparation to Inference: How Alluxio Speeds Up AI
Best Practice for LLM Serving in the Cloud
How Coupang Leverages Distributed Cache to Accelerate ML Model Training
Alluxio Webinar | Inside Deepseek 3FS: A Deep Dive into AI-Optimized Distribu...
AI/ML Infra Meetup | Building Production Platform for Large-Scale Recommendat...
AI/ML Infra Meetup | How Uber Optimizes LLM Training and Finetune
AI/ML Infra Meetup | Optimizing ML Data Access with Alluxio: Preprocessing, ...
AI/ML Infra Meetup | Deployment, Discovery and Serving of LLMs at Uber Scale
AI/ML Infra Meetup | A Faster and More Cost Efficient LLM Inference Stack
AI/ML Infra Meetup | Balancing Cost, Performance, and Scale - Running GPU/CPU...
AI/ML Infra Meetup | RAYvolution - The Last Mile: Mastering AI Deployment wit...
AI/ML Infra Meetup | The power of Ray in the era of LLM and multi-modality AI
AI/ML Infra Meetup | Exploring Distributed Caching for Faster GPU Training wi...
AI/ML Infra Meetup | Big Data and AI, Zoom Developers
AI/ML Infra Meetup | TorchTitan, One-stop PyTorch native solution for product...
AI/ML Infra Meetup | Scaling Experimentation Platform in Digital Marketplaces...
Ad

Recently uploaded (20)

PDF
Top 10 Project Management Software for Small Teams in 2025.pdf
PDF
Sanket Mhaiskar Resume - Senior Software Engineer (Backend, AI)
PPTX
Odoo ERP for Injection Molding Industry – Optimize Production & Reduce Scrap
PPTX
Bandicam Screen Recorder 8.2.1 Build 2529 Crack
PDF
Practical Indispensable Project Management Tips for Delivering Successful Exp...
PPTX
Human-Computer Interaction for Lecture 2
PDF
Cloud Native Aachen Meetup - Aug 21, 2025
PPT
3.Software Design for software engineering
PDF
MiniTool Power Data Recovery 12.6 Crack + Portable (Latest Version 2025)
PDF
What Makes a Great Data Visualization Consulting Service.pdf
PDF
IDM Crack 6.42 Build 42 Patch Serial Key 2025 Free New Version
PPTX
ERP Manufacturing Modules & Consulting Solutions : Contetra Pvt Ltd
PPTX
ROI from Efficient Content & Campaign Management in the Digital Media Industry
PDF
Building an Inclusive Web Accessibility Made Simple with Accessibility Analyzer
PDF
Lumion Pro Crack New latest version Download 2025
PDF
Website Design & Development_ Professional Web Design Services.pdf
PPTX
HackYourBrain__UtrechtJUG__11092025.pptx
PPTX
ROI Analysis for Newspaper Industry with Odoo ERP
PPTX
Lecture 5 Software Requirement Engineering
PDF
Workplace Software and Skills - OpenStax
Top 10 Project Management Software for Small Teams in 2025.pdf
Sanket Mhaiskar Resume - Senior Software Engineer (Backend, AI)
Odoo ERP for Injection Molding Industry – Optimize Production & Reduce Scrap
Bandicam Screen Recorder 8.2.1 Build 2529 Crack
Practical Indispensable Project Management Tips for Delivering Successful Exp...
Human-Computer Interaction for Lecture 2
Cloud Native Aachen Meetup - Aug 21, 2025
3.Software Design for software engineering
MiniTool Power Data Recovery 12.6 Crack + Portable (Latest Version 2025)
What Makes a Great Data Visualization Consulting Service.pdf
IDM Crack 6.42 Build 42 Patch Serial Key 2025 Free New Version
ERP Manufacturing Modules & Consulting Solutions : Contetra Pvt Ltd
ROI from Efficient Content & Campaign Management in the Digital Media Industry
Building an Inclusive Web Accessibility Made Simple with Accessibility Analyzer
Lumion Pro Crack New latest version Download 2025
Website Design & Development_ Professional Web Design Services.pdf
HackYourBrain__UtrechtJUG__11092025.pptx
ROI Analysis for Newspaper Industry with Odoo ERP
Lecture 5 Software Requirement Engineering
Workplace Software and Skills - OpenStax

Product Update: Alluxio AI 3.7 Now with Sub-Millisecond Latency

  • 1. Alluxio Confidential AI 3.7 Launch Universal S3 + POSIX cache for end-to-end AI workloads. AI at Full Throughput and Ultra Low Latency Jingwen Ouyang Senior Product Manager
  • 2. Alluxio Confidential AI Data Life Cycle Data Collection Data Preprocessing Model Training Model Verification Model Loading Inference Data Archiving Data is everywhere in every stage of the journey. Data needs to be accessed fast, friction free, and low cost.
  • 3. Alluxio Confidential Alluxio makes it easy to share and manage data from any storage to any compute engine in any environment with high performance and low cost.
  • 4. Alluxio Confidential Alluxio as the Universal Cache for S3 and POSIX Workloads AI 3.7 Highlight
  • 5. Alluxio Confidential Alluxio as the Universal Cache for S3 and POSIX Workloads THROUGHPUT LATENCY Simplified example: application throughput =100 MB/s, T_setup=200 ms ● Large dataset: 1 GB → ~10.2 s (throughput dominates) ● Small dataset: 128 KB → ~0.201 s (latency dominates). Training & rollout love throughput; inference loves latency! T_total ≈ T_setup (≈ TTFB) + (data_size / application throughput)
  • 6. Alluxio Confidential THROUGHPUT Alluxio has always been a leader in high throughput. Enables customers to rapidly load massive quantities of data into GPU memory for AI training and model deployment/cold starts. NEW in AI 3.7… Alluxio also delivers Ultra Low Latency Caching for data stored on cloud storage (e.g. AWS S3). LATENCY Alluxio as the Universal Cache for S3 and POSIX Workloads
  • 7. Alluxio Confidential Alluxio as the Universal Cache for S3 and POSIX Workloads ● A Single Alluxio Worker achieved a high throughput comparing to HPC Storage Solutions: ○ Up to 81.6 Gbps (or 9.5 GiB/s) w/ 100 Gbps network - 2.5 GiB/s(@1 thread) to 9.5 GiB/s(@32 threads) ○ Up to 352.2 Gbps (or 41 GiB/s) w/ 400 Gbps network Setup ● Alluxio: 1 Alluxio worker (i3en.metal) ● FIO Benchmark: Sequential Read bs = 256KB Note: an Alluxio fuse client (c5n.metal) co-located with training servers is responsible for POSIX API access to Alluxio Workers which actually cache the data Throughput Microbenchmark: Reads from A Single Worker
  • 8. Alluxio Confidential Alluxio is the industry-leading sub-ms time to first byte (TTFB) solution on S3-class storage How much better is Alluxio? (Details next slide) ➔ 45x Lower Latency than S3 Standard ➔ 5x Lower Latency than S3 Express One Zone ➔ Unlimited, linear scalability Alluxio as the Universal Cache for S3 and POSIX Workloads
  • 9. Alluxio Confidential Test environment references Alluxio EE ● Version/Spec: Alluxio Enterprise AI 3.6 (50TB cache) ● Test env: 1 FUSE (C5n.metal, 100Gbps network) and 1 Worker (i3en.metal) AWS S3 ● Version/Spec: AWS S3 bucket (Standard Class) ● Test env: 1 FUSE (C5n.metal, 100Gbps network) AWS S3 Express One Zone ● Version/Spec: AWS bucket (S3 Express One Zone Class) ● Test env: 1 FUSE (C5n.metal, 100Gbps network) Alluxio as the Universal Cache for S3 and POSIX Workloads ➔ 45x Lower Latency than S3 Standard ➔ 5x Lower Latency than S3 Express One Zone
  • 10. Alluxio Confidential But that’s not all… ➔ Alluxio is 100% transparent to AI workloads ➔ S3 API & POSIX API ◆ Broad application support PyTorch, Python, AWS SDK, Boto3 … ◆ NO code changes ◆ NO workflow changes ➔ NO data imports or migrations Drop in Alluxio as the transparent S3 caching layer for faster and more scalable AI! Alluxio as the Universal Cache for S3 and POSIX Workloads
  • 11. Alluxio Confidential What the Public Results Show — MLPerf Training v2.0 Alluxio Achieves Exceptional GPU Utilization Alluxio DDN Nutanix Hammerspace HPE Source: MLCommons MLPerf Training v2.0 (retrieved 08/12/2025). MLPerf®/MLCommons® are trademarks of MLCommons.
  • 12. Alluxio Confidential What the Public Results Show — MLPerf Training v2.0 Linear scale-out: clients↑ workers↑ throughput↑ Source: MLCommons MLPerf Training v2.0 (retrieved 08/12/2025). MLPerf®/MLCommons® are trademarks of MLCommons.
  • 13. Alluxio Confidential AI 3.7 Key Features: Operations & Management Deploy & Configure Alluxio in WebUI After installing the Alluxio K8s Operator, the intuitive WebUI can be used to configure cluster parameters, allocate resources, and customize deployments, making Alluxio deployment faster, simpler, and more accurate than ever before.
  • 14. Alluxio Confidential AI 3.7 Key Features: Operations & Management FUSE Non-Disruptive Upgrade: Rolling, K8s-native FUSE upgrades; mounts stay live. Alluxio's innovative FUSE Online Upgrades feature allows admins to upgrade FUSE services without interrupting active AI workloads Alluxio's FUSE Online Upgrade capability maintains data accessibility throughout the upgrade process by: ● Preserving active file handles and connections ● Queuing operations during the brief transition period ● Automatically resuming operations within tens of seconds In this release, read operations (read, stat) are fully retained, while write operations will be supported in future updates. This ensures your critical AI workloads keep running even during necessary system maintenance.
  • 15. Alluxio Confidential AI 3.7 Key Features: Security & Compliance Role Based Access Controls for S3 Data ● Define who can access which data ● Define what they can do with that data ● Integrations with Authentication & Authorization providers ○ OIDC based Providers (Okta, Cognito, Microsoft AD) ○ Apache Ranger, Open Policy Agent (OPA) Audit & Analyze User Data Access & Operations Now, every interaction is automatically recorded with: ● User identities and authentication details ● Operations performed (read, write, delete, etc.) ● Precise timestamps ● Accessed resources and paths Enables security teams to detect anomalies, investigate incidents, and demonstrate compliance with regulatory requirements, which are essential for enterprise AI deployments handling sensitive data.
  • 17. Alluxio Confidential Demo: Transparent S3 Cache Link 4:23 This demo showcases how to seamlessly accelerate a PyTorch data loading script using the Alluxio S3 API without any code modifications. The presenter first runs a Python script that uses the s3torchconnectorlibrary to read data directly from an S3 bucket, establishing a baseline performance of about 25 seconds. Then, by simply setting the S3_ENDPOINT_URLenvironment variable to point to the Alluxio S3 API endpoint, the exact same script is run again. This time, Alluxio serves the data from its cache, reducing the read time to about 15 seconds. The performance improvement is visually confirmed using a Grafana dashboard, which shows the data is now cached in Alluxio and the cache hit rate is 100%. ● Objective (0:01): Accelerate PyTorch data pipelines that use S3 by leveraging the Alluxio S3 API, without requiring any changes to the existing application code. ● The Mechanism (0:26): The integration works by redirecting S3 traffic to the Alluxio cluster. This is achieved by setting the S3_ENDPOINT_URLenvironment variable to the Alluxio S3 endpoint address. ● Baseline Performance - Cold Read (1:03): The first run of the script with the endpoint URL unset reads data directly from AWS S3, establishing a "cold read" benchmark of approximately 25 seconds. ● Accelerated Performance - Hot Read (1:52): For the second run, the S3_ENDPOINT_URLis set to the Alluxio endpoint. The exact same script now reads the data from the Alluxio cache in about 15 seconds, demonstrating a significant performance improvement. ● Verification with Metrics (3:25): The performance gain is visually confirmed using a Grafana dashboard, which shows that 3 GiB of data is now cached in Alluxio and the cache hit rate for the second run is 100%.
  • 18. Alluxio Confidential Demo: WebUI - Cluster Deployment link 4:17 This demo introduces the new features of the Alluxio Management Console in version 3.7. The presenter walks through how to access the console and use its new graphical user interface to deploy and manage Alluxio clusters. He demonstrates the two methods for cluster creation: a manual, form-based configuration and a streamlined YAML upload option. The demo highlights how the console provides a detailed, real-time status view of cluster resources, including workloads and individual pods, and shows how to easily access pod logs for quick troubleshooting. ● Accessing the Console (0:11): The Alluxio Operator now deploys a dedicated alluxio-consolepod and service, which can be accessed locally by port-forwarding its service to your machine. ● Cluster Creation (0:55): The console offers two ways to deploy a new Alluxio cluster: a step-by-step "Manual Configuration" form or by simply uploading an existing YAML configuration file. ● Live Status Monitoring (2:56): The "View Status" page provides a real-time dashboard showing the creation progress and health of all cluster workloads and their associated pods. ● Simplified Troubleshooting (3:29): Users can now view the logs of any pod directly from the console's UI with a single click, making it easy to diagnose issues like a pod in an error state. ● Integrated Cluster Management (3:52): For existing clusters, the "View Console" button provides a direct link to that specific cluster's traditional Alluxio UI, allowing for more detailed management operations like preloading data and managing storage.