2
Most read
5
Most read
6
Most read
Ceph
Object Storage at Spreadshirt
June 2015
Jens Hadlich
Chief Architect
About Spreadshirt
2
Spread it with Spreadshirt
A global e-commerce platform for everyone to create, sell and buy
ideas on clothing and accessories across many points of sale.
•  12 languages, 11 currencies
•  19 markets
•  150+ shipping regions
•  community of >70.000 active sellers
•  € 72M revenue (2014)
•  >3.3M items shipped (2014)
Object Storage at Spreadshirt
•  What?
–  Store and read primarily user generated content, mostly images
•  Typical sizes:
–  a few dozen KB, a few MB
•  Some 10s of terabyte (TB) of data
•  Read > Write
•  „Never change a running system“?
–  Currently solution from the early days with big storage + lots of files /
directories doesn‘t work anymore
•  Regular UNIX tools get unusable in practice
•  Not designed for „the cloud“ (e.g. replication is an issue)
–  Growing number of users à more content
–  Build a truly global platform (multiple regions and data centers)
3
Ceph
•  Why Ceph?
–  Vendor independent
–  Open source
–  Runs on commodity hardware
–  Local installation for minimal latency
–  Existing knowledge and experience
–  S3-API
•  Simple bucket-to-bucket replication
–  A good fit also for < Petabyte
–  Easy to add more storage
–  (Can be used later for block storage)
4
Ceph Object Storage Architecture
5
Overview
Ceph Object Gateway
Monitor
Cluster Network
Public Network
OSDOSD OSDOSDOSD
MonitorMonitor
A lot of
nodes and
disks
Client
HTTP (S3 or SWIFT API)
RADOS
(reliable autonomic distributed object store)
Ceph Object Storage Architecture
6
A little more detailled
Monitor
Cluster Network
Public Network
Client
RadosGW
HTTP (S3 or SWIFT API)
MonitorMonitor
Some SSDs
(for journals)
More HDDs
JBOD
(no RAID)
OSD node
Ceph Object
Gateway
librados
Odd number
(Quorum)
OSD node OSD node OSD node OSD node
1G
10G
(the more the
better)
...
RADOS
(reliable autonomic distributed object store)
OSD node
Ceph Object Storage Architecture
7
Initial Setup (planned)
Cluster Network (OSD Replication)
Cluster nodes
3 x SSD
(journal / index)
9 x HDD (data)
3 Monitors
2 x 1G, IPv4
2 x 10G, IPv6
Public Network
Client
HTTP (S3 or SWIFT API)
HAProxy
RadosGW
Monitor
RadosGW
Monitor
RadosGW
Monitor
RadosGWRadosGW
2 x 10G, IPv6Cluster Network
RadosGW
on each node
Ceph Object Storage Performance
8
Some smoke tests
•  How fast is RadosGW? Get an impression.
–  Response times (read / write)
•  Average?
•  Percentiles (P99)?
–  Compared to AWS S3?
•  A very minimalistic test setup
–  3 VMs (KVM) all with RadosGW, Monitor and OSD
•  2 Cores, 4GB RAM, 1 OSD each (15 GB + 5GB), 10G Network
between nodes, HAProxy (round-robin), LAN, HTTP
–  No further optimizations
Ceph Object Storage Performance
9
Some smoke tests
•  How fast is RadosGW?
–  Random read and write
–  Object size: 4 KB
•  Results: Pretty promising!
–  E.g. 16 parallel threads, read:
•  Avg 9 ms
•  P99 49 ms
•  > 1.300 requests/s
Ceph Object Storage Performance
10
Some smoke tests
•  Compared to Amazon S3?
–  Comparing apples and oranges (unfair, but interresting)
•  http vs. https, LAN vs. WAN etc.
•  Reponse times
–  Random read, object size: 4KB, 4 parallel threads, location: Leipzig
Ceph S3 AWS S3
eu-central-1 eu-west-1
Location Leipzig Frankfurt Ireland
Avg 6 ms 25 ms 56 ms
P99 47 ms 128 ms 374 ms
Requests/s 405 143 62
Global Availability
11
•  1 Ceph cluster per data center
•  S3 bucket-to-bucket replication
•  Multiple regions, local delivery
To be continued ...
+ = ?
Thank You!
jns@spreadshirt.com

More Related Content

PPTX
Introduction to Azure monitor
PDF
BlueStore: a new, faster storage backend for Ceph
PDF
PromQL Deep Dive - The Prometheus Query Language
PDF
Ceph scale testing with 10 Billion Objects
PDF
Airflow Best Practises & Roadmap to Airflow 2.0
PDF
Ceph Object Storage at Spreadshirt (July 2015, Ceph Berlin Meetup)
PDF
Blockchain and IOT and the GxP Lab Slides
Introduction to Azure monitor
BlueStore: a new, faster storage backend for Ceph
PromQL Deep Dive - The Prometheus Query Language
Ceph scale testing with 10 Billion Objects
Airflow Best Practises & Roadmap to Airflow 2.0
Ceph Object Storage at Spreadshirt (July 2015, Ceph Berlin Meetup)
Blockchain and IOT and the GxP Lab Slides

What's hot (20)

PPTX
Los beneficios de migrar sus cargas de trabajo de big data a AWS
PPTX
Presto query optimizer: pursuit of performance
PDF
Batch and Stream Graph Processing with Apache Flink
PDF
Prometheus - basics
PDF
대용량 로그분석 Bigquery로 간단히 사용하기 20160930
PDF
Mastering GC.pdf
PDF
Observability for Data Pipelines With OpenLineage
PDF
cLoki: Like Loki but for ClickHouse
PPTX
What is Object storage ?
PDF
2019.06.27 Intro to Ceph
PDF
Near Real-Time Netflix Recommendations Using Apache Spark Streaming with Nit...
PPTX
CQRS and Event Sourcing with Axon Framework
PDF
[db tech showcase Tokyo 2016] D13: NVMeフラッシュストレージを用いた高性能高拡張高可用なデータベースシステムの実現方...
PDF
Cassandra serving netflix @ scale
PPTX
Google Cloud Dataproc - Easier, faster, more cost-effective Spark and Hadoop
PDF
A crash course in CRUSH
PDF
VictoriaLogs: Open Source Log Management System - Preview
PDF
Ceph as software define storage
PDF
Building a Logical Data Fabric using Data Virtualization (ASEAN)
PDF
Log aggregation: using Elasticsearch, Fluentd/Fluentbit and Kibana (EFK)
Los beneficios de migrar sus cargas de trabajo de big data a AWS
Presto query optimizer: pursuit of performance
Batch and Stream Graph Processing with Apache Flink
Prometheus - basics
대용량 로그분석 Bigquery로 간단히 사용하기 20160930
Mastering GC.pdf
Observability for Data Pipelines With OpenLineage
cLoki: Like Loki but for ClickHouse
What is Object storage ?
2019.06.27 Intro to Ceph
Near Real-Time Netflix Recommendations Using Apache Spark Streaming with Nit...
CQRS and Event Sourcing with Axon Framework
[db tech showcase Tokyo 2016] D13: NVMeフラッシュストレージを用いた高性能高拡張高可用なデータベースシステムの実現方...
Cassandra serving netflix @ scale
Google Cloud Dataproc - Easier, faster, more cost-effective Spark and Hadoop
A crash course in CRUSH
VictoriaLogs: Open Source Log Management System - Preview
Ceph as software define storage
Building a Logical Data Fabric using Data Virtualization (ASEAN)
Log aggregation: using Elasticsearch, Fluentd/Fluentbit and Kibana (EFK)
Ad

Viewers also liked (11)

PDF
Ceph at Spreadshirt (June 2016)
PDF
optimizing_ceph_flash
PPTX
Owncloud
PPTX
Implementation of Dense Storage Utilizing HDDs with SSDs and PCIe Flash Acc...
PDF
Spreadshirt Platform - An Architectural Overview
PPTX
Ceph at Work in Bloomberg: Object Store, RBD and OpenStack
PPTX
What you need to know about ceph
PDF
Data Structures in and on IPFS
PDF
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...
PPTX
Ceph Introduction 2017
PPTX
Ceph Intro and Architectural Overview by Ross Turk
Ceph at Spreadshirt (June 2016)
optimizing_ceph_flash
Owncloud
Implementation of Dense Storage Utilizing HDDs with SSDs and PCIe Flash Acc...
Spreadshirt Platform - An Architectural Overview
Ceph at Work in Bloomberg: Object Store, RBD and OpenStack
What you need to know about ceph
Data Structures in and on IPFS
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...
Ceph Introduction 2017
Ceph Intro and Architectural Overview by Ross Turk
Ad

Similar to Ceph Object Storage at Spreadshirt (20)

PDF
The Future of Cloud Software Defined Storage with Ceph: Andrew Hatfield, Red Hat
PDF
OSDC 2015: John Spray | The Ceph Storage System
PDF
Ceph - A distributed storage system
PDF
SUSE Storage: Sizing and Performance (Ceph)
PDF
2015 open storage workshop ceph software defined storage
PDF
Red Hat Storage 2014 - Product(s) Overview
PDF
New use cases for Ceph, beyond OpenStack, Luis Rico
PDF
Quick-and-Easy Deployment of a Ceph Storage Cluster
PDF
Ceph Overview for Distributed Computing Denver Meetup
PDF
Open Source Storage at Scale: Ceph @ GRNET
PDF
adp.ceph.openstack.talk
PDF
Ceph openstack-jun-2015-meetup
PDF
Ceph & OpenStack talk given @ OpenStack Meetup @ Bangalore, June 2015
PDF
Ceph, Now and Later: Our Plan for Open Unified Cloud Storage
PDF
Linux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
PPTX
Ceph Day Santa Clara: Ceph Fundamentals
PDF
Modeling, estimating, and predicting Ceph (Linux Foundation - Vault 2015)
PDF
Storage tiering and erasure coding in Ceph (SCaLE13x)
ODP
Ceph Day SF 2015 - Keynote
PDF
Ceph Day London 2014 - Ceph Ecosystem Overview
The Future of Cloud Software Defined Storage with Ceph: Andrew Hatfield, Red Hat
OSDC 2015: John Spray | The Ceph Storage System
Ceph - A distributed storage system
SUSE Storage: Sizing and Performance (Ceph)
2015 open storage workshop ceph software defined storage
Red Hat Storage 2014 - Product(s) Overview
New use cases for Ceph, beyond OpenStack, Luis Rico
Quick-and-Easy Deployment of a Ceph Storage Cluster
Ceph Overview for Distributed Computing Denver Meetup
Open Source Storage at Scale: Ceph @ GRNET
adp.ceph.openstack.talk
Ceph openstack-jun-2015-meetup
Ceph & OpenStack talk given @ OpenStack Meetup @ Bangalore, June 2015
Ceph, Now and Later: Our Plan for Open Unified Cloud Storage
Linux Stammtisch Munich: Ceph - Overview, Experiences and Outlook
Ceph Day Santa Clara: Ceph Fundamentals
Modeling, estimating, and predicting Ceph (Linux Foundation - Vault 2015)
Storage tiering and erasure coding in Ceph (SCaLE13x)
Ceph Day SF 2015 - Keynote
Ceph Day London 2014 - Ceph Ecosystem Overview

Recently uploaded (20)

PPTX
DATA STRCUTURE LABORATORY -BCSL305(PRG1)
PPTX
INTERNET OF THINGS - EMBEDDED SYSTEMS AND INTERNET OF THINGS
DOCX
ENVIRONMENTAL PROTECTION AND MANAGEMENT (18CVL756)
PPTX
Software-Development-Life-Cycle-SDLC.pptx
PPTX
Agentic Artificial Intelligence (Agentic AI).pptx
PPTX
CS6006 - CLOUD COMPUTING - Module - 1.pptx
PPTX
Unit IImachinemachinetoolopeartions.pptx
PDF
MLpara ingenieira CIVIL, meca Y AMBIENTAL
PPTX
Solar energy pdf of gitam songa hemant k
PPTX
ARCHITECTURE AND PROGRAMMING OF EMBEDDED SYSTEMS
DOCX
An investigation of the use of recycled crumb rubber as a partial replacement...
PPTX
chapter 1.pptx dotnet technology introduction
PDF
Beginners-Guide-to-Artificial-Intelligence.pdf
PPTX
AI-Reporting for Emerging Technologies(BS Computer Engineering)
PDF
ASPEN PLUS USER GUIDE - PROCESS SIMULATIONS
PPTX
Environmental studies, Moudle 3-Environmental Pollution.pptx
PDF
LS-6-Digital-Literacy (1) K12 CURRICULUM .pdf
PDF
VTU IOT LAB MANUAL (BCS701) Computer science and Engineering
PDF
Using Technology to Foster Innovative Teaching Practices (www.kiu.ac.ug)
PDF
[jvmmeetup] next-gen integration with apache camel and quarkus.pdf
DATA STRCUTURE LABORATORY -BCSL305(PRG1)
INTERNET OF THINGS - EMBEDDED SYSTEMS AND INTERNET OF THINGS
ENVIRONMENTAL PROTECTION AND MANAGEMENT (18CVL756)
Software-Development-Life-Cycle-SDLC.pptx
Agentic Artificial Intelligence (Agentic AI).pptx
CS6006 - CLOUD COMPUTING - Module - 1.pptx
Unit IImachinemachinetoolopeartions.pptx
MLpara ingenieira CIVIL, meca Y AMBIENTAL
Solar energy pdf of gitam songa hemant k
ARCHITECTURE AND PROGRAMMING OF EMBEDDED SYSTEMS
An investigation of the use of recycled crumb rubber as a partial replacement...
chapter 1.pptx dotnet technology introduction
Beginners-Guide-to-Artificial-Intelligence.pdf
AI-Reporting for Emerging Technologies(BS Computer Engineering)
ASPEN PLUS USER GUIDE - PROCESS SIMULATIONS
Environmental studies, Moudle 3-Environmental Pollution.pptx
LS-6-Digital-Literacy (1) K12 CURRICULUM .pdf
VTU IOT LAB MANUAL (BCS701) Computer science and Engineering
Using Technology to Foster Innovative Teaching Practices (www.kiu.ac.ug)
[jvmmeetup] next-gen integration with apache camel and quarkus.pdf

Ceph Object Storage at Spreadshirt

  • 1. Ceph Object Storage at Spreadshirt June 2015 Jens Hadlich Chief Architect
  • 2. About Spreadshirt 2 Spread it with Spreadshirt A global e-commerce platform for everyone to create, sell and buy ideas on clothing and accessories across many points of sale. •  12 languages, 11 currencies •  19 markets •  150+ shipping regions •  community of >70.000 active sellers •  € 72M revenue (2014) •  >3.3M items shipped (2014)
  • 3. Object Storage at Spreadshirt •  What? –  Store and read primarily user generated content, mostly images •  Typical sizes: –  a few dozen KB, a few MB •  Some 10s of terabyte (TB) of data •  Read > Write •  „Never change a running system“? –  Currently solution from the early days with big storage + lots of files / directories doesn‘t work anymore •  Regular UNIX tools get unusable in practice •  Not designed for „the cloud“ (e.g. replication is an issue) –  Growing number of users à more content –  Build a truly global platform (multiple regions and data centers) 3
  • 4. Ceph •  Why Ceph? –  Vendor independent –  Open source –  Runs on commodity hardware –  Local installation for minimal latency –  Existing knowledge and experience –  S3-API •  Simple bucket-to-bucket replication –  A good fit also for < Petabyte –  Easy to add more storage –  (Can be used later for block storage) 4
  • 5. Ceph Object Storage Architecture 5 Overview Ceph Object Gateway Monitor Cluster Network Public Network OSDOSD OSDOSDOSD MonitorMonitor A lot of nodes and disks Client HTTP (S3 or SWIFT API) RADOS (reliable autonomic distributed object store)
  • 6. Ceph Object Storage Architecture 6 A little more detailled Monitor Cluster Network Public Network Client RadosGW HTTP (S3 or SWIFT API) MonitorMonitor Some SSDs (for journals) More HDDs JBOD (no RAID) OSD node Ceph Object Gateway librados Odd number (Quorum) OSD node OSD node OSD node OSD node 1G 10G (the more the better) ... RADOS (reliable autonomic distributed object store) OSD node
  • 7. Ceph Object Storage Architecture 7 Initial Setup (planned) Cluster Network (OSD Replication) Cluster nodes 3 x SSD (journal / index) 9 x HDD (data) 3 Monitors 2 x 1G, IPv4 2 x 10G, IPv6 Public Network Client HTTP (S3 or SWIFT API) HAProxy RadosGW Monitor RadosGW Monitor RadosGW Monitor RadosGWRadosGW 2 x 10G, IPv6Cluster Network RadosGW on each node
  • 8. Ceph Object Storage Performance 8 Some smoke tests •  How fast is RadosGW? Get an impression. –  Response times (read / write) •  Average? •  Percentiles (P99)? –  Compared to AWS S3? •  A very minimalistic test setup –  3 VMs (KVM) all with RadosGW, Monitor and OSD •  2 Cores, 4GB RAM, 1 OSD each (15 GB + 5GB), 10G Network between nodes, HAProxy (round-robin), LAN, HTTP –  No further optimizations
  • 9. Ceph Object Storage Performance 9 Some smoke tests •  How fast is RadosGW? –  Random read and write –  Object size: 4 KB •  Results: Pretty promising! –  E.g. 16 parallel threads, read: •  Avg 9 ms •  P99 49 ms •  > 1.300 requests/s
  • 10. Ceph Object Storage Performance 10 Some smoke tests •  Compared to Amazon S3? –  Comparing apples and oranges (unfair, but interresting) •  http vs. https, LAN vs. WAN etc. •  Reponse times –  Random read, object size: 4KB, 4 parallel threads, location: Leipzig Ceph S3 AWS S3 eu-central-1 eu-west-1 Location Leipzig Frankfurt Ireland Avg 6 ms 25 ms 56 ms P99 47 ms 128 ms 374 ms Requests/s 405 143 62
  • 11. Global Availability 11 •  1 Ceph cluster per data center •  S3 bucket-to-bucket replication •  Multiple regions, local delivery
  • 12. To be continued ... + = ?