Tugdual Grall (@tgrall)
Alain Hélaïli (@AlainHelaili)
#MongoDBBasics @MongoDB
Construire une application avec MongoDB
Déploiement de l’application
2
• Résumé des épisodes précédents
• Replication
• Sharding
Agenda
3
• Virtual Genius Bar
– Utilisez la fenêtre de chat
– Tug & Alain dispo
pendant, et après…
• MUGs à Paris, Toulouse,
Bordeaux, Rennes, Lyon
• Groupes LinkedIn
« MongoDB France » et
« MongoDB » sur
Viadeo
Q & A
@tgrall, tug@mongodb.com - @AlainHelaili, alain.helaili@mongodb.cm
Résumé des épisodes
précédents…
5
• Agrégation de données…
– Map Reduce
– Hadoop
– Rapports Pré-Agrégés
– Aggregation Framework
• Tuning avec Explain
• Calcul à la volée ou calcul/stocke
• Geospatial
• Text Search
Résumé
Replication
Why Replication?
• How many have faced node failures?
• How many have been woken up from sleep to do a
fail-over(s)?
• How many have experienced issues due to network
latency?
• Different uses for data
– Normal processing
– Simple analytics
Why Replication?
• Replication is designed for
– High Availability (HA)
– Disaster Recovery (DR)
• Not designed for scaling reads
– You can but there are drawbacks: eventual
consistency, etc.
– Use sharding for scaling!
Replica Set – Creation
Replica Set – Initialize
Replica Set – Failure
Replica Set – Failover
Replica Set – Recovery
Replica Set – Recovered
Replica Set Roles &
Configuration
Replica Set Roles
Example with 2 data nodes + 1 arbiter
> conf = { // 5 data nodes
_id : "mySet",
members : [
{_id : 0, host : "A”, priority : 3},
{_id : 1, host : "B", priority : 2},
{_id : 2, host : "C”},
{_id : 3, host : "D", hidden : true},
{_id : 4, host : "E", hidden : true, slaveDelay : 3600}
]
}
> rs.initiate(conf)
Configuration Options
Developing with
Replica Sets
Strong Consistency
Delayed / Eventual Consistency
Write Concerns
• Network acknowledgement (w = 0)
• Wait for return info/error (w = 1)
• Wait for journal sync (j = 1)
• Wait for replication (w >=2)
Tagging
• Control where data is written to, and read from
• Each member can have one or more tags
– tags: {dc: "ny"}
– tags: {dc: "ny",
 subnet: "192.168",
 rack:
"row3rk7"}
• Replica set defines rules for write concerns
• Rules can change without changing app code
{
_id : "mySet",
members : [
{_id : 0, host : "A", tags : {"dc": "ny"}},
{_id : 1, host : "B", tags : {"dc": "ny"}},
{_id : 2, host : "C", tags : {"dc": "sf"}},
{_id : 3, host : "D", tags : {"dc": "sf"}},
{_id : 4, host : "E", tags : {"dc": "cloud"}}],
settings : {
getLastErrorModes : {
allDCs : {"dc" : 3},
someDCs : {"dc" : 2}} }
}
> db.blogs.insert({...})
> db.runCommand({getLastError : 1, w : "someDCs"})
Tagging Example
Read Preference Modes
• 5 modes
– primary (only) - Default
– primaryPreferred
– secondary
– secondaryPreferred
– Nearest
When more than one node is possible, closest node is used
for reads (all modes but primary)
Tagged Read Preference
• Custom read preferences
• Control where you read from by (node) tags
– E.g. { "disk": "ssd", "use": "reporting" }
• Use in conjunction with standard read
preferences
– Except primary
Sharding
Read/Write Throughput Exceeds I/O
Working Set Exceeds Physical
Memory
Vertical Scalability (Scale Up)
Horizontal Scalability (Scale Out)
Partitioning
• User defines shard key
• Shard key defines range of data
• Key space is like points on a line
• Range is a segment of that line (chunk), smaller than
64MB
• Chunks are migrated from one shard to another to
maintain a balanced state
Shard Key
• Shard key is immutable
• Shard key values are immutable
• Shard key must be indexed
• Shard key limited to 512 bytes in size
• Shard key used to route queries
– Choose a field commonly used in queries
• Only shard key can be unique across shards
Shard Key Considerations
• Cardinality
• Write Distribution
• Query Isolation
• Reliability
• Index Locality
Initially 1 chunk
Default max chunk size: 64mb
MongoDB automatically splits & migrates chunks
when max reached
Data Distribution
Queries routed to specific shards
MongoDB balances cluster
MongoDB migrates data to new nodes
Routing and Balancing
Partitioning
- ∞ + ∞ shard 2 shard 3
Partitioning
- ∞ + ∞
- ∞ { x : 1}, { x : 3} …. { x : 99} +
∞
shard 2 shard 3
shard 2 shard 3
Partitioning
- ∞ + ∞
- ∞ { x : 1}, { x : 3} …. { x : 99} +
∞
- ∞ { x : 1} …. { x : 55}
{ x : 56} …. { x : 110} +
∞ shard 2 shard 3
shard 2 shard 3
shard 2 shard 3
Partitioning
- ∞ + ∞
- ∞ { x : 1}, { x : 3} …. { x : 99} +
∞
- ∞ { x : 1} …. { x : 55}
{ x : 56} …. { x : 110} +
∞ shard 2 shard 3
shard 2 shard 3
shard 2 shard 3
Partitioning
- ∞ + ∞
- ∞ { x : 1}, { x : 3} …. { x : 99} +
∞
- ∞ { x : 1} …. { x : 55}
{ x : 56} …. { x : 110} +
∞ shard 2 shard 3
shard 2 shard 3
shard 2 shard 3
MongoDB Auto-Sharding
• Minimal effort required
– Same interface as single mongod
• Two steps
– Enable Sharding for a database
– Shard collection within database
Architecture
What is a Shard?
• Shard is a node of the cluster
• Shard can be a single mongod or a replica set
Meta Data Storage
• Config Server
– Stores cluster chunk ranges and locations
– Can have only 1 or 3 (production must have 3)
– Not a replica set
Routing and Managing Data
• Mongos
– Acts as a router / balancer
– No local data (persists to config database)
– Can have 1 or many
Sharding infrastructure
Configuration
Example Cluster
mongod --configsvr
Starts a configuration server on the default port
(27019)
Starting the Configuration Server
mongos --configdb <hostname>:27019
For 3 configuration servers:
mongos --configdb<host1>:<port1>,<host2>:<port2>,<host3>:<port3>
Thisis always how to start a new mongos, even if the cluster is already running
Start the mongos Router
mongod --shardsvr
Starts a mongod with the default shard port (27018)
Shard is not yet connected to the rest of the cluster
Shard may have already been running in production
Start the shard database
On mongos:
– sh.addShard(„<host>:27018‟)
Adding a replica set:
– sh.addShard(„<rsname>/<seedlist>‟)
Add the Shard
db.runCommand({ listshards:1 })
{ "shards" :
[{"_id”:"shard0000”,"host”:”<hostname>:27018”} ],
"ok" : 1
}
Verify that the shard was added
Enabling Sharding
• Enable sharding on a database
sh.enableSharding(“<dbname>”)
• Shard a collection with the given key
sh.shardCollection(“<dbname>.people”,{“country”:1})
• Use a compound shard key to prevent duplicates
sh.shardCollection(“<dbname>.cars”,{“year”:1,”uniqueid”:1})
Tag Aware Sharding
• Tag aware sharding allows you to control the
distribution of your data
• Tag a range of shard keys
– sh.addTagRange(<collection>,<min>,<max>,<tag>)
• Tag a shard
– sh.addShardTag(<shard>,<tag>)
Routing Requests
Cluster Request Routing
• Targeted Queries
• Scatter Gather Queries
• Scatter Gather Queries with Sort
Cluster Request Routing: Targeted
Query
Routable request received
Request routed to appropriate
shard
Shard returns results
Mongos returns results to client
Cluster Request Routing: Non-Targeted
Query
Non-Targeted Request Received
Request sent to all shards
Shards return results to mongos
Mongos returns results to client
Cluster Request Routing: Non-Targeted
Query with Sort
Non-Targeted request with sort
received
Request sent to all shards
Query and sort performed locally
Shards return results to mongos
Mongos merges sorted results
Mongos returns results to client
Résumé
76
• Replica set pour haute disponibilité
• Sharding pour montée en charge
• Write concern
• Clé de sharding
Résumé
77
– Backup
– Reprise sur incident
Prochaine session – 3 Juin
2014 05-07-fr - add dev series - session 6 - deploying your application-2

2014 05-07-fr - add dev series - session 6 - deploying your application-2

Editor's Notes

  • #10 Basic explanation2 or more nodes form the setQuorum
  • #11 Initialize -&gt; ElectionPrimary + data replication from primary to secondaryHeartbeat every 2 seconds, timeout 10 seconds
  • #12 Primary down/network failureAutomatic election of new primary if majority existsFailover usually takes a couple of seconds. Depending on your application code and configuration, this can be seamless/transparent.
  • #13 New primary electedReplication established from new primary
  • #14 Down node comes upRejoins setsRecovery and then secondary
  • #15 Note that replication doesn’t always need to pull from the primary. Will pull from secondary if it is faster (less ping time).
  • #17 PrimaryData memberSecondaryHot standbyArbitersVoting member
  • #18 PriorityFloating point number between 0..1000Highest member that is up to date wins Up to date == within 10 seconds of primaryIf a higher priority member catches up, it will force election and win Slave DelayLags behind master by configurable time delay Automatically hidden from clientsProtects against operator errorsFat fingeringApplication corrupts data
  • #19 ConsistencyWrite preferencesRead preferences
  • #24 Using &apos;someDCs&apos; so that in the event of an outage, at least a majority of the DCs would receive the change. This favors availability over durability.
  • #29 Indexes should be contained in working set.
  • #30 From mainframes, to RAC Oracle servers... People solved problems by adding more resources to a single machine.
  • #31 Large scale operation can be combined with high performance on commodity hardware through horizontal scalingBuild - Document oriented database maps perfectly to object oriented languagesScale - MongoDB presents clear path to scalability that isn&apos;t ops intensive - Provides same interface for sharded cluster as single instance
  • #33 _id could be unique across shards if used as shard key.we could only guarantee uniqueness of (any) attributes if the keys are used as shard keys with unique attribute equals true
  • #34 Cardinality – Can your data be broken down enough?Query Isolation - query targeting to a specific shardReliability – shard outagesA good shard key can:Optimize routingMinimize (unnecessary) trafficAllow best scaling
  • #49 Don’t use this setup in production!Only one Config server (No Fault Tolerance)Shard not in a replica set (Low Availability)Only one mongos and shard (No Performance Improvement)Useful for development or demonstrating configuration mechanics
  • #53 MongoDB 2.2 and later only need &lt;host&gt; and &lt;port&gt; for one member of the replica set
  • #56 This can be skipped for the intro talk, but might be good to include if you’re doing the combined sharding talk. Totally optional, you don’t really have enough time to do this topic justice but it might be worth a mention.
  • #74 The mongos does not have to load the whole set into memory since each shard sorts locally. The mongos can just getMore from the shards as needed and incrementally return the results to the client.