SlideShare a Scribd company logo
How to Manage Large Amounts of Data
with Salesforce
1
Data & Integration Considerations for Architects
Paul McCollum
Cloud Architect
pmccollum@sensecorp.com
2
www.sensecorp.com/contact
@SenseCorp
facebook.com/SenseCorpSC/
linkedin.com/company/sense-corp
It’s just Data, right?
• Map my data scale (Migration and Ongoing)
• Will my application fit (long term) in my Target Architecture?
• Do I have any “Design Skews”?
• Where will my pain points be?
• Pain points and solution patterns for:
– Lots of data (LDV*)
– Lots of connections
– Lots of connections to lots of data sources
• In this Session, we will be focusing on identifying data and integration
issues in the design phase, discussing some solution platforms specifically.
3
GOAL: Learn Architect Design Planning Patterns
“A ‘large data volume’ is an imprecise, elastic
term. If your deployment has tens of thousands of
users, tens of millions of records, or hundreds of
gigabytes of total record storage, you have a large
data volume…”
~Salesforce
https://developer.salesforce.com/docs/atlas.en-us.salesforce_large_data_volumes_bp.meta/salesforce_large_data_volumes_bp/ldv_deployments_introduction.htm
The Science of Architecture
• As soon as you build it, (sometimes before you are even done) applications
take on lives of their own.
• Architects need to see the present and predict the future.
• Good Architecture is responsible for the ultimate success of
application/project.
• In relative dollars, it costs:
 $1 to effectively design and architect a solution
 $10 to re-design it due to technical debt if not architected well
 $100 to completely scrap and re-tool
 $____ to repair your reputation after a write-off
4
It’s just Data, right?
• We create:
– Pages
– Buttons
– Flows
– Validation Rules*
• Talking to:
– Objects
– Fields
• Combined with:
– Authentication: Active Directory, SSO
– HR Profile Data, Medical History, Order History
5
We’re all used to building Applications in Salesforce.
*Salesforce has imposed a limit of 100 Validation Rules per Object.
Ever try to create test data on an object with 100 Validation Rules?
Building Applications Based on Data Scale
1:1:1 or 101 : 101 : 101*
• We try to understand our applications’ consumption of
data relative to the complexity/magnitude of
downstream Data. NOT because data is bad but because
handling lots of anything has a cost.
• For our purposes “Design Skew” is defined as having
components out of proportion with each other.
• Skew can cause both System and Operational
degradation.
• The following slides show examples of how to start
documenting your data interactions and Predict and
Mitigate possible Skew complications.
6
*Obviously, there’s more than 3 elements of solutions, the key is balance.
Master Data
Be aware if you are building or
defining relationships for Master
Data, you may also be creating
duplicate “Master Logic”.
Rules, Logic and Relationships to data
that is not your application’s Master
Data should be conferred upstream
to the Master Data owner.
For Example: Product Price and Tax
Rate calculation should not be done
in multiple systems. Differences in
rounding rules can lead to
discrepancies that can be hard to
reconcile.
Common Types of Design Skews
7
How do I map my application?
8
Salesforce
Data
Start with this simple diagram
• Sketch User counts
• User Types
• Functions
• Connectivity
• Firewall
• Domains
• Data Sources
Objects Services
User
Share it with All teams and Stakeholders.
• Attach it to every deploy and code review
• Any changes must be ratified
Architecture Worksheet
9
Salesforce
Data
Objects Services
User
Users Functions Objects Records Connections Dependencies
Data
Sources
Master
Logic
Master
Data
Reference Architectures
Salesforce
Growth Enabled Company Infrastructure
Scalable
Regular Patterns
Repeatable
Mature Support
Any time the Application has a 1:>1:>1
Objects
Kafka
(Streaming)
NoSQL
ESB
DAM
SFTP
ETL
Main
Frame
MuleSoft
Jitterbit
Data Lake
(Object Storage) Data
Data
Data
Data
Snowflake
Step 1: Dream Big
Architects: Visualize Success
• How big will this get?
• How many people will use this?
• If it gets to size X, will I have to rebuild?
• Define Growth Boundaries with recommended changes.
• Create Transactional Warning systems inside triggers and functions.
11
If(DML.size() > 70% of limit){
SendEmail(to=setting.PerfAdmin);
}
Step 2: Dream Bigger
• Does this data have value?
– The answer is always: Yes.
• The larger the volume of data created, the more likely there will be
value in analyzing that data.
– Trends
– Next Best Action
– Inefficiencies
– Agent Performance
– ROI
• As you create data, think of the many ways it could be used.
– Plan accordingly!
12
Too Many Records
Salesforce Data
Schema Overload
• Trying to keep the data model a perfect twin to reality
• Not having a plan to make data “Mature Out” of the system
• Avoiding Data Skew
• Is every application responsible for Forensics?
Objects
Services
Objects
Objects
Objects
Objects
https://developer.salesforce.com/blogs/engineering/2012/04/avoid-account-data-skew-for-peak-performance.html
14
When to use Big Objects
https://trailhead.salesforce.com/content/learn/modules/big-data-strategy/choose-the-right-big-data-solution
Salesforce Object Storage
15 https://developer.salesforce.com/docs/atlas.en-us.bigobjects.meta/bigobjects/big_object.htm
Big Objects (NoSQL) Data Lake/Cloud Data Warehouse
When to use? When data is only recorded for forensics or
broad reporting and not accessed very often.
When data assets from Salesforce need to be mashed-
up with other enterprise data.
• When companies don’t already have a
reporting ODS available.
• Many 3rd party integrations.
• Data can be reused or resold.
• Organizations that have a strong/evolving cloud
analytics strategy.
• When ETL or external data is required for context.
• Do not bring data into Salesforce JUST to
perform calculations on it.
• Do not store data in Big Objects, if you
expect them to grow infinitely.
• Read/Write Optimized
• Read and Write optimizations.
• They reduce security headaches for consumers.
• Once you get to 1+ TB of Enterprise data
• Transact and Read/Write Optimized
What you need to
be successful?
Operational and Administrative diligence
managing size, load, license cost.
Healthy Cloud management presence or engagement.
Licensing? Licensing through Salesforce is ‘extra’ but
reasonable.
Varies but can be much cheaper from a bulk storage
perspective.
Objects  Big Objects  Data Lake/CDW
When to use? • If your application relies on a service or data provider that is also changing, consider
strongly an investment in an API framework like MuleSoft or Jitterbit.
• These frameworks allow for Services and ETL steps to be performed at the
translation layer and can be versioned and tracked.
Vitally important if your application needs to be UP while it is being changed or source
systems are changing.
What you need to be
successful?
Still new and bespoke management required. Dividends clearly paid in uptime and
efficiency.
Licensing? Varies
Leading API Orchestration Technology Platforms
16
Leading Analytics Technology Platforms
• Dream Big: What is the future of your data scale?
• If Consider Architecting for Data offload to an ODS like
Snowflake early.
• Snowflake is great for Near Real Time Process using the
Snowpipe Feature
• Storage and Compute Resources are Separated.
• It can Scale quickly if you need more Compute Power to
process Queries
• Handles both Structured and Semi Structured Data
17
When to use? Sales trend, customer demographics, Area
Performance, Product Returns, Sales,
discounts and bundles. All data that provides
insight once accumulated.
If you are going to run heavy calculation with
near-real-time data.
Ability to Share Information Across different
Organizations without having to transfer files.
Reporting and Dashboarding against a Data Lake
or Cloud Data Warehouse
Your needs are subject to change and
scalability is a concern.
Multiple data sources
What you need to
be successful?
If metrics and data are going to play a
strategic role in your business, invest early.
Competency in modern cloud analytics
architecture.
Data Literacy
BI skillset
Data Visualization and UI/UX skills
Licensing? Cloud consumption-based model Traditional or Consumption based.
On the
Horizon?
• With the acquisition of Tableau and closer dealings with Snowflake, things
could rapidly change in this area.
• Most likely this will tip the scales towards investments in both, if you are
planning to continue/grow your Salesforce investment.
18
Thanks For Joining Us
We hope you enjoyed the presentation.
If you’d like to learn more about how we use Salesforce
to help transform your organization, contact us.
https://sensecorp.com/salesforce-sense-corp/
CONTACT US
www.sensecorp.com | marketing@sensecorp.com
Questions?

More Related Content

What's hot (20)

PPTX
MuleSoft Kochi Meetup #3– Integration with Web Sockets
sumitahuja94
 
PPTX
Nyc mule soft_meetup_13_march_2021
NeerajKumar1965
 
PDF
Mule soft mcia-level-1 Dumps
Armstrongsmith
 
PDF
MuleSoft Nashik Virtual Meetup#2 - API Led Connectivity Integration:SAGA
Jitendra Bafna
 
PPTX
NYC MuleSoft Meetup 2019 Q2- MuleSoft for Mobile Applications
Gean Martinez
 
PDF
apidays LIVE Australia 2020 - Building an Enterprise Eventing Platform by Gna...
apidays
 
PDF
MuleSoft Surat Virtual Meetup#16 - Anypoint Deployment Option, API and Operat...
Jitendra Bafna
 
PPTX
Mulesoft Solutions for SOA
kumar gaurav
 
PDF
apidays LIVE Australia 2020 - Data with a Mission by Matt McLarty
apidays
 
PDF
MuleSoft Meetup Singapore - Reliable Messaging & RTF Operations
Julian Douch
 
PPTX
DevOps and APIs: Great Alone, Better Together
MuleSoft
 
PDF
Microservices & anypoint service mesh calgary mule soft meetup
Jimmy Attia
 
PDF
Optimizing TAS Usage at Ford Motor Company
VMware Tanzu
 
PDF
MuleSoft Surat Virtual Meetup#24 - MuleSoft and Salesforce Integration and De...
Jitendra Bafna
 
PPTX
Meet up slides_mumbai_21032020_final
Akshata Sawant
 
PDF
Nashik MuleSoft Virtual Meetup#1 - Shared and Dedicated Load Balancer
Jitendra Bafna
 
PPTX
Meetup_Bangalore_Rajesh
D.Rajesh Kumar
 
PDF
Sustainability Challenge, Postman, Rest sheet and Anypoint provider : MuleSof...
Angel Alberici
 
PDF
MuleSoft Meetup Singapore June 2021
Julian Douch
 
PPTX
Manchester Meetup #3
Francis Edwards
 
MuleSoft Kochi Meetup #3– Integration with Web Sockets
sumitahuja94
 
Nyc mule soft_meetup_13_march_2021
NeerajKumar1965
 
Mule soft mcia-level-1 Dumps
Armstrongsmith
 
MuleSoft Nashik Virtual Meetup#2 - API Led Connectivity Integration:SAGA
Jitendra Bafna
 
NYC MuleSoft Meetup 2019 Q2- MuleSoft for Mobile Applications
Gean Martinez
 
apidays LIVE Australia 2020 - Building an Enterprise Eventing Platform by Gna...
apidays
 
MuleSoft Surat Virtual Meetup#16 - Anypoint Deployment Option, API and Operat...
Jitendra Bafna
 
Mulesoft Solutions for SOA
kumar gaurav
 
apidays LIVE Australia 2020 - Data with a Mission by Matt McLarty
apidays
 
MuleSoft Meetup Singapore - Reliable Messaging & RTF Operations
Julian Douch
 
DevOps and APIs: Great Alone, Better Together
MuleSoft
 
Microservices & anypoint service mesh calgary mule soft meetup
Jimmy Attia
 
Optimizing TAS Usage at Ford Motor Company
VMware Tanzu
 
MuleSoft Surat Virtual Meetup#24 - MuleSoft and Salesforce Integration and De...
Jitendra Bafna
 
Meet up slides_mumbai_21032020_final
Akshata Sawant
 
Nashik MuleSoft Virtual Meetup#1 - Shared and Dedicated Load Balancer
Jitendra Bafna
 
Meetup_Bangalore_Rajesh
D.Rajesh Kumar
 
Sustainability Challenge, Postman, Rest sheet and Anypoint provider : MuleSof...
Angel Alberici
 
MuleSoft Meetup Singapore June 2021
Julian Douch
 
Manchester Meetup #3
Francis Edwards
 

Similar to Managing Large Amounts of Data with Salesforce (20)

PDF
Diagramming Salesforce Solutions: Matthew Morris - Jacksonville Architects - ...
A. Engin Utkan
 
PDF
Manage Massive Datasets with Big Objects & Async SOQL
Salesforce Developers
 
PDF
Bringing the Power of Big Data Computation to Salesforce
Salesforce Developers
 
PDF
Unleash the Potential of Big Data on Salesforce
Dreamforce
 
PDF
Extreme Salesforce Data Volumes Webinar (with Speaker Notes)
Salesforce Developers
 
PDF
Development Strategies for Enterprise Scale From the Salesforce.com Platform
Salesforce Developers
 
PPTX
SF Architect Interview questions v1.3.pptx
AnkitJain429819
 
PDF
Data engineering design patterns
Valdas Maksimavičius
 
PDF
Admin Tips, Tricks & Strategies for Data Quality in Salesforce - Francis Pind...
Salesforce Admins
 
PDF
Salesforce1 data gov lunch toronto deck
Beth Fitzpatrick
 
PPTX
LDV.pptx
Shams Pirzada
 
PPTX
2020 Barcelona Administrators Group - Virtual Event Dec 15
animuscrm
 
PPTX
[DSC Europe 24] Josip Saban - Buidling cloud data platforms in enterprises
DataScienceConferenc1
 
PDF
The Need for Speed: Building Reports That Fly
Salesforce Developers
 
PDF
Handling of Large Data by Salesforce
Thinqloud
 
PDF
Salesforce1 data gov lunch anaheim deck
Beth Fitzpatrick
 
PDF
Avoid Growing Pains: Scale Your App for the Enterprise (October 14, 2014)
Salesforce Partners
 
PDF
High Scale Relational Storage at Salesforce Built with Apache HBase and Apach...
Salesforce Engineering
 
PDF
Manage Development in Your Org with Salesforce Governance Framework
Salesforce Developers
 
PDF
Foundational Strategies for Trust in Big Data Part 1: Getting Data to the Pla...
Precisely
 
Diagramming Salesforce Solutions: Matthew Morris - Jacksonville Architects - ...
A. Engin Utkan
 
Manage Massive Datasets with Big Objects & Async SOQL
Salesforce Developers
 
Bringing the Power of Big Data Computation to Salesforce
Salesforce Developers
 
Unleash the Potential of Big Data on Salesforce
Dreamforce
 
Extreme Salesforce Data Volumes Webinar (with Speaker Notes)
Salesforce Developers
 
Development Strategies for Enterprise Scale From the Salesforce.com Platform
Salesforce Developers
 
SF Architect Interview questions v1.3.pptx
AnkitJain429819
 
Data engineering design patterns
Valdas Maksimavičius
 
Admin Tips, Tricks & Strategies for Data Quality in Salesforce - Francis Pind...
Salesforce Admins
 
Salesforce1 data gov lunch toronto deck
Beth Fitzpatrick
 
LDV.pptx
Shams Pirzada
 
2020 Barcelona Administrators Group - Virtual Event Dec 15
animuscrm
 
[DSC Europe 24] Josip Saban - Buidling cloud data platforms in enterprises
DataScienceConferenc1
 
The Need for Speed: Building Reports That Fly
Salesforce Developers
 
Handling of Large Data by Salesforce
Thinqloud
 
Salesforce1 data gov lunch anaheim deck
Beth Fitzpatrick
 
Avoid Growing Pains: Scale Your App for the Enterprise (October 14, 2014)
Salesforce Partners
 
High Scale Relational Storage at Salesforce Built with Apache HBase and Apach...
Salesforce Engineering
 
Manage Development in Your Org with Salesforce Governance Framework
Salesforce Developers
 
Foundational Strategies for Trust in Big Data Part 1: Getting Data to the Pla...
Precisely
 
Ad

More from Sense Corp (8)

PPTX
The Future of the Digital Experience: How to Embrace the New Order of Busines...
Sense Corp
 
PPTX
Achieve New Heights with Modern Analytics
Sense Corp
 
PPTX
Why Data Science Projects Fail
Sense Corp
 
PPTX
Small Investments, Big Returns: Three Successful Data Science Use Cases
Sense Corp
 
PPTX
10 Steps to Develop a Data Literate Workforce
Sense Corp
 
PPTX
Why Data Science Projects Fail
Sense Corp
 
PPTX
The Data Warehouse is NOT Dead
Sense Corp
 
PDF
Infographic data
Sense Corp
 
The Future of the Digital Experience: How to Embrace the New Order of Busines...
Sense Corp
 
Achieve New Heights with Modern Analytics
Sense Corp
 
Why Data Science Projects Fail
Sense Corp
 
Small Investments, Big Returns: Three Successful Data Science Use Cases
Sense Corp
 
10 Steps to Develop a Data Literate Workforce
Sense Corp
 
Why Data Science Projects Fail
Sense Corp
 
The Data Warehouse is NOT Dead
Sense Corp
 
Infographic data
Sense Corp
 
Ad

Recently uploaded (20)

PDF
Merits and Demerits of DBMS over File System & 3-Tier Architecture in DBMS
MD RIZWAN MOLLA
 
PDF
JavaScript - Good or Bad? Tips for Google Tag Manager
📊 Markus Baersch
 
PPTX
Aict presentation on dpplppp sjdhfh.pptx
vabaso5932
 
PDF
How to Connect Your On-Premises Site to AWS Using Site-to-Site VPN.pdf
Tamanna
 
PPTX
Module-5-Measures-of-Central-Tendency-Grouped-Data-1.pptx
lacsonjhoma0407
 
PDF
apidays Helsinki & North 2025 - How (not) to run a Graphql Stewardship Group,...
apidays
 
PPTX
Dr djdjjdsjsjsjsjsjsjjsjdjdjdjdjjd1.pptx
Nandy31
 
PDF
WEF_Future_of_Global_Fintech_Second_Edition_2025.pdf
AproximacionAlFuturo
 
PDF
Copia de Strategic Roadmap Infographics by Slidesgo.pptx (1).pdf
ssuserd4c6911
 
PDF
Context Engineering for AI Agents, approaches, memories.pdf
Tamanna
 
PPTX
apidays Helsinki & North 2025 - APIs at Scale: Designing for Alignment, Trust...
apidays
 
PDF
R Cookbook - Processing and Manipulating Geological spatial data with R.pdf
OtnielSimopiaref2
 
PPTX
ER_Model_with_Diagrams_Presentation.pptx
dharaadhvaryu1992
 
PDF
Avatar for apidays apidays PRO June 07, 2025 0 5 apidays Helsinki & North 2...
apidays
 
PDF
apidays Helsinki & North 2025 - REST in Peace? Hunting the Dominant Design fo...
apidays
 
PPT
deep dive data management sharepoint apps.ppt
novaprofk
 
PPTX
apidays Helsinki & North 2025 - API access control strategies beyond JWT bear...
apidays
 
PPT
Growth of Public Expendituuure_55423.ppt
NavyaDeora
 
PPTX
apidays Singapore 2025 - From Data to Insights: Building AI-Powered Data APIs...
apidays
 
PDF
Early_Diabetes_Detection_using_Machine_L.pdf
maria879693
 
Merits and Demerits of DBMS over File System & 3-Tier Architecture in DBMS
MD RIZWAN MOLLA
 
JavaScript - Good or Bad? Tips for Google Tag Manager
📊 Markus Baersch
 
Aict presentation on dpplppp sjdhfh.pptx
vabaso5932
 
How to Connect Your On-Premises Site to AWS Using Site-to-Site VPN.pdf
Tamanna
 
Module-5-Measures-of-Central-Tendency-Grouped-Data-1.pptx
lacsonjhoma0407
 
apidays Helsinki & North 2025 - How (not) to run a Graphql Stewardship Group,...
apidays
 
Dr djdjjdsjsjsjsjsjsjjsjdjdjdjdjjd1.pptx
Nandy31
 
WEF_Future_of_Global_Fintech_Second_Edition_2025.pdf
AproximacionAlFuturo
 
Copia de Strategic Roadmap Infographics by Slidesgo.pptx (1).pdf
ssuserd4c6911
 
Context Engineering for AI Agents, approaches, memories.pdf
Tamanna
 
apidays Helsinki & North 2025 - APIs at Scale: Designing for Alignment, Trust...
apidays
 
R Cookbook - Processing and Manipulating Geological spatial data with R.pdf
OtnielSimopiaref2
 
ER_Model_with_Diagrams_Presentation.pptx
dharaadhvaryu1992
 
Avatar for apidays apidays PRO June 07, 2025 0 5 apidays Helsinki & North 2...
apidays
 
apidays Helsinki & North 2025 - REST in Peace? Hunting the Dominant Design fo...
apidays
 
deep dive data management sharepoint apps.ppt
novaprofk
 
apidays Helsinki & North 2025 - API access control strategies beyond JWT bear...
apidays
 
Growth of Public Expendituuure_55423.ppt
NavyaDeora
 
apidays Singapore 2025 - From Data to Insights: Building AI-Powered Data APIs...
apidays
 
Early_Diabetes_Detection_using_Machine_L.pdf
maria879693
 

Managing Large Amounts of Data with Salesforce

  • 1. How to Manage Large Amounts of Data with Salesforce 1 Data & Integration Considerations for Architects
  • 3. It’s just Data, right? • Map my data scale (Migration and Ongoing) • Will my application fit (long term) in my Target Architecture? • Do I have any “Design Skews”? • Where will my pain points be? • Pain points and solution patterns for: – Lots of data (LDV*) – Lots of connections – Lots of connections to lots of data sources • In this Session, we will be focusing on identifying data and integration issues in the design phase, discussing some solution platforms specifically. 3 GOAL: Learn Architect Design Planning Patterns “A ‘large data volume’ is an imprecise, elastic term. If your deployment has tens of thousands of users, tens of millions of records, or hundreds of gigabytes of total record storage, you have a large data volume…” ~Salesforce https://developer.salesforce.com/docs/atlas.en-us.salesforce_large_data_volumes_bp.meta/salesforce_large_data_volumes_bp/ldv_deployments_introduction.htm
  • 4. The Science of Architecture • As soon as you build it, (sometimes before you are even done) applications take on lives of their own. • Architects need to see the present and predict the future. • Good Architecture is responsible for the ultimate success of application/project. • In relative dollars, it costs:  $1 to effectively design and architect a solution  $10 to re-design it due to technical debt if not architected well  $100 to completely scrap and re-tool  $____ to repair your reputation after a write-off 4
  • 5. It’s just Data, right? • We create: – Pages – Buttons – Flows – Validation Rules* • Talking to: – Objects – Fields • Combined with: – Authentication: Active Directory, SSO – HR Profile Data, Medical History, Order History 5 We’re all used to building Applications in Salesforce. *Salesforce has imposed a limit of 100 Validation Rules per Object. Ever try to create test data on an object with 100 Validation Rules?
  • 6. Building Applications Based on Data Scale 1:1:1 or 101 : 101 : 101* • We try to understand our applications’ consumption of data relative to the complexity/magnitude of downstream Data. NOT because data is bad but because handling lots of anything has a cost. • For our purposes “Design Skew” is defined as having components out of proportion with each other. • Skew can cause both System and Operational degradation. • The following slides show examples of how to start documenting your data interactions and Predict and Mitigate possible Skew complications. 6 *Obviously, there’s more than 3 elements of solutions, the key is balance. Master Data Be aware if you are building or defining relationships for Master Data, you may also be creating duplicate “Master Logic”. Rules, Logic and Relationships to data that is not your application’s Master Data should be conferred upstream to the Master Data owner. For Example: Product Price and Tax Rate calculation should not be done in multiple systems. Differences in rounding rules can lead to discrepancies that can be hard to reconcile.
  • 7. Common Types of Design Skews 7
  • 8. How do I map my application? 8 Salesforce Data Start with this simple diagram • Sketch User counts • User Types • Functions • Connectivity • Firewall • Domains • Data Sources Objects Services User Share it with All teams and Stakeholders. • Attach it to every deploy and code review • Any changes must be ratified
  • 9. Architecture Worksheet 9 Salesforce Data Objects Services User Users Functions Objects Records Connections Dependencies Data Sources Master Logic Master Data
  • 10. Reference Architectures Salesforce Growth Enabled Company Infrastructure Scalable Regular Patterns Repeatable Mature Support Any time the Application has a 1:>1:>1 Objects Kafka (Streaming) NoSQL ESB DAM SFTP ETL Main Frame MuleSoft Jitterbit Data Lake (Object Storage) Data Data Data Data Snowflake
  • 11. Step 1: Dream Big Architects: Visualize Success • How big will this get? • How many people will use this? • If it gets to size X, will I have to rebuild? • Define Growth Boundaries with recommended changes. • Create Transactional Warning systems inside triggers and functions. 11 If(DML.size() > 70% of limit){ SendEmail(to=setting.PerfAdmin); }
  • 12. Step 2: Dream Bigger • Does this data have value? – The answer is always: Yes. • The larger the volume of data created, the more likely there will be value in analyzing that data. – Trends – Next Best Action – Inefficiencies – Agent Performance – ROI • As you create data, think of the many ways it could be used. – Plan accordingly! 12
  • 13. Too Many Records Salesforce Data Schema Overload • Trying to keep the data model a perfect twin to reality • Not having a plan to make data “Mature Out” of the system • Avoiding Data Skew • Is every application responsible for Forensics? Objects Services Objects Objects Objects Objects https://developer.salesforce.com/blogs/engineering/2012/04/avoid-account-data-skew-for-peak-performance.html
  • 14. 14 When to use Big Objects https://trailhead.salesforce.com/content/learn/modules/big-data-strategy/choose-the-right-big-data-solution
  • 15. Salesforce Object Storage 15 https://developer.salesforce.com/docs/atlas.en-us.bigobjects.meta/bigobjects/big_object.htm Big Objects (NoSQL) Data Lake/Cloud Data Warehouse When to use? When data is only recorded for forensics or broad reporting and not accessed very often. When data assets from Salesforce need to be mashed- up with other enterprise data. • When companies don’t already have a reporting ODS available. • Many 3rd party integrations. • Data can be reused or resold. • Organizations that have a strong/evolving cloud analytics strategy. • When ETL or external data is required for context. • Do not bring data into Salesforce JUST to perform calculations on it. • Do not store data in Big Objects, if you expect them to grow infinitely. • Read/Write Optimized • Read and Write optimizations. • They reduce security headaches for consumers. • Once you get to 1+ TB of Enterprise data • Transact and Read/Write Optimized What you need to be successful? Operational and Administrative diligence managing size, load, license cost. Healthy Cloud management presence or engagement. Licensing? Licensing through Salesforce is ‘extra’ but reasonable. Varies but can be much cheaper from a bulk storage perspective. Objects  Big Objects  Data Lake/CDW
  • 16. When to use? • If your application relies on a service or data provider that is also changing, consider strongly an investment in an API framework like MuleSoft or Jitterbit. • These frameworks allow for Services and ETL steps to be performed at the translation layer and can be versioned and tracked. Vitally important if your application needs to be UP while it is being changed or source systems are changing. What you need to be successful? Still new and bespoke management required. Dividends clearly paid in uptime and efficiency. Licensing? Varies Leading API Orchestration Technology Platforms 16
  • 17. Leading Analytics Technology Platforms • Dream Big: What is the future of your data scale? • If Consider Architecting for Data offload to an ODS like Snowflake early. • Snowflake is great for Near Real Time Process using the Snowpipe Feature • Storage and Compute Resources are Separated. • It can Scale quickly if you need more Compute Power to process Queries • Handles both Structured and Semi Structured Data 17 When to use? Sales trend, customer demographics, Area Performance, Product Returns, Sales, discounts and bundles. All data that provides insight once accumulated. If you are going to run heavy calculation with near-real-time data. Ability to Share Information Across different Organizations without having to transfer files. Reporting and Dashboarding against a Data Lake or Cloud Data Warehouse Your needs are subject to change and scalability is a concern. Multiple data sources What you need to be successful? If metrics and data are going to play a strategic role in your business, invest early. Competency in modern cloud analytics architecture. Data Literacy BI skillset Data Visualization and UI/UX skills Licensing? Cloud consumption-based model Traditional or Consumption based.
  • 18. On the Horizon? • With the acquisition of Tableau and closer dealings with Snowflake, things could rapidly change in this area. • Most likely this will tip the scales towards investments in both, if you are planning to continue/grow your Salesforce investment. 18
  • 19. Thanks For Joining Us We hope you enjoyed the presentation. If you’d like to learn more about how we use Salesforce to help transform your organization, contact us. https://sensecorp.com/salesforce-sense-corp/ CONTACT US www.sensecorp.com | [email protected]

Editor's Notes

  • #2: The ability for Salesforce to handle large workloads and participate in high-consumption, mobile-application-powering technologies continues to evolve. Pub/sub models and the investment in adjacent properties like Snowflake, Kafka and MuleSoft has broadened the development scope of Salesforce. Solutions now range from internal and in-platform applications to fueling world-scale mobile applications and integrations. Unfortunately, guidance on the extended capabilities still is not well understood or well documented. Knowing when your solution needs to move to a higher order solution is an important Architect skill. In this webinar, Paul McCollum, UXMC and Technical Architect at Sense Corp, will share an overview of data and architecture considerations. Attend to learn how to identify reasons and guidelines for updating your solutions to larger scale modern reference infrastructures as well as when to introduce products like Big Objects, Kafka, MuleSoft and Snowflake.
  • #4: It’s time to Level-Up and start designing for full lifecycle and fit. http://virtualdreamin.com/top-9-considerations-when-transitioning-from-a-developer-to-an-architect/ https://developer.salesforce.com/docs/atlas.en-us.salesforce_large_data_volumes_bp.meta/salesforce_large_data_volumes_bp/ldv_deployments_introduction.htm
  • #5: Consequences! Let’s look at some of the typical types of design and potential problems. Why this is important! Worst agile design oversight ever: Continuous slight changes to the architecture without foresight led to a Surprise 10x increase in cost to client. Which led to a Surprise “escort out of the building” and “termination of contract”.
  • #6: Maximum of 100 validation rules on an object. Ever try to create test data on an object with 100 validation rules? Think the designer PLANNED for 100 rules? Probably not. Audience participation: on what you would have to do to work around 100 validation rules. Let’s take phone numbers. You need some validation rules 10 digits Xxx-xxx-xxxx Area code matches valid area code for Account.State lookup. Number not on no call list Number not clearly fake: repeated digits. I’m exhausted at 5, can you imagine what it would take to hit 101? You know why there’s a maximum of 100? Because some maniac tried to do business with 200!
  • #7: https://developer.salesforce.com/docs/atlas.en-us.salesforce_large_data_volumes_bp.meta/salesforce_large_data_volumes_bp/ldv_deployments_introduction.htm
  • #8: Is the schema too big to understand? Is the application too complex to be supported?
  • #9: Intro: simplification of Enterprise solution components Next: Patterns and Pitfalls
  • #10: Intro: simplification of Enterprise solution components Next: Patterns and Pitfalls
  • #11: Mature, Manageable, Scalable Architecture We start with this architecture in mind and very seldom devolve to simple (albeit cheaper) models. Our goal is to build a long-term thriving set of applications inside Salesforce and other subscription-based systems. Licensing a suite with only 1 application is a massive waste of funds. Build for scale in systems that are Built and Billed to scale.
  • #13: Story: Slashdotting
  • #14: When you let a business user design the schema
  • #15: Salesforce has put forth their own decision tree on big objects.
  • #16: Data Lakes store and allow access to multiple copies of source data in a single location. (often with cheaper per byte options.) spillover visual They reduce security headaches for consumers. Consider Data Lakes a pre-requisite if your organization is in any phase of a cloud migration. Cloud Data Lakes provide cloud consumer ready access for all platforms. If anything is going to ‘the cloud’ your integration data will probably need to go too. Best to execute one time and not reintegrate later, again.
  • #17: Table
  • #18: Mention data exchange and data marketplace as moving across boundaries. The second you want to use more than one of source data, go snowflake (and a terabyte)
  • #19: Courtesy Chris Rosser