SlideShare a Scribd company logo
Multicore Computers
Hardware Performance Issues
• Microprocessors have seen an exponential increase in performance
• Improved organization
• Increased clock frequency
• Increase in Parallelism
• Pipelining
• Superscalar
• Simultaneous multithreading (SMT)
• Diminishing returns
• More complexity requires more logic
• Increasing chip area for coordinating and signal transfer logic
• Harder to design, make and debug
Alternative Chip
Organizations
Intel Hardware
Trends
Increased Complexity
• Power requirements grow exponentially with chip density
and clock frequency
• Can use more chip area for cache
• Smaller
• Order of magnitude lower power requirements
• By 2015
• 100 billion transistors on 300mm2
die
• Cache of 100MB
• 1 billion transistors for logic
• Pollack’s rule:
• Performance is roughly proportional to square root of increase in
complexity
• Double complexity gives 40% more performance
• Multicore has potential for near-linear improvement
• Unlikely that one core can use all cache effectively
Power and Memory Considerations
Chip Utilization of Transistors
Software Performance Issues
• Performance benefits dependent on effective exploitation of parallel
resources
• Even small amounts of serial code impact performance
• 10% inherently serial on 8 processor system gives only 4.7 times
performance
• Communication, distribution of work and cache coherence
overheads
• Some applications effectively exploit multicore processors
Effective Applications for Multicore Processors
• Database
• Servers handling independent transactions
• Multi-threaded native applications
• Lotus Domino, Siebel CRM
• Multi-process applications
• Oracle, SAP, PeopleSoft
• Java applications
• Java VM is multi-thread with scheduling and memory management
• Sun’s Java Application Server, BEA’s Weblogic, IBM Websphere, Tomcat
• Multi-instance applications
• One application running multiple times
• E.g. Value Game Software
Multicore Organization• Number of core processors on chip
• Number of levels of cache on chip
• Amount of shared cache
• Next slide examples of each organization:
• (a) ARM11 MPCore
• (b) AMD Opteron
• (c) Intel Core Duo
• (d) Intel Core i7
Multicore Organization Alternatives
Advantages of shared L2 Cache
• Constructive interference reduces overall miss rate
• Data shared by multiple cores not replicated at cache level
• With proper frame replacement algorithms mean amount of
shared cache dedicated to each core is dynamic
• Threads with less locality can have more cache
• Easy inter-process communication through shared memory
• Cache coherency confined to L1
• Dedicated L2 cache gives each core more rapid access
• Good for threads with strong locality
• Shared L3 cache may also improve performance
Individual Core Architecture
• Intel Core Duo uses superscalar cores
• Intel Core i7 uses simultaneous multi-threading (SMT)
• Scales up number of threads supported
• 4 SMT cores, each supporting 4 threads appears as 16 core
Intel x86 Multicore Organization -
Core Duo (1)
• 2006
• Two x86 superscalar, shared L2 cache
• Dedicated L1 cache per core
• 32KB instruction and 32KB data
• Thermal control unit per core
• Manages chip heat dissipation
• Maximize performance within constraints
• Improved ergonomics
• Advanced Programmable Interrupt Controlled (APIC)
• Inter-process interrupts between cores
• Routes interrupts to appropriate core
• Includes timer so OS can interrupt core
Intel x86 Multicore Organization -
Core Duo (2)
• Power Management Logic
• Monitors thermal conditions and CPU activity
• Adjusts voltage and power consumption
• Can switch individual logic subsystems
• 2MB shared L2 cache
• Dynamic allocation
• MESI support for L1 caches
• Extended to support multiple Core Duo in SMP
• L2 data shared between local cores or external
• Bus interface
Intel x86 Multicore Organization -
Core i7
• November 2008
• Four x86 SMT processors
• Dedicated L2, shared L3 cache
• Speculative pre-fetch for caches
• On chip DDR3 memory controller
• Three 8 byte channels (192 bits) giving 32GB/s
• No front side bus
• QuickPath Interconnection
• Cache coherent point-to-point link
• High speed communications between processor chips
• 6.4G transfers per second, 16 bits per transfer
• Dedicated bi-directional pairs
• Total bandwidth 25.6GB/s
ARM11 MPCore
• Up to 4 processors each with own L1 instruction and data cache
• Distributed interrupt controller
• Timer per CPU
• Watchdog
• Warning alerts for software failures
• Counts down from predetermined values
• Issues warning at zero
• CPU interface
• Interrupt acknowledgement, masking and completion acknowledgement
• CPU
• Single ARM11 called MP11
• Vector floating-point unit
• FP co-processor
• L1 cache
• Snoop control unit
• L1 cache coherency
ARM11
MPCore
Block
Diagram
ARM11 MPCore Interrupt Handling
• Distributed Interrupt Controller (DIC) collates from many sources
• Masking
• Prioritization
• Distribution to target MP11 CPUs
• Status tracking
• Software interrupt generation
• Number of interrupts independent of MP11 CPU design
• Memory mapped
• Accessed by CPUs via private interface through SCU
• Can route interrupts to single or multiple CPUs
• Provides inter-process communication
• Thread on one CPU can cause activity by thread on another CPU
DIC Routing
• Direct to specific CPU
• To defined group of CPUs
• To all CPUs
• OS can generate interrupt to:
• All but self
• Self
• Other specific CPU
• Typically combined with shared memory for inter-process
communication
• 16 interrupt ids available for inter-process communication
Interrupt States
• Inactive
• Non-asserted
• Completed by that CPU but pending or active in others
• Pending
• Asserted
• Processing not started on that CPU
• Active
• Started on that CPU but not complete
• Can be pre-empted by higher priority interrupt
Interrupt Sources
• Inter-process Interrupts (IPI)
• Private to CPU
• ID0-ID15
• Software triggered
• Priority depends on target CPU not source
• Private timer and/or watchdog interrupt
• ID29 and ID30
• Legacy FIQ line
• Legacy FIQ pin, per CPU, bypasses interrupt distributor
• Directly drives interrupts to CPU
• Hardware
• Triggered by programmable events on associated interrupt lines
• Up to 224 lines
• Start at ID32
ARM11 MPCore Interrupt Distributor
Cache Coherency
• Snoop Control Unit (SCU) resolves most shared data bottleneck
issues
• L1 cache coherency based on MESI
• Direct data Intervention
• Copying clean entries between L1 caches without accessing external
memory
• Reduces read after write from L1 to L2
• Can resolve local L1 miss from rmote L1 rather than L2
• Duplicated tag RAMs
• Cache tags implemented as separate block of RAM
• Same length as number of lines in cache
• Duplicates used by SCU to check data availability before sending
coherency commands
• Only send to CPUs that must update coherent data cache
• Migratory lines
• Allows moving dirty data between CPUs without writing to L2 and
reading back from external memory
Recommended Reading
• Stallings chapter 18
• ARM web site
Intel Core i& Block Diagram
Intel Core Duo Block Diagram
Performance Effect of Multiple Cores
Recommended Reading
• Multicore Association web site
• ARM web site

More Related Content

What's hot (20)

PPTX
Multi core processor
Muhammad Ishaq
 
PDF
Introduction to Parallel Computing
Akhila Prabhakaran
 
PDF
ARM Architecture
Dwight Sabio
 
PPT
Branch prediction
Aneesh Raveendran
 
PDF
Linux-Internals-and-Networking
Emertxe Information Technologies Pvt Ltd
 
PDF
File System Implementation - Part1
Amir Payberah
 
PPTX
CXL Consortium Update: Advancing Coherent Connectivity
Memory Fabric Forum
 
PPTX
Single &Multi Core processor
Justify Shadap
 
PPTX
Multi processor scheduling
Shashank Kapoor
 
PPT
Intel Core i7 Processors
Anagh Vijayvargia
 
PPTX
Multicore processor by Ankit Raj and Akash Prajapati
Ankit Raj
 
PPTX
Processor types
Amr Aboelgood
 
PPT
Multi-core architectures
nextlib
 
PPT
6 multiprogramming & time sharing
myrajendra
 
PPTX
Message Passing, Remote Procedure Calls and Distributed Shared Memory as Com...
Sehrish Asif
 
PPTX
Single and Multi core processor
Munaam Munawar
 
PPTX
Cache memory principles
bit allahabad
 
PPTX
parallel language and compiler
Vignesh Tamil
 
PPTX
Compare Performance-power of Arm Cortex vs RISC-V for AI applications_oct_2021
Deepak Shankar
 
PDF
Unit II - 2 - Operating System - Threads
cscarcas
 
Multi core processor
Muhammad Ishaq
 
Introduction to Parallel Computing
Akhila Prabhakaran
 
ARM Architecture
Dwight Sabio
 
Branch prediction
Aneesh Raveendran
 
Linux-Internals-and-Networking
Emertxe Information Technologies Pvt Ltd
 
File System Implementation - Part1
Amir Payberah
 
CXL Consortium Update: Advancing Coherent Connectivity
Memory Fabric Forum
 
Single &Multi Core processor
Justify Shadap
 
Multi processor scheduling
Shashank Kapoor
 
Intel Core i7 Processors
Anagh Vijayvargia
 
Multicore processor by Ankit Raj and Akash Prajapati
Ankit Raj
 
Processor types
Amr Aboelgood
 
Multi-core architectures
nextlib
 
6 multiprogramming & time sharing
myrajendra
 
Message Passing, Remote Procedure Calls and Distributed Shared Memory as Com...
Sehrish Asif
 
Single and Multi core processor
Munaam Munawar
 
Cache memory principles
bit allahabad
 
parallel language and compiler
Vignesh Tamil
 
Compare Performance-power of Arm Cortex vs RISC-V for AI applications_oct_2021
Deepak Shankar
 
Unit II - 2 - Operating System - Threads
cscarcas
 

Similar to Multicore computers (20)

PPT
chap 18 multicore computers
Sher Shah Merkhel
 
PPTX
Intel® hyper threading technology
Amirali Sharifian
 
PPTX
Architecture of high end processors
University of Gujrat, Pakistan
 
PPTX
CPU Caches
shinolajla
 
PPTX
trends of microprocessor field
Ramya SK
 
PDF
5_Embedded Systems مختصر.pdf
aliamjd
 
PDF
Motivation for multithreaded architectures
Young Alista
 
PDF
Challenges in Embedded Computing
Pradeep Kumar TS
 
PPTX
Computer System Architecture Lecture Note 8.1 primary Memory
Budditha Hettige
 
PPTX
Central Processing Unit
Alaka Acharya
 
PPTX
Computer Organization: Introduction to Microprocessor and Microcontroller
AmrutaMehata
 
PDF
Exaflop In 2018 Hardware
Jacob Wu
 
PDF
This is Unit 1 of High Performance Computing For SRM students
cegafen778
 
PPTX
Microprocessor
Anand Tiwari
 
PPT
Sparc t3 2 technical presentation
solarisyougood
 
PDF
Multithreaded processors ppt
Siddhartha Anand
 
PPT
Computer Evolution.ppt
VivekTrial
 
PDF
MK Sistem Operasi.pdf
wisard1
 
PPTX
Microprocessor.ppt
safia kalwar
 
PPT
Chapter4 Data Processing
Muhammad Waqas
 
chap 18 multicore computers
Sher Shah Merkhel
 
Intel® hyper threading technology
Amirali Sharifian
 
Architecture of high end processors
University of Gujrat, Pakistan
 
CPU Caches
shinolajla
 
trends of microprocessor field
Ramya SK
 
5_Embedded Systems مختصر.pdf
aliamjd
 
Motivation for multithreaded architectures
Young Alista
 
Challenges in Embedded Computing
Pradeep Kumar TS
 
Computer System Architecture Lecture Note 8.1 primary Memory
Budditha Hettige
 
Central Processing Unit
Alaka Acharya
 
Computer Organization: Introduction to Microprocessor and Microcontroller
AmrutaMehata
 
Exaflop In 2018 Hardware
Jacob Wu
 
This is Unit 1 of High Performance Computing For SRM students
cegafen778
 
Microprocessor
Anand Tiwari
 
Sparc t3 2 technical presentation
solarisyougood
 
Multithreaded processors ppt
Siddhartha Anand
 
Computer Evolution.ppt
VivekTrial
 
MK Sistem Operasi.pdf
wisard1
 
Microprocessor.ppt
safia kalwar
 
Chapter4 Data Processing
Muhammad Waqas
 
Ad

More from Syed Zaid Irshad (20)

PDF
Data Structures & Algorithms - Spring 2025.pdf
Syed Zaid Irshad
 
PDF
Operating System.pdf
Syed Zaid Irshad
 
PDF
DBMS_Lab_Manual_&_Solution
Syed Zaid Irshad
 
PPTX
Data Structure and Algorithms.pptx
Syed Zaid Irshad
 
PPTX
Design and Analysis of Algorithms.pptx
Syed Zaid Irshad
 
PPTX
Professional Issues in Computing
Syed Zaid Irshad
 
PDF
Reduce course notes class xi
Syed Zaid Irshad
 
PDF
Reduce course notes class xii
Syed Zaid Irshad
 
PDF
Introduction to Database
Syed Zaid Irshad
 
PDF
C Language
Syed Zaid Irshad
 
PDF
Flowchart
Syed Zaid Irshad
 
PDF
Algorithm Pseudo
Syed Zaid Irshad
 
PDF
Computer Programming
Syed Zaid Irshad
 
PDF
ICS 2nd Year Book Introduction
Syed Zaid Irshad
 
PDF
Security, Copyright and the Law
Syed Zaid Irshad
 
PDF
Computer Architecture
Syed Zaid Irshad
 
PDF
Data Communication
Syed Zaid Irshad
 
PDF
Information Networks
Syed Zaid Irshad
 
PDF
Basic Concept of Information Technology
Syed Zaid Irshad
 
PDF
Introduction to ICS 1st Year Book
Syed Zaid Irshad
 
Data Structures & Algorithms - Spring 2025.pdf
Syed Zaid Irshad
 
Operating System.pdf
Syed Zaid Irshad
 
DBMS_Lab_Manual_&_Solution
Syed Zaid Irshad
 
Data Structure and Algorithms.pptx
Syed Zaid Irshad
 
Design and Analysis of Algorithms.pptx
Syed Zaid Irshad
 
Professional Issues in Computing
Syed Zaid Irshad
 
Reduce course notes class xi
Syed Zaid Irshad
 
Reduce course notes class xii
Syed Zaid Irshad
 
Introduction to Database
Syed Zaid Irshad
 
C Language
Syed Zaid Irshad
 
Flowchart
Syed Zaid Irshad
 
Algorithm Pseudo
Syed Zaid Irshad
 
Computer Programming
Syed Zaid Irshad
 
ICS 2nd Year Book Introduction
Syed Zaid Irshad
 
Security, Copyright and the Law
Syed Zaid Irshad
 
Computer Architecture
Syed Zaid Irshad
 
Data Communication
Syed Zaid Irshad
 
Information Networks
Syed Zaid Irshad
 
Basic Concept of Information Technology
Syed Zaid Irshad
 
Introduction to ICS 1st Year Book
Syed Zaid Irshad
 
Ad

Recently uploaded (20)

PPTX
Hashing Introduction , hash functions and techniques
sailajam21
 
PPTX
GitOps_Without_K8s_Training_detailed git repository
DanialHabibi2
 
PPTX
GitOps_Repo_Structure for begeinner(Scaffolindg)
DanialHabibi2
 
PPTX
265587293-NFPA 101 Life safety code-PPT-1.pptx
chandermwason
 
PPTX
Lecture 1 Shell and Tube Heat exchanger-1.pptx
mailforillegalwork
 
PPTX
Shinkawa Proposal to meet Vibration API670.pptx
AchmadBashori2
 
PPTX
Thermal runway and thermal stability.pptx
godow93766
 
PPTX
Product Development & DevelopmentLecture02.pptx
zeeshanwazir2
 
PPTX
Depth First Search Algorithm in 🧠 DFS in Artificial Intelligence (AI)
rafeeqshaik212002
 
PPTX
Element 11. ELECTRICITY safety and hazards
merrandomohandas
 
PDF
Set Relation Function Practice session 24.05.2025.pdf
DrStephenStrange4
 
PPTX
Day2 B2 Best.pptx
helenjenefa1
 
PDF
Reasons for the succes of MENARD PRESSUREMETER.pdf
majdiamz
 
PPTX
VITEEE 2026 Exam Details , Important Dates
SonaliSingh127098
 
PPTX
Types of Bearing_Specifications_PPT.pptx
PranjulAgrahariAkash
 
PDF
Introduction to Productivity and Quality
মোঃ ফুরকান উদ্দিন জুয়েল
 
PDF
Design Thinking basics for Engineers.pdf
CMR University
 
PPTX
Mechanical Design of shell and tube heat exchangers as per ASME Sec VIII Divi...
shahveer210504
 
PDF
GTU Civil Engineering All Semester Syllabus.pdf
Vimal Bhojani
 
PPTX
Damage of stability of a ship and how its change .pptx
ehamadulhaque
 
Hashing Introduction , hash functions and techniques
sailajam21
 
GitOps_Without_K8s_Training_detailed git repository
DanialHabibi2
 
GitOps_Repo_Structure for begeinner(Scaffolindg)
DanialHabibi2
 
265587293-NFPA 101 Life safety code-PPT-1.pptx
chandermwason
 
Lecture 1 Shell and Tube Heat exchanger-1.pptx
mailforillegalwork
 
Shinkawa Proposal to meet Vibration API670.pptx
AchmadBashori2
 
Thermal runway and thermal stability.pptx
godow93766
 
Product Development & DevelopmentLecture02.pptx
zeeshanwazir2
 
Depth First Search Algorithm in 🧠 DFS in Artificial Intelligence (AI)
rafeeqshaik212002
 
Element 11. ELECTRICITY safety and hazards
merrandomohandas
 
Set Relation Function Practice session 24.05.2025.pdf
DrStephenStrange4
 
Day2 B2 Best.pptx
helenjenefa1
 
Reasons for the succes of MENARD PRESSUREMETER.pdf
majdiamz
 
VITEEE 2026 Exam Details , Important Dates
SonaliSingh127098
 
Types of Bearing_Specifications_PPT.pptx
PranjulAgrahariAkash
 
Introduction to Productivity and Quality
মোঃ ফুরকান উদ্দিন জুয়েল
 
Design Thinking basics for Engineers.pdf
CMR University
 
Mechanical Design of shell and tube heat exchangers as per ASME Sec VIII Divi...
shahveer210504
 
GTU Civil Engineering All Semester Syllabus.pdf
Vimal Bhojani
 
Damage of stability of a ship and how its change .pptx
ehamadulhaque
 

Multicore computers

  • 2. Hardware Performance Issues • Microprocessors have seen an exponential increase in performance • Improved organization • Increased clock frequency • Increase in Parallelism • Pipelining • Superscalar • Simultaneous multithreading (SMT) • Diminishing returns • More complexity requires more logic • Increasing chip area for coordinating and signal transfer logic • Harder to design, make and debug
  • 5. Increased Complexity • Power requirements grow exponentially with chip density and clock frequency • Can use more chip area for cache • Smaller • Order of magnitude lower power requirements • By 2015 • 100 billion transistors on 300mm2 die • Cache of 100MB • 1 billion transistors for logic • Pollack’s rule: • Performance is roughly proportional to square root of increase in complexity • Double complexity gives 40% more performance • Multicore has potential for near-linear improvement • Unlikely that one core can use all cache effectively
  • 6. Power and Memory Considerations
  • 7. Chip Utilization of Transistors
  • 8. Software Performance Issues • Performance benefits dependent on effective exploitation of parallel resources • Even small amounts of serial code impact performance • 10% inherently serial on 8 processor system gives only 4.7 times performance • Communication, distribution of work and cache coherence overheads • Some applications effectively exploit multicore processors
  • 9. Effective Applications for Multicore Processors • Database • Servers handling independent transactions • Multi-threaded native applications • Lotus Domino, Siebel CRM • Multi-process applications • Oracle, SAP, PeopleSoft • Java applications • Java VM is multi-thread with scheduling and memory management • Sun’s Java Application Server, BEA’s Weblogic, IBM Websphere, Tomcat • Multi-instance applications • One application running multiple times • E.g. Value Game Software
  • 10. Multicore Organization• Number of core processors on chip • Number of levels of cache on chip • Amount of shared cache • Next slide examples of each organization: • (a) ARM11 MPCore • (b) AMD Opteron • (c) Intel Core Duo • (d) Intel Core i7
  • 12. Advantages of shared L2 Cache • Constructive interference reduces overall miss rate • Data shared by multiple cores not replicated at cache level • With proper frame replacement algorithms mean amount of shared cache dedicated to each core is dynamic • Threads with less locality can have more cache • Easy inter-process communication through shared memory • Cache coherency confined to L1 • Dedicated L2 cache gives each core more rapid access • Good for threads with strong locality • Shared L3 cache may also improve performance
  • 13. Individual Core Architecture • Intel Core Duo uses superscalar cores • Intel Core i7 uses simultaneous multi-threading (SMT) • Scales up number of threads supported • 4 SMT cores, each supporting 4 threads appears as 16 core
  • 14. Intel x86 Multicore Organization - Core Duo (1) • 2006 • Two x86 superscalar, shared L2 cache • Dedicated L1 cache per core • 32KB instruction and 32KB data • Thermal control unit per core • Manages chip heat dissipation • Maximize performance within constraints • Improved ergonomics • Advanced Programmable Interrupt Controlled (APIC) • Inter-process interrupts between cores • Routes interrupts to appropriate core • Includes timer so OS can interrupt core
  • 15. Intel x86 Multicore Organization - Core Duo (2) • Power Management Logic • Monitors thermal conditions and CPU activity • Adjusts voltage and power consumption • Can switch individual logic subsystems • 2MB shared L2 cache • Dynamic allocation • MESI support for L1 caches • Extended to support multiple Core Duo in SMP • L2 data shared between local cores or external • Bus interface
  • 16. Intel x86 Multicore Organization - Core i7 • November 2008 • Four x86 SMT processors • Dedicated L2, shared L3 cache • Speculative pre-fetch for caches • On chip DDR3 memory controller • Three 8 byte channels (192 bits) giving 32GB/s • No front side bus • QuickPath Interconnection • Cache coherent point-to-point link • High speed communications between processor chips • 6.4G transfers per second, 16 bits per transfer • Dedicated bi-directional pairs • Total bandwidth 25.6GB/s
  • 17. ARM11 MPCore • Up to 4 processors each with own L1 instruction and data cache • Distributed interrupt controller • Timer per CPU • Watchdog • Warning alerts for software failures • Counts down from predetermined values • Issues warning at zero • CPU interface • Interrupt acknowledgement, masking and completion acknowledgement • CPU • Single ARM11 called MP11 • Vector floating-point unit • FP co-processor • L1 cache • Snoop control unit • L1 cache coherency
  • 19. ARM11 MPCore Interrupt Handling • Distributed Interrupt Controller (DIC) collates from many sources • Masking • Prioritization • Distribution to target MP11 CPUs • Status tracking • Software interrupt generation • Number of interrupts independent of MP11 CPU design • Memory mapped • Accessed by CPUs via private interface through SCU • Can route interrupts to single or multiple CPUs • Provides inter-process communication • Thread on one CPU can cause activity by thread on another CPU
  • 20. DIC Routing • Direct to specific CPU • To defined group of CPUs • To all CPUs • OS can generate interrupt to: • All but self • Self • Other specific CPU • Typically combined with shared memory for inter-process communication • 16 interrupt ids available for inter-process communication
  • 21. Interrupt States • Inactive • Non-asserted • Completed by that CPU but pending or active in others • Pending • Asserted • Processing not started on that CPU • Active • Started on that CPU but not complete • Can be pre-empted by higher priority interrupt
  • 22. Interrupt Sources • Inter-process Interrupts (IPI) • Private to CPU • ID0-ID15 • Software triggered • Priority depends on target CPU not source • Private timer and/or watchdog interrupt • ID29 and ID30 • Legacy FIQ line • Legacy FIQ pin, per CPU, bypasses interrupt distributor • Directly drives interrupts to CPU • Hardware • Triggered by programmable events on associated interrupt lines • Up to 224 lines • Start at ID32
  • 23. ARM11 MPCore Interrupt Distributor
  • 24. Cache Coherency • Snoop Control Unit (SCU) resolves most shared data bottleneck issues • L1 cache coherency based on MESI • Direct data Intervention • Copying clean entries between L1 caches without accessing external memory • Reduces read after write from L1 to L2 • Can resolve local L1 miss from rmote L1 rather than L2 • Duplicated tag RAMs • Cache tags implemented as separate block of RAM • Same length as number of lines in cache • Duplicates used by SCU to check data availability before sending coherency commands • Only send to CPUs that must update coherent data cache • Migratory lines • Allows moving dirty data between CPUs without writing to L2 and reading back from external memory
  • 25. Recommended Reading • Stallings chapter 18 • ARM web site
  • 26. Intel Core i& Block Diagram
  • 27. Intel Core Duo Block Diagram
  • 28. Performance Effect of Multiple Cores
  • 29. Recommended Reading • Multicore Association web site • ARM web site