SlideShare une entreprise Scribd logo
1  sur  22
PARALLEL
AGORITHMS
ABILITIES AND LIMITATIONS
MURTADHA AL-SABBAGH
PARALLEL ALGORITHMS ABILITES
THROGHPUT AND
LATENCY
• Throughput: Is the number of operations done per time
unit.
• Latency : Is the time needed to complete one operation.
EXAMPLE (THROUGHPUT
AND LATENCY)
Suppose we have two vehicles a car that can carry two
persons with an average speed of 200 km/h and a bus that can
carry 40 persons with a speed of 50 km/h , find the latency and
throughput if they both used to carry a group of people for a
distance of 4500 km ?
Car
• Latency =22.5
• Throughput=0.089
Bus
• Latency =90
• Throughput=0.45
THROUGHPUT AND
LATENCY
• We can use parallelism to increase throughput by using
a larger number of lower clocked processing units (as
in the GPU) which is well suited for computation intensive
applications ( applications with need of large number of
calculations such as image processing applications).
• Also we can use it in latency sensitive applications which
are applications with need of fast response (e.g.
Windows applications which many parts running together
at the same time). In this case we use a more powerful
processing units such as (Intel core i7).
• Throughput and latency often contradictory .
WHY USE PARALLEL
COMPUTING?
1) The Real World is Massively Parallel:
In the natural world, many complex, interrelated events are
happening at the same time, yet within a temporal sequence.
Compared to serial computing, parallel computing is much better
suited for modeling, simulating and understanding complex, real
world phenomena.
For example, imagine modeling these serially:
WHY USE PARALLEL
COMPUTING?
2)SAVE TIME AND/OR MONEY:
• In theory, throwing more resources at a task will shorten its time to
completion, with potential cost savings.
• Parallel computers can be built from cheap, commodity
components.
3) SOLVE LARGER / MORE COMPLEX PROBLEMS:
• Many problems are so large and/or complex that it is impractical
or impossible to solve them on a single computer, especially
given limited computer memory.
• Example: Web search engines/databases processing millions of
transactions per second
• Using compute resources on a wide area network, or even the
Internet when local compute resources are scarce or insufficient.
4) TAKE ADVANTAGE OF NON-LOCAL RESOURCES:
PARALLEL ALGORITHMS LIMITS
THE LIMITATIONS
We Face the following limitations when designing a parallel
program:
1. Amadahl’s law.
2. Complexity.
3. Portability.
4. Resource Requirements.
5. Scalability.
6. Parallel Slowdown
AMDAHL’S LAW
Amdahl’s law provides an upper bound on the speedup that can be
obtained by a parallel program
Let us suppose that a processor need an ‘n’ time units to complete ‘n’
operations .
And we have a program that have a ‘p’ part of parallizable code and ‘1-p’
or ‘s’ un-paralleizable part of code.
Then the processing time for a ‘N’ number of processors is :
𝑻 = 𝒏𝒔 +
𝒏𝒑
𝑵
=𝒏 𝟏 − 𝒑 +
𝒏𝒑
𝑵
And hence the speed up ratio will be calculated as a ratio between a
single processor time ‘Ts’ and an ‘N’ processors processing time ‘Tp’.
𝑺𝒑𝒆𝒆𝒅 𝒖𝒑 =
𝒏
𝒏 𝟏 − 𝒑 +
𝒏𝒑
𝑵
=
𝟏
𝟏 − 𝒑 +
𝒑
𝑵
AMDAHL’S LAW
Introducing the number of processors performing the parallel
fraction of work, the relationship can be modeled by:
𝑺𝒑𝒆𝒆𝒅 𝒖𝒑 =
𝟏
(𝟏 − 𝒑) +
𝒑
𝑵
N = number of processors and S = serial fraction
AMDAHL’S LAW
If we consider N is very large then :
𝑺𝒑𝒆𝒆𝒅 𝒖𝒑 = 𝟏
(𝟏 − 𝑷)
COMPLEXITY
The costs of complexity are measured in every aspect of the
software development cycle:
1. Design
2. Coding
• Data dependency
• Race conditions
3. Debugging
4. Tuning
5. Maintenance
DATA DEPENDENCY
Results from multiple use of the same location(s) in storage by
different tasks.
e.g.
for (int i=0;i<100;i++)
array[i]=array[i-1]*20;
How to Handle Data Dependencies:
• Distributed memory architectures - communicate required data at
synchronization points.
• Shared memory architectures -synchronize read/write operations
between tasks.
RACE CONDITION
Thread A Thread B
1A: Read variable V 1B: Read variable V
2A: Add 1 to variable V 2B: Add 1 to variable V
3A: Write back to variable V 3B: Write back to variable V
If instruction 1B is executed between 1A and 3A, or if
instruction 1A is executed between 1B and 3B, the program
will produce incorrect data. This is known as a race
condition.
SYCNCHRONIZATION
• Our program often requires "serialization" of segments of
the program.
• Synchronization is the solution for the above problems.
• To implement synchronization we use:
• Barrier .
• Locks .
• Synchronous communication operations .
PORTABILITY
Although there is several standardization APIs, such as MPI,
POSIX threads, and OpenMP, portability issues with parallel
programs are not as serious as in years past. However...
Operating systems can play a key role in code portability issues.
Hardware architectures are characteristically highly variable and
can affect portability.
RESOURCE
REQUIREMENTS
• The primary intent of parallel programming is to decrease
execution wall clock time, however in order to accomplish
this, more CPU time is required. For example, a parallel code
that runs in 1 hour on 8 processors actually uses 8 hours of
CPU time.
• The amount of memory required can be greater for parallel
codes than serial codes, due to the need to replicate data and
for overheads associated with parallel support libraries and
subsystems.
• For short running parallel programs, there can actually be a
decrease in performance compared to a similar serial
implementation. The overhead costs associated with setting
up the parallel environment, task creation, communications
and task termination can comprise a significant portion of the
total execution time for short runs.
SCALABILITY
Two types of scaling based on time to solution:
• Strong scaling: The total problem size stays fixed as more
processors are added.
• Weak scaling: The problem size per processor stays fixed as
more processors are added.
Hardware factors play a significant role in scalability. Examples:
• Memory-CPU bus bandwidth
• Amount of memory available on any given machine or set
of machines
• Processor clock speed
PARALLEL
SLOWDOWN
• Not all parallelization results in speed-up.
• Once task split up into multiple threads those threads
spend a large amount of time communicating among each
other resulting degradation in the system.
• This is known as parallel slowdown.
THANK YOU ~_^

Contenu connexe

Tendances

Tendances (20)

Memory management
Memory managementMemory management
Memory management
 
Lecture 4 principles of parallel algorithm design updated
Lecture 4   principles of parallel algorithm design updatedLecture 4   principles of parallel algorithm design updated
Lecture 4 principles of parallel algorithm design updated
 
Building Aneka clouds.ppt
Building Aneka clouds.pptBuilding Aneka clouds.ppt
Building Aneka clouds.ppt
 
Distributed database
Distributed databaseDistributed database
Distributed database
 
Parallel and distributed Computing
Parallel and distributed Computing Parallel and distributed Computing
Parallel and distributed Computing
 
Distributed System ppt
Distributed System pptDistributed System ppt
Distributed System ppt
 
Underlying principles of parallel and distributed computing
Underlying principles of parallel and distributed computingUnderlying principles of parallel and distributed computing
Underlying principles of parallel and distributed computing
 
Transaction management DBMS
Transaction  management DBMSTransaction  management DBMS
Transaction management DBMS
 
Computer system architecture
Computer system architectureComputer system architecture
Computer system architecture
 
Unit 4 -Cloud Computing and security
Unit 4 -Cloud Computing and securityUnit 4 -Cloud Computing and security
Unit 4 -Cloud Computing and security
 
Black & White Box testing
Black & White Box testingBlack & White Box testing
Black & White Box testing
 
Database Performance Tuning
Database Performance Tuning Database Performance Tuning
Database Performance Tuning
 
Types of cloud computing
Types of cloud computingTypes of cloud computing
Types of cloud computing
 
Parallel Algorithms
Parallel AlgorithmsParallel Algorithms
Parallel Algorithms
 
Parallel Database
Parallel DatabaseParallel Database
Parallel Database
 
Virtualization in cloud computing
Virtualization in cloud computingVirtualization in cloud computing
Virtualization in cloud computing
 
Parallel computing
Parallel computingParallel computing
Parallel computing
 
distributed Computing system model
distributed Computing system modeldistributed Computing system model
distributed Computing system model
 
Virtualization in green computing
Virtualization in green computingVirtualization in green computing
Virtualization in green computing
 
Unit 5 Advanced Computer Architecture
Unit 5 Advanced Computer ArchitectureUnit 5 Advanced Computer Architecture
Unit 5 Advanced Computer Architecture
 

Similaire à Parallel Algorithms Advantages and Disadvantages

System models for distributed and cloud computing
System models for distributed and cloud computingSystem models for distributed and cloud computing
System models for distributed and cloud computing
purplesea
 
Performance Tuning by Dijesh P
Performance Tuning by Dijesh PPerformance Tuning by Dijesh P
Performance Tuning by Dijesh P
PlusOrMinusZero
 
Task allocation on many core-multi processor distributed system
Task allocation on many core-multi processor distributed systemTask allocation on many core-multi processor distributed system
Task allocation on many core-multi processor distributed system
Deepak Shankar
 

Similaire à Parallel Algorithms Advantages and Disadvantages (20)

System models for distributed and cloud computing
System models for distributed and cloud computingSystem models for distributed and cloud computing
System models for distributed and cloud computing
 
Simulation of Heterogeneous Cloud Infrastructures
Simulation of Heterogeneous Cloud InfrastructuresSimulation of Heterogeneous Cloud Infrastructures
Simulation of Heterogeneous Cloud Infrastructures
 
Cloud computing system models for distributed and cloud computing
Cloud computing system models for distributed and cloud computingCloud computing system models for distributed and cloud computing
Cloud computing system models for distributed and cloud computing
 
Cloud Computing System models for Distributed and cloud computing & Performan...
Cloud Computing System models for Distributed and cloud computing & Performan...Cloud Computing System models for Distributed and cloud computing & Performan...
Cloud Computing System models for Distributed and cloud computing & Performan...
 
Scalable analytics for iaas cloud availability
Scalable analytics for iaas cloud availabilityScalable analytics for iaas cloud availability
Scalable analytics for iaas cloud availability
 
Lecture1
Lecture1Lecture1
Lecture1
 
ICS 2410.Parallel.Sytsems.Lecture.Week 3.week5.pptx
ICS 2410.Parallel.Sytsems.Lecture.Week 3.week5.pptxICS 2410.Parallel.Sytsems.Lecture.Week 3.week5.pptx
ICS 2410.Parallel.Sytsems.Lecture.Week 3.week5.pptx
 
Unit 1 Computer organization and Instructions
Unit 1 Computer organization and InstructionsUnit 1 Computer organization and Instructions
Unit 1 Computer organization and Instructions
 
Multicore_Architecture Book.pdf
Multicore_Architecture Book.pdfMulticore_Architecture Book.pdf
Multicore_Architecture Book.pdf
 
Resisting skew accumulation
Resisting skew accumulationResisting skew accumulation
Resisting skew accumulation
 
5.7 Parallel Processing - Reactive Programming.pdf.pptx
5.7 Parallel Processing - Reactive Programming.pdf.pptx5.7 Parallel Processing - Reactive Programming.pdf.pptx
5.7 Parallel Processing - Reactive Programming.pdf.pptx
 
Parallel Computing-Part-1.pptx
Parallel Computing-Part-1.pptxParallel Computing-Part-1.pptx
Parallel Computing-Part-1.pptx
 
01-MessagePassingFundamentals.ppt
01-MessagePassingFundamentals.ppt01-MessagePassingFundamentals.ppt
01-MessagePassingFundamentals.ppt
 
Parallel Computing - Lec 6
Parallel Computing - Lec 6Parallel Computing - Lec 6
Parallel Computing - Lec 6
 
Parallel Computing - Lec 5
Parallel Computing - Lec 5Parallel Computing - Lec 5
Parallel Computing - Lec 5
 
Performance Tuning by Dijesh P
Performance Tuning by Dijesh PPerformance Tuning by Dijesh P
Performance Tuning by Dijesh P
 
Task allocation on many core-multi processor distributed system
Task allocation on many core-multi processor distributed systemTask allocation on many core-multi processor distributed system
Task allocation on many core-multi processor distributed system
 
Chap2 slides
Chap2 slidesChap2 slides
Chap2 slides
 
3. Potential Benefits, Limits and Costs of Parallel Programming.pdf
3. Potential Benefits, Limits and Costs of Parallel Programming.pdf3. Potential Benefits, Limits and Costs of Parallel Programming.pdf
3. Potential Benefits, Limits and Costs of Parallel Programming.pdf
 
Chap 2 classification of parralel architecture and introduction to parllel p...
Chap 2  classification of parralel architecture and introduction to parllel p...Chap 2  classification of parralel architecture and introduction to parllel p...
Chap 2 classification of parralel architecture and introduction to parllel p...
 

Dernier

+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
?#DUbAI#??##{{(☎️+971_581248768%)**%*]'#abortion pills for sale in dubai@
 

Dernier (20)

Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 
Navi Mumbai Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Navi Mumbai Call Girls 🥰 8617370543 Service Offer VIP Hot ModelNavi Mumbai Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Navi Mumbai Call Girls 🥰 8617370543 Service Offer VIP Hot Model
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Ransomware_Q4_2023. The report. [EN].pdf
Ransomware_Q4_2023. The report. [EN].pdfRansomware_Q4_2023. The report. [EN].pdf
Ransomware_Q4_2023. The report. [EN].pdf
 
Manulife - Insurer Transformation Award 2024
Manulife - Insurer Transformation Award 2024Manulife - Insurer Transformation Award 2024
Manulife - Insurer Transformation Award 2024
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
AXA XL - Insurer Innovation Award Americas 2024
AXA XL - Insurer Innovation Award Americas 2024AXA XL - Insurer Innovation Award Americas 2024
AXA XL - Insurer Innovation Award Americas 2024
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?
 
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
Apidays New York 2024 - Accelerating FinTech Innovation by Vasa Krishnan, Fin...
 
DBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor Presentation
 
Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
 
A Beginners Guide to Building a RAG App Using Open Source Milvus
A Beginners Guide to Building a RAG App Using Open Source MilvusA Beginners Guide to Building a RAG App Using Open Source Milvus
A Beginners Guide to Building a RAG App Using Open Source Milvus
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 

Parallel Algorithms Advantages and Disadvantages

  • 3. THROGHPUT AND LATENCY • Throughput: Is the number of operations done per time unit. • Latency : Is the time needed to complete one operation.
  • 4. EXAMPLE (THROUGHPUT AND LATENCY) Suppose we have two vehicles a car that can carry two persons with an average speed of 200 km/h and a bus that can carry 40 persons with a speed of 50 km/h , find the latency and throughput if they both used to carry a group of people for a distance of 4500 km ? Car • Latency =22.5 • Throughput=0.089 Bus • Latency =90 • Throughput=0.45
  • 5. THROUGHPUT AND LATENCY • We can use parallelism to increase throughput by using a larger number of lower clocked processing units (as in the GPU) which is well suited for computation intensive applications ( applications with need of large number of calculations such as image processing applications). • Also we can use it in latency sensitive applications which are applications with need of fast response (e.g. Windows applications which many parts running together at the same time). In this case we use a more powerful processing units such as (Intel core i7). • Throughput and latency often contradictory .
  • 6. WHY USE PARALLEL COMPUTING? 1) The Real World is Massively Parallel: In the natural world, many complex, interrelated events are happening at the same time, yet within a temporal sequence. Compared to serial computing, parallel computing is much better suited for modeling, simulating and understanding complex, real world phenomena.
  • 7. For example, imagine modeling these serially:
  • 8. WHY USE PARALLEL COMPUTING? 2)SAVE TIME AND/OR MONEY: • In theory, throwing more resources at a task will shorten its time to completion, with potential cost savings. • Parallel computers can be built from cheap, commodity components. 3) SOLVE LARGER / MORE COMPLEX PROBLEMS: • Many problems are so large and/or complex that it is impractical or impossible to solve them on a single computer, especially given limited computer memory. • Example: Web search engines/databases processing millions of transactions per second • Using compute resources on a wide area network, or even the Internet when local compute resources are scarce or insufficient. 4) TAKE ADVANTAGE OF NON-LOCAL RESOURCES:
  • 10. THE LIMITATIONS We Face the following limitations when designing a parallel program: 1. Amadahl’s law. 2. Complexity. 3. Portability. 4. Resource Requirements. 5. Scalability. 6. Parallel Slowdown
  • 11. AMDAHL’S LAW Amdahl’s law provides an upper bound on the speedup that can be obtained by a parallel program Let us suppose that a processor need an ‘n’ time units to complete ‘n’ operations . And we have a program that have a ‘p’ part of parallizable code and ‘1-p’ or ‘s’ un-paralleizable part of code. Then the processing time for a ‘N’ number of processors is : 𝑻 = 𝒏𝒔 + 𝒏𝒑 𝑵 =𝒏 𝟏 − 𝒑 + 𝒏𝒑 𝑵 And hence the speed up ratio will be calculated as a ratio between a single processor time ‘Ts’ and an ‘N’ processors processing time ‘Tp’. 𝑺𝒑𝒆𝒆𝒅 𝒖𝒑 = 𝒏 𝒏 𝟏 − 𝒑 + 𝒏𝒑 𝑵 = 𝟏 𝟏 − 𝒑 + 𝒑 𝑵
  • 12. AMDAHL’S LAW Introducing the number of processors performing the parallel fraction of work, the relationship can be modeled by: 𝑺𝒑𝒆𝒆𝒅 𝒖𝒑 = 𝟏 (𝟏 − 𝒑) + 𝒑 𝑵 N = number of processors and S = serial fraction
  • 13. AMDAHL’S LAW If we consider N is very large then : 𝑺𝒑𝒆𝒆𝒅 𝒖𝒑 = 𝟏 (𝟏 − 𝑷)
  • 14. COMPLEXITY The costs of complexity are measured in every aspect of the software development cycle: 1. Design 2. Coding • Data dependency • Race conditions 3. Debugging 4. Tuning 5. Maintenance
  • 15. DATA DEPENDENCY Results from multiple use of the same location(s) in storage by different tasks. e.g. for (int i=0;i<100;i++) array[i]=array[i-1]*20; How to Handle Data Dependencies: • Distributed memory architectures - communicate required data at synchronization points. • Shared memory architectures -synchronize read/write operations between tasks.
  • 16. RACE CONDITION Thread A Thread B 1A: Read variable V 1B: Read variable V 2A: Add 1 to variable V 2B: Add 1 to variable V 3A: Write back to variable V 3B: Write back to variable V If instruction 1B is executed between 1A and 3A, or if instruction 1A is executed between 1B and 3B, the program will produce incorrect data. This is known as a race condition.
  • 17. SYCNCHRONIZATION • Our program often requires "serialization" of segments of the program. • Synchronization is the solution for the above problems. • To implement synchronization we use: • Barrier . • Locks . • Synchronous communication operations .
  • 18. PORTABILITY Although there is several standardization APIs, such as MPI, POSIX threads, and OpenMP, portability issues with parallel programs are not as serious as in years past. However... Operating systems can play a key role in code portability issues. Hardware architectures are characteristically highly variable and can affect portability.
  • 19. RESOURCE REQUIREMENTS • The primary intent of parallel programming is to decrease execution wall clock time, however in order to accomplish this, more CPU time is required. For example, a parallel code that runs in 1 hour on 8 processors actually uses 8 hours of CPU time. • The amount of memory required can be greater for parallel codes than serial codes, due to the need to replicate data and for overheads associated with parallel support libraries and subsystems. • For short running parallel programs, there can actually be a decrease in performance compared to a similar serial implementation. The overhead costs associated with setting up the parallel environment, task creation, communications and task termination can comprise a significant portion of the total execution time for short runs.
  • 20. SCALABILITY Two types of scaling based on time to solution: • Strong scaling: The total problem size stays fixed as more processors are added. • Weak scaling: The problem size per processor stays fixed as more processors are added. Hardware factors play a significant role in scalability. Examples: • Memory-CPU bus bandwidth • Amount of memory available on any given machine or set of machines • Processor clock speed
  • 21. PARALLEL SLOWDOWN • Not all parallelization results in speed-up. • Once task split up into multiple threads those threads spend a large amount of time communicating among each other resulting degradation in the system. • This is known as parallel slowdown.

Notes de l'éditeur

  1. The Same law in the next page except 8