Choose a topic to test your knowledge and improve your High Performance Computing (HPC) skills
wimpleat way to send p-1 messages from source to the other p-1 processors
In a eight node ring, node ____ is source of broadcast
The processors compute ______ product of the vector element and the loval matrix
one to all broadcast use
In a broadcast and reduction on a balanced binary tree reduction is done in ______
if "X" is the message to broadcast it initially resides at the source node
logical operators used in algorithm are
Generalization of broadcast in Which each processor is
The algorithm terminates in _____ steps
Each node first sends to one of its neighbours the data it need to....
The second communication phase is a columnwise ______ broadcast of consolidated
All nodes collects _____ message corresponding to โp nodes to their respectively
It is not possible to port ____ for higher dimensional network
If we port algorithm to higher dimemsional network it would cause
In the scatter operation ____ node send message to every other node
The gather Operation is exactly the inverse of _____
Similar communication pattern to all-to-all broadcast except in the_____
Group communication operations are built using which primitives?
__ can be performed in an identical fashion by inverting the process.
Broadcast and reduction operations on a mesh is performed
Cost Analysis on a ring is
Cost Analysis on a mesh is
Communication between two directly link nodes
All-to-one communication (reduction) is the dual of ______ broadcast.
Which is known as Reduction?
Which is known as Broadcast?
The dual of all-to-all broadcast is
All-to-all broadcast algorithm for the 2D mesh is based on the
In the first phase of 2D Mesh All to All, the message size is ___
In the second phase of 2D Mesh All to All, the message size is ___
In All to All on Hypercube, The size of the message to be transmitted at the next step is ____ by concatenating the received message with their current data
The all-to-all broadcast on Hypercube needs ____ steps
One-to-All Personalized Communication operation is commonly called ___
The dual of the scatter operation is the
In Scatter Operation on Hypercube, on each step, the size of the messages communicated is ____
Which is also called "Total Exchange" ?
All-to-all personalized communication can be used in ____
In collective communication operations, collective means
efficiency of data parallel algorithm depends on the
All processes participate in a single ______ interaction operation.
subsets of processes in ______ interaction.
Goal of good algorithm is to implement commonly used _____ pattern.
Reduction can be used to find the sum, product, maximum, minimum of _____ of numbers.
source ____ is bottleneck.
only connections between single pairs of nodes are used at a time is
all processes that have the data can send it again is
The ____ do not snoop the messages going through them.
accumulate results and send with the same pattern is...
every node on the linear array has the data and broadcast on the columns with the linear array algorithm in _____
using different links every time and forwarding in parallel again is
In a balanced binary tree processing nodes is equal to
In one -to- all broadcast there is
For sake of simplicity, the number of nodes is a power of
Nides with zero in i least significant bits participate in _______
every node has to know when to communicate that is
the procedure is disturbed and require only point-to-point _______
Renaming relative to the source is _____ the source.
Task dependency graph is ------------------
In task dependency graph longest directed path between any pair of start and finish node is called as --------------
which of the following is not a granularity type
which of the following is a an example of data decomposition
which problems can be handled by recursive decomposition
In this decomposition problem decomposition goes hand in hand with its execution
the procedure is disturbed and require only point-to-point _______
Renaming relative to the source is _____ the source.
which of the following is not a granularity type
which of the following is not an example of explorative decomposition
In ------------task are defined before starting the execution of the algorithmting?
which of the following is not the array distribution method of data partitioning
blocking optimization is used to improve temmporal locality for reduce
CUDA thought that 'unifying theme' of every form of parallelism is
threads being block altogether and being executed in the sets of 32 threads called a
When the topological sort of a graph is unique?
What is a high performance multi-core processor that can be used to accelerate a wide variety of applications using parallel computing.
A good mapping does not depends on which following factor
Which of the following is not a form of parallelism supported by CUDA
The style of parallelism supported on GPUs is best described as
Which of the following correctly describes a GPU kernel
kernel may contain only host code
which of the following is not an parallel algorithm model