Wednesday 15 May 2013

AP9222 COMPUTER ARCHITECTURE AND PARALLEL PROCESSING ANNA UNIVERSITY QUESTION PAPERS

IMPORTANT QUESTIONS

AP9222 COMPUTER ARCHITECTURE AND PARALLEL PROCESSING


ANNA UNIVERSITY QUESTION PAPERS



Reg. No. :
M.E. DEGREE EXAMINATION, NOVEMBER/DECEMBER 2010
Second Semester
Applied Electronics
AP 9222 — COMPUTER ARCHITECTURE AND PARALLEL PROCESSING
(Common to M.E - Computer and Communication, M.E. - VLSI Design and
M.E.- Embedded System Technologies)
(Regulation 2009)
Time : Three hours Maximum : 100 Marks
Answer ALL questions
PART A — (10 × 2 = 20 Marks)
1. Compare temporal parallelism and data parallelism.
2. What is the major difference between multiprocessor and multicomputer?
3. Define anti dependence and output dependence with respect to parallelism and
dependence relations.
4. Compare control flow and data flow computers in terms of flow mechanism.
5. Why is memory hierarchy needed?
6. Define any four scalability merits for an application.
7. What is a symbolic processor?
8. State the instruction format used in VLIW process.
9. Name any two languages supporting parallelism.
10. Name the phases of parallel compiler.

PART B — (5 × 16 = 80 Marks)
11. (a) (i) Describe Flynn's classification of computers. (10)
(ii) Explain the super computer architecture for vector processing. (6)
Or
(b) (i) Explain the various theoretical models of parallel computers. How
are they useful in scalability and programmability analysis? (8)
(ii) Distinguish between multiprocessors and multi computers based on
their structures, resource sharing and inter process
communications. (8)
12. (a) (i) Explain the functional structure of a crossbar network and how I/O
connections are achieved? (8)
(ii) Explain the static data flow computer architecture with a diagram.
(8)
Or
(b) (i) Compare the features of static network with dynamic networks.
Explain how the interconnection is established for parallel
processing. (10)
(ii) Define the various speed up performance laws. (6)
13. (a) A two level memory system has five virtual pages on a disk to be mapped
into
(i) Two page frames
(ii) Three page frames in the main memory. A certain program
generated the following page trace.
0, 1, 2, 3, 4, 4, 4, 2, 1, 0, 2, 2, 2, 3, 0.
(1) Calculate the hit ratios using FIFO, LRU, circular FIFO and
LFU replacement policies on both cases.
(2) Indicate which page replacement policy is optimal on both
cases.
(3) Calculate the improvement or reduction in hit ratio on
increase in page frames in main memory. (16)
Or
(b) (i) Describe a four level memory hierarchy technology. (8)
(ii) Explain temporal, spacial and sequential locality associated with
program/data access in a memory hierarchy. (8)
14. (a) (i) Design a pipe line multiplier to multiply a stream of input numbers
X0, X2, X3.... by a fixed number Y. Assume all X's and Y's are 16 bit
numbers. Draw a neat schematic diagram of your design. (12)
(ii) Compare linear and non linear pipelines. (4)
Or
(b) (i) Explain any one scalable multithreaded architecture with a neat
diagram. (8)
(ii) Write about the features of data flow architectures. (8)
15. (a) (i) Explain the message based communications in MACH and 0SF/1
systems. (8)
(ii) Write a program for matrix multiplication in FORTRAN using its
parallel constructs. (8)
Or
(b) Draw and explain the phases of parallel language compilers. (16)




Reg. No. :
M.E. DEGREE EXAMINATION, APRIL/MAY 2011
Second Semester
Applied Electronics
AP 9222 — COMPUTER ARCHITECTURE AND PARALLEL PROCESSING
(Common to M.E. Computer and Communication, M.E. VLSI Design and
M.E. Embedded System and Technologies)
(Regulation 2009)
Time : Three hours Maximum : 100 marks
Answer ALL questions
PART A — (10 × 2 = 20 marks)
1. In parallel computing, what are shared memory-multiprocessors and
distributed-memory multicomputers?
2. Mention an approach to solve the mismatch problem between software
parallelism and hardware parallelism.
3. What is grain packing and scheduling in parallel programming that can help
achieve a short schedule for fast program execution?
4. What scalability metrics affect the scalability of a computer system for a given
application?
5. List the characteristics of CISC, RISC and Superscalar processors.
6. How are sequential consistency and weak consistency memory models
characterized?
7. Distinguish between asynchronous and synchronous pipeline models.
8. Distinguish between static dataflow and dynamic dataflow computers.
9. What are macrotasking, microtasking and autotasking levels of multitasking
employed for parallel execution on Cray X-MP or Y-MP multiprocessors?
10. Differentiate time-sharing and space-sharing operating systems.

PART B — (5 × 16 = 80 marks)
11. (a) Discuss how the instruction set, compiler technology. CPU
implementation and control, cache and memory hierarchy affects CPU
performance. Justify their effects in terms of program length, clock rate
and effective cycles per instruction (CPI). (16)
Or
(b) Compare the PRAM and VLSI models of parallel computers and mention
how these models facilitate the study of asymptotic behavior of
algorithms implementable on parallel computers. (16)
12. (a) Discuss the different network architectures used for interconnecting
multiprocessor or multicomputer systems highlighting their merits and
demerits. (16)
Or
(b) (i) Discuss the three speedup performance models used under different
computing objectives and resource constraints. (10)
(ii) A web server application running on a single processor is enhanced
by a 10 processor parallel computer. The parallel processor is
20 times faster than the original processor, but is available only for
40% of the time. What is the overall speedup gained by
incorporating the enhancement? (6)
13. (a) Discuss the design of a memory hierarchy for a computer system
satisfying the three properties of inclusion, coherence and locality with
due consideration to capacity planning and cost optimization. (16)
Or
(b) Discuss briefly the different shared- memory organization techniques
employed in computer system to meet the memory design goal of
matching memory bandwidth to processor bandwidth. (16)
14. (a) (i) Briefly discuss the instruction pipeline design with reference to
instruction processing, data forwarding, hazard avoidance,
scheduling and branch handling. (10)
(ii) A non pipelined processor X has a clock rate of 25 MHz and an
average CPI (Cycles Per Instruction) of 4. Processor Y. an improved
version of X, is designed with a live-stage linear instruction
pipeline. However, due to latch delay and clock skew effects, the
clock rate of Y is only 20 MHz. If a program containing 100
instructions is executed on both processors, what is the speed up of
processor Y compared with that of processor X? (6)
Or
(b) Compare the four context switching policies employed in different
multithreaded architectures: switch on cache miss, switch on every load,
switch on instruction (cycle by cycle) and switch on block of instructions.
(i) What are the advantages and shortcomings of each policy?
(ii) How can an optimal choice be made among the four policies? (16)
15. (a) Parallel programming models are specifically designed for
multiprocessors, multicomputers, or vector/SIMD computers. Briefly
discuss these models that exploit parallelism with different execution
paradigms for these computers. (16)
Or
(b) Discuss the use of spin locks, suspend locks, binary and counting
semaphores, deadlock prevention, avoidance, detection and recovery
schemes and monitors for shared-variable programming to implement
various synchronization methods among concurrent processes. (16)




No comments:

Post a Comment