Even More Fun
•
Some of that overhead can
be avoided when the
destination of the GPU’s data is graphics
•
Texture memory can be shared between general
purpose computations and normal
rendering
•
For post-processing effects or visualizing particles, the
pixel/vertex data never needs to leave the GPU
Conclusions
Certain classes of problem appear in many different
fields, and involve very data-parallel
operations such
as filtering, sorting, or integration
Taking advantage of the architecture decisions behind
graphics processing units such as their multiprocessing
and
native vector operations, these problems can be
solved quickly
and cheaply
References
•
1.
Ziegler, Grenot. Introduction to the CUDA Architecture. [Online] 2009.
http://www.cse.scitech.ac.uk/disco/workshops/200907/Day1_01_Intro_CUDA_Architecture.pdf.
•
2.
NVIDIA Corporation. NVIDIA Compute PTX: Parallel Thread Execution ISA Version 1.1. 2007.
•
3.
Göddeke, Dominik. Fast and Accurate Finite-Element Multigrid Solvers for PDE Simulations on GPU Clusters. Berlin : Logos Verlag, 2010.
978-3-8325-2768-6.
•
4.
Accellerating molecular modeling application swith graphics processors. John E Stone, James C Phillips, Peter L Freddolino, David J
Dostları ilə paylaş: