Author Topic: CUDA, Supercomputing for the Masses: Part 12  (Read 5567 times)

0 Members and 1 Guest are viewing this topic.


  • Consigliere
  • ****
  • Posts: 619
CUDA, Supercomputing for the Masses: Part 12
« on: May 14, 2009, 06:36:13 PM »
Prior to CUDA 2.2, CUDA kernels could not access host system memory directly. For that reason, CUDA programmers used the design pattern introduced in Part 1 and Part 2:

   1. Move data to the GPU.
   2. Perform calculation on GPU.
   3. Move result(s) from the GPU to host.

This paradigm has now changed as CUDA 2.2 has introduced new APIs that allow host memory to be mapped into device memory via a new function called cudaHostAlloc (or cuMemHostAlloc in the CUDA driver API).

Read full story at DDJ.
Systeminfo: CPU-Z - GPU-Z - GPUCapsViewer