Skip to Content
The CUDA Handbook: A Comprehensive Guide to GPU Programming
book

The CUDA Handbook: A Comprehensive Guide to GPU Programming

by Nicholas Wilt
June 2013
Intermediate to advanced
528 pages
13h 11m
English
Addison-Wesley Professional
Content preview from The CUDA Handbook: A Comprehensive Guide to GPU Programming

Chapter 5. Memory

To maximize performance, CUDA uses different types of memory, depending on the expected usage. Host memory refers to the memory attached to the CPU(s) in the system. CUDA provides APIs that enable faster access to host memory by page-locking and mapping it for the GPU(s). Device memory is attached to the GPU and accessed by a dedicated memory controller, and, as every beginning CUDA developer knows, data must be copied explicitly between host and device memory in order to be processed by the GPU.

Device memory can be allocated and accessed in a variety of ways.

Global memory may be allocated statically or dynamically and accessed via pointers in CUDA kernels, which translate to global load/store instructions.

Constant memory ...

Become an O’Reilly member and get unlimited access to this title plus top books and audiobooks from O’Reilly and nearly 200 top publishers, thousands of courses curated by job role, 150+ live events each month,
and much more.
Start your free trial

You might also like

CUDA Programming

CUDA Programming

Shane Cook
Professional CUDA C Programming

Professional CUDA C Programming

John Cheng, Max Grossman, Ty McKercher

Publisher Resources

ISBN: 9780133261516Purchase book