- Hands-On GPU:Accelerated Computer Vision with OpenCV and CUDA
- Bhaumik Vaidya
- 96字
- 2021-08-13 15:48:26
Cache memory
On the latest GPUs, there is an L1 cache per multiprocessor and an L2 cache, which is shared between all multiprocessors. Both global and local memories use these caches. As L1 is near to thread execution, it is very fast. As shown in the diagram for memory architecture earlier, the L1 cache and shared memory use the same 64 KB. Both can be configured for how many bytes they will use out of the 64 KB. All global memory access goes through an L2 cache. Texture memory and constant memory have their separate caches.
推薦閱讀
- JavaScript+jQuery網頁特效設計任務驅動教程(第2版)
- UI智能化與前端智能化:工程技術、實現方法與編程思想
- 你必須知道的204個Visual C++開發問題
- Python算法指南:程序員經典算法分析與實現
- 劍指大數據:企業級數據倉庫項目實戰(在線教育版)
- OpenResty完全開發指南:構建百萬級別并發的Web應用
- Programming with CodeIgniterMVC
- Building Serverless Architectures
- .NET 4.5 Parallel Extensions Cookbook
- 零基礎C#學習筆記
- Continuous Delivery and DevOps:A Quickstart Guide Second Edition
- 會當凌絕頂:Java開發修行實錄
- C/C++代碼調試的藝術(第2版)
- Clojure Data Structures and Algorithms Cookbook
- JavaScript高級程序設計(第4版)