Reputation: 51
I have noticed that the latency access to both cache and shared memory are the same in CUDA. Given that fact how are they different? How do we use them in different ways?
Upvotes: 1
Views: 1060
Reputation: 2916
CUDA shared memory usage is explicit with the __shared__
keyword. You have full control on it. L1 cache on the other hand is managed by hardware. Performance and caching strategy of L1 cache depends on hardware architecture.
Upvotes: 2