Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

You can in CUDA. You can have shared memory which is basically L1 cache you have full control over. It's called shared memory because all threads within a block (which reside on a common SM) have fast access to it. The downside: you now have less regular L1 cache.
 help



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: