Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
3abiton
on July 11, 2024
|
parent
|
context
|
favorite
| on:
FlashAttention-3: Fast and Accurate Attention with...
To add to the discussion, from a practical perspective, AMD hardware totally sucks and yet to have proper implementation with flash-attention-2. ROCm is moving to usable slowly, but not close to being even comparable with cuda.
LarsDu88
on July 12, 2024
[–]
Whi os it so hard to port FA2 to the m1300 instinct?
latchkey
on July 12, 2024
|
parent
[–]
https://github.com/kailums/flash-attention-rocm
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: