(no title)
the__alchemist | 7 days ago
Generally, you use an ASIC to perform a specific task. In this case, I think the takeaway is the LLM functionality here is performance-sensitive, and has enough utility as-is to choose ASIC.
the__alchemist | 7 days ago
Generally, you use an ASIC to perform a specific task. In this case, I think the takeaway is the LLM functionality here is performance-sensitive, and has enough utility as-is to choose ASIC.
RobotToaster|7 days ago
yunohn|7 days ago
AI being static weights is already challenged with the frequent model updates we already see - but may even be a relic once we find a new architecture.
hkt|7 days ago
GTP|7 days ago
dogma1138|7 days ago
FPGAs don’t scale if they did all GPUs would’ve been replaced by FPGAs for graphics a long time ago.
You use an FPGA when spinning a custom ASIC doesn’t makes financial sense and generic processor such as a CPU or GPU is overkill.
Arguably the middle ground here are TPUs, just taking the most efficient parts of a “GPU” when it comes to these workloads but still relying on memory access in every step of the computation.