23-02-2024, 12:46 PM
A generative AI firm has built a new chip designed to deliver blistering AI inference performance with large language models (LLMs).
Groq achieves this by creating a processing unit known as the Tensor Streaming Processor (TSP), which is designed to deliver deterministic performance for AI computations, eschewing the use of GPUs.
https://www.cdotrends.com/story/3823/gro...-inference
Groq achieves this by creating a processing unit known as the Tensor Streaming Processor (TSP), which is designed to deliver deterministic performance for AI computations, eschewing the use of GPUs.
https://www.cdotrends.com/story/3823/gro...-inference
“Be who you are and say what you feel, because those who mind don't matter and those who matter don't mind"