23-02-2024, 12:46 PM
A generative AI firm has built a new chip designed to deliver blistering AI inference performance with large language models (LLMs).
Groq achieves this by creating a processing unit known as the Tensor Streaming Processor (TSP), which is designed to deliver deterministic performance for AI computations, eschewing the use of GPUs.
https://www.cdotrends.com/story/3823/gro...-inference
Groq achieves this by creating a processing unit known as the Tensor Streaming Processor (TSP), which is designed to deliver deterministic performance for AI computations, eschewing the use of GPUs.
https://www.cdotrends.com/story/3823/gro...-inference