ElastixAI solves the systemic inefficiencies of GenAI inference through innovative software-ML-hardware co-design, delivering the next generation of scalable, sustainable AI. The founding team brings ...
Inception, the company behind the first commercial diffusion large language models (dLLMs), today announced the launch of Mercury 2, the fastest reasoning LLM and first reasoning dLLM. Mercury 2 ...
Taalas has launched an AI accelerator that puts the entire AI model into silicon, delivering 1-2 orders of magnitude greater ...
Researchers at the Tokyo-based startup Sakana AI have developed a new technique that enables language models to use memory more efficiently, helping enterprises cut the costs of building applications ...
With reported 3x speed gains and limited degradation in output quality, the method targets one of the biggest pain points in production AI systems: latency at scale.
A new technical paper titled “Combating the Memory Walls: Optimization Pathways for Long-Context Agentic LLM Inference” was published by researchers at University of Cambridge, Imperial College London ...
If mHC scales the way early benchmarks suggest, it could reshape how we think about model capacity, compute budgets and the ...
AWS Premier Tier Partner leverages its AI Services Competency and expertise to help founders cut LLM costs using ...
BEIJING--(BUSINESS WIRE)--On January 4th, the inaugural ceremony for the 2024 ASC Student Supercomputer Challenge (ASC24) unfolded in Beijing. With a global interest, ASC24 has garnered the ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results