Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Training a large language model (LLM) is ...
Nvidia researchers developed dynamic memory sparsification (DMS), a technique that compresses the KV cache in large language models by up to 8x while maintaining reasoning accuracy — and it can be ...
Health & Wellness Design Assistant Professor Edlin Garcia, Ph.D., is co-principal investigator (PI) on a research project titled " Designing Accountable Mental Health Large Language Model Therapy ...
A recent warning from OpenAI about the potential ramifications of a stringent copyright crackdown on artificial intelligence (AI) development has sparked a complex legal debate about the balance ...
ByteDance's Doubao AI team has open-sourced COMET, a Mixture of Experts (MoE) optimization framework that improves large language model (LLM) training efficiency while reducing costs. Already ...
Test-time Adaptive Optimization can be used to increase the efficiency of inexpensive models, such as Llama, the company said. Data lakehouse provider Databricks has unveiled a new large language ...
Apertus was released in early September 2025. It is a multilingual model developed by the Swiss Federal Institutes of Technology in Zurich (ETH) and Lausanne (EPFL). The model was pretrained with 60% ...
AI remains top of mind for many business and industry leaders, and the sector shows no signs of slowing down regarding investment and funding. Most recently, the generative AI startup Writer announced ...
As recently as 2022, just building a large language model (LLM) was a feat at the cutting edge of artificial-intelligence (AI) engineering. Three years on, experts are harder to impress. To really ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results