Large language models (LLMs) aren’t actually giant computer brains. Instead, they are massive vector spaces in which the ...
To improve data center efficiency, multiple storage devices are often pooled together over a network so many applications can share them. But even with pooling, significant device capacity remains ...
Intel and Nvidia show off how textures -- which take up a large chunk of PC games -- could be compressed to save you money ...
Intel TSNC brings neural texture compression with up to 18x reduction, faster decoding, and flexible SDK support for modern ...
NVIDIA showcases Neural Texture Compression at GTC 2026, cutting VRAM usage by up to 85% with real-time AI reconstruction.
Neural Texture Compression (NTC) could be a game-changer on par with DLSS if it can reduce the VRAM requirement for textures ...
In its "Tuscan Wheels" demo, the company showed VRAM usage dropping from roughly 6.5GB with traditional BCN-compressed ...
A new compression technique from Google Research threatens to shrink the memory footprint of large AI models so dramatically ...
If Google’s AI researchers had a sense of humor, they would have called TurboQuant, the new, ultra-efficient AI memory compression algorithm announced Tuesday, “Pied Piper” — or, at least that’s what ...
Researchers at Argonne National Laboratory and SLAC have designed a detector chip that compresses X-ray data by factors of 100 to 250 in real time, directly on the silicon that captures each frame.
TurboQuant compresses AI model vectors from 32 bits down to as few as 3 bits by mapping high-dimensional data onto an efficient quantized grid. (Image: Google Research) The AI industry loves a big ...
Nvidia researchers have introduced a new technique that dramatically reduces how much memory large language models need to track conversation history — by as much as 20x — without modifying the model ...