@ibmresearch
  @ibmresearch
IBM Research | ICLR Paper: Learn Step Size Quantization @ibmresearch | Uploaded June 2020 | Updated October 2024, 4 days ago.
As deep networks are increasingly deployed in memory-constrained and throughput-critical systems, there is a need to create AI models that can maintain accuracy – and, as a result, trust – while also consuming fewer resources. Researchers at IBM’s Almaden Research Laboratory have reached a new milestone in AI precision and developed an algorithm that matches the inference accuracy of a 32-bit network while using only three bits.

The researchers achieved this level of energy efficiency using a new process called “learned step size quantization,” which improves parameter change estimates in a low-precision network during training, to produce better performance. The research also uncovered evidence that AI systems seeking to optimize performance on a given system might run with as few as 2 bits. This advance means AI systems are steadily coming closer to the low levels of energy consumed by the human brain, while maintaining performance.
ICLR Paper: Learn Step Size QuantizationIn-memory physical superposition meets few-shot continual learningThe Short: AI chips come to UAlbany, protecting AI models from attack, and tiny magnetic moleculesCapturing and transforming CO2 to mitigate climate changeGenerative AI for businessWhats Next in AI : AI We Can TrustUnderstanding the NIST standards and IBMs contributions to post-quantum cryptographyAI and fully homomorphic encryption at IBM Research - HaifaFour years of quantum computing on the IBM CloudThe Short: Tiny benchmarks for LLMs, upending automation with gen AI and remembering Bob DennardLaying the Groundwork for Quantum Powered Use CasesEffects of qubit frequency crowding on scalable quantum processors*

ICLR Paper: Learn Step Size Quantization @ibmresearch

SHARE TO X SHARE TO REDDIT SHARE TO FACEBOOK WALLPAPER