
Efficient

How to build AI scaling laws for efficient LLM training and budget maximization
When researchers are building large language models (LLMs), they aim to maximize performance under a particular computational and financial budget. Since training a model can amount to millions of dollars, developers need to be judicious with cost-impacting decisions about, for instance, the model architecture, optimizers, and training datasets before committing to a model. To anticipate…

Arm Unveils Lumex Compute Subsystem For Powerful, Efficient On-Device AI
As the smartphone market has matured, the workloads that consumers expect from their tiny in-pocket mobile computers has increased drastically. Fortunately, chip designers continue to build faster processors that do perform well with varied workloads, without completely tanking battery life. Tonight, Arm introduced its Lumex Compute Subsystem (CSS) platform, which drives big improvements in not…

Power Tips #144: Designing an efficient, cost-effective micro DC/DC converter with high output accuracy for automotive applications
The ongoing electrification of cars brings new trends and requirements with every new design cycle. One trend for battery electric vehicles is reducing the size of the low-voltage batteries, which power either 12-V or 48-V systems. Some auto manufacturers are even investigating whether it’s possible to eliminate low-voltage batteries completely. Regardless, you’ll need isolated high-…

Meet SmallThinker: A Family of Efficient Large Language Models LLMs Natively Trained for Local Deployment
The generative AI landscape is dominated by massive language models, often designed for the vast capacities of cloud data centers. These models, while powerful, make it difficult or impossible for everyday users to deploy advanced AI privately and efficiently on local devices like laptops, smartphones, or embedded systems. Instead of compressing cloud-scale models for the…

Announcing Gemma 3n preview: powerful, efficient, mobile-first AI- Google Developers Blog
Following the exciting launches of Gemma 3 and Gemma 3 QAT, our family of state-of-the-art open models capable of running on a single cloud or desktop accelerator, we’re pushing our vision for accessible AI even further. Gemma 3 delivered powerful capabilities for developers, and we’re now extending that vision to highly capable, real-time AI operating…

This AI Paper from NVIDIA and SUTD Singapore Introduces TANGOFLUX and CRPO: Efficient and High-Quality Text-to-Audio Generation with Flow Matching
Text-to-audio generation has transformed how audio content is created, automating processes that traditionally required significant expertise and time. This technology enables the conversion of textual prompts into diverse and expressive audio, streamlining workflows in audio production and creative industries. Bridging textual input with realistic audio outputs has opened possibilities in applications like multimedia storytelling, music,…