How to build AI scaling laws for efficient LLM training and budget maximization

How to build AI scaling laws for efficient LLM training and budget maximization

When researchers are building large language models (LLMs), they aim to maximize performance under a particular computational and financial budget. Since training a model can amount to millions of dollars, developers need to be judicious with cost-impacting decisions about, for instance, the model architecture, optimizers, and training datasets before committing to a model. To anticipate…

Read More
Arm Unveils Lumex Compute Subsystem For Powerful, Efficient On-Device AI

Arm Unveils Lumex Compute Subsystem For Powerful, Efficient On-Device AI

As the smartphone market has matured, the workloads that consumers expect from their tiny in-pocket mobile computers has increased drastically. Fortunately, chip designers continue to build faster processors that do perform well with varied workloads, without completely tanking battery life. Tonight, Arm introduced its Lumex Compute Subsystem (CSS) platform, which drives big improvements in not…

Read More
Power Tips #144: Designing an efficient, cost-effective micro DC/DC converter with high output accuracy for automotive applications

Power Tips #144: Designing an efficient, cost-effective micro DC/DC converter with high output accuracy for automotive applications

The ongoing electrification of cars brings new trends and requirements with every new design cycle. One trend for battery electric vehicles is reducing the size of the low-voltage batteries, which power either 12-V or 48-V systems. Some auto manufacturers are even investigating whether it’s possible to eliminate low-voltage batteries completely. Regardless, you’ll need isolated high-…

Read More
Meet SmallThinker: A Family of Efficient Large Language Models LLMs Natively Trained for Local Deployment

Meet SmallThinker: A Family of Efficient Large Language Models LLMs Natively Trained for Local Deployment

The generative AI landscape is dominated by massive language models, often designed for the vast capacities of cloud data centers. These models, while powerful, make it difficult or impossible for everyday users to deploy advanced AI privately and efficiently on local devices like laptops, smartphones, or embedded systems. Instead of compressing cloud-scale models for the…

Read More
Announcing Gemma 3n preview: powerful, efficient, mobile-first AI- Google Developers Blog

Announcing Gemma 3n preview: powerful, efficient, mobile-first AI- Google Developers Blog

Following the exciting launches of Gemma 3 and Gemma 3 QAT, our family of state-of-the-art open models capable of running on a single cloud or desktop accelerator, we’re pushing our vision for accessible AI even further. Gemma 3 delivered powerful capabilities for developers, and we’re now extending that vision to highly capable, real-time AI operating…

Read More
This AI Paper from NVIDIA and SUTD Singapore Introduces TANGOFLUX and CRPO: Efficient and High-Quality Text-to-Audio Generation with Flow Matching

This AI Paper from NVIDIA and SUTD Singapore Introduces TANGOFLUX and CRPO: Efficient and High-Quality Text-to-Audio Generation with Flow Matching

Text-to-audio generation has transformed how audio content is created, automating processes that traditionally required significant expertise and time. This technology enables the conversion of textual prompts into diverse and expressive audio, streamlining workflows in audio production and creative industries. Bridging textual input with realistic audio outputs has opened possibilities in applications like multimedia storytelling, music,…

Read More