Australia’s Large Language Model Landscape: Technical Assessment

Australia’s Large Language Model Landscape: Technical Assessment

Key Points No flagship, globally competitive, locally developed LLM (such as GPT-4, Claude 3.5, LLaMA 3.1) has yet emerged from Australia. Australian research and commerce currently rely primarily on international LLMs, which are frequently used but have measurable limitations on Australian English and cultural context. Kangaroo LLM is the only major open-source, locally developed LLM…

Read More
SEA-LION v4: Multimodal Language Modeling for Southeast Asia

SEA-LION v4: Multimodal Language Modeling for Southeast Asia

AI Singapore (AISG) has released SEA-LION v4, an open-source multimodal language model developed in collaboration with Google and based on the Gemma 3 (27B) architecture. The model is designed to support Southeast Asian languages, including those with limited digital resources, and provides both text and image understanding capabilities. SEA-LION v4 uses a commercially permissive license…

Read More
Large Language Models LLMs vs. Small Language Models SLMs for Financial Institutions: A 2025 Practical Enterprise AI Guide

Large Language Models LLMs vs. Small Language Models SLMs for Financial Institutions: A 2025 Practical Enterprise AI Guide

No single solution universally wins between Large Language Models (LLMs, ≥30B parameters, often via APIs) and Small Language Models (SLMs, ~1–15B, typically open-weights or proprietary specialist models). For banks, insurers, and asset managers in 2025, your selection should be governed by regulatory risk, data sensitivity, latency and cost requirements, and the complexity of the use…

Read More
Meet SmallThinker: A Family of Efficient Large Language Models LLMs Natively Trained for Local Deployment

Meet SmallThinker: A Family of Efficient Large Language Models LLMs Natively Trained for Local Deployment

The generative AI landscape is dominated by massive language models, often designed for the vast capacities of cloud data centers. These models, while powerful, make it difficult or impossible for everyday users to deploy advanced AI privately and efficiently on local devices like laptops, smartphones, or embedded systems. Instead of compressing cloud-scale models for the…

Read More
Implementing Self-Refine Technique Using Large Language Models LLMs

Implementing Self-Refine Technique Using Large Language Models LLMs

This tutorial demonstrates how to implement the Self-Refine technique using Large Language Models (LLMs) with Mirascope, a powerful framework for building structured prompt workflows. Self-Refine is a prompt engineering strategy where the model evaluates its own output, generates feedback, and iteratively improves its response based on that feedback. This refinement loop can be repeated multiple…

Read More
You Don’t Need to Share Data to Train a Language Model Anymore—FlexOlmo Demonstrates How

You Don’t Need to Share Data to Train a Language Model Anymore—FlexOlmo Demonstrates How

The development of large-scale language models (LLMs) has historically required centralized access to extensive datasets, many of which are sensitive, copyrighted, or governed by usage restrictions. This constraint severely limits the participation of data-rich organizations operating in regulated or proprietary environments. FlexOlmo—introduced by researchers at the Allen Institute for AI and collaborators—proposes a modular training…

Read More
Mistral AI Releases Devstral 2507 for Code-Centric Language Modeling

Mistral AI Releases Devstral 2507 for Code-Centric Language Modeling

Mistral AI, in collaboration with All Hands AI, has released updated versions of its developer-focused large language models under the Devstral 2507 label. The release includes two models—Devstral Small 1.1 and Devstral Medium 2507—designed to support agent-based code reasoning, program synthesis, and structured task execution across large software repositories. These models are optimized for performance…

Read More