Australia’s Large Language Model Landscape: Technical Assessment

Australia’s Large Language Model Landscape: Technical Assessment

Key Points No flagship, globally competitive, locally developed LLM (such as GPT-4, Claude 3.5, LLaMA 3.1) has yet emerged from Australia. Australian research and commerce currently rely primarily on international LLMs, which are frequently used but have measurable limitations on Australian English and cultural context. Kangaroo LLM is the only major open-source, locally developed LLM…

Read More
Large Language Models LLMs vs. Small Language Models SLMs for Financial Institutions: A 2025 Practical Enterprise AI Guide

Large Language Models LLMs vs. Small Language Models SLMs for Financial Institutions: A 2025 Practical Enterprise AI Guide

No single solution universally wins between Large Language Models (LLMs, ≥30B parameters, often via APIs) and Small Language Models (SLMs, ~1–15B, typically open-weights or proprietary specialist models). For banks, insurers, and asset managers in 2025, your selection should be governed by regulatory risk, data sensitivity, latency and cost requirements, and the complexity of the use…

Read More
Analysis of large data acquisitions

Analysis of large data acquisitions

Digital acquisition instruments like oscilloscopes and digitizers are incorporating increasingly large acquisition memories. Acquisition memories with lengths in the gigasample (GS) range are commonly available. The advantage of long-acquisition memories is that they can capture longer-time records. They also support higher sampling rates at any given record duration, providing better time resolution. The downside of…

Read More
Meet SmallThinker: A Family of Efficient Large Language Models LLMs Natively Trained for Local Deployment

Meet SmallThinker: A Family of Efficient Large Language Models LLMs Natively Trained for Local Deployment

The generative AI landscape is dominated by massive language models, often designed for the vast capacities of cloud data centers. These models, while powerful, make it difficult or impossible for everyday users to deploy advanced AI privately and efficiently on local devices like laptops, smartphones, or embedded systems. Instead of compressing cloud-scale models for the…

Read More
Implementing Self-Refine Technique Using Large Language Models LLMs

Implementing Self-Refine Technique Using Large Language Models LLMs

This tutorial demonstrates how to implement the Self-Refine technique using Large Language Models (LLMs) with Mirascope, a powerful framework for building structured prompt workflows. Self-Refine is a prompt engineering strategy where the model evaluates its own output, generates feedback, and iteratively improves its response based on that feedback. This refinement loop can be repeated multiple…

Read More
REST: A Stress-Testing Framework for Evaluating Multi-Problem Reasoning in Large Reasoning Models

REST: A Stress-Testing Framework for Evaluating Multi-Problem Reasoning in Large Reasoning Models

Large Reasoning Models (LRMs) have rapidly advanced, exhibiting impressive performance in complex problem-solving tasks across domains like mathematics, coding, and scientific reasoning. However, current evaluation approaches primarily focus on single-question testing, which reveals significant limitations. This article introduces REST (Reasoning Evaluation through Simultaneous Testing) — a novel multi-problem stress-testing framework designed to push LRMs beyond isolated problem-solving…

Read More
Thought Anchors: A Machine Learning Framework for Identifying and Measuring Key Reasoning Steps in Large Language Models with Precision

Thought Anchors: A Machine Learning Framework for Identifying and Measuring Key Reasoning Steps in Large Language Models with Precision

Understanding the Limits of Current Interpretability Tools in LLMs AI models, such as DeepSeek and GPT variants, rely on billions of parameters working together to handle complex reasoning tasks. Despite their capabilities, one major challenge is understanding which parts of their reasoning have the greatest influence on the final output. This is especially crucial for…

Read More