SRIRAM RAGHAVAN
TECH & AI
“ Whether interpreting spoken language, analysing images, or processing textual information, it delivers highly efficient, low-latency inference – all while optimising for on-device execution and reduced computational overhead.”
Phi-4-mini is a 3.8B-parameter model specialised for fast, accurate text-based tasks such as reasoning, maths and code generation. It supports token sequences up to 128,000, making it adept at processing lengthy documents.
Both models offer high accuracy and scalability in a compact form, and their lower latency and cost make them ideal for analytical tasks in resourceconstrained environments. Their structure also not only improves sustainability, but enhances privacy and security by enabling local, on-device processing.
SRIRAM RAGHAVAN
TITLE: VICE PRESIDENT OF AI RESEARCH
COMPANY: IBM INDUSTRY: TECHNOLOGY
Sriram joined IBM in 2004 as a researcher and has moved around the world with the company, working in information and analytics in India, South Asia and the US.
IBM’ s Granite 3.2 models IBM, another AI leader, is leveraging decades of AI innovation to offer the Granite 3.2 model family. These new models are designed specifically for business use, providing robust language capabilities without the overhead associated with larger competitors. The Granite 3.2 series integrates advanced features such as“ chain of thought” reasoning, enabling step-bystep problem solving. This reasoning capability can be toggled, allowing organisations to save on resources for simpler tasks while deploying advanced logic only when necessary.
“ The next era of AI is about efficiency, integration and real-world impact – where enterprises can achieve powerful outcomes without excessive spend on compute,” says Sriram Raghavan, Vice President of IBM AI Research.
“ IBM’ s latest Granite developments focus on open solutions demonstrate another step forward in making AI more accessible, cost-effective and valuable for modern enterprises.”
A highlight of the Granite 3.2 launch is the Granite Vision 3.2 2B, a compact vision-language model built for enterprise document processing. Trained on more than 85 million PDFs using IBM’ s Docling toolkit, it rivals much larger models – including Meta’ s Llama 3.2 11B – by efficiently extracting, classifying and reasoning over complex documents.
172 August 2025