What happened
Over the past two weeks, a cluster of vendor announcements and updates showed AI shifting from large cloud models to smaller, multimodal models and agent tooling that run on device or tightly inside products. Highlights include Microsoft’s Phi‐4 family (12‐18‐2024) for on‐device and Azure use, Apple research on running large models efficiently on iPhones/iPads, broader safety tooling from Meta for smaller multimodal models, plus parallel moves in platform engineering, biotech pipelines, quantum roadmaps, market‐surveillance ML, and education/coding tool adoption.
Why this matters
Macro takeaway — Infrastructure embedding: AI is moving from standalone chat apps into core infrastructure layers (developer platforms, in‐product agents, automated drug‐discovery loops, exchange surveillance and quantum stacks).
This matters because:
- Scale and locality: Smaller multimodal models + safety tooling enable agentic workflows that can run close to user data (on device or inside enterprise VPCs), reducing reliance on external APIs and lowering data‐exposure risk.
- Platformization: Platform engineering products (e.g., GitHub Copilot Workspace, Terraform Cloud AI features, Backstage adopters) are turning AI into embedded capabilities for CI/pipelines, IDPs, and policy‐as‐code, changing how teams operationalize automation.
- Domain impact: In biotech, companies such as Absci and Generate:Biomedicines report AI‐designed antibodies/proteins entering preclinical pipelines, signalling long‐term, multi‐asset pharma deals rather than experimental uses.
- Measurement shift: In quantum, vendors and analysts are reframing progress around logical qubits and error rates instead of raw qubit counts, making timelines for practical workloads more concrete.
- Workforce effect: Education and tools show AI‐assisted coding becoming a baseline skill, shifting emphasis to system design, evaluation, and governance.
Risks/opportunities are practical and operational (data privacy, safety tooling, reproducibility, and the need to wire models into observable, governable systems) rather than purely technical novelty.
Sources
- Microsoft, “Introducing Phi‐4: Small, Open Models for Language and Vision” (12‐18‐2024): https://azure.microsoft.com/en-us/blog/introducing-phi-4-small-open-models-for-language-and-vision/
- Apple Machine Learning Research, “LLM in a flash” (2024): https://machinelearning.apple.com/research/llm-in-a-flash
- GitHub, “Introducing GitHub Copilot Workspace (preview)” (12‐16‐2024): https://github.blog/news-insights/product-news/introducing-github-copilot-workspace/
- Absci, “Absci and AstraZeneca expand AI drug creation partnership” (12‐17‐2024): https://investors.absci.com/news-releases/news-release-details/absci-and-astrazeneca-expand-ai-drug-creation-partnership
- IBM Research, “Demonstrating Logical Qubits with Error Correction” (11‐04‐2024; updates Dec 2024): https://research.ibm.com/blog/logical-qubits-error-correction