InfoQ Homepage AI Architecture Content on InfoQ
-
Tracking and Controlling Data Flows at Scale in GenAI: Meta’s Privacy-Aware Infrastructure
Meta has revealed how it scales its Privacy-Aware Infrastructure (PAI) to support generative AI development while enforcing privacy across complex data flows. Using large-scale lineage tracking, PrivacyLib instrumentation, and runtime policy controls, the system enables consistent privacy enforcement for AI workloads like Meta AI glasses without introducing manual bottlenecks.
-
Google and Retail Leaders Launch Universal Commerce Protocol to Power Next‑Generation AI Shopping
Google launched the Universal Commerce Protocol (UCP), an open standard co-developed with Shopify, Target, and others, enabling AI-driven shopping agents to complete tasks end-to-end from product discovery to checkout and post-purchase management. UCP aims to standardize commerce capabilities, support multiple payment providers, and expand globally. Shaping the next generation of agentic commerce.
-
TanStack Releases Framework Agnostic AI Toolkit
Introducing TanStack AI: a revolutionary, framework-agnostic toolkit empowering developers with unparalleled control over their AI stack. This open-source release features a unified interface across multiple providers and ensures type safety with innovative isomorphic tools. Say goodbye to vendor lock-in and hello to freedom in AI development!
-
CNCF Launches Certified Kubernetes AI Conformance Program to Standardise Workloads
The CNCF has launched the Certified Kubernetes AI Conformance program to standardise artificial intelligence workloads. By establishing a technical baseline for GPU management, networking, and gang scheduling, the initiative ensures portability across cloud providers. It aims to reduce technical debt and prevent vendor lock-in as enterprises move generative AI models into production.
-
SIMA 2 Uses Gemini and Self-Improvement to Generalize across Unseen 3D and Photorealistic Worlds
Google DeepMind researchers introduced SIMA 2 (Scalable Instructable Multiworld Agent), a generalist agent built on the Gemini foundation model that can understand and act across multiple 3D virtual game environments. The SIMA 2 architecture uses a Gemini Flash-Lite model trained on a mixture of gameplay and Gemini pretraining data.
-
Meta Details GEM Ads Model Using LLM-Scale Training, Hybrid Parallelism, and Knowledge Transfer
Meta released details about its Generative Ads Model (GEM), a foundation model designed to improve ads recommendation across its platforms. The model addresses core challenges in recommendation systems (RecSys) by processing billions of daily user-ad interactions where meaningful signals such as clicks and conversions are very sparse.
-
InfoQ Announces January Online Architect Cohort Focused on Socio-Technical Leadership
InfoQ announces the January 2026 intake for its Certified Architect Program. Facilitated by Luca Mezzalira, this 5-week online cohort focuses on socio-technical leadership, helping senior architects bridge the gap between technical design and organizational influence. Participants engage in weekly applied learning and peer collaboration to earn the ICSAET certification.
-
Private AI Compute Enables Google Inference with Hardware Isolation and Ephemeral Data Design
Google announced Private AI Compute, a system designed to process AI requests using Gemini cloud models while aiming to keep user data private. The announcement positions Private AI Compute as Google's approach to addressing privacy concerns while providing cloud-based AI capabilities, building on what the company calls privacy-enhancing technologies it has developed for AI use cases.
-
Amazon Adds A2A Protocol to Bedrock AgentCore for Interoperable Multi-Agent Workflows
Amazon announced support for the Agent-to-Agent (A2A) protocol in Amazon Bedrock AgentCore Runtime, enabling communication between agents built on different frameworks. The protocol allows agents developed with Strands Agents, OpenAI Agents SDK, LangGraph, Google ADK, or Claude Agents SDK to "share context, capabilities, and reasoning in a common, verifiable format."
-
Kimi's K2 Opensource Language Model Supports Dynamic Resource Availability and New Optimizer
Kimi released K2, a Mixture-of-Experts large language model with 32 billion activated parameters and 1.04 trillion total parameters, trained on 15.5 trillion tokens. The release introduces MuonClip, a new optimizer that builds on the Muon optimizer by adding a QK-clip technique designed to address training instability, which the team reports resulted in "zero loss spike" during pre-training.
-
Anthropic Adds Sandboxing and Web Access to Claude Code for Safer AI-Powered Coding
Anthropic released sandboxing capabilities for Claude Code and launched a web-based version of the tool that runs in isolated cloud environments. The company introduced these features to address security risks that arise when Claude Code writes, tests, and debugs code with broad access to developer codebases and files.
-
New Claude Haiku 4.5 Model Promises Faster Performance at One-Third the Cost
Anthropic released Claude Haiku 4.5, making the model available to all users as its latest entry in the small, fast model category. The company positions the new model as delivering performance levels comparable to Claude Sonnet 4, which launched five months ago as a state-of-the-art model, but at "one-third the cost and more than twice the speed."
-
QCon London 2026 Announces Tracks: AI Engineering, Building Teams, Tech of Finance, and More
The QCon London 2026 tracks are live: 15 practitioner-curated deep dives on AI adoption, resilient architectures, distributed systems, performance, modern languages, data, security, and Staff+ leadership, rooted in real production lessons.
-
Inside the Architectures Powering Modern AI Systems: QCon San Francisco 2025
Senior engineers face fast-moving AI adoption without clear patterns. QCon SF 2025 brings real-world lessons from teams at Netflix, Meta, Intuit, Anthropic & more, showing how to build reliable AI systems at scale. Early bird ends Nov 11.
-
Open Practices for Architecture and AI Adoption
Andrea Magnorsky presented on Byte-Sized Architecture at Cloud Native Summit 2025, as a format for building shared understanding through small, recurrent workshops. Ahilan Ponnusamy and Andreas Spanner discussed the Technology Operating Model for AI adoption. Both approaches drew on the Open Practice Library for human-centred collaboration and driving architectural evolution.