InfoQ Homepage Large language models Content on InfoQ
-
How Observability Can Improve the UX of LLM Based Systems: Insights of Honeycomb's CEO at KubeCon EU
During her KubeCon Europe keynote, Christine Yen, CEO and co-founder of Honeycomb, provided insights on how observability can help cope with the rapid shifts introduced by the integration of LLMs in software systems, which transformed not only the way we develop software but also the release methodology. She explained how to adapt your development feedback loop based on production observations.
-
OpenAI Introduces New Speech Models for Transcription and Voice Generation
OpenAI has introduced new speech-to-text and text-to-speech models in its API, focusing on improving transcription accuracy and offering more control over AI-generated voices. These updates aim to enhance automated speech applications, making them more adaptable to different environments and use cases.
-
Google DeepMind Launches TxGemma: Advancing AI-Driven Drug Discovery and Development
Designed to enhance the efficiency of drug discovery and clinical trial predictions. Built on the Gemma model family, TxGemma aims to streamline the drug development process and accelerate the discovery of new treatments.
-
How Airbnb Used LLMs to Accelerate Test Migration
Thanks to the right mix of workflow automation and large language models, Airbnb significantly accelerated the process of updating their codebase to adopt React Testing Library (RTL) and converted nearly 3.5K React test files originally using Enzyme.
-
Nvidia Unveils AI, GPU, and Quantum Computing Innovations at GTC 2025
Nvidia presented several new technologies at its GTC 2025 event, covering advancements in GPUs, AI, robotics, and quantum computing.
-
Roblox Releases Cube 3D, an AI Open-Source Model for 3D Model Generation
Roblox has introduced Cube 3D, a generative AI system designed for creating 3D and 4D objects and environments.
-
Dapr Agents: Scalable AI Workflows with LLMs, Kubernetes & Multi-Agent Coordination
Introducing Dapr Agents—a groundbreaking framework for creating scalable AI agents using Large Language Models (LLMs). With robust workflows, multi-agent coordination, and cloud-neutral architecture, it enables enterprises to deploy thousands of resilient agents. Built on Dapr’s proven infrastructure, Dapr Agents ensures reliability and observability in AI-driven applications.
-
Google Launches Gemma 3 1B for Mobile and Web Apps
Requiring a "mere" 529MB, Gemma 3 1B is a small language model (SLM) specifically meant for distribution across mobile and Web apps, where models must download quickly and be responsive to keep user engagement high.
-
Google Report Reveals How Threat Actors Are Currently Using Generative AI
Google's Threat Intelligence Group (GTIG) recently released a report on the adversarial misuse of generative AI. The team investigated prompts used by advanced persistent threat (APT) and coordinated information operations (IO) actors, finding that they have so far achieved productivity gains but have not yet developed novel capabilities.
-
Google Introduces AI Co-Scientist System to Aid Scientific Research
Google has announced the development of an AI co-scientist system designed to assist scientists in generating hypotheses and research proposals. Built using Gemini 2.0, the system aims to accelerate scientific and biomedical discoveries by emulating the scientific method and fostering collaboration between humans and AI.
-
OpenAI Introduces Software Engineering Benchmark
OpenAI has introduced the SWE-Lancer benchmark, to evaluate the capabilities of advanced AI language models in real-world freelance software engineering tasks.
-
Google's Image Generation Model Imagen 3 Now Available in Vertex AI in Firebase as a Preview
Google's most advanced GenAI image generation model, Imagen 3, is now available in preview through Vertex AI in Firebase enabling seamless integration into Android and iOS apps through its Kotlin and Swift SDKs.
-
instructlab.ai Uses Synthetic Data to Reduce Complexity of Fine-Tuning LLMs
InstructLab.ai implements the large-scale alignment for the chatbots concept(LAB), which intends to overcome the scalability challenges in the instruction-tuning phase of a large language model (LLM). Its approach leverages a synthetic data-based alignment tuning method for LLMs. Crafted taxonomies deliver the synthesization seeds for training data, reducing the need for human-annotated data.
-
Mistral AI Introduces Saba: Regional Language Model for Arabic and South Indian Language
Mistral AI has introduced Mistral Saba, a 24-billion-parameter language model designed to improve AI performance in Arabic and several Indian-origin languages, particularly South Indian languages like Tamil.
-
Hugging Face Publishes Guide on Efficient LLM Training across GPUs
Hugging Face has published the Ultra-Scale Playbook: Training LLMs on GPU Clusters, an open-source guide that provides a detailed exploration of the methodologies and technologies involved in training LLMs across GPU clusters.