InfoQ Homepage Large language models Content on InfoQ
-
Hugging Face Smolagents is a Simple Library to Build LLM-Powered Agents
Smolagents is a library created at Hugging Face to build agents based on large language models (LLMs). Hugging Faces says its new library aims to be simple and LLM-agnostic. It supports secure "agents that write their actions in code" and is integrated with Hugging Face Hub.
-
LLaMA-Mesh: NVIDIA’s Breakthrough in Unifying 3D Mesh Generation and Language Models
NVIDIA researchers have introduced LLaMA-Mesh, a groundbreaking approach that extends large language models (LLMs) to generate and interpret 3D mesh data in a unified, text-based framework. LLaMA-Mesh tokenizes 3D meshes as plain text, enabling the seamless integration of spatial and textual information.
-
DeepThought-8B Leverages LLaMA-3.1 8B to Create a Compact Reasoning Model
DeepThought-8B is a small "reasoning" model built on LLaMA-3.1 8B that can carry through decision-making processes step by step, similarly to how OpenAI o1 does but in a much smaller package.
-
Qwen Team Unveils QwQ-32B-Preview: Advancing AI Reasoning and Analytics
Qwen Team introduced QwQ-32B-Preview, an experimental research model designed to improve AI reasoning and analytical capabilities. Featuring a 32,768-token context and cutting-edge transformer architecture, it excels in math, programming, and scientific benchmarks like GPQA and MATH-500. Available on Hugging Face, it invites researchers to explore its features and contribute to its development.
-
InstaDeep Open-Sources Genomics AI Model Nucleotide Transformers
Researchers from InstaDeep and NVIDIA have open-sourced Nucleotide Transformers (NT), a set of foundation models for genomics data. The largest NT model has 2.5 billion parameters and was trained on genetic sequence data from 850 species. It outperforms other state-of-the-art genomics foundation models on several genomics benchmarks.
-
AWS Adds News Amazon Q Developer Agent Capabilities: Doc Generation, Code Reviews, and Unit Tests
AWS has enhanced its generative AI-powered Amazon Q Developer, streamlining software development with new agent capabilities. Key features include automated documentation, code reviews, and unit test generation, allowing developers to focus on coding. Available in all AWS Regions, Amazon Q Developer simplifies processes in IDEs like Visual Studio Code and IntelliJ IDEA.
-
EuroLLM-9B Aims to Improve State of the Art LLM Support for European Languages
EuroLLM-9B is an open-source large language model built in Europe and tailored to European languages, including all the official EU languages as well as 11 other non-official albeit commercially important languages. According to the team behind it, its performance makes it one of the best European-made LLM of this size.
-
Anthropic Publishes Model Context Protocol Specification for LLM App Integration
Anthropic recently released their Model Context Protocol (MCP), an open standard describing a protocol for integrating external resources and tools with LLM apps. The release includes SDKs implementing the protocol, as well as an open-source repository of reference implementations of MCP.
-
Recap of OpenAI Highlights Key Updates in 12-Day "Shipmas"
OpenAI's "12 Days of Shipmas" event featured daily announcements of new AI features and tools. Below is a summary of the key developments.
-
OpenAI Releases Sora and Full Version of O1 Reasoning Model with Fine-Tuning
OpenAI has unveiled its advanced o1 reasoning model and the video generation model Sora, enhancing complex reasoning and video creation capabilities. Sora produces high-quality videos using innovative diffusion techniques, while o1 excels in nuanced reasoning and safety. Together, they signal a transformative leap in AI, bridging creativity and rigorous reasoning.
-
Meta Releases Llama 3.3: a Multilingual Model with Enhanced Performance and Efficiency
Meta has released Llama 3.3, a multilingual large language model aimed at supporting a range of AI applications in research and industry. Featuring a 128k-token context window and architectural improvements for efficiency, the model demonstrates strong performance in benchmarks for reasoning, coding, and multilingual tasks. It is available under a community license on Hugging Face.
-
Google AI Agent Jules Aims at Helping Developers with Their GitHub-Based Workflows
Part of Gemini 2.0, Google has launched its new AI-based coding assistant in closed preview. Dubbed "Jules", the assistant aims at helping developers to work with Python and JavaScript issues and pull requests, handle bug fixes, and other related tasks.
-
New LangChain Report Reveals Growing Adoption of AI Agents
LangChain presented the State of AI Agents where they examined the current state of AI agent adoption across industries, gathering insights from over 1,300 professionals, including engineers, product managers, and executives. The findings provide a detailed view of how AI agents are being integrated into workflows and the challenges companies face in deploying these systems effectively.
-
Amazon Introduces Amazon Nova, a Series of Foundation Models
Amazon has announced Amazon Nova, a family of foundation models designed for generative AI tasks. The announcement, made during AWS re:Invent, highlights the models' capabilities in tasks such as document and video analysis, chart comprehension, video content generation, and AI agent development.
-
Micro Metrics for LLM System Evaluation at QCon SF 2024
Denys Linkov's QCon San Francisco 2024 talk dissected the complexities of evaluating large language models (LLMs). He advocated for nuanced micro-metrics, robust observability, and alignment with business objectives to enhance model performance. Linkov’s insights highlight the need for multidimensional evaluation and actionable metrics that drive meaningful decisions.