InfoQ Homepage Artificial Intelligence Content on InfoQ
-
Measuring and Reducing the Environmental Impact of Software
Software applications often manage big amounts of data; most of them are internet-based applications, and incorporate artificial intelligence. According to Coral Calero, these three aspects improve the capabilities and functionalities provided by software but they have also increased the amount of energy needed. We need to measure energy consumption of software to control its environmental impact.
-
Anthropic Unveils Contextual Retrieval for Enhanced AI Data Handling
Anthropic has announced Contextual Retrieval, a significant advancement in AI systems' interaction with extensive knowledge bases. This technique addresses the challenge of context loss in Retrieval-Augmented Generation (RAG) systems by enriching text chunks with contextual information before embedding or indexing.
-
Microsoft Launches Azure AI Inference SDK for .NET
Microsoft launched Azure AI Inference SDK for .NET, streamlining access to generative AI models in the Azure AI Studio model catalog. This catalog includes models from providers like Azure OpenAI Service, Mistral, Meta, Cohere, NVIDIA, and Hugging Face, organized into three collections: Curated by Azure AI, Azure OpenAI Models, and Open Models from Hugging Face Hub.
-
Stability AI Announces Integration of Top Text-to-Image Models with Amazon Bedrock
Stability AI has introduced three new text-to-image models to Amazon Bedrock: Stable Image Ultra, Stable Diffusion 3 Large, and Stable Image Core. These models focus on improving performance in multi-subject prompts, image quality, and typography. They are designed to generate high-quality visuals for various use cases in marketing, advertising, media, entertainment, retail, and more.
-
AWS Announces General Availability of EC2 P5e Instances, Powered by NVIDIA H100 Tensor Core GPUs
Amazon Web Services (AWS) has launched EC2 P5e instances featuring NVIDIA H100 Tensor Core GPUs, substantially boosting AI and HPC performance. With enhanced memory bandwidth, these instances reduce latency for real-time applications. Ideal for tasks like LLM training and simulations, they offer improved scalability and cost-efficiency, making them pivotal for modern cloud computing.
-
Leveraging the Transformer Architecture for Music Recommendation on YouTube
Google has described an approach to use transformer models, which ignited the current generative AI boom, for music recommendation. This approach, which is currently being applied experimentally on YouTube, aims to build a recommender that can understand sequences of user actions when listening to music to better predict user preferences based on their context.
-
The Value of Using Timeless Testing Tools
According to Benjamin Bischoff, developers find new tools much more interesting than old ones, as they offer an opportunity to learn new technologies and approaches and to expand their tool belt. Using tools that have been around for decades, however, can save time and budget. When evaluating tools, it is more important to understand the problem to be solved than to jump straight into the tools.
-
Grok-2 Beta Version Released on X Platform
The Grok-2 language model has been released in beta on the X platform, introduced alongside Grok-2 mini. The model, tested under the designation "sus-column-r" on the LMSYS leaderboard, has achieved a higher Elo Score compared to Claude 3.5 Sonnet and GPT-4-Turbo. Grok-2 mini, a smaller variant, is also part of the beta release, designed to offer a balance between speed and performance.
-
Microsoft Launches Open-Source Phi-3.5 Models for Advanced AI Development
Microsoft launched three new open-source AI models in its Phi-3.5 series: Phi-3.5-mini-instruct, Phi-3.5-MoE-instruct, and Phi-3.5-vision-instruct. Available under a permissive MIT license, these models offer developers powerful tools for various tasks, including reasoning, multilingual processing, and image and video analysis.
-
Microsoft Releases Prompty: New VS Code Extension for Integrating LLMs into .NET Development
Microsoft has released a new Visual Studio Code extension called Prompty, designed to integrate Large Language Models (LLMs) like GPT-4o directly into .NET development workflows. This free tool aims to simplify the process of adding AI-driven capabilities to applications. The official release post includes a practical example demonstrating how Prompty can be used in real-world scenarios.
-
Google Cloud Launches C4 Machine Series: High-Performance Computing and Data Analytics
Google Cloud recently announced the general availability of its new C4 machine series, powered by 4th Gen Intel Xeon Scalable Processors (Sapphire Rapids). The series offers a range of configurations tailored to meet the needs of demanding applications such as high-performance computing (HPC), large-scale simulations, and data analytics.
-
Advancing System Reliability: Meta's AI-Driven Approach to Root Cause Analysis
Meta recently shared how they are enhancing their system reliability through advanced investigation tools, including the AI-assisted Hawkeye, which aids in debugging machine learning workflows. By integrating Artificial Intelligence, Meta has developed a new investigation system that combines heuristic-based retrieval with large language model (LLM) ranking to assist in root cause analysis.
-
LLM-Powered DevOps Assistant Clio Launches to Help Engineers Manage Cloud Infrastructure
Acorn, a company founded to help organisations implement AI-based technology solutions, has recently introduced Clio, an AI-powered command-line assistant designed to help engineers manage DevOps workflows. In a blog post and a YouTube live stream, Acorn describes Clio as a "slightly grumpy but friendly assistant" which can handle various DevOps-related tasks through natural language.
-
Pinterest Modernises Machine Learning Infrastructure with Ray
Pinterest, the visual discovery platform, has revealed details about its journey to modernise its machine learning infrastructure using Ray, an open-source distributed computing framework. In a recent blog post, the company shared insights into the challenges faced and solutions implemented as they integrated Ray into their large-scale production environment.
-
NVIDIA NIM Now Available on Hugging Face with Inference-as-a-Service
Hugging Face has announced the launch of an inference-as-a-service capability powered by NVIDIA NIM. This new service will provide developers easy access to NVIDIA-accelerated inference for popular AI models.