InfoQ Homepage Artificial Intelligence Content on InfoQ
-
Google Cloud Launches C4 Machine Series: High-Performance Computing and Data Analytics
Google Cloud recently announced the general availability of its new C4 machine series, powered by 4th Gen Intel Xeon Scalable Processors (Sapphire Rapids). The series offers a range of configurations tailored to meet the needs of demanding applications such as high-performance computing (HPC), large-scale simulations, and data analytics.
-
Advancing System Reliability: Meta's AI-Driven Approach to Root Cause Analysis
Meta recently shared how they are enhancing their system reliability through advanced investigation tools, including the AI-assisted Hawkeye, which aids in debugging machine learning workflows. By integrating Artificial Intelligence, Meta has developed a new investigation system that combines heuristic-based retrieval with large language model (LLM) ranking to assist in root cause analysis.
-
LLM-Powered DevOps Assistant Clio Launches to Help Engineers Manage Cloud Infrastructure
Acorn, a company founded to help organisations implement AI-based technology solutions, has recently introduced Clio, an AI-powered command-line assistant designed to help engineers manage DevOps workflows. In a blog post and a YouTube live stream, Acorn describes Clio as a "slightly grumpy but friendly assistant" which can handle various DevOps-related tasks through natural language.
-
Pinterest Modernises Machine Learning Infrastructure with Ray
Pinterest, the visual discovery platform, has revealed details about its journey to modernise its machine learning infrastructure using Ray, an open-source distributed computing framework. In a recent blog post, the company shared insights into the challenges faced and solutions implemented as they integrated Ray into their large-scale production environment.
-
NVIDIA NIM Now Available on Hugging Face with Inference-as-a-Service
Hugging Face has announced the launch of an inference-as-a-service capability powered by NVIDIA NIM. This new service will provide developers easy access to NVIDIA-accelerated inference for popular AI models.
-
GitHub Models Brings New AI Playground and Tight Integration with Other GitHub Tools
GitHub has launched GitHub Models, a free capability aimed at letting developers explore various AI models from within the GitHub tool ecosystem and make it easier to deploy AI-based services using Azure AI. GitHub Models includes both private and public models and is currently in closed preview.
-
Increasing Productivity by Becoming a Dual-Purpose Stream Aligned and Platform Software Team
To manage their increased workload effectively and maintain quality and efficiency, a software team decided to become dual-purpose: stream-aligned and platform. They rewrote their main application to be API-first and implemented micro releases with their customer-facing products, to provide value to their end users quickly and maintain a steady flow of accomplishments for the team.
-
Meta Releases Llama 3.1 405B, Largest Open-Source Model to Date
Meta recently unveiled its latest language model, Llama 3.1 405B. This AI model is the largest of the new Llama models, which also include 8B and 70B versions. With 405 billion parameters, 15 trillion tokens, and 16,000 GPUs, Llama 3.1 405B offers a range of impressive features.
-
AWS Introduces Amazon Q Developer in SageMaker Studio to Streamline ML Workflows
AWS announced that Amazon SageMaker Studio now includes Amazon Q Developer as a new capability. This generative AI-powered assistant is built natively into SageMaker’s JupyterLab experience and provides recommendations for the best tools for each task, step-by-step guidance, code generation, and troubleshooting assistance.
-
Amazon Brings AI Assistant to Software Development as Part of Amazon Q Suite
Amazon has recently released Amazon Q Developer Agent, an AI-powered assistant that uses natural language input from developers to generate features, bug fixes, and unit tests within an integrated development environment (IDE). It employs large language models and generative AI to understand a developer's natural language request, and then generate the necessary code changes.
-
InfoQ Dev Summit Boston: Being a Responsible Developer in the Age of AI Hype
At InfoQ Dev Summit in Boston, Justin Sheehy of Akamai delivered an insightful opening keynote on being a responsible developer in the age of artificial intelligence (AI) hype. The talk was aimed at software practitioners who might be feeling overwhelmed by the rapid developments and inflated expectations surrounding AI.
-
Generative AI Capabilities for Logic Apps Standard with Azure OpenAI and AI Search Connectors
Microsoft has announced that the Azure OpenAI and Azure AI Search connectors for Logic Apps Standard are now generally available, following an earlier public preview. These connectors are fully integrated into Azure Integration Services, providing developers with powerful tools to enhance application functionality with advanced AI capabilities.
-
Mistral Introduces AI Code Generation Model Codestral
Mistral AI has unveiled Codestral, its first code-focused AI model. Codestral helps the developers with coding tasks offering efficiency and accuracy in code generation.
-
Recap of MSBuild 2024: Copilot AI Agents, Phi-3, GPT-4o on Azure AI
Microsoft recently held its annual MSBuild developer conference, where it made several significant announcements, including updates to its AI capabilities, focusing on Copilot AI Agents, Phi-3, and GPT-4o now available on Azure AI.
-
University of Washington AI-Powered Headphones Let Users Listen to a Single Person in a Crowd
"Target speech hearing" is a new deep-learning algorithm developed at the University of Washington to allow users to "enroll" a speaker and cancel all environmental noise surrounding their voice.