InfoQ Homepage Artificial Intelligence Content on InfoQ
-
QCon London: Lessons Learned from Building LinkedIn’s AI/ML Data Platform
At the QCon London 2024 conference, Félix GV from LinkedIn discussed the AI/ML platform powering the company’s products. He specifically delved into Venice DB, the NoSQL data store used for feature persistence. The presenter shared the lessons learned from evolving and operating the platform, including cluster management and library versioning.
-
QCon London: gRPC Migration Automation at LinkedIn
At QCon London 2024, Karthik Ramgopal and Min Chen described how AI helped LinkedIn change the remote procedure calls (RPC) protocol for 50,000 production endpoints from Rest.li to Google's gRPC. A planned 2-3 year manual migration turned into an AI-supported migration lasting 2-3 quarters. It changed 20 million lines of code across 2000 services – without business interruption.
-
Large Language Models for Code by Loubna Ben Allal at QCon London
At QCon London, Loubna Ben Allal discussed Large Language Models (LLMs) for code. She discussed the lifecycle of code completion models, which consists of pre-training on vast codebases and finetuning and continuous adaptation. She specifically discussed open-source models, which are powered by platforms like Hugging Face.
-
Efficient DevSecOps Workflows with a Little Help from AI: Q&A with Michael Friedrich
At QCon London, Michael Friedrich, senior developer advocate at GitLab, discussed how AI can help in DevSecOps workflows. His session was part of the Cloud-Native Engineering track on the first day of the conference. InfoQ interviewed Friedrich after the session.
-
Navigating LLM Deployment: Tips, Tricks and Techniques by Meryem Arik at QCon London
At QCon London, Meryem Arik discussed deploying Large Language Models (LLMs). While initial proofs of concept benefit from hosted solutions, scaling demands self-hosting to cut costs, enhance performance with tailored models, and meet privacy and security requirements. She emphasized understanding deployment limits, quantization for efficiency, and optimizing inference to fully use GPU resources.
-
Nvidia Announces Robotics-Oriented AI Foundational Model
At its recent GTC 2024 event, Nvidia announced a new foundational model to build intelligent humanoid robots. Dubbed GR00T, short for Generalist Robot 00 Technology, the model will understand natural language and be able to observe human actions and emulate human movements.
-
Microsoft AI-Driven Security Tool Copilot for Security is Now GA
Microsoft recently announced the general availability of Copilot for Security, a generative Artificial Intelligence (AI) security product designed to help security and IT teams with the capabilities to protect their digital assets.
-
KubeCon EU Keynotes: a Call to Action to Innovate Responsibly with Generative AI
The KubeCon EU morning keynotes were a veritable call to action encouraging the cloud-native community's involvement in building the scalable infrastructure needed by generative AI. This call was balanced with encouragement to make a cloud-native platform’s “golden path” green and sustainable, ensuring that any innovation is also responsible.
-
Java News Roundup: Jakarta Data and Jakarta NoSQL Milestones, Class-File API Targeted for JDK 23
This week's Java roundup for March 25th, 2024, features news highlighting: JEP 466, Class-File API (Second Preview), targeted for JDK 23; milestone releases of Jakarta Data and Jakarta NoSQL specifications; the second release candidate for JobRunr 7.0.0; and point releases for Spring projects, Quarkus, Helidon and LangChain4j.
-
Databrix Announces DBRX, an Open Source General Purpose LLM
Databricks launched DBRX, a new open-source large language model (LLM) that aims to redefine the standards of open models and outperform well-known competitors on industry benchmarks.
-
Apple Researchers Detail Method to Combine Different LLMs to Achieve State-of-the-Art Performance
Many large language models (LLMs) have become available recently, both closed and open source further leading to the creation of combined models known as Multimodal LLMs (MLLMs). Yet, few or none of them unveil what design choices were made to create them, say Apple researchers who distilled principles and lessons to design state-of-the-art (SOTA) Multimodal LLMs.
-
xAI Releases Grok as an Open-Source Large Language Model
Elon Musk announced that xAI would make its AI chatbot Grok open source, and now the release is accessible on GitHub and Hugging Face. This move enables researchers and developers to expand upon the model, influencing how xAI evolves Grok in the face of competition from tech giants like OpenAI, Meta, Google, Microsoft, and others.
-
Netflix Uses Metaflow to Manage Hundreds of AI/ML Applications at Scale
Netflix recently published how its Machine Learning Platform (MLP) team provides an ecosystem around Metaflow, an open-source machine learning infrastructure framework. By creating various integrations for Metaflow, Netflix already has hundreds of Metaflow projects maintained by multiple engineering teams.
-
Copilot in Azure SQL Database in Private Preview
Microsoft has announced a private preview of Copilot for SQL Azure, which offers a natural language for SQL conversion and self-help for database administration.
-
Researchers Open-Source LLM Jailbreak Defense Algorithm SafeDecoding
Researchers from the University of Washington, the Pennsylvania State University, and Allen Institute for AI have open-sourced SafeDecoding, a technique for protecting large language models (LLMs) against jailbreak attacks. SafeDecoding outperforms baseline jailbreak defenses without incurring significant computational overhead.