InfoQ Homepage Large language models Content on InfoQ
Articles
RSS Feed-
Efficient Resource Management with Small Language Models (SLMs) in Edge Computing
Small Language Models (SLMs) bring AI inference to the edge without overwhelming the resource-constrained devices. In this article, author Suruchi Shah dives into how SLMs can be used in edge computing applications for learning and adapting to patterns in real-time, reducing the computational burden and making edge devices smarter.
-
Virtual Panel: What to Consider When Adopting Large Language Models
Four experts discuss some issues people should think about when adopting LLMs and how they can make the best choice for their specific use case. Topics include how to choose between an API-based vs. self-hosted LLM, when to fine-tune an LLM, how to mitigate LLM risks, and what non-technical changes organizations need to make when adopting LLMs.
-
Navigating LLM Deployment: Tips, Tricks, and Techniques
This article focuses on self-hosted LLMs and how to get the best performance from them. The author provides best practices on how to overcome challenges due to model size, GPU scarcity, and a rapidly evolving field.
-
Article Series: Practical Applications of Generative AI
Generative AI (GenAI) has become a major component of the artificial intelligence (AI) and machine learning (ML) industry. However, using GenAI comes with challenges and risks. In the InfoQ "Practical Applications of Generative AI" article series, we present real-world solutions and hands-on practices from leading GenAI practitioners.
-
Llama 3 in Action: Deployment Strategies and Advanced Functionality for Real-World Applications
This article details the enhanced capabilities of the open-source Llama 3 LLM, and how businesses can adopt the model in their applications. The author gives step-by-step instructions for deploying Llama 3 in the cloud or on-premise, and how to leverage fine-tuned versions for specific tasks.
-
InfoQ AI, ML and Data Engineering Trends Report - September 2024
InfoQ editorial staff and friends of InfoQ are discussing the current trends in the domain of AI, ML and Data Engineering as part of the process of creating our annual trends report.
-
The AI Revolution Will Not Be Monopolized
Large language models have significantly transformed the field of artificial intelligence. The fundamental innovation behind this change is surprisingly straightforward: make the models a lot bigger. With each new iteration, the capabilities of these models expand, prompting a critical question: are we moving toward a black box era where AI is controlled by a few tech monopolies?
-
Experimenting with LLMs for Developer Productivity
This article describes an experiment that sought to determine if no-cost LLM-based code generation tools can improve developer productivity. The experiment evaluated several LLMs by generating unit tests for some open-source code and measuring the code coverage as well as the manual rework necessary to make the tests work.
-
InfoQ Software Architecture and Design Trends Report - April 2024
The InfoQ Trends Reports offer InfoQ readers a comprehensive overview of key topics worthy of attention. The reports also guide the InfoQ editorial team towards cutting-edge technologies in our reporting. In conjunction with the report and trends graph, our accompanying podcast features insightful discussions among the editors digging deeper into some of the trends.
-
Maximizing the Utility of Large Language Models (LLMs) through Prompting
In this article, authors Numa Dhamani and Maggie Engler discuss how prompt engineering techniques can help use the large language models (LLMs) more effectively to achieve better results. Prompting techniques discussed include few-shot, chain-of-thought, self-consistency, and tree-of-thoughts prompting.
-
InfoQ AI, ML, and Data Engineering Trends Report - September 2023
In this annual report, the InfoQ editors discuss the current state of AI, ML, and data engineering and what emerging trends you as a software engineer, architect, or data scientist should watch. We curate our discussions into a technology adoption curve with supporting commentary to help you understand how things are evolving.
-
AI-Based Prose Programming for Subject Matter Experts: Will This Work?
In this article, author Markus Völter discusses the future of programming using Large Language Model (LLM) tools like ChatGPT and GitHub’s Copilot for prose-to-code generation. He also talks about what new approaches and language changes need to be in place to help non-programmers take advantage of the "program in prose" techniques.