InfoQ Homepage Performance Tuning Content on InfoQ
-
QCon London: Scaling Microservices Architecture and Technology Organization at Trainline
During the recent QCon London conference, Trainline’s CTO spoke about the evolution of the company’s system architecture and organizational structure over the last five years. The company had to adapt to market changes and growing customer expectations by improving the performance and reliability of its technology platform.
-
QCon London: Lessons Learned from Building LinkedIn’s AI/ML Data Platform
At the QCon London 2024 conference, Félix GV from LinkedIn discussed the AI/ML platform powering the company’s products. He specifically delved into Venice DB, the NoSQL data store used for feature persistence. The presenter shared the lessons learned from evolving and operating the platform, including cluster management and library versioning.
-
Expedia Speeds up Flights Search with Micro Frontends and GraphQL Optimizations
Expedia made flight search faster by up to 52% (page usable time) by applying a range of optimizations to web and mobile applications. To support these improvements, the company improved the observability of its applications. Expedia Flights web application has been migrated to Micro Frontend Architecture (MFA) to allow flexibility, reusability, and better optimization.
-
Discord Scales to 1 Million+ Online MidJourney Users in a Single Server
Discord optimized its platform to serve over one million online users in a single server while maintaining a responsive user experience. The company evolved the guild component, which is responsible for fanning out billions of message notifications, in a series of performance and scalability improvements supported by system observability and performance tuning.
-
Why LinkedIn chose gRPC+Protobuf over REST+JSON: Q&A with Karthik Ramgopal and Min Chen
LinkedIn announced that it would be moving to gRPC with Protocol Buffers for the inter-service communication in its microservices platform, where previously an open-source Rest.li framework was used with JSON as a primary serialization format. InfoQ contacted Karthik Ramgopal and Min Chen to learn more about the decision and company motivations behind it.
-
LinkedIn Migrates Espresso to HTTP2 and Reduces Connections by 88% and Latency by 75%
LinkedIn was able to dramatically improve the scalability and performance of its Espresso database by migrating it from HTTP1.1 to HTTP2, resulting in a reduction in the number of connections, latency, and garbage collection times. To achieve these gains, the team had to optimize the Netty’s default HTTP2 stack to make it fit their needs.
-
LinkedIn Adopts Protocol Buffers for Microservices Integration and Reduces Latency by up to 60%
LinkedIn adopted Protocol Buffers for exchanging data between microservices more efficiently across its platform and integrated it with Rest.li, their open-source REST framework. After the company-wide rollout, they reduced the latency by up to 60% and improved resource utilization at the same time.
-
Spring Boot and Azul JDK Support Java Startup Time Reducer CRaC
The OpenJDK project CRaC drastically reduces the startup time of a Java application and its Time to Peak performance. It does so by taking a memory snapshot at runtime and restoring it in later runs. Azul, the creator of CRaC, now ships an OpenJDK 17 distribution with built-in support for CRaC. Micronaut and Quarkus already support CRaC, and Spring Framework will do so in November 2023.
-
ETTrace is an Open-Source Profiler for iOS Aiming to Simplify Performance Optimization
Recently open-sourced by Emerge Tools, maker of several analysis tools for iOS apps, ETTrace aims to simplify iOS performance profiling by providing intuitive visualizations and straightforward operation.
-
From Extinct Computers to Statistical Nightmares: Adventures in Performance
Thomas Dullien, distinguished software engineer at Elastic, shared at QCon London some lessons learned from analyzing the performance of large-scale compute systems.
-
Google Cloud Introduces Startup CPU Boost for Cloud Run and Cloud Functions 2nd Gen
Google Cloud recently introduced startup CPU boost for Cloud Run and Cloud Functions 2nd gen, a new feature that allows developers to significantly reduce the cold start time of Cloud Run and Cloud Functions. The new capability is currently in preview.
-
How Lyft Reduced its Android App Launch Time by 21% in One Month
Based on the insights provided by Google's Android App Vitals, Lyft Android team improved their Android app's startup time by 21% and increased driver sessions by 5%.
-
Uber's Engineering Manages to Cut 70k CPUs by Tuning Go GC
In an effort to help the company become profitable, Uber’s engineering department has focused their efforts on making their infrastructure more efficient. As an outcome of this effort, they managed to develop a semi-automated GO Garbage Collection tuning mechanism which in turn saved 70K CPU cores across 30 mission critical services.
-
Amazon Elasticsearch Service Introduces Auto-Tune
Amazon has recently announced the Auto-Tune feature in Amazon Elasticsearch Service, a closed-loop control system that adapts the Elasticsearch cluster to the running workload. The new automated memory management provides better ingestion throughput for log analytics workloads and reduced tail latencies for search queries.
-
.NET 5 Runtime Improvements: from Functional to Performant Implementations
During a no-slides presentation at .NET Conf 2020, software architects from the .NET runtime team presented several .NET 5 runtime improvements and how they achieved them, including ARM64 support, HTTP/3, and single-file applications.