InfoQ Homepage Performance & Scalability Content on InfoQ
-
Unlimited Load Testing for MSDN Subscribers
Visual Studio Ultimate with MSDN subscribers will receive the new Visual Studio 2010 Load Test Feature Pack with Visual Studio 2010 Service Pack 1. Developers are now able to load-test applications with no limit on virtual users and no need to purchase extra user packs.
-
Palladio provides Version 3.2 of its Software Architecture Simulator
Palladio Bench supports architects and developers in deriving quality of software estimations such as performance, reliability, maintainability and development costs from models. Its new version 3.2 is based on Eclipse 3.5 (Galileo) and adds several stabilizations and fixes, a reliability analysis, a new graphical editor and a new workflow engine, as well as a probe specification framework.
-
Google Page Speed Goes Online and Mobile
Google has made Page Speed available online, enhancing it for analyzing web pages targeted at smartphones.
-
Visual C++ gets GPU-accelerated graphics and animations, Intrinsics
The new Visual Studio 2010 SP1 improves Visual C++ with GPU-accelerated graphics and animations support and Compiler-Intrinsics, which allow for highly efficient computing.
-
Google Revamps Native Client, but will it be taking it all the way?
About a year after it’s original announcement, Google released a new version of the Native Client (NaCl) SDK, which allows the safe execution of native code from a web browser. It is unclear though, if this ambitious project will be making it to production, or have the fate of other projects like Wave or Gears.
-
Partitioning and Sharding Options for SQL Server and SQL Azure
Horizontal partitioning and sharding are important tools for developers working with extremely large datasets. We take a brief look at what’s available in SQL Server and what’s coming down the road.
-
NASA’s OODT selected as an Apache Top Level Project
The Apache Software Foundation has selected the Object Oriented Data Technology architecture to become one of its Top-Level-Projects (TLP). Originally created by NASA’s Jet Propulsion Laboratory, Pasadena, OODT allows transparent integration of geographically distributed and disparate computing and data resources via metadata middleware.
-
Targeting the GPU with GPU.NET
GPU.NET is a managed solution integrated with Visual Studio 2010 for .NET developers and aimed at creating calculation intensive applications for GPU.
-
Yahoo! Releases S4, a Real Time, Distributed Stream Computing Platform
This month, Yahoo! released a new open source framework for "processing continuous, unbounded streams of data." The framework, named S4, allows for massively distributed computations over data that is constantly changing. InfoQ examines some of the examples and compares S4 to other technologies.
-
MySQL/HandlerSocket and VoltDB: Contenders to NoSQL
NoSQL systems are considered by some as performing better than traditional SQL ones. Two SQL solutions, one based on MySQL plus a NoSQL layer used as a plug-in and VoltDB claim SQL still is a viable solution for large applications with high scalability needs.
-
Azul’s Zing Elastic Java Runtime for x86 is Generally Available from Today
Azul’s Zing is generally available from today, bringing their highly-scalable Java architecture to x86-based servers. InfoQ spoke to George Gould and Gil Tene about the launch, performance figures and licensing costs.
-
JINSPIRED Releases New Version of Lightweight Java Monitoring Tool: OpenCore
OpenCore, a lightweight Java application performance monitor by JINSPIRED, released version 6.0 this month. InfoQ reviews what lightweight monitoring is and some of the terms and concepts involved
-
New Relic Has Released RPM for .NET and PHP
New Relic has released two new variants of its performance tool: RPM for .NET and RPM for PHP. RPM offers performance monitoring and analysis for web applications running on premises or in the cloud.
-
Percolator: a System for Incrementally Processing Updates to a Large Data Set
Google's Daniel Peng and Frank Dabek published a paper on "Large-scale Incremental Processing Using Distributed Transactions and Notifications” explaining that databases do not meet the storage or throughput requirements for Google's indexing system which stores tens of petabytes of data and processes billions of updates per day on thousands of machines.
-
Google WebP - Creating Smaller Images for Faster Pages
Google wants to shrink images transferred over the Internet by proposing a new lossy format called WebP. They claim they have achieved 39% reduction in image byte size leading to speedier page load.