InfoQ Homepage S3 Content on InfoQ
-
Designing for Failure in the BBC's Analytics Platform
Last week at InfoQ Live, Blanca Garcia-Gil, principal systems engineer at BBC, gave a session on Evolving Analytics in the Data Platform. During this session, Garcia-Gil focused on how her team prepared and designed for two types of failure - "known unknowns" and "unknown unknowns."
-
AWS Outposts Now Supports Amazon Elastic Block Store Local Snapshots
Recently AWS announced that its Outposts service now supports Amazon Elastic Block Store (ESB) local snapshots. With Amazon EBS Local Snapshots on Outposts, customers can store snapshots of their Amazon EBS volumes locally on Amazon S3 on Outposts to meet their data residency and local data processing needs.
-
Amazon S3 Now Delivers Strong Read-After-Write Consistency
To guarantee higher availability and better performances, S3 has for years relied on an eventual consistency model. During the first week of re:invent, AWS announced that S3 now supports strong read-after-write consistency.
-
AWS Announces New Service: Amazon S3 Storage Lens
AWS announces a new service called Amazon S3 Storage Lens, which can provide customers with organization-wide visibility into their object storage usage and activity trends. With the service, they can understand what happens across their S3 Object Storage installations.
-
AWS Announces New S3 Intelligent-Tiering Archive Access Tiers
S3 Intelligent-Tiering is a storage class, which AWS added to Amazon S3 for optimization of storage costs for its users. The company recently announced it would add two new archive access tiers designed for asynchronous access optimized for rare access at a very low cost: Archive Access tier and Deep Archive Access tier.
-
Amazon Updates S3 Service with Same-Region Replication
Recently, Amazon introduced a new option to its cloud storage service S3 - Same Region Replication (SRR). With this new option in S3, customers can now create a replica of their uploaded data in the same region yet in a different S3 bucket.
-
Amazon Offers Sustainability Datasets for Analysis
Amazon Web Services Open Data (AWSOD) and Amazon Sustainability (AS) are working together to make sustainability datasets available on the AWS Simple Storage Service (S3), and they are removing the undifferentiated heavy lifting by pre-processing the datasets for optimal retrieval. Sustainable datasets are commonly from satellites, geological studies, weather radars, agricultural studies, etc.
-
Amazon Introduces Intelligent-Tiering for S3 Storage to Automatically Optimize Costs
Amazon introduces a new S3 Storage tier called Intelligent-Tiering which optimizes storage costs by automatically selecting the most cost-effective storage tier based on usage patterns. This new tier can help reduce costs for customers infrequently accessing their data in S3.
-
Amazon S3 Increases Request Rate Performance and Drops Randomized Prefix Requirement
Amazon Web Services (AWS) recently announced significantly increased S3 request rate performance and the ability to parallelize requests to scale to the desired throughput. Notably this performance increase also "removes any previous guidance to randomize object prefixes" and enables the use of "logical or sequential naming patterns in S3 object naming without any performance implications".
-
Instana Extends AI Application Monitoring to AWS Lambda
Instana, a cloud-native provider of artificial intelligence based monitoring tools for dynamic containerized microservice applications, has extended support to include AWS Lambda, a serverless computing platform and also announced availability through the AWS Marketplace.
-
Amazon S3 Receives New Updates: New Storage Class and General Availability of S3 Select
Amazon S3 recently received new updates to its service: a new storage class S3 One Zone-Infrequent Access (Z-IA) Storage Class and S3 Select capability, now generally available for customers. With S3 Select, customers can get a subset of data faster, while the new storage One Zone IA class is a low-cost option for retrieving infrequent data.
-
A Human Error Took Down AWS S3 US-EAST-1
A mistake took down more S3 servers than it should, including two subsystems essential to S3 operation. This resulted in S3 failure, affecting the S3 service and other services depending on it. Normal functioning was restored in about four hours.
-
Fatigue, Spam, and Lack of Backups Take down GitLab.com
What started out as an attempt to protect GitLab.com from spammers turned sour as engineer fatigue and a lack of backups took the site down for nearly 18 hours and the loss of six hours worth of production data.
-
Netflix Details Evolution of Keystone Data Pipeline
Netflix has shed light on how the company uses the latest version of their Keystone Data Pipeline, a petabyte-scale real-time event stream processing system for business and product analytics. This news summarizes the three major versions of the pipeline, now used by almost every application at Netflix.
-
AWS Simplifies Resource Access with VPC Endpoints, Initially Supporting S3
Amazon Web Services recently introduced VPC endpoints to enable a "private connection between your VPC and another AWS service without requiring access over the Internet, through a NAT instance, a VPN connection, or AWS Direct Connect". VPC endpoint policies provide granular access control to other service's resources. Initially available are connections to S3, other services will be added later.