AWS Storage Blog
Tag: Amazon Simple Storage Service (Amazon S3)
Migrating Google Cloud Storage to Amazon S3 using AWS DataSync
Update (4/17/2024): The solution presented in this post using AWS DataSync for migration to Amazon S3 from Google Cloud Storage works best if you are looking for a secure managed service for your data transfer workflow that provides data validation, integrated auditing and monitoring capabilities, and the ability to transfer changed data. If you are […]
Restore data from Amazon S3 Glacier storage classes starting with partial object keys
When managing data storage, it is important to optimize for cost by storing data in the most cost-effective manner based on how often data is used or accessed. For many enterprises, this means using some form of cold storage or archiving for data that is less frequently accessed or used while keeping more frequently used […]
Reduce storage costs with fewer noncurrent versions using Amazon S3 Lifecycle
Keeping multiple copies and versions of data is a tried-and-true security and data protection protocol. In the event that one version is harmed or corrupted, another is ready as a backup. While increased security with multiple versions and copies is a plus, the added storage costs of (purposefully) redundant data must be carefully considered. It […]
Addressing I/O latency when restoring Amazon EBS volumes from EBS Snapshots
From a storage volume perspective, latency is the time elapsed between sending an I/O request to a volume and receiving an acknowledgement from the volume that the I/O read or write is complete. Latency is a key measurement for applications which are sensitive to the round trip time (RTT) of I/O operation. An example of […]
Updating Amazon S3 object ACLs at scale with S3 Batch Operations
Update (4/27/2023): Amazon S3 now automatically enables S3 Block Public Access and disables S3 access control lists (ACLs) for all new S3 buckets in all AWS Regions. Access control lists (ACLs) are permission sets associated with data or other system resources that dictate access permissions, and they have been a staple of data security for decades. […]
How Discover accelerates data ingestion using AWS PrivateLink for Amazon S3
Discover Financial Services (NYSE: DFS) is a digital banking and payment services company with one of the most recognizable brands in US financial services. Since its inception in 1986, Discover has become one of the largest card issuers in the United States. We are proud members of the platform team at Discover, where we are responsible […]
How to move and store your genomics sequencing data with AWS DataSync
Genomics data is expanding at a rate exceeding Moore’s law according to the National Human Genome Research Institute. As more sequencing data is produced and researchers move from genotyping to whole genome sequencing, the amount of data produced is outpacing on-premises capacity. Organizations need cloud solutions that help manage data movement, storage, and analysis. The […]
Optimize storage costs by analyzing API operations on Amazon S3
The demand for data storage has increased with the advent of a fast-paced data environment – creating, sharing, and replicating data at a large scale. Most organizations are looking for the optimal way to store their data cost-effectively, giving them everything they need from their data but without breaking the bank. Cloud storage provides flexible […]
How Simon Data reduced encryption costs by using Amazon S3 Bucket Keys on existing objects
As more organizations look to operate faster and at scale, they need ways to meet critical compliance requirements and improve data security. Encryption is a critical component of a defense in depth strategy, and when used correctly, can provide an additional layer of protection above basic access control. However, workloads that access millions or billions […]
Analytical processing of millions of cell images using Amazon EFS and Amazon S3
Analytical workloads such as batch processing, high performance computing, or machine learning inference often have high IOPS and low latency requirements but operate at irregular intervals on subsets of large datasets. Typically, data is manually copied between storage tiers in preparation of processing, which can be cumbersome and error-prone. Given this, IT teams want to […]