AWS Storage Blog
Category: Thought Leadership
How Trend Micro uses Amazon S3 Object Lambda to help keep sensitive data secure
Does your application handle data that is uploaded by hundreds of thousands of end users? Is that same underlying data then shared across the same magnitude of users? Being able to scan data for malware before it’s returned to an application helps keep sensitive data secure, provides protection regardless of when the data was initially […]
Compressing and archiving logs to the Amazon S3 Glacier storage classes
In distributed architectures, there is often a need to preserve application logs, and for AWS customers preservation is often done via an Amazon S3 bucket. The logs may contain information on runtime transactions, error/failure states, or application metrics and statistics. These logs are later used in business intelligence to provide useful insights and generate dashboards, […]
How Cohesity uses Amazon EBS direct APIs to accelerate application backup and recovery times
When backing up applications, AWS Backup and Restore Partners seek methods that minimize complexity and reduce costs for their customers. Most backup applications protect Amazon Elastic Block Store (EBS) volumes using EBS snapshots as a part of Amazon EC2 protection feature. For backups with long term retention, backup applications offer additional streaming backup capabilities that […]
Automate and centrally manage data protection for Amazon S3 with AWS Backup
Customers globally, especially in regulated industries, require centralized protection and demonstrable compliance for their application data. Centralized data protection and enhanced visibility across backup operations can reduce the risks of costly disasters and accidents, improve business continuity, and simplify the auditing process. With AWS Backup for Amazon S3 now being generally available, you can centralize […]
Caching data using Amazon FSx for NetApp ONTAP
Accessing network attached storage (NAS) over long distances can introduce latency that can potentially impact business processes, slow down engineering jobs, and increase costs. Often only a small percent of the dataset needs to be accessed at any given time, allowing caching data locally to solve these challenges without replicating the full dataset. This is […]
Considering four different replication options for data in Amazon S3
UPDATE (2/10/2022): Amazon S3 Batch Replication launched on 2/8/2022, allowing you to replicate existing S3 objects and synchronize your S3 buckets. See the S3 User Guide for additional details. UPDATE (5/1/2023): Updated the comparison table to reflect new Amazon S3 features and capabilities. As your business grows and accumulates more data over time, you may […]
Optimizing SAS Grid on AWS with Amazon FSx for Lustre
Many customers run complex analytics and high performance SAS-based applications on premises using the SAS Grid platform to perform large-scale analytics. Customers with a strategy to move to open-source or cloud-native solutions will often consider refactoring applications to Python or R to lower their total cost of ownership, however refactoring these applications as part of […]
Achieving data consistency with AWS Elastic Disaster Recovery
When designing a disaster recovery (DR) plan, data consistency is an important factor to consider. This is especially important when protecting certain database applications, such as Microsoft SQL, Oracle, and SAP Hana. These database workloads must be restored to a consistent state to avoid database corruption. Failing to do so could result in loss of […]
AWS SAN in the Cloud: Millions of IOPs and tens of GB/s to any Amazon EC2 instance
Over the years, traditional on-premises applications have relied on ever more powerful (and expensive) storage arrays to scale application IO performance and provide a single server with millions of IOPs. Meanwhile, AWS was perfecting the art of linear, elastic horizontal scalability of storage performance in the cloud. Until now, it’s been difficult to meet the […]
How CineSend manages their media content using S3 Intelligent-Tiering
Is your organization managing terabytes (or even petabytes) of data stored as objects across hundreds if not thousands of buckets on Amazon S3? What are the chances that the access patterns and application requirements for all of these objects are the same? For most companies out there – slim to none. We operate in a […]