AWS Database Blog

Category: Technical How-to

Create a Knowledge Graph application with metaphactory and Amazon Neptune

In a previous post, we described how to connect Amazon Neptune to metaphactory, securely, and then how to explore and search the Neptune graph data using metaphactory. In this post, we show how you can use metaphactory to build an end user application using its dynamic model driven components, driven by SPARQL queries.

Configure SSL encryption on an SAP ASE source endpoint in AWS DMS

In this post, we walk you through how to configure Secure Sockets Layer (SSL) encryption between the source endpoints in AWS DMS and an on-premises SAP ASE source for secure data transfer. We also show you the steps for enabling SSL on an on-premises SAP ASE database. Configuring SSL encryption on source endpoints enables encrypting data in transit during the database migration process for enhanced security.

Adding real-time ML predictions for your Amazon Aurora database: Part 2

In this post, we discuss how to implement Aurora ML performance optimizations to perform real-time inference against a SageMaker endpoint at a large scale. More specifically, we simulate an OLTP workload against the database, where multiple clients are making simultaneous calls against the database and are putting the SageMaker endpoint under stress to respond to thousands of requests in a short time window. Moreover, we show how to use SQL triggers to create an automatic orchestration pipeline for your predictive workload without using additional services.

Automate cross-account backup of Amazon RDS for Oracle including database parameter groups, option groups and security groups

In this post, we showcase AWS Backup and CloudFormation support feature of AWS Backup to automate the backup of Amazon RDS for Oracle, including customized database resources such as database parameter group, option group, and security group across AWS accounts.

Perform a two-step database migration from an on-premises Oracle database to Amazon RDS for Oracle using RMAN

In this post, we discuss how to perform a homogeneous migration from an on-premises Oracle database to Amazon Relational Database Service (Amazon RDS) for Oracle. For our solution, we use a two-step approach to migrate the source database to Amazon RDS for Oracle. First, we use RMAN to restore the RMAN backup on an EC2 instance, then we use Data Pump to export data to Amazon S3 and restore that in the RDS for Oracle database.

Use Spring Cloud to capture Amazon DynamoDB changes through Amazon Kinesis Data Streams

In this post, we demonstrate how you can use Spring Cloud to interact with Amazon DynamoDB and capture table-level changes using Kinesis Data Streams through familiar Spring constructs. We run you through a basic implementation and configuration that will help you get started.

Application Continuity for Oracle workloads with Amazon RDS Custom for Oracle

In this post, we show you how to implement Application Continuity in an RDS Custom for Oracle environment using a sample application. We also show you how to test the implementation to see that, when an outage occurs at the database tier, the application recovers and resumes without any data loss—automatically and transparently—along with the database failover. Finally, we show you how to verify the results before cleaning up the environment.

Optimize Amazon RDS costs for predictable workloads with automated IOPS and throughput scaling

In this post, we explain how you can use Amazon RDS IOPS and throughput provisioned settings, automate scaling around monthly and seasonal peaks, and decrease settings during slower weeks. By right-sizing IOPS and throughput levels to your workload’s typical cycles, you can reduce Amazon RDS spend while still getting great performance when you need it most.

Privileged Database User Activity Monitoring using Database Activity Streams(DAS) and Amazon OpenSearch Service

In this post, we demonstrate how to create a centralized monitoring solution using Database Activity Streams and Amazon OpenSearch Service to meet audit requirements. The solution enables the security team to gather audit data from several Kinesis data streams, enrich, process, and store it with retention to meet compliance requirements, and produce relevant alarms and dashboards.