AWS Database Blog
Triple your knowledge graph speed with RDF linked data and openCypher using Amazon Neptune Analytics
There are numerous publicly available Resource Description Framework (RDF) datasets that cover a wide range of fields, including geography, life sciences, cultural heritage, and government data. Many of these public datasets can be linked together by loading them into an RDF-compatible database. In this post, we demonstrate how to build knowledge graphs with RDF linked data and openCypher using Amazon Neptune Analytics.
Optimizing costs on Amazon DocumentDB using event-driven architecture and the AWS EventBridge Terraform module
A primary reason companies move their workloads to AWS is because of cost. With AWS, cloud migration and application modernization plans are based on your business needs and not agreements or licensing. You can acquire technology on an as-needed basis, only paying for the resources you use. You can build modern, scalable applications on AWS […]
Build multi-tenant architectures on Amazon Neptune
In this post, we explore approaches that address operating Amazon Neptune in a multi-tenant SaaS environment, as well as the considerations that may influence how and when to apply these strategies depending on your tenant needs.
Build a custom HTTP client in Amazon Aurora PostgreSQL and Amazon RDS for PostgreSQL: An alternative to Oracle’s UTL_HTTP
Some customers use Oracle UTL_HTTP package to write PL/SQL programs that communicate with web (HTTP) servers and invoke third-party APIs. When migrating to Amazon Aurora PostgreSQL-Compatible Edition or Amazon Relational Database Service (Amazon RDS) for PostgreSQL, these customers need to perform a custom conversion of their SQL code since PostgreSQL does not offer a similar […]
Validate database object consistency after migrating from IBM Db2 z/OS to Amazon RDS for Db2
In this post, we delve into the best practices for migrating database objects from IBM Db2 z/OS to RDS for Db2 and walk you through how to validate these migrated database objects.
Improve speed and reduce cost for generative AI workloads with a persistent semantic cache in Amazon MemoryDB
In this post, we present the concepts needed to use a persistent semantic cache in MemoryDB with Knowledge Bases for Amazon Bedrock, and the steps to create a chatbot application that uses the cache. We use MemoryDB as the caching layer for this use case because it delivers the fastest vector search performance at the highest recall rates among popular vector databases on AWS. We use Knowledge Bases for Amazon Bedrock as a vector database because it implements and maintains the RAG functionality for our application without the need of writing additional code.
Build and deploy knowledge graphs faster with RDF and openCypher
Amazon Neptune Analytics now supports openCypher queries over RDF graphs. When you build an application that uses a graph database such as Amazon Neptune, you’re typically faced with a technology choice at the start: There are two different types of graphs, Resource Description Framework (RDF) graphs and labeled property graphs (LPGs), and your choice of […]
Monitor Amazon DynamoDB operation counts with Amazon CloudWatch
Amazon DynamoDB continuously sends metrics about its behavior to Amazon CloudWatch. Something I’ve heard customers ask for is how to get a count of successful requests of each operation type (for example, how many GetItem or DeleteItem calls were made) in order to better understand usage and costs. In this post, I show you how to retrieve this metric.
How to deploy Stacks blockchain nodes on AWS with the AWS Blockchain Node Runners Stacks blueprint
In this post, we demonstrate how to swiftly deploy Stacks blockchain nodes on AWS with the AWS Blockchain Node Runners blueprint.
Stream change data in a multicloud environment using AWS DMS, Amazon MSK, and Amazon Managed Service for Apache Flink
When workloads and their corresponding transactional databases are distributed across multiple cloud providers, it can create challenges in using the data in near real time for advanced analytics. In this post, we discuss architecture, approaches, and considerations for streaming data changes from the transactional databases deployed in other cloud providers to a streaming data solution deployed on AWS.