Remove Database Remove Kafka Remove PostgreSQL
article thumbnail

End-to-End Data Engineering System on Real Data with Kafka, Spark, Airflow, Postgres, and Docker

Towards Data Science

This involves getting data from an API and storing it in a PostgreSQL database. In the second phase, we’ll develop an application that uses a language model to interact with this database. The second article, which will come later, will delve into creating agents using tools like LangChain to communicate with external databases.

Kafka 98
article thumbnail

Easier Stream Processing On Kafka With ksqlDB

Data Engineering Podcast

The ksqlDB project was created to address this state of affairs by building a unified layer on top of the Kafka ecosystem for stream processing. Developers can work with the SQL constructs that they are familiar with while automatically getting the durability and reliability that Kafka offers. How is ksqlDB architected?

Kafka 100
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

TimescaleDB: The Timeseries Database Built For SQL And Scale - Episode 65

Data Engineering Podcast

How has the market for timeseries databases changed since we last spoke? How have the improvements and new features in the recent releases of PostgreSQL impacted the Timescale product? How has the market for timeseries databases changed since we last spoke? Can you refresh our memory about what TimescaleDB is?

Database 100
article thumbnail

Change Data Capture For All Of Your Databases With Debezium

Data Engineering Podcast

Summary Databases are useful for inspecting the current state of your application, but inspecting the history of that data can get messy without a way to track changes as they happen. How has the tight coupling with Kafka impacted the direction and capabilities of Debezium? What, if any, other substrates does Debezium support (e.g.

Database 100
article thumbnail

Streaming Heterogeneous Databases with Kafka Connect – The Easy Way

Confluent

However, not all databases can be in the […]. Building a Cloud ETL Pipeline on Confluent Cloud shows you how to build and deploy a data pipeline entirely in the cloud.

Database 102
article thumbnail

Getting Started with Rust and Apache Kafka

Confluent

We’ll also take a look at some performance tests to see if Rust might be a viable alternative for Java applications using Apache Kafka ®. In this case, that means a command is created for a particular action, which will be assigned to a Kafka topic specific for that action. On May 15, 2015, the Core Kafka team released version 1.0

Kafka 18
article thumbnail

Kafka Connect Deep Dive – JDBC Source Connector

Confluent

One of the most common integrations that people want to do with Apache Kafka ® is getting data in from a database. That is because relational databases are a rich source of events. The existing data in a database, and any changes to that data, can be streamed into a Kafka topic. Setting the Kafka message key.

Kafka 89