Remove Data Ingestion Remove Events Remove Metadata
article thumbnail

Level Up Your Data Platform With Active Metadata

Data Engineering Podcast

Summary Metadata is the lifeblood of your data platform, providing information about what is happening in your systems. In order to level up their value a new trend of active metadata is being implemented, allowing use cases like keeping BI reports up to date, auto-scaling your warehouses, and automated data governance.

Metadata 130
article thumbnail

Data Engineering Zoomcamp – Data Ingestion (Week 2)

Hepta Analytics

DE Zoomcamp 2.2.1 – Introduction to Workflow Orchestration Following last weeks blog , we move to data ingestion. We already had a script that downloaded a csv file, processed the data and pushed the data to postgres database. This week, we got to think about our data ingestion design.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Tame The Entropy In Your Data Stack And Prevent Failures With Sifflet

Data Engineering Podcast

Instead of trapping data in a black box, they enable you to easily collect customer data from the entire stack and build an identity graph on your warehouse, giving you full visibility and control. Sign up free at dataengineeringpodcast.com/rudder Data teams are increasingly under pressure to deliver. In fact, while only 3.5%

Data Lake 130
article thumbnail

The Data Integration Solution Checklist: Top 10 Considerations

Precisely

A true enterprise-grade integration solution calls for source and target connectors that can accommodate: VSAM files COBOL copybooks open standards like JSON modern platforms like Amazon Web Services ( AWS ), Confluent , Databricks , or Snowflake Questions to ask each vendor: Which enterprise data sources and targets do you support?

article thumbnail

New Snowflake Features Released in January 2024

Snowflake

Snowpark Updates Model management with the Snowpark Model Registry – public preview Snowpark Model Registry is an integrated solution to register, manage and use models and their metadata natively in Snowflake. As a result, you should not rely on any forward-looking statements as predictions of future events. Learn more here.

article thumbnail

An Engineering Guide to Data Quality - A Data Contract Perspective - Part 2

Data Engineering Weekly

WAP [Write-Audit-Publish] Pattern The WAP pattern follows a three-step process Write Phase The write phase results from a data ingestion or data transformation step. In the 'Write' stage, we capture the computed data in a log or a staging area. Event Routers typically don’t alter the payload.

article thumbnail

Sysmon Security Event Processing in Real Time with KSQL and HELK

Confluent

During a recent talk titled Hunters ATT&CKing with the Right Data , which I presented with my brother Jose Luis Rodriguez at ATT&CKcon, we talked about the importance of documenting and modeling security event logs before developing any data analytics while preparing for a threat hunting engagement. Why KSQL and HELK?

Process 83