📄️ About Lakehouse
Lakehouse provides a unified platform for storing diverse data (structured, semi-structured, unstructured) and performing advanced analytics. Key capabilities include:
📄️ Lakehouse workflow
1. Connecting to data Sources: The journey begins with establishing connections to various data sources. Think of these as the starting points where your raw data lives. These could be relational databases like SQL Server, PostgreSQL, cloud data warehouses like Redshift, or even file storage systems such as S3 or Azure blob.. The crucial function here is to enable the Lakehouse to access the data that needs to be processed and analyzed.
🗃️ Data pipeline
7 items
🗃️ Data sources
5 items
🗃️ Orchestrations
11 items
📄️ Schema change detection
DataGOL's Schema Change Detection feature monitors the source tables involved in your pipelines for any modifications to their structure (e.g., new columns, deleted columns, data type changes). This helps you stay informed about potential impacts on your data pipelines and warehouse. When a schema change is detected, it is recorded, indicating the table, the modified columns, and the nature of the change (e.g., data type, size).
📄️ Jobs
Navigate to Jobs: Under the Lakehouse section, go to Jobs.