How to Create a Data Source for Data Streamed from Kafka
Create a data source for data streamed from Kafka in DataPancake.
Last updated
Create a data source for data streamed from Kafka in DataPancake.
Last updated
Open a new snowflake worksheet and run the necessary statements for the database, schema, and table which contain the Kafka data. For example:
These are simply used for filtering / searching. For example: dev
, prod
, api
, csv
If checked, two dynamic tables will be produced for the root attributes in the semi-structured data source. The first dynamic table will be used to flatten the root attributes. The second dynamic table will be used to filter the flattened rows using a window function (configured in the next step) to produce the most recent message for each primary key.
The sql expression used to deduplicate the source stream messages based on a primary key and sort order to produce the most recent message. This value is required if the SQL Code Generation feature is selected and the Deduplicate Messages option is enabled.
This column name can be a sql expression that refers to a specific path of the semi-structured data source. See the example below.
How to configure dynamic table settings (include metadata How to configure the semantic layer How to configure schema consolidation How to configure schema filters
The save button is near the bottom of the page.