|
1. How can I create a kafka table that can use headers and map them to columns? Currently, I am using KafkaDeserilizationSchema to create a DataStream, and then I convert that DataStream into a Table. I would like to use a more direct approach.
2. What is the recommended way to enrich a kafka table or data-stream with data-from postgres?
a) kafka table and JDBC temporal dimension table with temporal join and lookup cache setup
b) data-stream with async io which connects via JDBC. (note that asycio does not support Keyed State cache)
c) data-stream rich function or process function that uses Keyed State.
3. When using a kafka told and JDBC temporal dimension table how do I prevent N + 1 queries per join row?
When I issued a query such as this:
SELECT k.name, t1.id, t2.metadata, SUM(k.cost)
FROM kafka_table AS k
JOIN jdbc_table_one AS t1 ON k.t1_id = t1.ID
LEFT JOIN jdbc_table_two FOR SYSTEM_TIME AS OF k.proc_time AS t2 ON t1.t2_id = t2.id AND t2.name = k.name
GROUP BY TUMBLE (k.proc_time, INTERVAL '3' MINUTE), k.name, t1.id, t2.metadata
My PostgreSQL sql logs show that jdbc_table_two has a query per each distinct t2.name.
In a real production system, that would be 200,000 queries!
4. When using a JDBC temporal dimension table does Flink retrieve the from the database asynchronously , or is it possible for Flink to multiple join rows at time with a IN (subquery) syntax?
|