Hi Expert,
Usually when we write Flink-SQL program, usually we need to use multiple tables to get the final result, this is due to sometimes it is not possible to implement complicated logic in one SQL, sometimes due to the clarity of logic. For example: create view A as select * from source where xxx; create view B as select * from A where xxx; create view C as select * from B where xxx; insert into sink select * from C where xxx; But when we write complicated logic, we may accomplish it step by step, make sure that the first step is correct, then go on with the next step. In batch program such as Hive or Spark, we usually write SQL like this, step by step. For example: create view A as select * from source where xxx; I want to check if the content in A is correct, if it is correct I go on to write another SQL. But I do not want to define a sink for each step, because it is not worthy just create a sink for such a “debug” step. So is there a solution or best practice for such a scenario? How do we easily debug or verify the correctness of a Flink SQL program? Best Henry |
Hi Henry,
such a feature is currently under discussion [1] feel free to participate here and give feedback. So far you need to have some intermediate store usually this could be Kafka or a filesystem. I would recommend to write little unit tests that test each SQL step like it is done here [2]. I hope this helps. Regards, Timo [1] http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Support-Interactive-Programming-in-Flink-Table-API-tt25372.html#a25666 [2] https://github.com/apache/flink/blob/master/flink-libraries/flink-table/src/test/scala/org/apache/flink/table/runtime/stream/sql/SqlITCase.scala Am 07.01.19 um 16:19 schrieb 徐涛: > Hi Expert, > Usually when we write Flink-SQL program, usually we need to use multiple tables to get the final result, this is due to sometimes it is not possible to implement complicated logic in one SQL, sometimes due to the clarity of logic. For example: > create view A as > select * from source where xxx; > > create view B as > select * from A where xxx; > > create view C as > select * from B where xxx; > > insert into sink > select * from C where xxx; > > But when we write complicated logic, we may accomplish it step by step, make sure that the first step is correct, then go on with the next step. In batch program such as Hive or Spark, we usually write SQL like this, step by step. > > For example: > create view A as > select * from source where xxx; > I want to check if the content in A is correct, if it is correct I go on to write another SQL. But I do not want to define a sink for each step, because it is not worthy just create a sink for such a “debug” step. > So is there a solution or best practice for such a scenario? How do we easily debug or verify the correctness of a Flink SQL program? > > Best > Henry |
In reply to this post by 徐涛
In our case, we wrote a console table sink which print everything on the
console, and use "insert into" to write the interim result to console. -- Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/ |
Hi, A print user-defined table sink is helpful. I think a print user-defined UDF is another workaround. Hope this helps. Best, Hequn On Tue, Jan 8, 2019 at 1:45 PM yinhua.dai <[hidden email]> wrote: In our case, we wrote a console table sink which print everything on the |
Free forum by Nabble | Edit this page |