Nikola created FLINK-16860:
------------------------------
Summary: TableException: Failed to push filter into table source! when upgrading flink to 1.9.2
Key: FLINK-16860
URL:
https://issues.apache.org/jira/browse/FLINK-16860 Project: Flink
Issue Type: Bug
Components: Connectors / ORC
Affects Versions: 1.10.0, 1.9.2
Environment: flink 1.8.2
flink 1.9.2
Reporter: Nikola
Attachments: flink-1.8.2.txt, flink-1.9.2.txt
We have a batch job which we currently have on a flink cluster running 1.8.2
The job runs fine. We wanted to upgrade to flink 1.10, but that yielded errors, so we started downgrading until we found that the issue is in flink 1.9.2
The job on 1.9.2 fails with:
{code:java}
Caused by: org.apache.flink.table.api.TableException: Failed to push filter into table source! table source with pushdown capability must override and change explainSource() API to explain the pushdown applied!{code}
Which is not happening on flink 1.8.2
I tried to narrow it down and it seems that this exception has been added in FLINK-12399 and there was a small discussion regarding the exception: [
https://github.com/apache/flink/pull/8468#discussion_r329876088]
Our code looks something like this:
{code:java}
String tempTableName = "tempTable";
String sql = SqlBuilder.buildSql(tempTableName);
BatchTableEnvironment tableEnv = BatchTableEnvironment.create(env);
OrcTableSource orcTableSource = OrcTableSource.builder()
.path(hdfsFolder, true)
.forOrcSchema(ORC.getSchema())
.withConfiguration(config)
.build();
tableEnv.registerTableSource(tempTableName, orcTableSource);
Table tempTable = tableEnv.sqlQuery(sql);
return tableEnv.toDataSet(tempTable, Row.class);
{code}
Where the sql build is nothing more than
{code:java}
SELECT * FROM table WHERE id IN (1,2,3) AND mid IN(4,5,6){code}
--
This message was sent by Atlassian Jira
(v8.3.4#803005)