FLINK-9752 (s3 recoverable writer) not actually fixed in 1.6.2?

classic Classic list List threaded Threaded
5 messages Options
Reply | Threaded
Open this post in threaded view
|

FLINK-9752 (s3 recoverable writer) not actually fixed in 1.6.2?

Addison Higham
Hi all,

Been hitting my head against a wall for the last few hours. The release
notes for 1.6.2 show https://issues.apache.org/jira/browse/FLINK-9752 as
resolved in 1.6.2. I am trying to upgrade and switch some things to use the
StreamingFileSink against s3. However, when doing so, I get the following
error:

Recoverable writers on Hadoop are only supported for HDFS and for
Hadoop version 2.7 or newer


I started digging into the code and looking deeper, I don't see the code at
https://github.com/apache/flink/pull/6795 as being backported to 1.6.2?

Was the Fix Version erroneous? If so, are there plans to backport it? If
not, seems like that should be fixed in the release notes.

I have been waiting for that functionality and we already build our own
flink, so I am tempted to backport it onto 1.6.2... anything tricky about
that?

Thanks!
Reply | Threaded
Open this post in threaded view
|

Re: FLINK-9752 (s3 recoverable writer) not actually fixed in 1.6.2?

Till Rohrmann
Hi Addison,

I think the idea was to also backport this feature to 1.6 since we
considered it a bug that S3 was not supported in 1.6. I've pulled in Kostas
who worked on the S3 writer. @Klou did we intentionally not backport this
feature?

I think there should be nothing special about backporting this feature and
building your own version of Flink.

Cheers,
Till

On Tue, Oct 30, 2018 at 10:54 PM Addison Higham <[hidden email]> wrote:

> Hi all,
>
> Been hitting my head against a wall for the last few hours. The release
> notes for 1.6.2 show https://issues.apache.org/jira/browse/FLINK-9752 as
> resolved in 1.6.2. I am trying to upgrade and switch some things to use the
> StreamingFileSink against s3. However, when doing so, I get the following
> error:
>
> Recoverable writers on Hadoop are only supported for HDFS and for
> Hadoop version 2.7 or newer
>
>
> I started digging into the code and looking deeper, I don't see the code at
> https://github.com/apache/flink/pull/6795 as being backported to 1.6.2?
>
> Was the Fix Version erroneous? If so, are there plans to backport it? If
> not, seems like that should be fixed in the release notes.
>
> I have been waiting for that functionality and we already build our own
> flink, so I am tempted to backport it onto 1.6.2... anything tricky about
> that?
>
> Thanks!
>
Reply | Threaded
Open this post in threaded view
|

Re: FLINK-9752 (s3 recoverable writer) not actually fixed in 1.6.2?

Mike Mintz
FWIW I also tried this on Flink 1.6.2 today and got the same error. This is
my full stack trace:

java.lang.UnsupportedOperationException: Recoverable writers on Hadoop
are only supported for HDFS and for Hadoop version 2.7 or newer
        at org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopRecoverableWriter.<init>(HadoopRecoverableWriter.java:57)
        at org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.createRecoverableWriter(HadoopFileSystem.java:202)
        at org.apache.flink.core.fs.SafetyNetWrapperFileSystem.createRecoverableWriter(SafetyNetWrapperFileSystem.java:69)
        at org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.<init>(Buckets.java:111)
        at org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink$RowFormatBuilder.createBuckets(StreamingFileSink.java:242)
        at org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink.initializeState(StreamingFileSink.java:327)
        at org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:178)
        at org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:160)
        at org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96)
        at org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:277)
        at org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:738)
        at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:289)
        at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711)
        at java.lang.Thread.run(Thread.java:748)


On Tue, Oct 30, 2018 at 4:12 PM Till Rohrmann <[hidden email]> wrote:

> Hi Addison,
>
> I think the idea was to also backport this feature to 1.6 since we
> considered it a bug that S3 was not supported in 1.6. I've pulled in Kostas
> who worked on the S3 writer. @Klou did we intentionally not backport this
> feature?
>
> I think there should be nothing special about backporting this feature and
> building your own version of Flink.
>
> Cheers,
> Till
>
> On Tue, Oct 30, 2018 at 10:54 PM Addison Higham <[hidden email]>
> wrote:
>
> > Hi all,
> >
> > Been hitting my head against a wall for the last few hours. The release
> > notes for 1.6.2 show https://issues.apache.org/jira/browse/FLINK-9752 as
> > resolved in 1.6.2. I am trying to upgrade and switch some things to use
> the
> > StreamingFileSink against s3. However, when doing so, I get the following
> > error:
> >
> > Recoverable writers on Hadoop are only supported for HDFS and for
> > Hadoop version 2.7 or newer
> >
> >
> > I started digging into the code and looking deeper, I don't see the code
> at
> > https://github.com/apache/flink/pull/6795 as being backported to 1.6.2?
> >
> > Was the Fix Version erroneous? If so, are there plans to backport it? If
> > not, seems like that should be fixed in the release notes.
> >
> > I have been waiting for that functionality and we already build our own
> > flink, so I am tempted to backport it onto 1.6.2... anything tricky about
> > that?
> >
> > Thanks!
> >
>
Reply | Threaded
Open this post in threaded view
|

Re: FLINK-9752 (s3 recoverable writer) not actually fixed in 1.6.2?

Addison Higham
I have backported this at
https://github.com/instructure/flink/tree/s3_recover_backport by
cherry-picking all the relevant code, I am not sure how backports are
usually done with Flink (if you squash and merge) but there were a few
minor conflicts and involved quite a few changes from master.

Going to try with my branch tomorrow and will report any issues.



On Tue, Oct 30, 2018 at 8:44 PM Mike Mintz <[hidden email]> wrote:

> FWIW I also tried this on Flink 1.6.2 today and got the same error. This is
> my full stack trace:
>
> java.lang.UnsupportedOperationException: Recoverable writers on Hadoop
> are only supported for HDFS and for Hadoop version 2.7 or newer
>         at
> org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopRecoverableWriter.<init>(HadoopRecoverableWriter.java:57)
>         at
> org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.createRecoverableWriter(HadoopFileSystem.java:202)
>         at
> org.apache.flink.core.fs.SafetyNetWrapperFileSystem.createRecoverableWriter(SafetyNetWrapperFileSystem.java:69)
>         at
> org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.<init>(Buckets.java:111)
>         at
> org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink$RowFormatBuilder.createBuckets(StreamingFileSink.java:242)
>         at
> org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink.initializeState(StreamingFileSink.java:327)
>         at
> org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:178)
>         at
> org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:160)
>         at
> org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96)
>         at
> org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:277)
>         at
> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:738)
>         at
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:289)
>         at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711)
>         at java.lang.Thread.run(Thread.java:748)
>
>
> On Tue, Oct 30, 2018 at 4:12 PM Till Rohrmann <[hidden email]>
> wrote:
>
> > Hi Addison,
> >
> > I think the idea was to also backport this feature to 1.6 since we
> > considered it a bug that S3 was not supported in 1.6. I've pulled in
> Kostas
> > who worked on the S3 writer. @Klou did we intentionally not backport this
> > feature?
> >
> > I think there should be nothing special about backporting this feature
> and
> > building your own version of Flink.
> >
> > Cheers,
> > Till
> >
> > On Tue, Oct 30, 2018 at 10:54 PM Addison Higham <[hidden email]>
> > wrote:
> >
> > > Hi all,
> > >
> > > Been hitting my head against a wall for the last few hours. The release
> > > notes for 1.6.2 show https://issues.apache.org/jira/browse/FLINK-9752
> as
> > > resolved in 1.6.2. I am trying to upgrade and switch some things to use
> > the
> > > StreamingFileSink against s3. However, when doing so, I get the
> following
> > > error:
> > >
> > > Recoverable writers on Hadoop are only supported for HDFS and for
> > > Hadoop version 2.7 or newer
> > >
> > >
> > > I started digging into the code and looking deeper, I don't see the
> code
> > at
> > > https://github.com/apache/flink/pull/6795 as being backported to
> 1.6.2?
> > >
> > > Was the Fix Version erroneous? If so, are there plans to backport it?
> If
> > > not, seems like that should be fixed in the release notes.
> > >
> > > I have been waiting for that functionality and we already build our own
> > > flink, so I am tempted to backport it onto 1.6.2... anything tricky
> about
> > > that?
> > >
> > > Thanks!
> > >
> >
>
Reply | Threaded
Open this post in threaded view
|

Re: FLINK-9752 (s3 recoverable writer) not actually fixed in 1.6.2?

Kostas Kloudas
Hi Addison,

This is definitely an error on my end, as the feature is going to be available from Flink 1.7 onwards.
I forgot to correctly update the Flink version when closing the JIRA issue.

I will update the release notes accordingly.

Sorry for the miscommunication,
Kostas


> On Oct 31, 2018, at 5:45 AM, Addison Higham <[hidden email]> wrote:
>
> I have backported this at https://github.com/instructure/flink/tree/s3_recover_backport <https://github.com/instructure/flink/tree/s3_recover_backport> by cherry-picking all the relevant code, I am not sure how backports are usually done with Flink (if you squash and merge) but there were a few minor conflicts and involved quite a few changes from master.
>
> Going to try with my branch tomorrow and will report any issues.
>
>
>
> On Tue, Oct 30, 2018 at 8:44 PM Mike Mintz <[hidden email] <mailto:[hidden email]>> wrote:
> FWIW I also tried this on Flink 1.6.2 today and got the same error. This is
> my full stack trace:
>
> java.lang.UnsupportedOperationException: Recoverable writers on Hadoop
> are only supported for HDFS and for Hadoop version 2.7 or newer
>         at org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopRecoverableWriter.<init>(HadoopRecoverableWriter.java:57)
>         at org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.createRecoverableWriter(HadoopFileSystem.java:202)
>         at org.apache.flink.core.fs.SafetyNetWrapperFileSystem.createRecoverableWriter(SafetyNetWrapperFileSystem.java:69)
>         at org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.<init>(Buckets.java:111)
>         at org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink$RowFormatBuilder.createBuckets(StreamingFileSink.java:242)
>         at org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink.initializeState(StreamingFileSink.java:327)
>         at org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:178)
>         at org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:160)
>         at org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96)
>         at org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:277)
>         at org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:738)
>         at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:289)
>         at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711)
>         at java.lang.Thread.run(Thread.java:748)
>
>
> On Tue, Oct 30, 2018 at 4:12 PM Till Rohrmann <[hidden email] <mailto:[hidden email]>> wrote:
>
> > Hi Addison,
> >
> > I think the idea was to also backport this feature to 1.6 since we
> > considered it a bug that S3 was not supported in 1.6. I've pulled in Kostas
> > who worked on the S3 writer. @Klou did we intentionally not backport this
> > feature?
> >
> > I think there should be nothing special about backporting this feature and
> > building your own version of Flink.
> >
> > Cheers,
> > Till
> >
> > On Tue, Oct 30, 2018 at 10:54 PM Addison Higham <[hidden email] <mailto:[hidden email]>>
> > wrote:
> >
> > > Hi all,
> > >
> > > Been hitting my head against a wall for the last few hours. The release
> > > notes for 1.6.2 show https://issues.apache.org/jira/browse/FLINK-9752 <https://issues.apache.org/jira/browse/FLINK-9752> as
> > > resolved in 1.6.2. I am trying to upgrade and switch some things to use
> > the
> > > StreamingFileSink against s3. However, when doing so, I get the following
> > > error:
> > >
> > > Recoverable writers on Hadoop are only supported for HDFS and for
> > > Hadoop version 2.7 or newer
> > >
> > >
> > > I started digging into the code and looking deeper, I don't see the code
> > at
> > > https://github.com/apache/flink/pull/6795 <https://github.com/apache/flink/pull/6795> as being backported to 1.6.2?
> > >
> > > Was the Fix Version erroneous? If so, are there plans to backport it? If
> > > not, seems like that should be fixed in the release notes.
> > >
> > > I have been waiting for that functionality and we already build our own
> > > flink, so I am tempted to backport it onto 1.6.2... anything tricky about
> > > that?
> > >
> > > Thanks!
> > >
> >