Hello All,
I am new to Flink community and am very excited about the project and work you all have been doing. Kudos!! I was looking to pickup some starter task. Robert recommended to pick up https://issues.apache.org/jira/browse/FLINK-1711. Thanks Robert for your guidance. Sorry for a dumb question. I am done with code changes but my "mvn verify" failing only for the scala module as follows flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: error: ambiguous reference to overloaded definition, [ERROR] both method checkNotNull in object Preconditions of type [T](x$1: T, x$2: String, x$3: <repeated...>[Object])T [ERROR] and method checkNotNull in object Preconditions of type [T](x$1: T, x$2: Any)T [ERROR] match argument types ((L, R) => O,String) [ERROR] Preconditions.checkNotNull(fun, "Join function must not be null.") Same error I see for all of the Scala classes I changed. Any pointers here will be very helpful for me to proceed further. Please let me know if you need more information. Thanks in advance for your help and support. Thanks, Lokesh |
Hi. There is some problems using Guava’s check method in Scala. (https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k <https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k>) You can solve this error simply with casting last argument to java.lang.Object. But I think we’d better use `require`, `assume`, `assert` method provided by Scala. (http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html <http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html>)
Because this changes affects many other codes, so we should discuss about changing Guava's method to Scala’s method. Regards. Chiwan Park (Sent with iPhone) > On May 10, 2015, at 11:49 AM, Lokesh Rajaram <[hidden email]> wrote: > > Hello All, > > I am new to Flink community and am very excited about the project and work > you all have been doing. Kudos!! > > I was looking to pickup some starter task. Robert recommended to pick up > https://issues.apache.org/jira/browse/FLINK-1711. Thanks Robert for your > guidance. > > Sorry for a dumb question. I am done with code changes but my "mvn verify" > failing only for the scala module as follows > > flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: > error: ambiguous reference to overloaded definition, > [ERROR] both method checkNotNull in object Preconditions of type [T](x$1: > T, x$2: String, x$3: <repeated...>[Object])T > [ERROR] and method checkNotNull in object Preconditions of type [T](x$1: > T, x$2: Any)T > [ERROR] match argument types ((L, R) => O,String) > [ERROR] Preconditions.checkNotNull(fun, "Join function must not be > null.") > > Same error I see for all of the Scala classes I changed. Any pointers here > will be very helpful for me to proceed further. Please let me know if you > need more information. > > Thanks in advance for your help and support. > > Thanks, > Lokesh |
Thank you for the reference links. Which approach should I take, casting or
use scala methods. If it's the latter option will the JIRA ticket FLINK-1711 <https://issues.apache.org/jira/browse/FLINK-1711> be updated to reflect it? Thanks, Lokesh On Sat, May 9, 2015 at 8:16 PM, Chiwan Park <[hidden email]> wrote: > Hi. There is some problems using Guava’s check method in Scala. ( > https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k < > https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k>) You > can solve this error simply with casting last argument to java.lang.Object. > But I think we’d better use `require`, `assume`, `assert` method provided > by Scala. ( > http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html < > http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html>) > Because this changes affects many other codes, so we should discuss about > changing Guava's method to Scala’s method. > > Regards. > Chiwan Park (Sent with iPhone) > > > > > On May 10, 2015, at 11:49 AM, Lokesh Rajaram <[hidden email]> > wrote: > > > > Hello All, > > > > I am new to Flink community and am very excited about the project and > work > > you all have been doing. Kudos!! > > > > I was looking to pickup some starter task. Robert recommended to pick up > > https://issues.apache.org/jira/browse/FLINK-1711. Thanks Robert for your > > guidance. > > > > Sorry for a dumb question. I am done with code changes but my "mvn > verify" > > failing only for the scala module as follows > > > > > flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: > > error: ambiguous reference to overloaded definition, > > [ERROR] both method checkNotNull in object Preconditions of type [T](x$1: > > T, x$2: String, x$3: <repeated...>[Object])T > > [ERROR] and method checkNotNull in object Preconditions of type [T](x$1: > > T, x$2: Any)T > > [ERROR] match argument types ((L, R) => O,String) > > [ERROR] Preconditions.checkNotNull(fun, "Join function must not be > > null.") > > > > Same error I see for all of the Scala classes I changed. Any pointers > here > > will be very helpful for me to proceed further. Please let me know if you > > need more information. > > > > Thanks in advance for your help and support. > > > > Thanks, > > Lokesh > > |
I would propose using the methods as Chiwan suggested. If everyone
agrees I can change the Jira issue. On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram <[hidden email]> wrote: > Thank you for the reference links. Which approach should I take, casting or > use scala methods. > If it's the latter option will the JIRA ticket FLINK-1711 > <https://issues.apache.org/jira/browse/FLINK-1711> be updated to reflect it? > > Thanks, > Lokesh > > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park <[hidden email]> wrote: > >> Hi. There is some problems using Guava’s check method in Scala. ( >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k < >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k>) You >> can solve this error simply with casting last argument to java.lang.Object. >> But I think we’d better use `require`, `assume`, `assert` method provided >> by Scala. ( >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html < >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html>) >> Because this changes affects many other codes, so we should discuss about >> changing Guava's method to Scala’s method. >> >> Regards. >> Chiwan Park (Sent with iPhone) >> >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram <[hidden email]> >> wrote: >> > >> > Hello All, >> > >> > I am new to Flink community and am very excited about the project and >> work >> > you all have been doing. Kudos!! >> > >> > I was looking to pickup some starter task. Robert recommended to pick up >> > https://issues.apache.org/jira/browse/FLINK-1711. Thanks Robert for your >> > guidance. >> > >> > Sorry for a dumb question. I am done with code changes but my "mvn >> verify" >> > failing only for the scala module as follows >> > >> > >> flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: >> > error: ambiguous reference to overloaded definition, >> > [ERROR] both method checkNotNull in object Preconditions of type [T](x$1: >> > T, x$2: String, x$3: <repeated...>[Object])T >> > [ERROR] and method checkNotNull in object Preconditions of type [T](x$1: >> > T, x$2: Any)T >> > [ERROR] match argument types ((L, R) => O,String) >> > [ERROR] Preconditions.checkNotNull(fun, "Join function must not be >> > null.") >> > >> > Same error I see for all of the Scala classes I changed. Any pointers >> here >> > will be very helpful for me to proceed further. Please let me know if you >> > need more information. >> > >> > Thanks in advance for your help and support. >> > >> > Thanks, >> > Lokesh >> >> |
1. I think I can use require for replacing Validate.isTrue
2. What about Validate.notNull? If require is used it would throw IllegalArgumentException, if assume or assert is used it would throw AssertionError which is not compatible with current implementation. Please let me know if my understanding is correct. Also, let me know your thoughts. Thanks, Lokesh On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek <[hidden email]> wrote: > I would propose using the methods as Chiwan suggested. If everyone > agrees I can change the Jira issue. > > On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram > <[hidden email]> wrote: > > Thank you for the reference links. Which approach should I take, casting > or > > use scala methods. > > If it's the latter option will the JIRA ticket FLINK-1711 > > <https://issues.apache.org/jira/browse/FLINK-1711> be updated to > reflect it? > > > > Thanks, > > Lokesh > > > > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park <[hidden email]> > wrote: > > > >> Hi. There is some problems using Guava’s check method in Scala. ( > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k < > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k>) You > >> can solve this error simply with casting last argument to > java.lang.Object. > >> But I think we’d better use `require`, `assume`, `assert` method > provided > >> by Scala. ( > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html < > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html>) > >> Because this changes affects many other codes, so we should discuss > about > >> changing Guava's method to Scala’s method. > >> > >> Regards. > >> Chiwan Park (Sent with iPhone) > >> > >> > >> > >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < > [hidden email]> > >> wrote: > >> > > >> > Hello All, > >> > > >> > I am new to Flink community and am very excited about the project and > >> work > >> > you all have been doing. Kudos!! > >> > > >> > I was looking to pickup some starter task. Robert recommended to pick > up > >> > https://issues.apache.org/jira/browse/FLINK-1711. Thanks Robert for > your > >> > guidance. > >> > > >> > Sorry for a dumb question. I am done with code changes but my "mvn > >> verify" > >> > failing only for the scala module as follows > >> > > >> > > >> > flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: > >> > error: ambiguous reference to overloaded definition, > >> > [ERROR] both method checkNotNull in object Preconditions of type > [T](x$1: > >> > T, x$2: String, x$3: <repeated...>[Object])T > >> > [ERROR] and method checkNotNull in object Preconditions of type > [T](x$1: > >> > T, x$2: Any)T > >> > [ERROR] match argument types ((L, R) => O,String) > >> > [ERROR] Preconditions.checkNotNull(fun, "Join function must not be > >> > null.") > >> > > >> > Same error I see for all of the Scala classes I changed. Any pointers > >> here > >> > will be very helpful for me to proceed further. Please let me know if > you > >> > need more information. > >> > > >> > Thanks in advance for your help and support. > >> > > >> > Thanks, > >> > Lokesh > >> > >> > |
I think you can replace Validate.NotNull(p) with require(p != null, "p
is null (or something like this)"). On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram <[hidden email]> wrote: > 1. I think I can use require for replacing Validate.isTrue > 2. What about Validate.notNull? If require is used it would throw > IllegalArgumentException, > if assume or assert is used it would throw AssertionError which is not > compatible with current implementation. > > Please let me know if my understanding is correct. Also, let me know your > thoughts. > > Thanks, > Lokesh > > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek <[hidden email]> > wrote: > >> I would propose using the methods as Chiwan suggested. If everyone >> agrees I can change the Jira issue. >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram >> <[hidden email]> wrote: >> > Thank you for the reference links. Which approach should I take, casting >> or >> > use scala methods. >> > If it's the latter option will the JIRA ticket FLINK-1711 >> > <https://issues.apache.org/jira/browse/FLINK-1711> be updated to >> reflect it? >> > >> > Thanks, >> > Lokesh >> > >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park <[hidden email]> >> wrote: >> > >> >> Hi. There is some problems using Guava’s check method in Scala. ( >> >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k < >> >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k>) You >> >> can solve this error simply with casting last argument to >> java.lang.Object. >> >> But I think we’d better use `require`, `assume`, `assert` method >> provided >> >> by Scala. ( >> >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html < >> >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html>) >> >> Because this changes affects many other codes, so we should discuss >> about >> >> changing Guava's method to Scala’s method. >> >> >> >> Regards. >> >> Chiwan Park (Sent with iPhone) >> >> >> >> >> >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < >> [hidden email]> >> >> wrote: >> >> > >> >> > Hello All, >> >> > >> >> > I am new to Flink community and am very excited about the project and >> >> work >> >> > you all have been doing. Kudos!! >> >> > >> >> > I was looking to pickup some starter task. Robert recommended to pick >> up >> >> > https://issues.apache.org/jira/browse/FLINK-1711. Thanks Robert for >> your >> >> > guidance. >> >> > >> >> > Sorry for a dumb question. I am done with code changes but my "mvn >> >> verify" >> >> > failing only for the scala module as follows >> >> > >> >> > >> >> >> flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: >> >> > error: ambiguous reference to overloaded definition, >> >> > [ERROR] both method checkNotNull in object Preconditions of type >> [T](x$1: >> >> > T, x$2: String, x$3: <repeated...>[Object])T >> >> > [ERROR] and method checkNotNull in object Preconditions of type >> [T](x$1: >> >> > T, x$2: Any)T >> >> > [ERROR] match argument types ((L, R) => O,String) >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join function must not be >> >> > null.") >> >> > >> >> > Same error I see for all of the Scala classes I changed. Any pointers >> >> here >> >> > will be very helpful for me to proceed further. Please let me know if >> you >> >> > need more information. >> >> > >> >> > Thanks in advance for your help and support. >> >> > >> >> > Thanks, >> >> > Lokesh >> >> >> >> >> |
Thanks Aljioscha. I was able to change as recommended and able to run the
entire test suite in local successfully. However Travis build is failing for pull request: https://github.com/apache/flink/pull/673. It's a compilation failure: [ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.1:compile (default-compile) on project flink-spargel: Compilation failure: Compilation failure: [ERROR] /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30] package com.google.common.base does not exist I can definitely see the package imported in the class, compiling and passing all tests in local. Anything I am missing here? Thanks, Lokesh On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek <[hidden email]> wrote: > I think you can replace Validate.NotNull(p) with require(p != null, "p > is null (or something like this)"). > > On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram > <[hidden email]> wrote: > > 1. I think I can use require for replacing Validate.isTrue > > 2. What about Validate.notNull? If require is used it would throw > > IllegalArgumentException, > > if assume or assert is used it would throw AssertionError which is not > > compatible with current implementation. > > > > Please let me know if my understanding is correct. Also, let me know your > > thoughts. > > > > Thanks, > > Lokesh > > > > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek <[hidden email]> > > wrote: > > > >> I would propose using the methods as Chiwan suggested. If everyone > >> agrees I can change the Jira issue. > >> > >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram > >> <[hidden email]> wrote: > >> > Thank you for the reference links. Which approach should I take, > casting > >> or > >> > use scala methods. > >> > If it's the latter option will the JIRA ticket FLINK-1711 > >> > <https://issues.apache.org/jira/browse/FLINK-1711> be updated to > >> reflect it? > >> > > >> > Thanks, > >> > Lokesh > >> > > >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park <[hidden email]> > >> wrote: > >> > > >> >> Hi. There is some problems using Guava’s check method in Scala. ( > >> >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k < > >> >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k>) > You > >> >> can solve this error simply with casting last argument to > >> java.lang.Object. > >> >> But I think we’d better use `require`, `assume`, `assert` method > >> provided > >> >> by Scala. ( > >> >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html < > >> >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html>) > >> >> Because this changes affects many other codes, so we should discuss > >> about > >> >> changing Guava's method to Scala’s method. > >> >> > >> >> Regards. > >> >> Chiwan Park (Sent with iPhone) > >> >> > >> >> > >> >> > >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < > >> [hidden email]> > >> >> wrote: > >> >> > > >> >> > Hello All, > >> >> > > >> >> > I am new to Flink community and am very excited about the project > and > >> >> work > >> >> > you all have been doing. Kudos!! > >> >> > > >> >> > I was looking to pickup some starter task. Robert recommended to > pick > >> up > >> >> > https://issues.apache.org/jira/browse/FLINK-1711. Thanks Robert > for > >> your > >> >> > guidance. > >> >> > > >> >> > Sorry for a dumb question. I am done with code changes but my "mvn > >> >> verify" > >> >> > failing only for the scala module as follows > >> >> > > >> >> > > >> >> > >> > flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: > >> >> > error: ambiguous reference to overloaded definition, > >> >> > [ERROR] both method checkNotNull in object Preconditions of type > >> [T](x$1: > >> >> > T, x$2: String, x$3: <repeated...>[Object])T > >> >> > [ERROR] and method checkNotNull in object Preconditions of type > >> [T](x$1: > >> >> > T, x$2: Any)T > >> >> > [ERROR] match argument types ((L, R) => O,String) > >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join function must > not be > >> >> > null.") > >> >> > > >> >> > Same error I see for all of the Scala classes I changed. Any > pointers > >> >> here > >> >> > will be very helpful for me to proceed further. Please let me know > if > >> you > >> >> > need more information. > >> >> > > >> >> > Thanks in advance for your help and support. > >> >> > > >> >> > Thanks, > >> >> > Lokesh > >> >> > >> >> > >> > |
@Robert, this seems like a problem with the Shading?
On Thu, May 14, 2015 at 5:41 AM, Lokesh Rajaram <[hidden email]> wrote: > Thanks Aljioscha. I was able to change as recommended and able to run the > entire test suite in local successfully. > However Travis build is failing for pull request: > https://github.com/apache/flink/pull/673. > > It's a compilation failure: > > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile > (default-compile) on project flink-spargel: Compilation failure: > Compilation failure: > [ERROR] > /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30] > package com.google.common.base does not exist > > I can definitely see the package imported in the class, compiling and > passing all tests in local. > Anything I am missing here? > > Thanks, > Lokesh > > On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek <[hidden email]> > wrote: > >> I think you can replace Validate.NotNull(p) with require(p != null, "p >> is null (or something like this)"). >> >> On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram >> <[hidden email]> wrote: >> > 1. I think I can use require for replacing Validate.isTrue >> > 2. What about Validate.notNull? If require is used it would throw >> > IllegalArgumentException, >> > if assume or assert is used it would throw AssertionError which is not >> > compatible with current implementation. >> > >> > Please let me know if my understanding is correct. Also, let me know your >> > thoughts. >> > >> > Thanks, >> > Lokesh >> > >> > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek <[hidden email]> >> > wrote: >> > >> >> I would propose using the methods as Chiwan suggested. If everyone >> >> agrees I can change the Jira issue. >> >> >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram >> >> <[hidden email]> wrote: >> >> > Thank you for the reference links. Which approach should I take, >> casting >> >> or >> >> > use scala methods. >> >> > If it's the latter option will the JIRA ticket FLINK-1711 >> >> > <https://issues.apache.org/jira/browse/FLINK-1711> be updated to >> >> reflect it? >> >> > >> >> > Thanks, >> >> > Lokesh >> >> > >> >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park <[hidden email]> >> >> wrote: >> >> > >> >> >> Hi. There is some problems using Guava’s check method in Scala. ( >> >> >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k < >> >> >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k>) >> You >> >> >> can solve this error simply with casting last argument to >> >> java.lang.Object. >> >> >> But I think we’d better use `require`, `assume`, `assert` method >> >> provided >> >> >> by Scala. ( >> >> >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html < >> >> >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html>) >> >> >> Because this changes affects many other codes, so we should discuss >> >> about >> >> >> changing Guava's method to Scala’s method. >> >> >> >> >> >> Regards. >> >> >> Chiwan Park (Sent with iPhone) >> >> >> >> >> >> >> >> >> >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < >> >> [hidden email]> >> >> >> wrote: >> >> >> > >> >> >> > Hello All, >> >> >> > >> >> >> > I am new to Flink community and am very excited about the project >> and >> >> >> work >> >> >> > you all have been doing. Kudos!! >> >> >> > >> >> >> > I was looking to pickup some starter task. Robert recommended to >> pick >> >> up >> >> >> > https://issues.apache.org/jira/browse/FLINK-1711. Thanks Robert >> for >> >> your >> >> >> > guidance. >> >> >> > >> >> >> > Sorry for a dumb question. I am done with code changes but my "mvn >> >> >> verify" >> >> >> > failing only for the scala module as follows >> >> >> > >> >> >> > >> >> >> >> >> >> flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: >> >> >> > error: ambiguous reference to overloaded definition, >> >> >> > [ERROR] both method checkNotNull in object Preconditions of type >> >> [T](x$1: >> >> >> > T, x$2: String, x$3: <repeated...>[Object])T >> >> >> > [ERROR] and method checkNotNull in object Preconditions of type >> >> [T](x$1: >> >> >> > T, x$2: Any)T >> >> >> > [ERROR] match argument types ((L, R) => O,String) >> >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join function must >> not be >> >> >> > null.") >> >> >> > >> >> >> > Same error I see for all of the Scala classes I changed. Any >> pointers >> >> >> here >> >> >> > will be very helpful for me to proceed further. Please let me know >> if >> >> you >> >> >> > need more information. >> >> >> > >> >> >> > Thanks in advance for your help and support. >> >> >> > >> >> >> > Thanks, >> >> >> > Lokesh >> >> >> >> >> >> >> >> >> |
I think flink-spargel is missing the guava dependency.
On Thu, May 14, 2015 at 8:18 AM, Aljoscha Krettek <[hidden email]> wrote: > @Robert, this seems like a problem with the Shading? > > On Thu, May 14, 2015 at 5:41 AM, Lokesh Rajaram > <[hidden email]> wrote: > > Thanks Aljioscha. I was able to change as recommended and able to run the > > entire test suite in local successfully. > > However Travis build is failing for pull request: > > https://github.com/apache/flink/pull/673. > > > > It's a compilation failure: > > > > [ERROR] Failed to execute goal > > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile > > (default-compile) on project flink-spargel: Compilation failure: > > Compilation failure: > > [ERROR] > > > /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30] > > package com.google.common.base does not exist > > > > I can definitely see the package imported in the class, compiling and > > passing all tests in local. > > Anything I am missing here? > > > > Thanks, > > Lokesh > > > > On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek <[hidden email]> > > wrote: > > > >> I think you can replace Validate.NotNull(p) with require(p != null, "p > >> is null (or something like this)"). > >> > >> On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram > >> <[hidden email]> wrote: > >> > 1. I think I can use require for replacing Validate.isTrue > >> > 2. What about Validate.notNull? If require is used it would throw > >> > IllegalArgumentException, > >> > if assume or assert is used it would throw AssertionError which is not > >> > compatible with current implementation. > >> > > >> > Please let me know if my understanding is correct. Also, let me know > your > >> > thoughts. > >> > > >> > Thanks, > >> > Lokesh > >> > > >> > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek < > [hidden email]> > >> > wrote: > >> > > >> >> I would propose using the methods as Chiwan suggested. If everyone > >> >> agrees I can change the Jira issue. > >> >> > >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram > >> >> <[hidden email]> wrote: > >> >> > Thank you for the reference links. Which approach should I take, > >> casting > >> >> or > >> >> > use scala methods. > >> >> > If it's the latter option will the JIRA ticket FLINK-1711 > >> >> > <https://issues.apache.org/jira/browse/FLINK-1711> be updated to > >> >> reflect it? > >> >> > > >> >> > Thanks, > >> >> > Lokesh > >> >> > > >> >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park <[hidden email] > > > >> >> wrote: > >> >> > > >> >> >> Hi. There is some problems using Guava’s check method in Scala. ( > >> >> >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > < > >> >> >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > >) > >> You > >> >> >> can solve this error simply with casting last argument to > >> >> java.lang.Object. > >> >> >> But I think we’d better use `require`, `assume`, `assert` method > >> >> provided > >> >> >> by Scala. ( > >> >> >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > < > >> >> >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > >) > >> >> >> Because this changes affects many other codes, so we should > discuss > >> >> about > >> >> >> changing Guava's method to Scala’s method. > >> >> >> > >> >> >> Regards. > >> >> >> Chiwan Park (Sent with iPhone) > >> >> >> > >> >> >> > >> >> >> > >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < > >> >> [hidden email]> > >> >> >> wrote: > >> >> >> > > >> >> >> > Hello All, > >> >> >> > > >> >> >> > I am new to Flink community and am very excited about the > project > >> and > >> >> >> work > >> >> >> > you all have been doing. Kudos!! > >> >> >> > > >> >> >> > I was looking to pickup some starter task. Robert recommended to > >> pick > >> >> up > >> >> >> > https://issues.apache.org/jira/browse/FLINK-1711. Thanks Robert > >> for > >> >> your > >> >> >> > guidance. > >> >> >> > > >> >> >> > Sorry for a dumb question. I am done with code changes but my > "mvn > >> >> >> verify" > >> >> >> > failing only for the scala module as follows > >> >> >> > > >> >> >> > > >> >> >> > >> >> > >> > flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: > >> >> >> > error: ambiguous reference to overloaded definition, > >> >> >> > [ERROR] both method checkNotNull in object Preconditions of type > >> >> [T](x$1: > >> >> >> > T, x$2: String, x$3: <repeated...>[Object])T > >> >> >> > [ERROR] and method checkNotNull in object Preconditions of type > >> >> [T](x$1: > >> >> >> > T, x$2: Any)T > >> >> >> > [ERROR] match argument types ((L, R) => O,String) > >> >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join function must > >> not be > >> >> >> > null.") > >> >> >> > > >> >> >> > Same error I see for all of the Scala classes I changed. Any > >> pointers > >> >> >> here > >> >> >> > will be very helpful for me to proceed further. Please let me > know > >> if > >> >> you > >> >> >> > need more information. > >> >> >> > > >> >> >> > Thanks in advance for your help and support. > >> >> >> > > >> >> >> > Thanks, > >> >> >> > Lokesh > >> >> >> > >> >> >> > >> >> > >> > |
Thanks Aljoscha, Robert. After adding guava dependency for flink-spargel I
was able to progress further but now it's failing in flink-streaming-connectors for the following test case: KafkaITCase.brokerFailureTest:936->tryExecute:352 Test failed with: Job execution failed. Any pointers would help me proceed further. Sorry for a lot of trivial questions, I am just getting started not familiar with the code base. I tried running locally, I am able to run it successfully, don't know why it's only failing in Travis build. Not sure if I am missing something in my local environment. Thanks, Lokesh On Thu, May 14, 2015 at 1:39 AM, Robert Metzger <[hidden email]> wrote: > I think flink-spargel is missing the guava dependency. > > On Thu, May 14, 2015 at 8:18 AM, Aljoscha Krettek <[hidden email]> > wrote: > > > @Robert, this seems like a problem with the Shading? > > > > On Thu, May 14, 2015 at 5:41 AM, Lokesh Rajaram > > <[hidden email]> wrote: > > > Thanks Aljioscha. I was able to change as recommended and able to run > the > > > entire test suite in local successfully. > > > However Travis build is failing for pull request: > > > https://github.com/apache/flink/pull/673. > > > > > > It's a compilation failure: > > > > > > [ERROR] Failed to execute goal > > > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile > > > (default-compile) on project flink-spargel: Compilation failure: > > > Compilation failure: > > > [ERROR] > > > > > > /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30] > > > package com.google.common.base does not exist > > > > > > I can definitely see the package imported in the class, compiling and > > > passing all tests in local. > > > Anything I am missing here? > > > > > > Thanks, > > > Lokesh > > > > > > On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek <[hidden email] > > > > > wrote: > > > > > >> I think you can replace Validate.NotNull(p) with require(p != null, "p > > >> is null (or something like this)"). > > >> > > >> On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram > > >> <[hidden email]> wrote: > > >> > 1. I think I can use require for replacing Validate.isTrue > > >> > 2. What about Validate.notNull? If require is used it would throw > > >> > IllegalArgumentException, > > >> > if assume or assert is used it would throw AssertionError which is > not > > >> > compatible with current implementation. > > >> > > > >> > Please let me know if my understanding is correct. Also, let me know > > your > > >> > thoughts. > > >> > > > >> > Thanks, > > >> > Lokesh > > >> > > > >> > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek < > > [hidden email]> > > >> > wrote: > > >> > > > >> >> I would propose using the methods as Chiwan suggested. If everyone > > >> >> agrees I can change the Jira issue. > > >> >> > > >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram > > >> >> <[hidden email]> wrote: > > >> >> > Thank you for the reference links. Which approach should I take, > > >> casting > > >> >> or > > >> >> > use scala methods. > > >> >> > If it's the latter option will the JIRA ticket FLINK-1711 > > >> >> > <https://issues.apache.org/jira/browse/FLINK-1711> be updated to > > >> >> reflect it? > > >> >> > > > >> >> > Thanks, > > >> >> > Lokesh > > >> >> > > > >> >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park < > [hidden email] > > > > > >> >> wrote: > > >> >> > > > >> >> >> Hi. There is some problems using Guava’s check method in Scala. > ( > > >> >> >> > https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > > < > > >> >> >> > https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > > >) > > >> You > > >> >> >> can solve this error simply with casting last argument to > > >> >> java.lang.Object. > > >> >> >> But I think we’d better use `require`, `assume`, `assert` method > > >> >> provided > > >> >> >> by Scala. ( > > >> >> >> > http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > > < > > >> >> >> > http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > > >) > > >> >> >> Because this changes affects many other codes, so we should > > discuss > > >> >> about > > >> >> >> changing Guava's method to Scala’s method. > > >> >> >> > > >> >> >> Regards. > > >> >> >> Chiwan Park (Sent with iPhone) > > >> >> >> > > >> >> >> > > >> >> >> > > >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < > > >> >> [hidden email]> > > >> >> >> wrote: > > >> >> >> > > > >> >> >> > Hello All, > > >> >> >> > > > >> >> >> > I am new to Flink community and am very excited about the > > project > > >> and > > >> >> >> work > > >> >> >> > you all have been doing. Kudos!! > > >> >> >> > > > >> >> >> > I was looking to pickup some starter task. Robert recommended > to > > >> pick > > >> >> up > > >> >> >> > https://issues.apache.org/jira/browse/FLINK-1711. Thanks > Robert > > >> for > > >> >> your > > >> >> >> > guidance. > > >> >> >> > > > >> >> >> > Sorry for a dumb question. I am done with code changes but my > > "mvn > > >> >> >> verify" > > >> >> >> > failing only for the scala module as follows > > >> >> >> > > > >> >> >> > > > >> >> >> > > >> >> > > >> > > > flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: > > >> >> >> > error: ambiguous reference to overloaded definition, > > >> >> >> > [ERROR] both method checkNotNull in object Preconditions of > type > > >> >> [T](x$1: > > >> >> >> > T, x$2: String, x$3: <repeated...>[Object])T > > >> >> >> > [ERROR] and method checkNotNull in object Preconditions of > type > > >> >> [T](x$1: > > >> >> >> > T, x$2: Any)T > > >> >> >> > [ERROR] match argument types ((L, R) => O,String) > > >> >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join function > must > > >> not be > > >> >> >> > null.") > > >> >> >> > > > >> >> >> > Same error I see for all of the Scala classes I changed. Any > > >> pointers > > >> >> >> here > > >> >> >> > will be very helpful for me to proceed further. Please let me > > know > > >> if > > >> >> you > > >> >> >> > need more information. > > >> >> >> > > > >> >> >> > Thanks in advance for your help and support. > > >> >> >> > > > >> >> >> > Thanks, > > >> >> >> > Lokesh > > >> >> >> > > >> >> >> > > >> >> > > >> > > > |
Hi,
don't worry, there are very few stupid questions. :D The KafkaITCase sometimes fails on Travis, this is a known problem currently. On travis you can restart the individual runs for a commit in the view of the failed run. Hope that helps. Cheers, Aljoscha On Thu, May 14, 2015 at 5:35 PM, Lokesh Rajaram <[hidden email]> wrote: > Thanks Aljoscha, Robert. After adding guava dependency for flink-spargel I > was able to progress further but now it's failing in > flink-streaming-connectors for the following test case: > > KafkaITCase.brokerFailureTest:936->tryExecute:352 Test failed with: Job > execution failed. > > Any pointers would help me proceed further. Sorry for a lot of trivial > questions, I am just getting started not familiar with the code base. > I tried running locally, I am able to run it successfully, don't know why > it's only failing in Travis build. Not sure if I am missing something in my > local environment. > > Thanks, > Lokesh > > On Thu, May 14, 2015 at 1:39 AM, Robert Metzger <[hidden email]> wrote: > >> I think flink-spargel is missing the guava dependency. >> >> On Thu, May 14, 2015 at 8:18 AM, Aljoscha Krettek <[hidden email]> >> wrote: >> >> > @Robert, this seems like a problem with the Shading? >> > >> > On Thu, May 14, 2015 at 5:41 AM, Lokesh Rajaram >> > <[hidden email]> wrote: >> > > Thanks Aljioscha. I was able to change as recommended and able to run >> the >> > > entire test suite in local successfully. >> > > However Travis build is failing for pull request: >> > > https://github.com/apache/flink/pull/673. >> > > >> > > It's a compilation failure: >> > > >> > > [ERROR] Failed to execute goal >> > > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile >> > > (default-compile) on project flink-spargel: Compilation failure: >> > > Compilation failure: >> > > [ERROR] >> > > >> > >> /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30] >> > > package com.google.common.base does not exist >> > > >> > > I can definitely see the package imported in the class, compiling and >> > > passing all tests in local. >> > > Anything I am missing here? >> > > >> > > Thanks, >> > > Lokesh >> > > >> > > On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek <[hidden email] >> > >> > > wrote: >> > > >> > >> I think you can replace Validate.NotNull(p) with require(p != null, "p >> > >> is null (or something like this)"). >> > >> >> > >> On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram >> > >> <[hidden email]> wrote: >> > >> > 1. I think I can use require for replacing Validate.isTrue >> > >> > 2. What about Validate.notNull? If require is used it would throw >> > >> > IllegalArgumentException, >> > >> > if assume or assert is used it would throw AssertionError which is >> not >> > >> > compatible with current implementation. >> > >> > >> > >> > Please let me know if my understanding is correct. Also, let me know >> > your >> > >> > thoughts. >> > >> > >> > >> > Thanks, >> > >> > Lokesh >> > >> > >> > >> > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek < >> > [hidden email]> >> > >> > wrote: >> > >> > >> > >> >> I would propose using the methods as Chiwan suggested. If everyone >> > >> >> agrees I can change the Jira issue. >> > >> >> >> > >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram >> > >> >> <[hidden email]> wrote: >> > >> >> > Thank you for the reference links. Which approach should I take, >> > >> casting >> > >> >> or >> > >> >> > use scala methods. >> > >> >> > If it's the latter option will the JIRA ticket FLINK-1711 >> > >> >> > <https://issues.apache.org/jira/browse/FLINK-1711> be updated to >> > >> >> reflect it? >> > >> >> > >> > >> >> > Thanks, >> > >> >> > Lokesh >> > >> >> > >> > >> >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park < >> [hidden email] >> > > >> > >> >> wrote: >> > >> >> > >> > >> >> >> Hi. There is some problems using Guava’s check method in Scala. >> ( >> > >> >> >> >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k >> > < >> > >> >> >> >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k >> > >) >> > >> You >> > >> >> >> can solve this error simply with casting last argument to >> > >> >> java.lang.Object. >> > >> >> >> But I think we’d better use `require`, `assume`, `assert` method >> > >> >> provided >> > >> >> >> by Scala. ( >> > >> >> >> >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html >> > < >> > >> >> >> >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html >> > >) >> > >> >> >> Because this changes affects many other codes, so we should >> > discuss >> > >> >> about >> > >> >> >> changing Guava's method to Scala’s method. >> > >> >> >> >> > >> >> >> Regards. >> > >> >> >> Chiwan Park (Sent with iPhone) >> > >> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < >> > >> >> [hidden email]> >> > >> >> >> wrote: >> > >> >> >> > >> > >> >> >> > Hello All, >> > >> >> >> > >> > >> >> >> > I am new to Flink community and am very excited about the >> > project >> > >> and >> > >> >> >> work >> > >> >> >> > you all have been doing. Kudos!! >> > >> >> >> > >> > >> >> >> > I was looking to pickup some starter task. Robert recommended >> to >> > >> pick >> > >> >> up >> > >> >> >> > https://issues.apache.org/jira/browse/FLINK-1711. Thanks >> Robert >> > >> for >> > >> >> your >> > >> >> >> > guidance. >> > >> >> >> > >> > >> >> >> > Sorry for a dumb question. I am done with code changes but my >> > "mvn >> > >> >> >> verify" >> > >> >> >> > failing only for the scala module as follows >> > >> >> >> > >> > >> >> >> > >> > >> >> >> >> > >> >> >> > >> >> > >> flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: >> > >> >> >> > error: ambiguous reference to overloaded definition, >> > >> >> >> > [ERROR] both method checkNotNull in object Preconditions of >> type >> > >> >> [T](x$1: >> > >> >> >> > T, x$2: String, x$3: <repeated...>[Object])T >> > >> >> >> > [ERROR] and method checkNotNull in object Preconditions of >> type >> > >> >> [T](x$1: >> > >> >> >> > T, x$2: Any)T >> > >> >> >> > [ERROR] match argument types ((L, R) => O,String) >> > >> >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join function >> must >> > >> not be >> > >> >> >> > null.") >> > >> >> >> > >> > >> >> >> > Same error I see for all of the Scala classes I changed. Any >> > >> pointers >> > >> >> >> here >> > >> >> >> > will be very helpful for me to proceed further. Please let me >> > know >> > >> if >> > >> >> you >> > >> >> >> > need more information. >> > >> >> >> > >> > >> >> >> > Thanks in advance for your help and support. >> > >> >> >> > >> > >> >> >> > Thanks, >> > >> >> >> > Lokesh >> > >> >> >> >> > >> >> >> >> > >> >> >> > >> >> > >> |
However, you can only restart runs in your travis account, not on the
apache account (also used for validating pull requests). I have opened a pull request a few minutes ago which will reduce the number of KafakITCase failures (there is still one other unresolved issue). On Thu, May 14, 2015 at 5:37 PM, Aljoscha Krettek <[hidden email]> wrote: > Hi, > don't worry, there are very few stupid questions. :D > > The KafkaITCase sometimes fails on Travis, this is a known problem > currently. On travis you can restart the individual runs for a commit > in the view of the failed run. > > Hope that helps. > > Cheers, > Aljoscha > > On Thu, May 14, 2015 at 5:35 PM, Lokesh Rajaram > <[hidden email]> wrote: > > Thanks Aljoscha, Robert. After adding guava dependency for flink-spargel > I > > was able to progress further but now it's failing in > > flink-streaming-connectors for the following test case: > > > > KafkaITCase.brokerFailureTest:936->tryExecute:352 Test failed with: Job > > execution failed. > > > > Any pointers would help me proceed further. Sorry for a lot of trivial > > questions, I am just getting started not familiar with the code base. > > I tried running locally, I am able to run it successfully, don't know why > > it's only failing in Travis build. Not sure if I am missing something in > my > > local environment. > > > > Thanks, > > Lokesh > > > > On Thu, May 14, 2015 at 1:39 AM, Robert Metzger <[hidden email]> > wrote: > > > >> I think flink-spargel is missing the guava dependency. > >> > >> On Thu, May 14, 2015 at 8:18 AM, Aljoscha Krettek <[hidden email]> > >> wrote: > >> > >> > @Robert, this seems like a problem with the Shading? > >> > > >> > On Thu, May 14, 2015 at 5:41 AM, Lokesh Rajaram > >> > <[hidden email]> wrote: > >> > > Thanks Aljioscha. I was able to change as recommended and able to > run > >> the > >> > > entire test suite in local successfully. > >> > > However Travis build is failing for pull request: > >> > > https://github.com/apache/flink/pull/673. > >> > > > >> > > It's a compilation failure: > >> > > > >> > > [ERROR] Failed to execute goal > >> > > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile > >> > > (default-compile) on project flink-spargel: Compilation failure: > >> > > Compilation failure: > >> > > [ERROR] > >> > > > >> > > >> > /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30] > >> > > package com.google.common.base does not exist > >> > > > >> > > I can definitely see the package imported in the class, compiling > and > >> > > passing all tests in local. > >> > > Anything I am missing here? > >> > > > >> > > Thanks, > >> > > Lokesh > >> > > > >> > > On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek < > [hidden email] > >> > > >> > > wrote: > >> > > > >> > >> I think you can replace Validate.NotNull(p) with require(p != > null, "p > >> > >> is null (or something like this)"). > >> > >> > >> > >> On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram > >> > >> <[hidden email]> wrote: > >> > >> > 1. I think I can use require for replacing Validate.isTrue > >> > >> > 2. What about Validate.notNull? If require is used it would throw > >> > >> > IllegalArgumentException, > >> > >> > if assume or assert is used it would throw AssertionError which > is > >> not > >> > >> > compatible with current implementation. > >> > >> > > >> > >> > Please let me know if my understanding is correct. Also, let me > know > >> > your > >> > >> > thoughts. > >> > >> > > >> > >> > Thanks, > >> > >> > Lokesh > >> > >> > > >> > >> > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek < > >> > [hidden email]> > >> > >> > wrote: > >> > >> > > >> > >> >> I would propose using the methods as Chiwan suggested. If > everyone > >> > >> >> agrees I can change the Jira issue. > >> > >> >> > >> > >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram > >> > >> >> <[hidden email]> wrote: > >> > >> >> > Thank you for the reference links. Which approach should I > take, > >> > >> casting > >> > >> >> or > >> > >> >> > use scala methods. > >> > >> >> > If it's the latter option will the JIRA ticket FLINK-1711 > >> > >> >> > <https://issues.apache.org/jira/browse/FLINK-1711> be > updated to > >> > >> >> reflect it? > >> > >> >> > > >> > >> >> > Thanks, > >> > >> >> > Lokesh > >> > >> >> > > >> > >> >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park < > >> [hidden email] > >> > > > >> > >> >> wrote: > >> > >> >> > > >> > >> >> >> Hi. There is some problems using Guava’s check method in > Scala. > >> ( > >> > >> >> >> > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > >> > < > >> > >> >> >> > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > >> > >) > >> > >> You > >> > >> >> >> can solve this error simply with casting last argument to > >> > >> >> java.lang.Object. > >> > >> >> >> But I think we’d better use `require`, `assume`, `assert` > method > >> > >> >> provided > >> > >> >> >> by Scala. ( > >> > >> >> >> > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > >> > < > >> > >> >> >> > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > >> > >) > >> > >> >> >> Because this changes affects many other codes, so we should > >> > discuss > >> > >> >> about > >> > >> >> >> changing Guava's method to Scala’s method. > >> > >> >> >> > >> > >> >> >> Regards. > >> > >> >> >> Chiwan Park (Sent with iPhone) > >> > >> >> >> > >> > >> >> >> > >> > >> >> >> > >> > >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < > >> > >> >> [hidden email]> > >> > >> >> >> wrote: > >> > >> >> >> > > >> > >> >> >> > Hello All, > >> > >> >> >> > > >> > >> >> >> > I am new to Flink community and am very excited about the > >> > project > >> > >> and > >> > >> >> >> work > >> > >> >> >> > you all have been doing. Kudos!! > >> > >> >> >> > > >> > >> >> >> > I was looking to pickup some starter task. Robert > recommended > >> to > >> > >> pick > >> > >> >> up > >> > >> >> >> > https://issues.apache.org/jira/browse/FLINK-1711. Thanks > >> Robert > >> > >> for > >> > >> >> your > >> > >> >> >> > guidance. > >> > >> >> >> > > >> > >> >> >> > Sorry for a dumb question. I am done with code changes but > my > >> > "mvn > >> > >> >> >> verify" > >> > >> >> >> > failing only for the scala module as follows > >> > >> >> >> > > >> > >> >> >> > > >> > >> >> >> > >> > >> >> > >> > >> > >> > > >> > flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: > >> > >> >> >> > error: ambiguous reference to overloaded definition, > >> > >> >> >> > [ERROR] both method checkNotNull in object Preconditions of > >> type > >> > >> >> [T](x$1: > >> > >> >> >> > T, x$2: String, x$3: <repeated...>[Object])T > >> > >> >> >> > [ERROR] and method checkNotNull in object Preconditions of > >> type > >> > >> >> [T](x$1: > >> > >> >> >> > T, x$2: Any)T > >> > >> >> >> > [ERROR] match argument types ((L, R) => O,String) > >> > >> >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join function > >> must > >> > >> not be > >> > >> >> >> > null.") > >> > >> >> >> > > >> > >> >> >> > Same error I see for all of the Scala classes I changed. > Any > >> > >> pointers > >> > >> >> >> here > >> > >> >> >> > will be very helpful for me to proceed further. Please let > me > >> > know > >> > >> if > >> > >> >> you > >> > >> >> >> > need more information. > >> > >> >> >> > > >> > >> >> >> > Thanks in advance for your help and support. > >> > >> >> >> > > >> > >> >> >> > Thanks, > >> > >> >> >> > Lokesh > >> > >> >> >> > >> > >> >> >> > >> > >> >> > >> > >> > >> > > >> > |
If I understand it correct, I have to wait for your pull request to be
merged, I can rebase and trigger build again. is that right? Thanks Robert, Aljoscha for super fast reply/help. Thanks, Lokesh On Thu, May 14, 2015 at 8:39 AM, Robert Metzger <[hidden email]> wrote: > However, you can only restart runs in your travis account, not on the > apache account (also used for validating pull requests). > > I have opened a pull request a few minutes ago which will reduce the number > of KafakITCase failures (there is still one other unresolved issue). > > On Thu, May 14, 2015 at 5:37 PM, Aljoscha Krettek <[hidden email]> > wrote: > > > Hi, > > don't worry, there are very few stupid questions. :D > > > > The KafkaITCase sometimes fails on Travis, this is a known problem > > currently. On travis you can restart the individual runs for a commit > > in the view of the failed run. > > > > Hope that helps. > > > > Cheers, > > Aljoscha > > > > On Thu, May 14, 2015 at 5:35 PM, Lokesh Rajaram > > <[hidden email]> wrote: > > > Thanks Aljoscha, Robert. After adding guava dependency for > flink-spargel > > I > > > was able to progress further but now it's failing in > > > flink-streaming-connectors for the following test case: > > > > > > KafkaITCase.brokerFailureTest:936->tryExecute:352 Test failed with: Job > > > execution failed. > > > > > > Any pointers would help me proceed further. Sorry for a lot of trivial > > > questions, I am just getting started not familiar with the code base. > > > I tried running locally, I am able to run it successfully, don't know > why > > > it's only failing in Travis build. Not sure if I am missing something > in > > my > > > local environment. > > > > > > Thanks, > > > Lokesh > > > > > > On Thu, May 14, 2015 at 1:39 AM, Robert Metzger <[hidden email]> > > wrote: > > > > > >> I think flink-spargel is missing the guava dependency. > > >> > > >> On Thu, May 14, 2015 at 8:18 AM, Aljoscha Krettek < > [hidden email]> > > >> wrote: > > >> > > >> > @Robert, this seems like a problem with the Shading? > > >> > > > >> > On Thu, May 14, 2015 at 5:41 AM, Lokesh Rajaram > > >> > <[hidden email]> wrote: > > >> > > Thanks Aljioscha. I was able to change as recommended and able to > > run > > >> the > > >> > > entire test suite in local successfully. > > >> > > However Travis build is failing for pull request: > > >> > > https://github.com/apache/flink/pull/673. > > >> > > > > >> > > It's a compilation failure: > > >> > > > > >> > > [ERROR] Failed to execute goal > > >> > > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile > > >> > > (default-compile) on project flink-spargel: Compilation failure: > > >> > > Compilation failure: > > >> > > [ERROR] > > >> > > > > >> > > > >> > > > /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30] > > >> > > package com.google.common.base does not exist > > >> > > > > >> > > I can definitely see the package imported in the class, compiling > > and > > >> > > passing all tests in local. > > >> > > Anything I am missing here? > > >> > > > > >> > > Thanks, > > >> > > Lokesh > > >> > > > > >> > > On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek < > > [hidden email] > > >> > > > >> > > wrote: > > >> > > > > >> > >> I think you can replace Validate.NotNull(p) with require(p != > > null, "p > > >> > >> is null (or something like this)"). > > >> > >> > > >> > >> On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram > > >> > >> <[hidden email]> wrote: > > >> > >> > 1. I think I can use require for replacing Validate.isTrue > > >> > >> > 2. What about Validate.notNull? If require is used it would > throw > > >> > >> > IllegalArgumentException, > > >> > >> > if assume or assert is used it would throw AssertionError which > > is > > >> not > > >> > >> > compatible with current implementation. > > >> > >> > > > >> > >> > Please let me know if my understanding is correct. Also, let me > > know > > >> > your > > >> > >> > thoughts. > > >> > >> > > > >> > >> > Thanks, > > >> > >> > Lokesh > > >> > >> > > > >> > >> > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek < > > >> > [hidden email]> > > >> > >> > wrote: > > >> > >> > > > >> > >> >> I would propose using the methods as Chiwan suggested. If > > everyone > > >> > >> >> agrees I can change the Jira issue. > > >> > >> >> > > >> > >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram > > >> > >> >> <[hidden email]> wrote: > > >> > >> >> > Thank you for the reference links. Which approach should I > > take, > > >> > >> casting > > >> > >> >> or > > >> > >> >> > use scala methods. > > >> > >> >> > If it's the latter option will the JIRA ticket FLINK-1711 > > >> > >> >> > <https://issues.apache.org/jira/browse/FLINK-1711> be > > updated to > > >> > >> >> reflect it? > > >> > >> >> > > > >> > >> >> > Thanks, > > >> > >> >> > Lokesh > > >> > >> >> > > > >> > >> >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park < > > >> [hidden email] > > >> > > > > >> > >> >> wrote: > > >> > >> >> > > > >> > >> >> >> Hi. There is some problems using Guava’s check method in > > Scala. > > >> ( > > >> > >> >> >> > > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > > >> > < > > >> > >> >> >> > > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > > >> > >) > > >> > >> You > > >> > >> >> >> can solve this error simply with casting last argument to > > >> > >> >> java.lang.Object. > > >> > >> >> >> But I think we’d better use `require`, `assume`, `assert` > > method > > >> > >> >> provided > > >> > >> >> >> by Scala. ( > > >> > >> >> >> > > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > > >> > < > > >> > >> >> >> > > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > > >> > >) > > >> > >> >> >> Because this changes affects many other codes, so we should > > >> > discuss > > >> > >> >> about > > >> > >> >> >> changing Guava's method to Scala’s method. > > >> > >> >> >> > > >> > >> >> >> Regards. > > >> > >> >> >> Chiwan Park (Sent with iPhone) > > >> > >> >> >> > > >> > >> >> >> > > >> > >> >> >> > > >> > >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < > > >> > >> >> [hidden email]> > > >> > >> >> >> wrote: > > >> > >> >> >> > > > >> > >> >> >> > Hello All, > > >> > >> >> >> > > > >> > >> >> >> > I am new to Flink community and am very excited about the > > >> > project > > >> > >> and > > >> > >> >> >> work > > >> > >> >> >> > you all have been doing. Kudos!! > > >> > >> >> >> > > > >> > >> >> >> > I was looking to pickup some starter task. Robert > > recommended > > >> to > > >> > >> pick > > >> > >> >> up > > >> > >> >> >> > https://issues.apache.org/jira/browse/FLINK-1711. Thanks > > >> Robert > > >> > >> for > > >> > >> >> your > > >> > >> >> >> > guidance. > > >> > >> >> >> > > > >> > >> >> >> > Sorry for a dumb question. I am done with code changes > but > > my > > >> > "mvn > > >> > >> >> >> verify" > > >> > >> >> >> > failing only for the scala module as follows > > >> > >> >> >> > > > >> > >> >> >> > > > >> > >> >> >> > > >> > >> >> > > >> > >> > > >> > > > >> > > > flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: > > >> > >> >> >> > error: ambiguous reference to overloaded definition, > > >> > >> >> >> > [ERROR] both method checkNotNull in object Preconditions > of > > >> type > > >> > >> >> [T](x$1: > > >> > >> >> >> > T, x$2: String, x$3: <repeated...>[Object])T > > >> > >> >> >> > [ERROR] and method checkNotNull in object Preconditions > of > > >> type > > >> > >> >> [T](x$1: > > >> > >> >> >> > T, x$2: Any)T > > >> > >> >> >> > [ERROR] match argument types ((L, R) => O,String) > > >> > >> >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join > function > > >> must > > >> > >> not be > > >> > >> >> >> > null.") > > >> > >> >> >> > > > >> > >> >> >> > Same error I see for all of the Scala classes I changed. > > Any > > >> > >> pointers > > >> > >> >> >> here > > >> > >> >> >> > will be very helpful for me to proceed further. Please > let > > me > > >> > know > > >> > >> if > > >> > >> >> you > > >> > >> >> >> > need more information. > > >> > >> >> >> > > > >> > >> >> >> > Thanks in advance for your help and support. > > >> > >> >> >> > > > >> > >> >> >> > Thanks, > > >> > >> >> >> > Lokesh > > >> > >> >> >> > > >> > >> >> >> > > >> > >> >> > > >> > >> > > >> > > > >> > > > |
No, you don't have to wait.
The KafkaITCase is not always failing. If you're lucky, it will pass with the next run. On Thu, May 14, 2015 at 5:48 PM, Lokesh Rajaram <[hidden email]> wrote: > If I understand it correct, I have to wait for your pull request to be > merged, I can rebase and trigger build again. is that right? > > Thanks Robert, Aljoscha for super fast reply/help. > > Thanks, > Lokesh > > On Thu, May 14, 2015 at 8:39 AM, Robert Metzger <[hidden email]> > wrote: > > > However, you can only restart runs in your travis account, not on the > > apache account (also used for validating pull requests). > > > > I have opened a pull request a few minutes ago which will reduce the > number > > of KafakITCase failures (there is still one other unresolved issue). > > > > On Thu, May 14, 2015 at 5:37 PM, Aljoscha Krettek <[hidden email]> > > wrote: > > > > > Hi, > > > don't worry, there are very few stupid questions. :D > > > > > > The KafkaITCase sometimes fails on Travis, this is a known problem > > > currently. On travis you can restart the individual runs for a commit > > > in the view of the failed run. > > > > > > Hope that helps. > > > > > > Cheers, > > > Aljoscha > > > > > > On Thu, May 14, 2015 at 5:35 PM, Lokesh Rajaram > > > <[hidden email]> wrote: > > > > Thanks Aljoscha, Robert. After adding guava dependency for > > flink-spargel > > > I > > > > was able to progress further but now it's failing in > > > > flink-streaming-connectors for the following test case: > > > > > > > > KafkaITCase.brokerFailureTest:936->tryExecute:352 Test failed with: > Job > > > > execution failed. > > > > > > > > Any pointers would help me proceed further. Sorry for a lot of > trivial > > > > questions, I am just getting started not familiar with the code base. > > > > I tried running locally, I am able to run it successfully, don't know > > why > > > > it's only failing in Travis build. Not sure if I am missing something > > in > > > my > > > > local environment. > > > > > > > > Thanks, > > > > Lokesh > > > > > > > > On Thu, May 14, 2015 at 1:39 AM, Robert Metzger <[hidden email] > > > > > wrote: > > > > > > > >> I think flink-spargel is missing the guava dependency. > > > >> > > > >> On Thu, May 14, 2015 at 8:18 AM, Aljoscha Krettek < > > [hidden email]> > > > >> wrote: > > > >> > > > >> > @Robert, this seems like a problem with the Shading? > > > >> > > > > >> > On Thu, May 14, 2015 at 5:41 AM, Lokesh Rajaram > > > >> > <[hidden email]> wrote: > > > >> > > Thanks Aljioscha. I was able to change as recommended and able > to > > > run > > > >> the > > > >> > > entire test suite in local successfully. > > > >> > > However Travis build is failing for pull request: > > > >> > > https://github.com/apache/flink/pull/673. > > > >> > > > > > >> > > It's a compilation failure: > > > >> > > > > > >> > > [ERROR] Failed to execute goal > > > >> > > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile > > > >> > > (default-compile) on project flink-spargel: Compilation failure: > > > >> > > Compilation failure: > > > >> > > [ERROR] > > > >> > > > > > >> > > > > >> > > > > > > /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30] > > > >> > > package com.google.common.base does not exist > > > >> > > > > > >> > > I can definitely see the package imported in the class, > compiling > > > and > > > >> > > passing all tests in local. > > > >> > > Anything I am missing here? > > > >> > > > > > >> > > Thanks, > > > >> > > Lokesh > > > >> > > > > > >> > > On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek < > > > [hidden email] > > > >> > > > > >> > > wrote: > > > >> > > > > > >> > >> I think you can replace Validate.NotNull(p) with require(p != > > > null, "p > > > >> > >> is null (or something like this)"). > > > >> > >> > > > >> > >> On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram > > > >> > >> <[hidden email]> wrote: > > > >> > >> > 1. I think I can use require for replacing Validate.isTrue > > > >> > >> > 2. What about Validate.notNull? If require is used it would > > throw > > > >> > >> > IllegalArgumentException, > > > >> > >> > if assume or assert is used it would throw AssertionError > which > > > is > > > >> not > > > >> > >> > compatible with current implementation. > > > >> > >> > > > > >> > >> > Please let me know if my understanding is correct. Also, let > me > > > know > > > >> > your > > > >> > >> > thoughts. > > > >> > >> > > > > >> > >> > Thanks, > > > >> > >> > Lokesh > > > >> > >> > > > > >> > >> > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek < > > > >> > [hidden email]> > > > >> > >> > wrote: > > > >> > >> > > > > >> > >> >> I would propose using the methods as Chiwan suggested. If > > > everyone > > > >> > >> >> agrees I can change the Jira issue. > > > >> > >> >> > > > >> > >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram > > > >> > >> >> <[hidden email]> wrote: > > > >> > >> >> > Thank you for the reference links. Which approach should I > > > take, > > > >> > >> casting > > > >> > >> >> or > > > >> > >> >> > use scala methods. > > > >> > >> >> > If it's the latter option will the JIRA ticket FLINK-1711 > > > >> > >> >> > <https://issues.apache.org/jira/browse/FLINK-1711> be > > > updated to > > > >> > >> >> reflect it? > > > >> > >> >> > > > > >> > >> >> > Thanks, > > > >> > >> >> > Lokesh > > > >> > >> >> > > > > >> > >> >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park < > > > >> [hidden email] > > > >> > > > > > >> > >> >> wrote: > > > >> > >> >> > > > > >> > >> >> >> Hi. There is some problems using Guava’s check method in > > > Scala. > > > >> ( > > > >> > >> >> >> > > > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > > > >> > < > > > >> > >> >> >> > > > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > > > >> > >) > > > >> > >> You > > > >> > >> >> >> can solve this error simply with casting last argument to > > > >> > >> >> java.lang.Object. > > > >> > >> >> >> But I think we’d better use `require`, `assume`, `assert` > > > method > > > >> > >> >> provided > > > >> > >> >> >> by Scala. ( > > > >> > >> >> >> > > > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > > > >> > < > > > >> > >> >> >> > > > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > > > >> > >) > > > >> > >> >> >> Because this changes affects many other codes, so we > should > > > >> > discuss > > > >> > >> >> about > > > >> > >> >> >> changing Guava's method to Scala’s method. > > > >> > >> >> >> > > > >> > >> >> >> Regards. > > > >> > >> >> >> Chiwan Park (Sent with iPhone) > > > >> > >> >> >> > > > >> > >> >> >> > > > >> > >> >> >> > > > >> > >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < > > > >> > >> >> [hidden email]> > > > >> > >> >> >> wrote: > > > >> > >> >> >> > > > > >> > >> >> >> > Hello All, > > > >> > >> >> >> > > > > >> > >> >> >> > I am new to Flink community and am very excited about > the > > > >> > project > > > >> > >> and > > > >> > >> >> >> work > > > >> > >> >> >> > you all have been doing. Kudos!! > > > >> > >> >> >> > > > > >> > >> >> >> > I was looking to pickup some starter task. Robert > > > recommended > > > >> to > > > >> > >> pick > > > >> > >> >> up > > > >> > >> >> >> > https://issues.apache.org/jira/browse/FLINK-1711. > Thanks > > > >> Robert > > > >> > >> for > > > >> > >> >> your > > > >> > >> >> >> > guidance. > > > >> > >> >> >> > > > > >> > >> >> >> > Sorry for a dumb question. I am done with code changes > > but > > > my > > > >> > "mvn > > > >> > >> >> >> verify" > > > >> > >> >> >> > failing only for the scala module as follows > > > >> > >> >> >> > > > > >> > >> >> >> > > > > >> > >> >> >> > > > >> > >> >> > > > >> > >> > > > >> > > > > >> > > > > > > flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: > > > >> > >> >> >> > error: ambiguous reference to overloaded definition, > > > >> > >> >> >> > [ERROR] both method checkNotNull in object > Preconditions > > of > > > >> type > > > >> > >> >> [T](x$1: > > > >> > >> >> >> > T, x$2: String, x$3: <repeated...>[Object])T > > > >> > >> >> >> > [ERROR] and method checkNotNull in object > Preconditions > > of > > > >> type > > > >> > >> >> [T](x$1: > > > >> > >> >> >> > T, x$2: Any)T > > > >> > >> >> >> > [ERROR] match argument types ((L, R) => O,String) > > > >> > >> >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join > > function > > > >> must > > > >> > >> not be > > > >> > >> >> >> > null.") > > > >> > >> >> >> > > > > >> > >> >> >> > Same error I see for all of the Scala classes I > changed. > > > Any > > > >> > >> pointers > > > >> > >> >> >> here > > > >> > >> >> >> > will be very helpful for me to proceed further. Please > > let > > > me > > > >> > know > > > >> > >> if > > > >> > >> >> you > > > >> > >> >> >> > need more information. > > > >> > >> >> >> > > > > >> > >> >> >> > Thanks in advance for your help and support. > > > >> > >> >> >> > > > > >> > >> >> >> > Thanks, > > > >> > >> >> >> > Lokesh > > > >> > >> >> >> > > > >> > >> >> >> > > > >> > >> >> > > > >> > >> > > > >> > > > > >> > > > > > > |
This almost worked. Of the 5 build jobs four passed and one failed.
What's the acceptance criteria for a pull request? Do I need to build again to get all 5 build jobs passing? Thanks, Lokesh On Thu, May 14, 2015 at 8:50 AM, Robert Metzger <[hidden email]> wrote: > No, you don't have to wait. > The KafkaITCase is not always failing. If you're lucky, it will pass with > the next run. > > On Thu, May 14, 2015 at 5:48 PM, Lokesh Rajaram <[hidden email]> > wrote: > > > If I understand it correct, I have to wait for your pull request to be > > merged, I can rebase and trigger build again. is that right? > > > > Thanks Robert, Aljoscha for super fast reply/help. > > > > Thanks, > > Lokesh > > > > On Thu, May 14, 2015 at 8:39 AM, Robert Metzger <[hidden email]> > > wrote: > > > > > However, you can only restart runs in your travis account, not on the > > > apache account (also used for validating pull requests). > > > > > > I have opened a pull request a few minutes ago which will reduce the > > number > > > of KafakITCase failures (there is still one other unresolved issue). > > > > > > On Thu, May 14, 2015 at 5:37 PM, Aljoscha Krettek <[hidden email] > > > > > wrote: > > > > > > > Hi, > > > > don't worry, there are very few stupid questions. :D > > > > > > > > The KafkaITCase sometimes fails on Travis, this is a known problem > > > > currently. On travis you can restart the individual runs for a commit > > > > in the view of the failed run. > > > > > > > > Hope that helps. > > > > > > > > Cheers, > > > > Aljoscha > > > > > > > > On Thu, May 14, 2015 at 5:35 PM, Lokesh Rajaram > > > > <[hidden email]> wrote: > > > > > Thanks Aljoscha, Robert. After adding guava dependency for > > > flink-spargel > > > > I > > > > > was able to progress further but now it's failing in > > > > > flink-streaming-connectors for the following test case: > > > > > > > > > > KafkaITCase.brokerFailureTest:936->tryExecute:352 Test failed with: > > Job > > > > > execution failed. > > > > > > > > > > Any pointers would help me proceed further. Sorry for a lot of > > trivial > > > > > questions, I am just getting started not familiar with the code > base. > > > > > I tried running locally, I am able to run it successfully, don't > know > > > why > > > > > it's only failing in Travis build. Not sure if I am missing > something > > > in > > > > my > > > > > local environment. > > > > > > > > > > Thanks, > > > > > Lokesh > > > > > > > > > > On Thu, May 14, 2015 at 1:39 AM, Robert Metzger < > [hidden email] > > > > > > > wrote: > > > > > > > > > >> I think flink-spargel is missing the guava dependency. > > > > >> > > > > >> On Thu, May 14, 2015 at 8:18 AM, Aljoscha Krettek < > > > [hidden email]> > > > > >> wrote: > > > > >> > > > > >> > @Robert, this seems like a problem with the Shading? > > > > >> > > > > > >> > On Thu, May 14, 2015 at 5:41 AM, Lokesh Rajaram > > > > >> > <[hidden email]> wrote: > > > > >> > > Thanks Aljioscha. I was able to change as recommended and able > > to > > > > run > > > > >> the > > > > >> > > entire test suite in local successfully. > > > > >> > > However Travis build is failing for pull request: > > > > >> > > https://github.com/apache/flink/pull/673. > > > > >> > > > > > > >> > > It's a compilation failure: > > > > >> > > > > > > >> > > [ERROR] Failed to execute goal > > > > >> > > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile > > > > >> > > (default-compile) on project flink-spargel: Compilation > failure: > > > > >> > > Compilation failure: > > > > >> > > [ERROR] > > > > >> > > > > > > >> > > > > > >> > > > > > > > > > > /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30] > > > > >> > > package com.google.common.base does not exist > > > > >> > > > > > > >> > > I can definitely see the package imported in the class, > > compiling > > > > and > > > > >> > > passing all tests in local. > > > > >> > > Anything I am missing here? > > > > >> > > > > > > >> > > Thanks, > > > > >> > > Lokesh > > > > >> > > > > > > >> > > On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek < > > > > [hidden email] > > > > >> > > > > > >> > > wrote: > > > > >> > > > > > > >> > >> I think you can replace Validate.NotNull(p) with require(p != > > > > null, "p > > > > >> > >> is null (or something like this)"). > > > > >> > >> > > > > >> > >> On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram > > > > >> > >> <[hidden email]> wrote: > > > > >> > >> > 1. I think I can use require for replacing Validate.isTrue > > > > >> > >> > 2. What about Validate.notNull? If require is used it would > > > throw > > > > >> > >> > IllegalArgumentException, > > > > >> > >> > if assume or assert is used it would throw AssertionError > > which > > > > is > > > > >> not > > > > >> > >> > compatible with current implementation. > > > > >> > >> > > > > > >> > >> > Please let me know if my understanding is correct. Also, > let > > me > > > > know > > > > >> > your > > > > >> > >> > thoughts. > > > > >> > >> > > > > > >> > >> > Thanks, > > > > >> > >> > Lokesh > > > > >> > >> > > > > > >> > >> > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek < > > > > >> > [hidden email]> > > > > >> > >> > wrote: > > > > >> > >> > > > > > >> > >> >> I would propose using the methods as Chiwan suggested. If > > > > everyone > > > > >> > >> >> agrees I can change the Jira issue. > > > > >> > >> >> > > > > >> > >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram > > > > >> > >> >> <[hidden email]> wrote: > > > > >> > >> >> > Thank you for the reference links. Which approach > should I > > > > take, > > > > >> > >> casting > > > > >> > >> >> or > > > > >> > >> >> > use scala methods. > > > > >> > >> >> > If it's the latter option will the JIRA ticket > FLINK-1711 > > > > >> > >> >> > <https://issues.apache.org/jira/browse/FLINK-1711> be > > > > updated to > > > > >> > >> >> reflect it? > > > > >> > >> >> > > > > > >> > >> >> > Thanks, > > > > >> > >> >> > Lokesh > > > > >> > >> >> > > > > > >> > >> >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park < > > > > >> [hidden email] > > > > >> > > > > > > >> > >> >> wrote: > > > > >> > >> >> > > > > > >> > >> >> >> Hi. There is some problems using Guava’s check method > in > > > > Scala. > > > > >> ( > > > > >> > >> >> >> > > > > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > > > > >> > < > > > > >> > >> >> >> > > > > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k > > > > >> > >) > > > > >> > >> You > > > > >> > >> >> >> can solve this error simply with casting last argument > to > > > > >> > >> >> java.lang.Object. > > > > >> > >> >> >> But I think we’d better use `require`, `assume`, > `assert` > > > > method > > > > >> > >> >> provided > > > > >> > >> >> >> by Scala. ( > > > > >> > >> >> >> > > > > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > > > > >> > < > > > > >> > >> >> >> > > > > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html > > > > >> > >) > > > > >> > >> >> >> Because this changes affects many other codes, so we > > should > > > > >> > discuss > > > > >> > >> >> about > > > > >> > >> >> >> changing Guava's method to Scala’s method. > > > > >> > >> >> >> > > > > >> > >> >> >> Regards. > > > > >> > >> >> >> Chiwan Park (Sent with iPhone) > > > > >> > >> >> >> > > > > >> > >> >> >> > > > > >> > >> >> >> > > > > >> > >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < > > > > >> > >> >> [hidden email]> > > > > >> > >> >> >> wrote: > > > > >> > >> >> >> > > > > > >> > >> >> >> > Hello All, > > > > >> > >> >> >> > > > > > >> > >> >> >> > I am new to Flink community and am very excited about > > the > > > > >> > project > > > > >> > >> and > > > > >> > >> >> >> work > > > > >> > >> >> >> > you all have been doing. Kudos!! > > > > >> > >> >> >> > > > > > >> > >> >> >> > I was looking to pickup some starter task. Robert > > > > recommended > > > > >> to > > > > >> > >> pick > > > > >> > >> >> up > > > > >> > >> >> >> > https://issues.apache.org/jira/browse/FLINK-1711. > > Thanks > > > > >> Robert > > > > >> > >> for > > > > >> > >> >> your > > > > >> > >> >> >> > guidance. > > > > >> > >> >> >> > > > > > >> > >> >> >> > Sorry for a dumb question. I am done with code > changes > > > but > > > > my > > > > >> > "mvn > > > > >> > >> >> >> verify" > > > > >> > >> >> >> > failing only for the scala module as follows > > > > >> > >> >> >> > > > > > >> > >> >> >> > > > > > >> > >> >> >> > > > > >> > >> >> > > > > >> > >> > > > > >> > > > > > >> > > > > > > > > > > flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: > > > > >> > >> >> >> > error: ambiguous reference to overloaded definition, > > > > >> > >> >> >> > [ERROR] both method checkNotNull in object > > Preconditions > > > of > > > > >> type > > > > >> > >> >> [T](x$1: > > > > >> > >> >> >> > T, x$2: String, x$3: <repeated...>[Object])T > > > > >> > >> >> >> > [ERROR] and method checkNotNull in object > > Preconditions > > > of > > > > >> type > > > > >> > >> >> [T](x$1: > > > > >> > >> >> >> > T, x$2: Any)T > > > > >> > >> >> >> > [ERROR] match argument types ((L, R) => O,String) > > > > >> > >> >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join > > > function > > > > >> must > > > > >> > >> not be > > > > >> > >> >> >> > null.") > > > > >> > >> >> >> > > > > > >> > >> >> >> > Same error I see for all of the Scala classes I > > changed. > > > > Any > > > > >> > >> pointers > > > > >> > >> >> >> here > > > > >> > >> >> >> > will be very helpful for me to proceed further. > Please > > > let > > > > me > > > > >> > know > > > > >> > >> if > > > > >> > >> >> you > > > > >> > >> >> >> > need more information. > > > > >> > >> >> >> > > > > > >> > >> >> >> > Thanks in advance for your help and support. > > > > >> > >> >> >> > > > > > >> > >> >> >> > Thanks, > > > > >> > >> >> >> > Lokesh > > > > >> > >> >> >> > > > > >> > >> >> >> > > > > >> > >> >> > > > > >> > >> > > > > >> > > > > > >> > > > > > > > > > > |
We'll look at why it failed and then decide whether it's good. So for
those KafkaITCase fails we no that it doesn't matter right now. On Fri, May 15, 2015 at 12:56 AM, Lokesh Rajaram <[hidden email]> wrote: > This almost worked. Of the 5 build jobs four passed and one failed. > What's the acceptance criteria for a pull request? Do I need to build again > to get all 5 build jobs passing? > > Thanks, > Lokesh > > On Thu, May 14, 2015 at 8:50 AM, Robert Metzger <[hidden email]> wrote: > >> No, you don't have to wait. >> The KafkaITCase is not always failing. If you're lucky, it will pass with >> the next run. >> >> On Thu, May 14, 2015 at 5:48 PM, Lokesh Rajaram <[hidden email]> >> wrote: >> >> > If I understand it correct, I have to wait for your pull request to be >> > merged, I can rebase and trigger build again. is that right? >> > >> > Thanks Robert, Aljoscha for super fast reply/help. >> > >> > Thanks, >> > Lokesh >> > >> > On Thu, May 14, 2015 at 8:39 AM, Robert Metzger <[hidden email]> >> > wrote: >> > >> > > However, you can only restart runs in your travis account, not on the >> > > apache account (also used for validating pull requests). >> > > >> > > I have opened a pull request a few minutes ago which will reduce the >> > number >> > > of KafakITCase failures (there is still one other unresolved issue). >> > > >> > > On Thu, May 14, 2015 at 5:37 PM, Aljoscha Krettek <[hidden email] >> > >> > > wrote: >> > > >> > > > Hi, >> > > > don't worry, there are very few stupid questions. :D >> > > > >> > > > The KafkaITCase sometimes fails on Travis, this is a known problem >> > > > currently. On travis you can restart the individual runs for a commit >> > > > in the view of the failed run. >> > > > >> > > > Hope that helps. >> > > > >> > > > Cheers, >> > > > Aljoscha >> > > > >> > > > On Thu, May 14, 2015 at 5:35 PM, Lokesh Rajaram >> > > > <[hidden email]> wrote: >> > > > > Thanks Aljoscha, Robert. After adding guava dependency for >> > > flink-spargel >> > > > I >> > > > > was able to progress further but now it's failing in >> > > > > flink-streaming-connectors for the following test case: >> > > > > >> > > > > KafkaITCase.brokerFailureTest:936->tryExecute:352 Test failed with: >> > Job >> > > > > execution failed. >> > > > > >> > > > > Any pointers would help me proceed further. Sorry for a lot of >> > trivial >> > > > > questions, I am just getting started not familiar with the code >> base. >> > > > > I tried running locally, I am able to run it successfully, don't >> know >> > > why >> > > > > it's only failing in Travis build. Not sure if I am missing >> something >> > > in >> > > > my >> > > > > local environment. >> > > > > >> > > > > Thanks, >> > > > > Lokesh >> > > > > >> > > > > On Thu, May 14, 2015 at 1:39 AM, Robert Metzger < >> [hidden email] >> > > >> > > > wrote: >> > > > > >> > > > >> I think flink-spargel is missing the guava dependency. >> > > > >> >> > > > >> On Thu, May 14, 2015 at 8:18 AM, Aljoscha Krettek < >> > > [hidden email]> >> > > > >> wrote: >> > > > >> >> > > > >> > @Robert, this seems like a problem with the Shading? >> > > > >> > >> > > > >> > On Thu, May 14, 2015 at 5:41 AM, Lokesh Rajaram >> > > > >> > <[hidden email]> wrote: >> > > > >> > > Thanks Aljioscha. I was able to change as recommended and able >> > to >> > > > run >> > > > >> the >> > > > >> > > entire test suite in local successfully. >> > > > >> > > However Travis build is failing for pull request: >> > > > >> > > https://github.com/apache/flink/pull/673. >> > > > >> > > >> > > > >> > > It's a compilation failure: >> > > > >> > > >> > > > >> > > [ERROR] Failed to execute goal >> > > > >> > > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile >> > > > >> > > (default-compile) on project flink-spargel: Compilation >> failure: >> > > > >> > > Compilation failure: >> > > > >> > > [ERROR] >> > > > >> > > >> > > > >> > >> > > > >> >> > > > >> > > >> > >> /home/travis/build/apache/flink/flink-staging/flink-spargel/src/main/java/org/apache/flink/spargel/java/VertexCentricIteration.java:[42,30] >> > > > >> > > package com.google.common.base does not exist >> > > > >> > > >> > > > >> > > I can definitely see the package imported in the class, >> > compiling >> > > > and >> > > > >> > > passing all tests in local. >> > > > >> > > Anything I am missing here? >> > > > >> > > >> > > > >> > > Thanks, >> > > > >> > > Lokesh >> > > > >> > > >> > > > >> > > On Mon, May 11, 2015 at 1:25 AM, Aljoscha Krettek < >> > > > [hidden email] >> > > > >> > >> > > > >> > > wrote: >> > > > >> > > >> > > > >> > >> I think you can replace Validate.NotNull(p) with require(p != >> > > > null, "p >> > > > >> > >> is null (or something like this)"). >> > > > >> > >> >> > > > >> > >> On Mon, May 11, 2015 at 12:27 AM, Lokesh Rajaram >> > > > >> > >> <[hidden email]> wrote: >> > > > >> > >> > 1. I think I can use require for replacing Validate.isTrue >> > > > >> > >> > 2. What about Validate.notNull? If require is used it would >> > > throw >> > > > >> > >> > IllegalArgumentException, >> > > > >> > >> > if assume or assert is used it would throw AssertionError >> > which >> > > > is >> > > > >> not >> > > > >> > >> > compatible with current implementation. >> > > > >> > >> > >> > > > >> > >> > Please let me know if my understanding is correct. Also, >> let >> > me >> > > > know >> > > > >> > your >> > > > >> > >> > thoughts. >> > > > >> > >> > >> > > > >> > >> > Thanks, >> > > > >> > >> > Lokesh >> > > > >> > >> > >> > > > >> > >> > On Sun, May 10, 2015 at 1:04 AM, Aljoscha Krettek < >> > > > >> > [hidden email]> >> > > > >> > >> > wrote: >> > > > >> > >> > >> > > > >> > >> >> I would propose using the methods as Chiwan suggested. If >> > > > everyone >> > > > >> > >> >> agrees I can change the Jira issue. >> > > > >> > >> >> >> > > > >> > >> >> On Sun, May 10, 2015 at 6:47 AM, Lokesh Rajaram >> > > > >> > >> >> <[hidden email]> wrote: >> > > > >> > >> >> > Thank you for the reference links. Which approach >> should I >> > > > take, >> > > > >> > >> casting >> > > > >> > >> >> or >> > > > >> > >> >> > use scala methods. >> > > > >> > >> >> > If it's the latter option will the JIRA ticket >> FLINK-1711 >> > > > >> > >> >> > <https://issues.apache.org/jira/browse/FLINK-1711> be >> > > > updated to >> > > > >> > >> >> reflect it? >> > > > >> > >> >> > >> > > > >> > >> >> > Thanks, >> > > > >> > >> >> > Lokesh >> > > > >> > >> >> > >> > > > >> > >> >> > On Sat, May 9, 2015 at 8:16 PM, Chiwan Park < >> > > > >> [hidden email] >> > > > >> > > >> > > > >> > >> >> wrote: >> > > > >> > >> >> > >> > > > >> > >> >> >> Hi. There is some problems using Guava’s check method >> in >> > > > Scala. >> > > > >> ( >> > > > >> > >> >> >> >> > > > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k >> > > > >> > < >> > > > >> > >> >> >> >> > > > >> https://groups.google.com/forum/#!topic/guava-discuss/juwovq26R3k >> > > > >> > >) >> > > > >> > >> You >> > > > >> > >> >> >> can solve this error simply with casting last argument >> to >> > > > >> > >> >> java.lang.Object. >> > > > >> > >> >> >> But I think we’d better use `require`, `assume`, >> `assert` >> > > > method >> > > > >> > >> >> provided >> > > > >> > >> >> >> by Scala. ( >> > > > >> > >> >> >> >> > > > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html >> > > > >> > < >> > > > >> > >> >> >> >> > > > >> http://daily-scala.blogspot.kr/2010/03/assert-require-assume.html >> > > > >> > >) >> > > > >> > >> >> >> Because this changes affects many other codes, so we >> > should >> > > > >> > discuss >> > > > >> > >> >> about >> > > > >> > >> >> >> changing Guava's method to Scala’s method. >> > > > >> > >> >> >> >> > > > >> > >> >> >> Regards. >> > > > >> > >> >> >> Chiwan Park (Sent with iPhone) >> > > > >> > >> >> >> >> > > > >> > >> >> >> >> > > > >> > >> >> >> >> > > > >> > >> >> >> > On May 10, 2015, at 11:49 AM, Lokesh Rajaram < >> > > > >> > >> >> [hidden email]> >> > > > >> > >> >> >> wrote: >> > > > >> > >> >> >> > >> > > > >> > >> >> >> > Hello All, >> > > > >> > >> >> >> > >> > > > >> > >> >> >> > I am new to Flink community and am very excited about >> > the >> > > > >> > project >> > > > >> > >> and >> > > > >> > >> >> >> work >> > > > >> > >> >> >> > you all have been doing. Kudos!! >> > > > >> > >> >> >> > >> > > > >> > >> >> >> > I was looking to pickup some starter task. Robert >> > > > recommended >> > > > >> to >> > > > >> > >> pick >> > > > >> > >> >> up >> > > > >> > >> >> >> > https://issues.apache.org/jira/browse/FLINK-1711. >> > Thanks >> > > > >> Robert >> > > > >> > >> for >> > > > >> > >> >> your >> > > > >> > >> >> >> > guidance. >> > > > >> > >> >> >> > >> > > > >> > >> >> >> > Sorry for a dumb question. I am done with code >> changes >> > > but >> > > > my >> > > > >> > "mvn >> > > > >> > >> >> >> verify" >> > > > >> > >> >> >> > failing only for the scala module as follows >> > > > >> > >> >> >> > >> > > > >> > >> >> >> > >> > > > >> > >> >> >> >> > > > >> > >> >> >> > > > >> > >> >> > > > >> > >> > > > >> >> > > > >> > > >> > >> flink/flink-scala/src/main/scala/org/apache/flink/api/scala/joinDataSet.scala:77: >> > > > >> > >> >> >> > error: ambiguous reference to overloaded definition, >> > > > >> > >> >> >> > [ERROR] both method checkNotNull in object >> > Preconditions >> > > of >> > > > >> type >> > > > >> > >> >> [T](x$1: >> > > > >> > >> >> >> > T, x$2: String, x$3: <repeated...>[Object])T >> > > > >> > >> >> >> > [ERROR] and method checkNotNull in object >> > Preconditions >> > > of >> > > > >> type >> > > > >> > >> >> [T](x$1: >> > > > >> > >> >> >> > T, x$2: Any)T >> > > > >> > >> >> >> > [ERROR] match argument types ((L, R) => O,String) >> > > > >> > >> >> >> > [ERROR] Preconditions.checkNotNull(fun, "Join >> > > function >> > > > >> must >> > > > >> > >> not be >> > > > >> > >> >> >> > null.") >> > > > >> > >> >> >> > >> > > > >> > >> >> >> > Same error I see for all of the Scala classes I >> > changed. >> > > > Any >> > > > >> > >> pointers >> > > > >> > >> >> >> here >> > > > >> > >> >> >> > will be very helpful for me to proceed further. >> Please >> > > let >> > > > me >> > > > >> > know >> > > > >> > >> if >> > > > >> > >> >> you >> > > > >> > >> >> >> > need more information. >> > > > >> > >> >> >> > >> > > > >> > >> >> >> > Thanks in advance for your help and support. >> > > > >> > >> >> >> > >> > > > >> > >> >> >> > Thanks, >> > > > >> > >> >> >> > Lokesh >> > > > >> > >> >> >> >> > > > >> > >> >> >> >> > > > >> > >> >> >> > > > >> > >> >> > > > >> > >> > > > >> >> > > > >> > > >> > >> |
Free forum by Nabble | Edit this page |