Current master broken?

classic Classic list List threaded Threaded
8 messages Options
Reply | Threaded
Open this post in threaded view
|

Current master broken?

Stephan Ewen
It seems that the current master is broken, with respect to the tests.

I see all build on Travis consistently failing, in the gelly project. Since
Travis is a bit behind in the "apache" account, I triggered a build in my
own account. The hash is the same, it should contain the master from
yesterday.

https://travis-ci.org/StephanEwen/incubator-flink/builds/54386416

In all executions it results in the stack trace below. I cannot reproduce
the problem locally, unfortunately.

This is a serious issue, it totally kills the testability.

Results :

Failed tests:
  DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcId:113
expected:<[The edge src/trg id could not be found within the
vertexIds]> but was:<[Failed to deploy the task Reduce(SUM(1), at
getDegrees(Graph.java:664) (30/32) - execution #0 to slot SimpleSlot
(2)(2) - 31624115d75feb2c387ae9043021d8e6 - ALLOCATED/ALIVE:
java.io.IOException: Insufficient number of network buffers: required
32, but only 2 available. The total number of network buffers is
currently set to 2048. You can increase this number by setting the
configuration key 'taskmanager.network.numberOfBuffers'.
        at org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
        at org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:163)
        at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
        at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
        at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
        at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
        at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
        at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
        at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
        at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
        at org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
        at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
        at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
        at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
        at akka.actor.ActorCell.invoke(ActorCell.scala:487)
        at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
        at akka.dispatch.Mailbox.run(Mailbox.scala:221)
        at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
        at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
        at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
        at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
        at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
        at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
]>
  DegreesWithExceptionITCase.testGetDegreesInvalidEdgeTrgId:92
expected:<[The edge src/trg id could not be found within the
vertexIds]> but was:<[Failed to deploy the task CoGroup (CoGroup at
inDegrees(Graph.java:655)) (29/32) - execution #0 to slot SimpleSlot
(1)(3) - 1735ca6f2fb76f9f0a0ab03ffd9c9f93 - ALLOCATED/ALIVE:
java.io.IOException: Insufficient number of network buffers: required
32, but only 8 available. The total number of network buffers is
currently set to 2048. You can increase this number by setting the
configuration key 'taskmanager.network.numberOfBuffers'.
        at org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
        at org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
        at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
        at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
        at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
        at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
        at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
        at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
        at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
        at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
        at org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
        at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
        at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
        at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
        at akka.actor.ActorCell.invoke(ActorCell.scala:487)
        at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
        at akka.dispatch.Mailbox.run(Mailbox.scala:221)
        at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
        at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
        at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
        at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
        at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
        at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
]>
  DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcTrgId:134
expected:<[The edge src/trg id could not be found within the
vertexIds]> but was:<[Failed to deploy the task CoGroup (CoGroup at
inDegrees(Graph.java:655)) (31/32) - execution #0 to slot SimpleSlot
(1)(3) - 3a465bdbeca9625e5d90572ed0959b1d - ALLOCATED/ALIVE:
java.io.IOException: Insufficient number of network buffers: required
32, but only 8 available. The total number of network buffers is
currently set to 2048. You can increase this number by setting the
configuration key 'taskmanager.network.numberOfBuffers'.
        at org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
        at org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
        at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
        at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
        at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
        at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
        at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
        at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
        at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
        at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
        at org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
        at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
        at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
        at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
        at akka.actor.ActorCell.invoke(ActorCell.scala:487)
        at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
        at akka.dispatch.Mailbox.run(Mailbox.scala:221)
        at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
        at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
        at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
        at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
        at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
        at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
]>

Tests run: 180, Failures: 3, Errors: 0, Skipped: 0

[INFO]
[INFO] --- maven-failsafe-plugin:2.17:verify (default) @ flink-gelly ---
[INFO] Failsafe report directory:
/home/travis/build/StephanEwen/incubator-flink/flink-staging/flink-gelly/target/failsafe-reports
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] flink .............................................. SUCCESS [  6.075 s]
[INFO] flink-shaded-hadoop ................................ SUCCESS [  1.827 s]
[INFO] flink-shaded-hadoop1 ............................... SUCCESS [  7.384 s]
[INFO] flink-core ......................................... SUCCESS [ 37.973 s]
[INFO] flink-java ......................................... SUCCESS [ 17.373 s]
[INFO] flink-runtime ...................................... SUCCESS [11:13 min]
[INFO] flink-compiler ..................................... SUCCESS [  7.149 s]
[INFO] flink-clients ...................................... SUCCESS [  9.130 s]
[INFO] flink-test-utils ................................... SUCCESS [  8.519 s]
[INFO] flink-scala ........................................ SUCCESS [ 36.171 s]
[INFO] flink-examples ..................................... SUCCESS [  0.370 s]
[INFO] flink-java-examples ................................ SUCCESS [  2.335 s]
[INFO] flink-scala-examples ............................... SUCCESS [ 25.139 s]
[INFO] flink-staging ...................................... SUCCESS [  0.093 s]
[INFO] flink-streaming .................................... SUCCESS [  0.315 s]
[INFO] flink-streaming-core ............................... SUCCESS [  9.560 s]
[INFO] flink-tests ........................................ SUCCESS [09:11 min]
[INFO] flink-avro ......................................... SUCCESS [ 17.307 s]
[INFO] flink-jdbc ......................................... SUCCESS [  3.715 s]
[INFO] flink-spargel ...................................... SUCCESS [  7.141 s]
[INFO] flink-hadoop-compatibility ......................... SUCCESS [ 19.508 s]
[INFO] flink-streaming-scala .............................. SUCCESS [ 14.936 s]
[INFO] flink-streaming-connectors ......................... SUCCESS [  2.784 s]
[INFO] flink-streaming-examples ........................... SUCCESS [ 18.787 s]
[INFO] flink-hbase ........................................ SUCCESS [  2.870 s]
[INFO] flink-gelly ........................................ FAILURE [ 58.548 s]
[INFO] flink-hcatalog ..................................... SKIPPED
[INFO] flink-expressions .................................. SKIPPED
[INFO] flink-quickstart ................................... SKIPPED
[INFO] flink-quickstart-java .............................. SKIPPED
[INFO] flink-quickstart-scala ............................. SKIPPED
[INFO] flink-contrib ...................................... SKIPPED
[INFO] flink-dist ......................................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
Reply | Threaded
Open this post in threaded view
|

Re: Current master broken?

Stephan Ewen
Cause of the Failures:

The tests in DegreesWithExceptionITCase use the context execution
environment without extending a test base. This context environment
instantiates a local excution environment with a parallelism equal to the
number of cores. Since on travis, build run in containers on big machines,
the number of cores may be very high 32/64 - this causes the tests to run
out of network buffers, with the default configuration.


IMPORTANT: Please make sure that all tests in the future either use one of
the test base classes (that define a reasonable parallelism), or define the
parallelism manually to be safe!

On Sun, Mar 15, 2015 at 3:43 PM, Stephan Ewen <[hidden email]> wrote:

> It seems that the current master is broken, with respect to the tests.
>
> I see all build on Travis consistently failing, in the gelly project.
> Since Travis is a bit behind in the "apache" account, I triggered a build
> in my own account. The hash is the same, it should contain the master from
> yesterday.
>
> https://travis-ci.org/StephanEwen/incubator-flink/builds/54386416
>
> In all executions it results in the stack trace below. I cannot reproduce
> the problem locally, unfortunately.
>
> This is a serious issue, it totally kills the testability.
>
> Results :
>
> Failed tests:
>   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcId:113 expected:<[The edge src/trg id could not be found within the vertexIds]> but was:<[Failed to deploy the task Reduce(SUM(1), at getDegrees(Graph.java:664) (30/32) - execution #0 to slot SimpleSlot (2)(2) - 31624115d75feb2c387ae9043021d8e6 - ALLOCATED/ALIVE: java.io.IOException: Insufficient number of network buffers: required 32, but only 2 available. The total number of network buffers is currently set to 2048. You can increase this number by setting the configuration key 'taskmanager.network.numberOfBuffers'.
> at org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> at org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:163)
> at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> at org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> ]>
>   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeTrgId:92 expected:<[The edge src/trg id could not be found within the vertexIds]> but was:<[Failed to deploy the task CoGroup (CoGroup at inDegrees(Graph.java:655)) (29/32) - execution #0 to slot SimpleSlot (1)(3) - 1735ca6f2fb76f9f0a0ab03ffd9c9f93 - ALLOCATED/ALIVE: java.io.IOException: Insufficient number of network buffers: required 32, but only 8 available. The total number of network buffers is currently set to 2048. You can increase this number by setting the configuration key 'taskmanager.network.numberOfBuffers'.
> at org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> at org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> at org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> ]>
>   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcTrgId:134 expected:<[The edge src/trg id could not be found within the vertexIds]> but was:<[Failed to deploy the task CoGroup (CoGroup at inDegrees(Graph.java:655)) (31/32) - execution #0 to slot SimpleSlot (1)(3) - 3a465bdbeca9625e5d90572ed0959b1d - ALLOCATED/ALIVE: java.io.IOException: Insufficient number of network buffers: required 32, but only 8 available. The total number of network buffers is currently set to 2048. You can increase this number by setting the configuration key 'taskmanager.network.numberOfBuffers'.
> at org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> at org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> at org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> ]>
>
> Tests run: 180, Failures: 3, Errors: 0, Skipped: 0
>
> [INFO]
> [INFO] --- maven-failsafe-plugin:2.17:verify (default) @ flink-gelly ---
> [INFO] Failsafe report directory: /home/travis/build/StephanEwen/incubator-flink/flink-staging/flink-gelly/target/failsafe-reports
> [INFO] ------------------------------------------------------------------------
> [INFO] Reactor Summary:
> [INFO]
> [INFO] flink .............................................. SUCCESS [  6.075 s]
> [INFO] flink-shaded-hadoop ................................ SUCCESS [  1.827 s]
> [INFO] flink-shaded-hadoop1 ............................... SUCCESS [  7.384 s]
> [INFO] flink-core ......................................... SUCCESS [ 37.973 s]
> [INFO] flink-java ......................................... SUCCESS [ 17.373 s]
> [INFO] flink-runtime ...................................... SUCCESS [11:13 min]
> [INFO] flink-compiler ..................................... SUCCESS [  7.149 s]
> [INFO] flink-clients ...................................... SUCCESS [  9.130 s]
> [INFO] flink-test-utils ................................... SUCCESS [  8.519 s]
> [INFO] flink-scala ........................................ SUCCESS [ 36.171 s]
> [INFO] flink-examples ..................................... SUCCESS [  0.370 s]
> [INFO] flink-java-examples ................................ SUCCESS [  2.335 s]
> [INFO] flink-scala-examples ............................... SUCCESS [ 25.139 s]
> [INFO] flink-staging ...................................... SUCCESS [  0.093 s]
> [INFO] flink-streaming .................................... SUCCESS [  0.315 s]
> [INFO] flink-streaming-core ............................... SUCCESS [  9.560 s]
> [INFO] flink-tests ........................................ SUCCESS [09:11 min]
> [INFO] flink-avro ......................................... SUCCESS [ 17.307 s]
> [INFO] flink-jdbc ......................................... SUCCESS [  3.715 s]
> [INFO] flink-spargel ...................................... SUCCESS [  7.141 s]
> [INFO] flink-hadoop-compatibility ......................... SUCCESS [ 19.508 s]
> [INFO] flink-streaming-scala .............................. SUCCESS [ 14.936 s]
> [INFO] flink-streaming-connectors ......................... SUCCESS [  2.784 s]
> [INFO] flink-streaming-examples ........................... SUCCESS [ 18.787 s]
> [INFO] flink-hbase ........................................ SUCCESS [  2.870 s]
> [INFO] flink-gelly ........................................ FAILURE [ 58.548 s]
> [INFO] flink-hcatalog ..................................... SKIPPED
> [INFO] flink-expressions .................................. SKIPPED
> [INFO] flink-quickstart ................................... SKIPPED
> [INFO] flink-quickstart-java .............................. SKIPPED
> [INFO] flink-quickstart-scala ............................. SKIPPED
> [INFO] flink-contrib ...................................... SKIPPED
> [INFO] flink-dist ......................................... SKIPPED
> [INFO] ------------------------------------------------------------------------
> [INFO] BUILD FAILURE
> [INFO] ------------------------------------------------------------------------
>
>
Reply | Threaded
Open this post in threaded view
|

Re: Current master broken?

Stephan Ewen
I am fixing this with a slight modification of
https://github.com/apache/flink/pull/475

On Sun, Mar 15, 2015 at 3:54 PM, Stephan Ewen <[hidden email]> wrote:

> Cause of the Failures:
>
> The tests in DegreesWithExceptionITCase use the context execution
> environment without extending a test base. This context environment
> instantiates a local excution environment with a parallelism equal to the
> number of cores. Since on travis, build run in containers on big machines,
> the number of cores may be very high 32/64 - this causes the tests to run
> out of network buffers, with the default configuration.
>
>
> IMPORTANT: Please make sure that all tests in the future either use one of
> the test base classes (that define a reasonable parallelism), or define the
> parallelism manually to be safe!
>
> On Sun, Mar 15, 2015 at 3:43 PM, Stephan Ewen <[hidden email]> wrote:
>
>> It seems that the current master is broken, with respect to the tests.
>>
>> I see all build on Travis consistently failing, in the gelly project.
>> Since Travis is a bit behind in the "apache" account, I triggered a build
>> in my own account. The hash is the same, it should contain the master from
>> yesterday.
>>
>> https://travis-ci.org/StephanEwen/incubator-flink/builds/54386416
>>
>> In all executions it results in the stack trace below. I cannot reproduce
>> the problem locally, unfortunately.
>>
>> This is a serious issue, it totally kills the testability.
>>
>> Results :
>>
>> Failed tests:
>>   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcId:113 expected:<[The edge src/trg id could not be found within the vertexIds]> but was:<[Failed to deploy the task Reduce(SUM(1), at getDegrees(Graph.java:664) (30/32) - execution #0 to slot SimpleSlot (2)(2) - 31624115d75feb2c387ae9043021d8e6 - ALLOCATED/ALIVE: java.io.IOException: Insufficient number of network buffers: required 32, but only 2 available. The total number of network buffers is currently set to 2048. You can increase this number by setting the configuration key 'taskmanager.network.numberOfBuffers'.
>> at org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
>> at org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:163)
>> at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
>> at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
>> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
>> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
>> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
>> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
>> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
>> at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
>> at org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
>> at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>> at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>> at akka.actor.ActorCell.invoke(ActorCell.scala:487)
>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
>> at akka.dispatch.Mailbox.run(Mailbox.scala:221)
>> at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>> at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
>> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
>> at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>> at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>> ]>
>>   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeTrgId:92 expected:<[The edge src/trg id could not be found within the vertexIds]> but was:<[Failed to deploy the task CoGroup (CoGroup at inDegrees(Graph.java:655)) (29/32) - execution #0 to slot SimpleSlot (1)(3) - 1735ca6f2fb76f9f0a0ab03ffd9c9f93 - ALLOCATED/ALIVE: java.io.IOException: Insufficient number of network buffers: required 32, but only 8 available. The total number of network buffers is currently set to 2048. You can increase this number by setting the configuration key 'taskmanager.network.numberOfBuffers'.
>> at org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
>> at org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
>> at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
>> at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
>> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
>> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
>> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
>> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
>> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
>> at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
>> at org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
>> at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>> at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>> at akka.actor.ActorCell.invoke(ActorCell.scala:487)
>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
>> at akka.dispatch.Mailbox.run(Mailbox.scala:221)
>> at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>> at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
>> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
>> at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>> at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>> ]>
>>   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcTrgId:134 expected:<[The edge src/trg id could not be found within the vertexIds]> but was:<[Failed to deploy the task CoGroup (CoGroup at inDegrees(Graph.java:655)) (31/32) - execution #0 to slot SimpleSlot (1)(3) - 3a465bdbeca9625e5d90572ed0959b1d - ALLOCATED/ALIVE: java.io.IOException: Insufficient number of network buffers: required 32, but only 8 available. The total number of network buffers is currently set to 2048. You can increase this number by setting the configuration key 'taskmanager.network.numberOfBuffers'.
>> at org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
>> at org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
>> at org.apache.flink.runtime.taskmanager.TaskManager.org$apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
>> at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
>> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
>> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
>> at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
>> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
>> at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
>> at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
>> at org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
>> at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>> at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>> at akka.actor.ActorCell.invoke(ActorCell.scala:487)
>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
>> at akka.dispatch.Mailbox.run(Mailbox.scala:221)
>> at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>> at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
>> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
>> at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>> at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>> ]>
>>
>> Tests run: 180, Failures: 3, Errors: 0, Skipped: 0
>>
>> [INFO]
>> [INFO] --- maven-failsafe-plugin:2.17:verify (default) @ flink-gelly ---
>> [INFO] Failsafe report directory: /home/travis/build/StephanEwen/incubator-flink/flink-staging/flink-gelly/target/failsafe-reports
>> [INFO] ------------------------------------------------------------------------
>> [INFO] Reactor Summary:
>> [INFO]
>> [INFO] flink .............................................. SUCCESS [  6.075 s]
>> [INFO] flink-shaded-hadoop ................................ SUCCESS [  1.827 s]
>> [INFO] flink-shaded-hadoop1 ............................... SUCCESS [  7.384 s]
>> [INFO] flink-core ......................................... SUCCESS [ 37.973 s]
>> [INFO] flink-java ......................................... SUCCESS [ 17.373 s]
>> [INFO] flink-runtime ...................................... SUCCESS [11:13 min]
>> [INFO] flink-compiler ..................................... SUCCESS [  7.149 s]
>> [INFO] flink-clients ...................................... SUCCESS [  9.130 s]
>> [INFO] flink-test-utils ................................... SUCCESS [  8.519 s]
>> [INFO] flink-scala ........................................ SUCCESS [ 36.171 s]
>> [INFO] flink-examples ..................................... SUCCESS [  0.370 s]
>> [INFO] flink-java-examples ................................ SUCCESS [  2.335 s]
>> [INFO] flink-scala-examples ............................... SUCCESS [ 25.139 s]
>> [INFO] flink-staging ...................................... SUCCESS [  0.093 s]
>> [INFO] flink-streaming .................................... SUCCESS [  0.315 s]
>> [INFO] flink-streaming-core ............................... SUCCESS [  9.560 s]
>> [INFO] flink-tests ........................................ SUCCESS [09:11 min]
>> [INFO] flink-avro ......................................... SUCCESS [ 17.307 s]
>> [INFO] flink-jdbc ......................................... SUCCESS [  3.715 s]
>> [INFO] flink-spargel ...................................... SUCCESS [  7.141 s]
>> [INFO] flink-hadoop-compatibility ......................... SUCCESS [ 19.508 s]
>> [INFO] flink-streaming-scala .............................. SUCCESS [ 14.936 s]
>> [INFO] flink-streaming-connectors ......................... SUCCESS [  2.784 s]
>> [INFO] flink-streaming-examples ........................... SUCCESS [ 18.787 s]
>> [INFO] flink-hbase ........................................ SUCCESS [  2.870 s]
>> [INFO] flink-gelly ........................................ FAILURE [ 58.548 s]
>> [INFO] flink-hcatalog ..................................... SKIPPED
>> [INFO] flink-expressions .................................. SKIPPED
>> [INFO] flink-quickstart ................................... SKIPPED
>> [INFO] flink-quickstart-java .............................. SKIPPED
>> [INFO] flink-quickstart-scala ............................. SKIPPED
>> [INFO] flink-contrib ...................................... SKIPPED
>> [INFO] flink-dist ......................................... SKIPPED
>> [INFO] ------------------------------------------------------------------------
>> [INFO] BUILD FAILURE
>> [INFO] ------------------------------------------------------------------------
>>
>>
>
Reply | Threaded
Open this post in threaded view
|

Re: Current master broken?

Andra Lungu
In reply to this post by Stephan Ewen
Hi Stephan,

The degree of parallelism was manually set there. MultipleProgramsTestBase
cannot be extended; Ufuk explained why.

But I see that for the latest travis check, that test passed.
https://github.com/apache/flink/pull/475

On Sun, Mar 15, 2015 at 3:54 PM, Stephan Ewen <[hidden email]> wrote:

> Cause of the Failures:
>
> The tests in DegreesWithExceptionITCase use the context execution
> environment without extending a test base. This context environment
> instantiates a local excution environment with a parallelism equal to the
> number of cores. Since on travis, build run in containers on big machines,
> the number of cores may be very high 32/64 - this causes the tests to run
> out of network buffers, with the default configuration.
>
>
> IMPORTANT: Please make sure that all tests in the future either use one of
> the test base classes (that define a reasonable parallelism), or define the
> parallelism manually to be safe!
>
> On Sun, Mar 15, 2015 at 3:43 PM, Stephan Ewen <[hidden email]> wrote:
>
> > It seems that the current master is broken, with respect to the tests.
> >
> > I see all build on Travis consistently failing, in the gelly project.
> > Since Travis is a bit behind in the "apache" account, I triggered a build
> > in my own account. The hash is the same, it should contain the master
> from
> > yesterday.
> >
> > https://travis-ci.org/StephanEwen/incubator-flink/builds/54386416
> >
> > In all executions it results in the stack trace below. I cannot reproduce
> > the problem locally, unfortunately.
> >
> > This is a serious issue, it totally kills the testability.
> >
> > Results :
> >
> > Failed tests:
> >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcId:113
> expected:<[The edge src/trg id could not be found within the vertexIds]>
> but was:<[Failed to deploy the task Reduce(SUM(1), at
> getDegrees(Graph.java:664) (30/32) - execution #0 to slot SimpleSlot (2)(2)
> - 31624115d75feb2c387ae9043021d8e6 - ALLOCATED/ALIVE: java.io.IOException:
> Insufficient number of network buffers: required 32, but only 2 available.
> The total number of network buffers is currently set to 2048. You can
> increase this number by setting the configuration key
> 'taskmanager.network.numberOfBuffers'.
> >       at
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> >       at
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:163)
> >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> >       at
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> >       at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> >       at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> >       at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> >       at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> >       at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> >       at
> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> >       at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> >       at
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> >       at
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> >       at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> >       at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> >       at
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> >       at
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > ]>
> >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeTrgId:92
> expected:<[The edge src/trg id could not be found within the vertexIds]>
> but was:<[Failed to deploy the task CoGroup (CoGroup at
> inDegrees(Graph.java:655)) (29/32) - execution #0 to slot SimpleSlot (1)(3)
> - 1735ca6f2fb76f9f0a0ab03ffd9c9f93 - ALLOCATED/ALIVE: java.io.IOException:
> Insufficient number of network buffers: required 32, but only 8 available.
> The total number of network buffers is currently set to 2048. You can
> increase this number by setting the configuration key
> 'taskmanager.network.numberOfBuffers'.
> >       at
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> >       at
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> >       at
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> >       at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> >       at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> >       at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> >       at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> >       at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> >       at
> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> >       at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> >       at
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> >       at
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> >       at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> >       at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> >       at
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> >       at
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > ]>
> >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcTrgId:134
> expected:<[The edge src/trg id could not be found within the vertexIds]>
> but was:<[Failed to deploy the task CoGroup (CoGroup at
> inDegrees(Graph.java:655)) (31/32) - execution #0 to slot SimpleSlot (1)(3)
> - 3a465bdbeca9625e5d90572ed0959b1d - ALLOCATED/ALIVE: java.io.IOException:
> Insufficient number of network buffers: required 32, but only 8 available.
> The total number of network buffers is currently set to 2048. You can
> increase this number by setting the configuration key
> 'taskmanager.network.numberOfBuffers'.
> >       at
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> >       at
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> >       at
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> >       at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> >       at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> >       at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> >       at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> >       at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> >       at
> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> >       at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> >       at
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> >       at
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> >       at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> >       at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> >       at
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> >       at
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > ]>
> >
> > Tests run: 180, Failures: 3, Errors: 0, Skipped: 0
> >
> > [INFO]
> > [INFO] --- maven-failsafe-plugin:2.17:verify (default) @ flink-gelly ---
> > [INFO] Failsafe report directory:
> /home/travis/build/StephanEwen/incubator-flink/flink-staging/flink-gelly/target/failsafe-reports
> > [INFO]
> ------------------------------------------------------------------------
> > [INFO] Reactor Summary:
> > [INFO]
> > [INFO] flink .............................................. SUCCESS [
> 6.075 s]
> > [INFO] flink-shaded-hadoop ................................ SUCCESS [
> 1.827 s]
> > [INFO] flink-shaded-hadoop1 ............................... SUCCESS [
> 7.384 s]
> > [INFO] flink-core ......................................... SUCCESS [
> 37.973 s]
> > [INFO] flink-java ......................................... SUCCESS [
> 17.373 s]
> > [INFO] flink-runtime ...................................... SUCCESS
> [11:13 min]
> > [INFO] flink-compiler ..................................... SUCCESS [
> 7.149 s]
> > [INFO] flink-clients ...................................... SUCCESS [
> 9.130 s]
> > [INFO] flink-test-utils ................................... SUCCESS [
> 8.519 s]
> > [INFO] flink-scala ........................................ SUCCESS [
> 36.171 s]
> > [INFO] flink-examples ..................................... SUCCESS [
> 0.370 s]
> > [INFO] flink-java-examples ................................ SUCCESS [
> 2.335 s]
> > [INFO] flink-scala-examples ............................... SUCCESS [
> 25.139 s]
> > [INFO] flink-staging ...................................... SUCCESS [
> 0.093 s]
> > [INFO] flink-streaming .................................... SUCCESS [
> 0.315 s]
> > [INFO] flink-streaming-core ............................... SUCCESS [
> 9.560 s]
> > [INFO] flink-tests ........................................ SUCCESS
> [09:11 min]
> > [INFO] flink-avro ......................................... SUCCESS [
> 17.307 s]
> > [INFO] flink-jdbc ......................................... SUCCESS [
> 3.715 s]
> > [INFO] flink-spargel ...................................... SUCCESS [
> 7.141 s]
> > [INFO] flink-hadoop-compatibility ......................... SUCCESS [
> 19.508 s]
> > [INFO] flink-streaming-scala .............................. SUCCESS [
> 14.936 s]
> > [INFO] flink-streaming-connectors ......................... SUCCESS [
> 2.784 s]
> > [INFO] flink-streaming-examples ........................... SUCCESS [
> 18.787 s]
> > [INFO] flink-hbase ........................................ SUCCESS [
> 2.870 s]
> > [INFO] flink-gelly ........................................ FAILURE [
> 58.548 s]
> > [INFO] flink-hcatalog ..................................... SKIPPED
> > [INFO] flink-expressions .................................. SKIPPED
> > [INFO] flink-quickstart ................................... SKIPPED
> > [INFO] flink-quickstart-java .............................. SKIPPED
> > [INFO] flink-quickstart-scala ............................. SKIPPED
> > [INFO] flink-contrib ...................................... SKIPPED
> > [INFO] flink-dist ......................................... SKIPPED
> > [INFO]
> ------------------------------------------------------------------------
> > [INFO] BUILD FAILURE
> > [INFO]
> ------------------------------------------------------------------------
> >
> >
>
Reply | Threaded
Open this post in threaded view
|

Re: Current master broken?

Robert Metzger
I think the issue is that our tests are executed on travis machines with
different physical CPU core counts.

I've pushed a 5 days old commit (
https://github.com/rmetzger/flink/commit/b4e8350f52c81704ffc726a1689bb0dc7180776d)
to travis, and it also failed with that issue:
https://travis-ci.org/rmetzger/flink/builds/54443951

Thanks for resolving the issue so quickly Stephan!

On Sun, Mar 15, 2015 at 4:06 PM, Andra Lungu <[hidden email]> wrote:

> Hi Stephan,
>
> The degree of parallelism was manually set there. MultipleProgramsTestBase
> cannot be extended; Ufuk explained why.
>
> But I see that for the latest travis check, that test passed.
> https://github.com/apache/flink/pull/475
>
> On Sun, Mar 15, 2015 at 3:54 PM, Stephan Ewen <[hidden email]> wrote:
>
> > Cause of the Failures:
> >
> > The tests in DegreesWithExceptionITCase use the context execution
> > environment without extending a test base. This context environment
> > instantiates a local excution environment with a parallelism equal to the
> > number of cores. Since on travis, build run in containers on big
> machines,
> > the number of cores may be very high 32/64 - this causes the tests to run
> > out of network buffers, with the default configuration.
> >
> >
> > IMPORTANT: Please make sure that all tests in the future either use one
> of
> > the test base classes (that define a reasonable parallelism), or define
> the
> > parallelism manually to be safe!
> >
> > On Sun, Mar 15, 2015 at 3:43 PM, Stephan Ewen <[hidden email]> wrote:
> >
> > > It seems that the current master is broken, with respect to the tests.
> > >
> > > I see all build on Travis consistently failing, in the gelly project.
> > > Since Travis is a bit behind in the "apache" account, I triggered a
> build
> > > in my own account. The hash is the same, it should contain the master
> > from
> > > yesterday.
> > >
> > > https://travis-ci.org/StephanEwen/incubator-flink/builds/54386416
> > >
> > > In all executions it results in the stack trace below. I cannot
> reproduce
> > > the problem locally, unfortunately.
> > >
> > > This is a serious issue, it totally kills the testability.
> > >
> > > Results :
> > >
> > > Failed tests:
> > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcId:113
> > expected:<[The edge src/trg id could not be found within the vertexIds]>
> > but was:<[Failed to deploy the task Reduce(SUM(1), at
> > getDegrees(Graph.java:664) (30/32) - execution #0 to slot SimpleSlot
> (2)(2)
> > - 31624115d75feb2c387ae9043021d8e6 - ALLOCATED/ALIVE:
> java.io.IOException:
> > Insufficient number of network buffers: required 32, but only 2
> available.
> > The total number of network buffers is currently set to 2048. You can
> > increase this number by setting the configuration key
> > 'taskmanager.network.numberOfBuffers'.
> > >       at
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > >       at
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:163)
> > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > >       at
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > >       at
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > >       at
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > >       at
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > >       at
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > >       at
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > >       at
> > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > >       at
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > >       at
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > >       at
> > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > >       at
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > >       at
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > >       at
> > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > >       at
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > ]>
> > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeTrgId:92
> > expected:<[The edge src/trg id could not be found within the vertexIds]>
> > but was:<[Failed to deploy the task CoGroup (CoGroup at
> > inDegrees(Graph.java:655)) (29/32) - execution #0 to slot SimpleSlot
> (1)(3)
> > - 1735ca6f2fb76f9f0a0ab03ffd9c9f93 - ALLOCATED/ALIVE:
> java.io.IOException:
> > Insufficient number of network buffers: required 32, but only 8
> available.
> > The total number of network buffers is currently set to 2048. You can
> > increase this number by setting the configuration key
> > 'taskmanager.network.numberOfBuffers'.
> > >       at
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > >       at
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > >       at
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > >       at
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > >       at
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > >       at
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > >       at
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > >       at
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > >       at
> > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > >       at
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > >       at
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > >       at
> > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > >       at
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > >       at
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > >       at
> > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > >       at
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > ]>
> > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcTrgId:134
> > expected:<[The edge src/trg id could not be found within the vertexIds]>
> > but was:<[Failed to deploy the task CoGroup (CoGroup at
> > inDegrees(Graph.java:655)) (31/32) - execution #0 to slot SimpleSlot
> (1)(3)
> > - 3a465bdbeca9625e5d90572ed0959b1d - ALLOCATED/ALIVE:
> java.io.IOException:
> > Insufficient number of network buffers: required 32, but only 8
> available.
> > The total number of network buffers is currently set to 2048. You can
> > increase this number by setting the configuration key
> > 'taskmanager.network.numberOfBuffers'.
> > >       at
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > >       at
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > >       at
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > >       at
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > >       at
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > >       at
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > >       at
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > >       at
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > >       at
> > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > >       at
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > >       at
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > >       at
> > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > >       at
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > >       at
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > >       at
> > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > >       at
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > ]>
> > >
> > > Tests run: 180, Failures: 3, Errors: 0, Skipped: 0
> > >
> > > [INFO]
> > > [INFO] --- maven-failsafe-plugin:2.17:verify (default) @ flink-gelly
> ---
> > > [INFO] Failsafe report directory:
> >
> /home/travis/build/StephanEwen/incubator-flink/flink-staging/flink-gelly/target/failsafe-reports
> > > [INFO]
> > ------------------------------------------------------------------------
> > > [INFO] Reactor Summary:
> > > [INFO]
> > > [INFO] flink .............................................. SUCCESS [
> > 6.075 s]
> > > [INFO] flink-shaded-hadoop ................................ SUCCESS [
> > 1.827 s]
> > > [INFO] flink-shaded-hadoop1 ............................... SUCCESS [
> > 7.384 s]
> > > [INFO] flink-core ......................................... SUCCESS [
> > 37.973 s]
> > > [INFO] flink-java ......................................... SUCCESS [
> > 17.373 s]
> > > [INFO] flink-runtime ...................................... SUCCESS
> > [11:13 min]
> > > [INFO] flink-compiler ..................................... SUCCESS [
> > 7.149 s]
> > > [INFO] flink-clients ...................................... SUCCESS [
> > 9.130 s]
> > > [INFO] flink-test-utils ................................... SUCCESS [
> > 8.519 s]
> > > [INFO] flink-scala ........................................ SUCCESS [
> > 36.171 s]
> > > [INFO] flink-examples ..................................... SUCCESS [
> > 0.370 s]
> > > [INFO] flink-java-examples ................................ SUCCESS [
> > 2.335 s]
> > > [INFO] flink-scala-examples ............................... SUCCESS [
> > 25.139 s]
> > > [INFO] flink-staging ...................................... SUCCESS [
> > 0.093 s]
> > > [INFO] flink-streaming .................................... SUCCESS [
> > 0.315 s]
> > > [INFO] flink-streaming-core ............................... SUCCESS [
> > 9.560 s]
> > > [INFO] flink-tests ........................................ SUCCESS
> > [09:11 min]
> > > [INFO] flink-avro ......................................... SUCCESS [
> > 17.307 s]
> > > [INFO] flink-jdbc ......................................... SUCCESS [
> > 3.715 s]
> > > [INFO] flink-spargel ...................................... SUCCESS [
> > 7.141 s]
> > > [INFO] flink-hadoop-compatibility ......................... SUCCESS [
> > 19.508 s]
> > > [INFO] flink-streaming-scala .............................. SUCCESS [
> > 14.936 s]
> > > [INFO] flink-streaming-connectors ......................... SUCCESS [
> > 2.784 s]
> > > [INFO] flink-streaming-examples ........................... SUCCESS [
> > 18.787 s]
> > > [INFO] flink-hbase ........................................ SUCCESS [
> > 2.870 s]
> > > [INFO] flink-gelly ........................................ FAILURE [
> > 58.548 s]
> > > [INFO] flink-hcatalog ..................................... SKIPPED
> > > [INFO] flink-expressions .................................. SKIPPED
> > > [INFO] flink-quickstart ................................... SKIPPED
> > > [INFO] flink-quickstart-java .............................. SKIPPED
> > > [INFO] flink-quickstart-scala ............................. SKIPPED
> > > [INFO] flink-contrib ...................................... SKIPPED
> > > [INFO] flink-dist ......................................... SKIPPED
> > > [INFO]
> > ------------------------------------------------------------------------
> > > [INFO] BUILD FAILURE
> > > [INFO]
> > ------------------------------------------------------------------------
> > >
> > >
> >
>
Reply | Threaded
Open this post in threaded view
|

Re: Current master broken?

Stephan Ewen
Waiting for travis to give me the green light, then I'll push the fix...

On Sun, Mar 15, 2015 at 5:04 PM, Robert Metzger <[hidden email]> wrote:

> I think the issue is that our tests are executed on travis machines with
> different physical CPU core counts.
>
> I've pushed a 5 days old commit (
>
> https://github.com/rmetzger/flink/commit/b4e8350f52c81704ffc726a1689bb0dc7180776d
> )
> to travis, and it also failed with that issue:
> https://travis-ci.org/rmetzger/flink/builds/54443951
>
> Thanks for resolving the issue so quickly Stephan!
>
> On Sun, Mar 15, 2015 at 4:06 PM, Andra Lungu <[hidden email]>
> wrote:
>
> > Hi Stephan,
> >
> > The degree of parallelism was manually set there.
> MultipleProgramsTestBase
> > cannot be extended; Ufuk explained why.
> >
> > But I see that for the latest travis check, that test passed.
> > https://github.com/apache/flink/pull/475
> >
> > On Sun, Mar 15, 2015 at 3:54 PM, Stephan Ewen <[hidden email]> wrote:
> >
> > > Cause of the Failures:
> > >
> > > The tests in DegreesWithExceptionITCase use the context execution
> > > environment without extending a test base. This context environment
> > > instantiates a local excution environment with a parallelism equal to
> the
> > > number of cores. Since on travis, build run in containers on big
> > machines,
> > > the number of cores may be very high 32/64 - this causes the tests to
> run
> > > out of network buffers, with the default configuration.
> > >
> > >
> > > IMPORTANT: Please make sure that all tests in the future either use one
> > of
> > > the test base classes (that define a reasonable parallelism), or define
> > the
> > > parallelism manually to be safe!
> > >
> > > On Sun, Mar 15, 2015 at 3:43 PM, Stephan Ewen <[hidden email]>
> wrote:
> > >
> > > > It seems that the current master is broken, with respect to the
> tests.
> > > >
> > > > I see all build on Travis consistently failing, in the gelly project.
> > > > Since Travis is a bit behind in the "apache" account, I triggered a
> > build
> > > > in my own account. The hash is the same, it should contain the master
> > > from
> > > > yesterday.
> > > >
> > > > https://travis-ci.org/StephanEwen/incubator-flink/builds/54386416
> > > >
> > > > In all executions it results in the stack trace below. I cannot
> > reproduce
> > > > the problem locally, unfortunately.
> > > >
> > > > This is a serious issue, it totally kills the testability.
> > > >
> > > > Results :
> > > >
> > > > Failed tests:
> > > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcId:113
> > > expected:<[The edge src/trg id could not be found within the
> vertexIds]>
> > > but was:<[Failed to deploy the task Reduce(SUM(1), at
> > > getDegrees(Graph.java:664) (30/32) - execution #0 to slot SimpleSlot
> > (2)(2)
> > > - 31624115d75feb2c387ae9043021d8e6 - ALLOCATED/ALIVE:
> > java.io.IOException:
> > > Insufficient number of network buffers: required 32, but only 2
> > available.
> > > The total number of network buffers is currently set to 2048. You can
> > > increase this number by setting the configuration key
> > > 'taskmanager.network.numberOfBuffers'.
> > > >       at
> > >
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > > >       at
> > >
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:163)
> > > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> > >
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > > >       at
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > > >       at
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > > >       at
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > > >       at
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > > >       at
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > > >       at
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > > >       at
> > > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > > >       at
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > > >       at
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > > >       at
> > > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > > >       at
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > > >       at
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > > >       at
> > >
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > > >       at
> > >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > > ]>
> > > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeTrgId:92
> > > expected:<[The edge src/trg id could not be found within the
> vertexIds]>
> > > but was:<[Failed to deploy the task CoGroup (CoGroup at
> > > inDegrees(Graph.java:655)) (29/32) - execution #0 to slot SimpleSlot
> > (1)(3)
> > > - 1735ca6f2fb76f9f0a0ab03ffd9c9f93 - ALLOCATED/ALIVE:
> > java.io.IOException:
> > > Insufficient number of network buffers: required 32, but only 8
> > available.
> > > The total number of network buffers is currently set to 2048. You can
> > > increase this number by setting the configuration key
> > > 'taskmanager.network.numberOfBuffers'.
> > > >       at
> > >
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > > >       at
> > >
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> > > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> > >
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > > >       at
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > > >       at
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > > >       at
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > > >       at
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > > >       at
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > > >       at
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > > >       at
> > > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > > >       at
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > > >       at
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > > >       at
> > > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > > >       at
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > > >       at
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > > >       at
> > >
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > > >       at
> > >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > > ]>
> > > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcTrgId:134
> > > expected:<[The edge src/trg id could not be found within the
> vertexIds]>
> > > but was:<[Failed to deploy the task CoGroup (CoGroup at
> > > inDegrees(Graph.java:655)) (31/32) - execution #0 to slot SimpleSlot
> > (1)(3)
> > > - 3a465bdbeca9625e5d90572ed0959b1d - ALLOCATED/ALIVE:
> > java.io.IOException:
> > > Insufficient number of network buffers: required 32, but only 8
> > available.
> > > The total number of network buffers is currently set to 2048. You can
> > > increase this number by setting the configuration key
> > > 'taskmanager.network.numberOfBuffers'.
> > > >       at
> > >
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > > >       at
> > >
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> > > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> > >
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > > >       at
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > > >       at
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > > >       at
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > > >       at
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > > >       at
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > > >       at
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > > >       at
> > > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > > >       at
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > > >       at
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > > >       at
> > > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > > >       at
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > > >       at
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > > >       at
> > >
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > > >       at
> > >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > > ]>
> > > >
> > > > Tests run: 180, Failures: 3, Errors: 0, Skipped: 0
> > > >
> > > > [INFO]
> > > > [INFO] --- maven-failsafe-plugin:2.17:verify (default) @ flink-gelly
> > ---
> > > > [INFO] Failsafe report directory:
> > >
> >
> /home/travis/build/StephanEwen/incubator-flink/flink-staging/flink-gelly/target/failsafe-reports
> > > > [INFO]
> > >
> ------------------------------------------------------------------------
> > > > [INFO] Reactor Summary:
> > > > [INFO]
> > > > [INFO] flink .............................................. SUCCESS [
> > > 6.075 s]
> > > > [INFO] flink-shaded-hadoop ................................ SUCCESS [
> > > 1.827 s]
> > > > [INFO] flink-shaded-hadoop1 ............................... SUCCESS [
> > > 7.384 s]
> > > > [INFO] flink-core ......................................... SUCCESS [
> > > 37.973 s]
> > > > [INFO] flink-java ......................................... SUCCESS [
> > > 17.373 s]
> > > > [INFO] flink-runtime ...................................... SUCCESS
> > > [11:13 min]
> > > > [INFO] flink-compiler ..................................... SUCCESS [
> > > 7.149 s]
> > > > [INFO] flink-clients ...................................... SUCCESS [
> > > 9.130 s]
> > > > [INFO] flink-test-utils ................................... SUCCESS [
> > > 8.519 s]
> > > > [INFO] flink-scala ........................................ SUCCESS [
> > > 36.171 s]
> > > > [INFO] flink-examples ..................................... SUCCESS [
> > > 0.370 s]
> > > > [INFO] flink-java-examples ................................ SUCCESS [
> > > 2.335 s]
> > > > [INFO] flink-scala-examples ............................... SUCCESS [
> > > 25.139 s]
> > > > [INFO] flink-staging ...................................... SUCCESS [
> > > 0.093 s]
> > > > [INFO] flink-streaming .................................... SUCCESS [
> > > 0.315 s]
> > > > [INFO] flink-streaming-core ............................... SUCCESS [
> > > 9.560 s]
> > > > [INFO] flink-tests ........................................ SUCCESS
> > > [09:11 min]
> > > > [INFO] flink-avro ......................................... SUCCESS [
> > > 17.307 s]
> > > > [INFO] flink-jdbc ......................................... SUCCESS [
> > > 3.715 s]
> > > > [INFO] flink-spargel ...................................... SUCCESS [
> > > 7.141 s]
> > > > [INFO] flink-hadoop-compatibility ......................... SUCCESS [
> > > 19.508 s]
> > > > [INFO] flink-streaming-scala .............................. SUCCESS [
> > > 14.936 s]
> > > > [INFO] flink-streaming-connectors ......................... SUCCESS [
> > > 2.784 s]
> > > > [INFO] flink-streaming-examples ........................... SUCCESS [
> > > 18.787 s]
> > > > [INFO] flink-hbase ........................................ SUCCESS [
> > > 2.870 s]
> > > > [INFO] flink-gelly ........................................ FAILURE [
> > > 58.548 s]
> > > > [INFO] flink-hcatalog ..................................... SKIPPED
> > > > [INFO] flink-expressions .................................. SKIPPED
> > > > [INFO] flink-quickstart ................................... SKIPPED
> > > > [INFO] flink-quickstart-java .............................. SKIPPED
> > > > [INFO] flink-quickstart-scala ............................. SKIPPED
> > > > [INFO] flink-contrib ...................................... SKIPPED
> > > > [INFO] flink-dist ......................................... SKIPPED
> > > > [INFO]
> > >
> ------------------------------------------------------------------------
> > > > [INFO] BUILD FAILURE
> > > > [INFO]
> > >
> ------------------------------------------------------------------------
> > > >
> > > >
> > >
> >
>
Reply | Threaded
Open this post in threaded view
|

Re: Current master broken?

Vasiliki Kalavri
Hi,

thanks a lot for fixing this so quickly!

Just to make sure I fully understand what happened, how come the travis
build was successful when we merged this, but failed later?
Is there a way to avoid such issues in the future?

Cheers,
Vasia.


On 15 March 2015 at 17:07, Stephan Ewen <[hidden email]> wrote:

> Waiting for travis to give me the green light, then I'll push the fix...
>
> On Sun, Mar 15, 2015 at 5:04 PM, Robert Metzger <[hidden email]>
> wrote:
>
> > I think the issue is that our tests are executed on travis machines with
> > different physical CPU core counts.
> >
> > I've pushed a 5 days old commit (
> >
> >
> https://github.com/rmetzger/flink/commit/b4e8350f52c81704ffc726a1689bb0dc7180776d
> > )
> > to travis, and it also failed with that issue:
> > https://travis-ci.org/rmetzger/flink/builds/54443951
> >
> > Thanks for resolving the issue so quickly Stephan!
> >
> > On Sun, Mar 15, 2015 at 4:06 PM, Andra Lungu <[hidden email]>
> > wrote:
> >
> > > Hi Stephan,
> > >
> > > The degree of parallelism was manually set there.
> > MultipleProgramsTestBase
> > > cannot be extended; Ufuk explained why.
> > >
> > > But I see that for the latest travis check, that test passed.
> > > https://github.com/apache/flink/pull/475
> > >
> > > On Sun, Mar 15, 2015 at 3:54 PM, Stephan Ewen <[hidden email]>
> wrote:
> > >
> > > > Cause of the Failures:
> > > >
> > > > The tests in DegreesWithExceptionITCase use the context execution
> > > > environment without extending a test base. This context environment
> > > > instantiates a local excution environment with a parallelism equal to
> > the
> > > > number of cores. Since on travis, build run in containers on big
> > > machines,
> > > > the number of cores may be very high 32/64 - this causes the tests to
> > run
> > > > out of network buffers, with the default configuration.
> > > >
> > > >
> > > > IMPORTANT: Please make sure that all tests in the future either use
> one
> > > of
> > > > the test base classes (that define a reasonable parallelism), or
> define
> > > the
> > > > parallelism manually to be safe!
> > > >
> > > > On Sun, Mar 15, 2015 at 3:43 PM, Stephan Ewen <[hidden email]>
> > wrote:
> > > >
> > > > > It seems that the current master is broken, with respect to the
> > tests.
> > > > >
> > > > > I see all build on Travis consistently failing, in the gelly
> project.
> > > > > Since Travis is a bit behind in the "apache" account, I triggered a
> > > build
> > > > > in my own account. The hash is the same, it should contain the
> master
> > > > from
> > > > > yesterday.
> > > > >
> > > > > https://travis-ci.org/StephanEwen/incubator-flink/builds/54386416
> > > > >
> > > > > In all executions it results in the stack trace below. I cannot
> > > reproduce
> > > > > the problem locally, unfortunately.
> > > > >
> > > > > This is a serious issue, it totally kills the testability.
> > > > >
> > > > > Results :
> > > > >
> > > > > Failed tests:
> > > > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcId:113
> > > > expected:<[The edge src/trg id could not be found within the
> > vertexIds]>
> > > > but was:<[Failed to deploy the task Reduce(SUM(1), at
> > > > getDegrees(Graph.java:664) (30/32) - execution #0 to slot SimpleSlot
> > > (2)(2)
> > > > - 31624115d75feb2c387ae9043021d8e6 - ALLOCATED/ALIVE:
> > > java.io.IOException:
> > > > Insufficient number of network buffers: required 32, but only 2
> > > available.
> > > > The total number of network buffers is currently set to 2048. You can
> > > > increase this number by setting the configuration key
> > > > 'taskmanager.network.numberOfBuffers'.
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:163)
> > > > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> > > >
> > >
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > > > >       at
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > > > >       at
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > > > >       at
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > > > >       at
> > > > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > > > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > > > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > > > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > > > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > > > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > > > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > > > >       at
> > > > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > > > >       at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > > > >       at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > > > >       at
> > > >
> > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > > > >       at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > > > ]>
> > > > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeTrgId:92
> > > > expected:<[The edge src/trg id could not be found within the
> > vertexIds]>
> > > > but was:<[Failed to deploy the task CoGroup (CoGroup at
> > > > inDegrees(Graph.java:655)) (29/32) - execution #0 to slot SimpleSlot
> > > (1)(3)
> > > > - 1735ca6f2fb76f9f0a0ab03ffd9c9f93 - ALLOCATED/ALIVE:
> > > java.io.IOException:
> > > > Insufficient number of network buffers: required 32, but only 8
> > > available.
> > > > The total number of network buffers is currently set to 2048. You can
> > > > increase this number by setting the configuration key
> > > > 'taskmanager.network.numberOfBuffers'.
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> > > > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> > > >
> > >
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > > > >       at
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > > > >       at
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > > > >       at
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > > > >       at
> > > > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > > > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > > > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > > > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > > > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > > > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > > > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > > > >       at
> > > > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > > > >       at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > > > >       at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > > > >       at
> > > >
> > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > > > >       at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > > > ]>
> > > > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcTrgId:134
> > > > expected:<[The edge src/trg id could not be found within the
> > vertexIds]>
> > > > but was:<[Failed to deploy the task CoGroup (CoGroup at
> > > > inDegrees(Graph.java:655)) (31/32) - execution #0 to slot SimpleSlot
> > > (1)(3)
> > > > - 3a465bdbeca9625e5d90572ed0959b1d - ALLOCATED/ALIVE:
> > > java.io.IOException:
> > > > Insufficient number of network buffers: required 32, but only 8
> > > available.
> > > > The total number of network buffers is currently set to 2048. You can
> > > > increase this number by setting the configuration key
> > > > 'taskmanager.network.numberOfBuffers'.
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> > > > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> > > >
> > >
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > > > >       at
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > > > >       at
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > > > >       at
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > > > >       at
> > > > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > > > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > > > >       at
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > > > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > > > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > > > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > > > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > > > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > > > >       at
> > > > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > > > >       at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > > > >       at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > > > >       at
> > > >
> > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > > > >       at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > > > ]>
> > > > >
> > > > > Tests run: 180, Failures: 3, Errors: 0, Skipped: 0
> > > > >
> > > > > [INFO]
> > > > > [INFO] --- maven-failsafe-plugin:2.17:verify (default) @
> flink-gelly
> > > ---
> > > > > [INFO] Failsafe report directory:
> > > >
> > >
> >
> /home/travis/build/StephanEwen/incubator-flink/flink-staging/flink-gelly/target/failsafe-reports
> > > > > [INFO]
> > > >
> > ------------------------------------------------------------------------
> > > > > [INFO] Reactor Summary:
> > > > > [INFO]
> > > > > [INFO] flink ..............................................
> SUCCESS [
> > > > 6.075 s]
> > > > > [INFO] flink-shaded-hadoop ................................
> SUCCESS [
> > > > 1.827 s]
> > > > > [INFO] flink-shaded-hadoop1 ...............................
> SUCCESS [
> > > > 7.384 s]
> > > > > [INFO] flink-core .........................................
> SUCCESS [
> > > > 37.973 s]
> > > > > [INFO] flink-java .........................................
> SUCCESS [
> > > > 17.373 s]
> > > > > [INFO] flink-runtime ...................................... SUCCESS
> > > > [11:13 min]
> > > > > [INFO] flink-compiler .....................................
> SUCCESS [
> > > > 7.149 s]
> > > > > [INFO] flink-clients ......................................
> SUCCESS [
> > > > 9.130 s]
> > > > > [INFO] flink-test-utils ...................................
> SUCCESS [
> > > > 8.519 s]
> > > > > [INFO] flink-scala ........................................
> SUCCESS [
> > > > 36.171 s]
> > > > > [INFO] flink-examples .....................................
> SUCCESS [
> > > > 0.370 s]
> > > > > [INFO] flink-java-examples ................................
> SUCCESS [
> > > > 2.335 s]
> > > > > [INFO] flink-scala-examples ...............................
> SUCCESS [
> > > > 25.139 s]
> > > > > [INFO] flink-staging ......................................
> SUCCESS [
> > > > 0.093 s]
> > > > > [INFO] flink-streaming ....................................
> SUCCESS [
> > > > 0.315 s]
> > > > > [INFO] flink-streaming-core ...............................
> SUCCESS [
> > > > 9.560 s]
> > > > > [INFO] flink-tests ........................................ SUCCESS
> > > > [09:11 min]
> > > > > [INFO] flink-avro .........................................
> SUCCESS [
> > > > 17.307 s]
> > > > > [INFO] flink-jdbc .........................................
> SUCCESS [
> > > > 3.715 s]
> > > > > [INFO] flink-spargel ......................................
> SUCCESS [
> > > > 7.141 s]
> > > > > [INFO] flink-hadoop-compatibility .........................
> SUCCESS [
> > > > 19.508 s]
> > > > > [INFO] flink-streaming-scala ..............................
> SUCCESS [
> > > > 14.936 s]
> > > > > [INFO] flink-streaming-connectors .........................
> SUCCESS [
> > > > 2.784 s]
> > > > > [INFO] flink-streaming-examples ...........................
> SUCCESS [
> > > > 18.787 s]
> > > > > [INFO] flink-hbase ........................................
> SUCCESS [
> > > > 2.870 s]
> > > > > [INFO] flink-gelly ........................................
> FAILURE [
> > > > 58.548 s]
> > > > > [INFO] flink-hcatalog ..................................... SKIPPED
> > > > > [INFO] flink-expressions .................................. SKIPPED
> > > > > [INFO] flink-quickstart ................................... SKIPPED
> > > > > [INFO] flink-quickstart-java .............................. SKIPPED
> > > > > [INFO] flink-quickstart-scala ............................. SKIPPED
> > > > > [INFO] flink-contrib ...................................... SKIPPED
> > > > > [INFO] flink-dist ......................................... SKIPPED
> > > > > [INFO]
> > > >
> > ------------------------------------------------------------------------
> > > > > [INFO] BUILD FAILURE
> > > > > [INFO]
> > > >
> > ------------------------------------------------------------------------
> > > > >
> > > > >
> > > >
> > >
> >
>
Reply | Threaded
Open this post in threaded view
|

Re: Current master broken?

Stephan Ewen
The reason is probably that Travis used different VMs then and now. The
test did not specify the parallelism itself, but picked it up from the
hardware. When the VMs changed, the parallelism changed (increased a lot)
to the point where the default environment has no more network buffers.

Simple fix is to use either test bases, or specify the parallelism directly.
Am 15.03.2015 19:24 schrieb "Vasiliki Kalavri" <[hidden email]>:

> Hi,
>
> thanks a lot for fixing this so quickly!
>
> Just to make sure I fully understand what happened, how come the travis
> build was successful when we merged this, but failed later?
> Is there a way to avoid such issues in the future?
>
> Cheers,
> Vasia.
>
>
> On 15 March 2015 at 17:07, Stephan Ewen <[hidden email]> wrote:
>
> > Waiting for travis to give me the green light, then I'll push the fix...
> >
> > On Sun, Mar 15, 2015 at 5:04 PM, Robert Metzger <[hidden email]>
> > wrote:
> >
> > > I think the issue is that our tests are executed on travis machines
> with
> > > different physical CPU core counts.
> > >
> > > I've pushed a 5 days old commit (
> > >
> > >
> >
> https://github.com/rmetzger/flink/commit/b4e8350f52c81704ffc726a1689bb0dc7180776d
> > > )
> > > to travis, and it also failed with that issue:
> > > https://travis-ci.org/rmetzger/flink/builds/54443951
> > >
> > > Thanks for resolving the issue so quickly Stephan!
> > >
> > > On Sun, Mar 15, 2015 at 4:06 PM, Andra Lungu <[hidden email]>
> > > wrote:
> > >
> > > > Hi Stephan,
> > > >
> > > > The degree of parallelism was manually set there.
> > > MultipleProgramsTestBase
> > > > cannot be extended; Ufuk explained why.
> > > >
> > > > But I see that for the latest travis check, that test passed.
> > > > https://github.com/apache/flink/pull/475
> > > >
> > > > On Sun, Mar 15, 2015 at 3:54 PM, Stephan Ewen <[hidden email]>
> > wrote:
> > > >
> > > > > Cause of the Failures:
> > > > >
> > > > > The tests in DegreesWithExceptionITCase use the context execution
> > > > > environment without extending a test base. This context environment
> > > > > instantiates a local excution environment with a parallelism equal
> to
> > > the
> > > > > number of cores. Since on travis, build run in containers on big
> > > > machines,
> > > > > the number of cores may be very high 32/64 - this causes the tests
> to
> > > run
> > > > > out of network buffers, with the default configuration.
> > > > >
> > > > >
> > > > > IMPORTANT: Please make sure that all tests in the future either use
> > one
> > > > of
> > > > > the test base classes (that define a reasonable parallelism), or
> > define
> > > > the
> > > > > parallelism manually to be safe!
> > > > >
> > > > > On Sun, Mar 15, 2015 at 3:43 PM, Stephan Ewen <[hidden email]>
> > > wrote:
> > > > >
> > > > > > It seems that the current master is broken, with respect to the
> > > tests.
> > > > > >
> > > > > > I see all build on Travis consistently failing, in the gelly
> > project.
> > > > > > Since Travis is a bit behind in the "apache" account, I
> triggered a
> > > > build
> > > > > > in my own account. The hash is the same, it should contain the
> > master
> > > > > from
> > > > > > yesterday.
> > > > > >
> > > > > >
> https://travis-ci.org/StephanEwen/incubator-flink/builds/54386416
> > > > > >
> > > > > > In all executions it results in the stack trace below. I cannot
> > > > reproduce
> > > > > > the problem locally, unfortunately.
> > > > > >
> > > > > > This is a serious issue, it totally kills the testability.
> > > > > >
> > > > > > Results :
> > > > > >
> > > > > > Failed tests:
> > > > > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcId:113
> > > > > expected:<[The edge src/trg id could not be found within the
> > > vertexIds]>
> > > > > but was:<[Failed to deploy the task Reduce(SUM(1), at
> > > > > getDegrees(Graph.java:664) (30/32) - execution #0 to slot
> SimpleSlot
> > > > (2)(2)
> > > > > - 31624115d75feb2c387ae9043021d8e6 - ALLOCATED/ALIVE:
> > > > java.io.IOException:
> > > > > Insufficient number of network buffers: required 32, but only 2
> > > > available.
> > > > > The total number of network buffers is currently set to 2048. You
> can
> > > > > increase this number by setting the configuration key
> > > > > 'taskmanager.network.numberOfBuffers'.
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:163)
> > > > > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> > > > >
> > > >
> > >
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > > > > >       at
> > > > > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > > > > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > > > > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > > > > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > > > > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > > > > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > > > > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > > > > >       at
> > > > >
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > > > > >       at
> > > > >
> > >
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > > > > ]>
> > > > > >   DegreesWithExceptionITCase.testGetDegreesInvalidEdgeTrgId:92
> > > > > expected:<[The edge src/trg id could not be found within the
> > > vertexIds]>
> > > > > but was:<[Failed to deploy the task CoGroup (CoGroup at
> > > > > inDegrees(Graph.java:655)) (29/32) - execution #0 to slot
> SimpleSlot
> > > > (1)(3)
> > > > > - 1735ca6f2fb76f9f0a0ab03ffd9c9f93 - ALLOCATED/ALIVE:
> > > > java.io.IOException:
> > > > > Insufficient number of network buffers: required 32, but only 8
> > > > available.
> > > > > The total number of network buffers is currently set to 2048. You
> can
> > > > > increase this number by setting the configuration key
> > > > > 'taskmanager.network.numberOfBuffers'.
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> > > > > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> > > > >
> > > >
> > >
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > > > > >       at
> > > > > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > > > > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > > > > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > > > > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > > > > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > > > > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > > > > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > > > > >       at
> > > > >
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > > > > >       at
> > > > >
> > >
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > > > > ]>
> > > > > >
>  DegreesWithExceptionITCase.testGetDegreesInvalidEdgeSrcTrgId:134
> > > > > expected:<[The edge src/trg id could not be found within the
> > > vertexIds]>
> > > > > but was:<[Failed to deploy the task CoGroup (CoGroup at
> > > > > inDegrees(Graph.java:655)) (31/32) - execution #0 to slot
> SimpleSlot
> > > > (1)(3)
> > > > > - 3a465bdbeca9625e5d90572ed0959b1d - ALLOCATED/ALIVE:
> > > > java.io.IOException:
> > > > > Insufficient number of network buffers: required 32, but only 8
> > > > available.
> > > > > The total number of network buffers is currently set to 2048. You
> can
> > > > > increase this number by setting the configuration key
> > > > > 'taskmanager.network.numberOfBuffers'.
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:158)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:135)
> > > > > >       at org.apache.flink.runtime.taskmanager.TaskManager.org
> > > > >
> > > >
> > >
> >
> $apache$flink$runtime$taskmanager$TaskManager$$submitTask(TaskManager.scala:454)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:237)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:37)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:30)
> > > > > >       at
> > > > > scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:30)
> > > > > >       at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:91)
> > > > > >       at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> > > > > >       at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> > > > > >       at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
> > > > > >       at akka.dispatch.Mailbox.run(Mailbox.scala:221)
> > > > > >       at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > > > > >       at
> > > > >
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > > > > >       at
> > > > >
> > >
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > > > > >       at
> > > > >
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > > > > ]>
> > > > > >
> > > > > > Tests run: 180, Failures: 3, Errors: 0, Skipped: 0
> > > > > >
> > > > > > [INFO]
> > > > > > [INFO] --- maven-failsafe-plugin:2.17:verify (default) @
> > flink-gelly
> > > > ---
> > > > > > [INFO] Failsafe report directory:
> > > > >
> > > >
> > >
> >
> /home/travis/build/StephanEwen/incubator-flink/flink-staging/flink-gelly/target/failsafe-reports
> > > > > > [INFO]
> > > > >
> > >
> ------------------------------------------------------------------------
> > > > > > [INFO] Reactor Summary:
> > > > > > [INFO]
> > > > > > [INFO] flink ..............................................
> > SUCCESS [
> > > > > 6.075 s]
> > > > > > [INFO] flink-shaded-hadoop ................................
> > SUCCESS [
> > > > > 1.827 s]
> > > > > > [INFO] flink-shaded-hadoop1 ...............................
> > SUCCESS [
> > > > > 7.384 s]
> > > > > > [INFO] flink-core .........................................
> > SUCCESS [
> > > > > 37.973 s]
> > > > > > [INFO] flink-java .........................................
> > SUCCESS [
> > > > > 17.373 s]
> > > > > > [INFO] flink-runtime ......................................
> SUCCESS
> > > > > [11:13 min]
> > > > > > [INFO] flink-compiler .....................................
> > SUCCESS [
> > > > > 7.149 s]
> > > > > > [INFO] flink-clients ......................................
> > SUCCESS [
> > > > > 9.130 s]
> > > > > > [INFO] flink-test-utils ...................................
> > SUCCESS [
> > > > > 8.519 s]
> > > > > > [INFO] flink-scala ........................................
> > SUCCESS [
> > > > > 36.171 s]
> > > > > > [INFO] flink-examples .....................................
> > SUCCESS [
> > > > > 0.370 s]
> > > > > > [INFO] flink-java-examples ................................
> > SUCCESS [
> > > > > 2.335 s]
> > > > > > [INFO] flink-scala-examples ...............................
> > SUCCESS [
> > > > > 25.139 s]
> > > > > > [INFO] flink-staging ......................................
> > SUCCESS [
> > > > > 0.093 s]
> > > > > > [INFO] flink-streaming ....................................
> > SUCCESS [
> > > > > 0.315 s]
> > > > > > [INFO] flink-streaming-core ...............................
> > SUCCESS [
> > > > > 9.560 s]
> > > > > > [INFO] flink-tests ........................................
> SUCCESS
> > > > > [09:11 min]
> > > > > > [INFO] flink-avro .........................................
> > SUCCESS [
> > > > > 17.307 s]
> > > > > > [INFO] flink-jdbc .........................................
> > SUCCESS [
> > > > > 3.715 s]
> > > > > > [INFO] flink-spargel ......................................
> > SUCCESS [
> > > > > 7.141 s]
> > > > > > [INFO] flink-hadoop-compatibility .........................
> > SUCCESS [
> > > > > 19.508 s]
> > > > > > [INFO] flink-streaming-scala ..............................
> > SUCCESS [
> > > > > 14.936 s]
> > > > > > [INFO] flink-streaming-connectors .........................
> > SUCCESS [
> > > > > 2.784 s]
> > > > > > [INFO] flink-streaming-examples ...........................
> > SUCCESS [
> > > > > 18.787 s]
> > > > > > [INFO] flink-hbase ........................................
> > SUCCESS [
> > > > > 2.870 s]
> > > > > > [INFO] flink-gelly ........................................
> > FAILURE [
> > > > > 58.548 s]
> > > > > > [INFO] flink-hcatalog .....................................
> SKIPPED
> > > > > > [INFO] flink-expressions ..................................
> SKIPPED
> > > > > > [INFO] flink-quickstart ...................................
> SKIPPED
> > > > > > [INFO] flink-quickstart-java ..............................
> SKIPPED
> > > > > > [INFO] flink-quickstart-scala .............................
> SKIPPED
> > > > > > [INFO] flink-contrib ......................................
> SKIPPED
> > > > > > [INFO] flink-dist .........................................
> SKIPPED
> > > > > > [INFO]
> > > > >
> > >
> ------------------------------------------------------------------------
> > > > > > [INFO] BUILD FAILURE
> > > > > > [INFO]
> > > > >
> > >
> ------------------------------------------------------------------------
> > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
>