java.lang.AssertionError: assertion failed: SpreadRDD failure - IPs are not equal: (1,169.254.183.61,-1,169.254.183.61) != (0, 169.254.3.199)

GitHub | tranlm | 5 months ago
  1. 0

    "IPs are not equal" error when starting H2OContext with Spark Context

    GitHub | 5 months ago | tranlm
    java.lang.AssertionError: assertion failed: SpreadRDD failure - IPs are not equal: (1,169.254.183.61,-1,169.254.183.61) != (0, 169.254.3.199)
  2. 0

    Heres the error message after doing a new H2OContext(sc).start() 16/03/24 15:20:08 WARN TaskSetManager: Lost task 2.0 in stage 1.0 (TID 43, mr-0xc7.0xdata.loc): java.lang.AssertionError: assertion failed: SpreadRDD failure - IPs are not equal: (2,172.16.2.148,-1,172.16.2.148) != (1, 172.16.2.147) at scala.Predef$.assert(Predef.scala:179) at org.apache.spark.h2o.H2OContextUtils$$anonfun$7.apply(H2OContextUtils.scala:107) at org.apache.spark.h2o.H2OContextUtils$$anonfun$7.apply(H2OContextUtils.scala:106) at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) at scala.collection.Iterator$class.foreach(Iterator.scala:727) at scala.collection.AbstractIterator.foreach(Iterator.scala:1157) at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48) at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103) at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47) at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273) at scala.collection.AbstractIterator.to(Iterator.scala:1157) at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265) at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157) at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252) at scala.collection.AbstractIterator.toArray(Iterator.scala:1157) at org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.apply(RDD.scala:927) at org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.apply(RDD.scala:927) at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858) at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:89) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)

    JIRA | 8 months ago | Mark Chan
    java.lang.AssertionError: assertion failed: SpreadRDD failure - IPs are not equal: (2,172.16.2.148,-1,172.16.2.148) != (1, 172.16.2.147)
  3. 0

    Heres the error message after doing a new H2OContext(sc).start() 16/03/24 15:20:08 WARN TaskSetManager: Lost task 2.0 in stage 1.0 (TID 43, mr-0xc7.0xdata.loc): java.lang.AssertionError: assertion failed: SpreadRDD failure - IPs are not equal: (2,172.16.2.148,-1,172.16.2.148) != (1, 172.16.2.147) at scala.Predef$.assert(Predef.scala:179) at org.apache.spark.h2o.H2OContextUtils$$anonfun$7.apply(H2OContextUtils.scala:107) at org.apache.spark.h2o.H2OContextUtils$$anonfun$7.apply(H2OContextUtils.scala:106) at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) at scala.collection.Iterator$class.foreach(Iterator.scala:727) at scala.collection.AbstractIterator.foreach(Iterator.scala:1157) at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48) at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103) at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47) at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273) at scala.collection.AbstractIterator.to(Iterator.scala:1157) at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265) at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157) at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252) at scala.collection.AbstractIterator.toArray(Iterator.scala:1157) at org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.apply(RDD.scala:927) at org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.apply(RDD.scala:927) at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858) at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:89) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)

    JIRA | 8 months ago | Mark Chan
    java.lang.AssertionError: assertion failed: SpreadRDD failure - IPs are not equal: (2,172.16.2.148,-1,172.16.2.148) != (1, 172.16.2.147)
  4. Speed up your debug routine!

    Automated exception search integrated into your IDE

  5. 0

    Cannot use coref

    GitHub | 3 years ago | schmmd
    java.lang.AssertionError: assertion failed: Could not find annotator for goal class cc.factorie.app.nlp.phrase.NounPhraseList, map includes class cc.factorie.app.nlp.phrase.NumberLabel, class cc.factorie.app.nlp.ner.NerTag, class cc.factorie.app.nlp.lemma.SimplifyDigitsTokenLemma, class cc.factorie.app.nlp.coref.mention.ParseBasedMentionList, class cc.factorie.app.nlp.pos.PennPosTag, class cc.factorie.app.nlp.segment.PlainNormalizedTokenString, class cc.factorie.app.nlp.Token, class cc.factorie.app.nlp.ner.BilouConllNerTag, class cc.factorie.app.nlp.coref.mention.MentionEntityType, class cc.factorie.util.coref.GenericEntityMap, class cc.factorie.app.nlp.lemma.CollapseDigitsTokenLemma, class cc.factorie.app.nlp.ner.BilouOntonotesNerTag, class cc.factorie.app.nlp.phrase.GenderLabel, class cc.factorie.app.nlp.lemma.WordNetTokenLemma, class cc.factorie.app.nlp.parse.ParseTree, class cc.factorie.app.nlp.Sentence, class cc.factorie.app.nlp.lemma.PorterTokenLemma, class cc.factorie.app.nlp.coref.mention.NerMentionList, class cc.factorie.app.nlp.lemma.LowercaseTokenLemma
  6. 0

    Error when running profile with Cypher query

    GitHub | 3 years ago | peterneubauer
    java.lang.AssertionError: assertion failed: Can't profile the same pipe twice: NullPipe(SymbolTable(Map(a -> Node, b -> Node)),Eager() Filter(pred="Property(b,name(0)) == Literal(Central)") NodeByLabel(label="Division", identifier="b") Filter(pred="Property(a,name(0)) == Literal(East)") NodeByLabel(label="Conference", identifier="a"))

  1. tyson925 1 times, last 20 hours ago
  2. rp 1 times, last 4 days ago
  3. poroszd 1 times, last 4 months ago
  4. rp 1 times, last 9 months ago
17 unregistered visitors
Not finding the right solution?
Take a tour to get the most out of Samebug.

Tired of useless tips?

Automated exception search integrated into your IDE

Root Cause Analysis

  1. java.lang.AssertionError

    assertion failed: SpreadRDD failure - IPs are not equal: (1,169.254.183.61,-1,169.254.183.61) != (0, 169.254.3.199)

    at scala.Predef$.assert()
  2. Scala
    Predef$.assert
    1. scala.Predef$.assert(Predef.scala:179)
    1 frame
  3. org.apache.spark
    H2OContextUtils$$anonfun$7.apply
    1. org.apache.spark.h2o.H2OContextUtils$$anonfun$7.apply(H2OContextUtils.scala:107)
    2. org.apache.spark.h2o.H2OContextUtils$$anonfun$7.apply(H2OContextUtils.scala:106)
    2 frames
  4. Scala
    AbstractIterator.toArray
    1. scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
    2. scala.collection.Iterator$class.foreach(Iterator.scala:727)
    3. scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
    4. scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
    5. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
    6. scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
    7. scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
    8. scala.collection.AbstractIterator.to(Iterator.scala:1157)
    9. scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
    10. scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
    11. scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
    12. scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
    12 frames
  5. Spark
    Executor$TaskRunner.run
    1. org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.apply(RDD.scala:927)
    2. org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.apply(RDD.scala:927)
    3. org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858)
    4. org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858)
    5. org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
    6. org.apache.spark.scheduler.Task.run(Task.scala:89)
    7. org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
    7 frames
  6. Java RT
    Thread.run
    1. java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    2. java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    3. java.lang.Thread.run(Thread.java:745)
    3 frames