Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via spark-issues by Sital Kedia (JIRA), 1 year ago
via Apache's JIRA Issue Tracker by Nathan McCarthy, 1 year ago
Comparison method violates its general contract!
via Apache's JIRA Issue Tracker by Yin Huai, 1 year ago
Comparison method violates its general contract!
java.lang.IllegalArgumentException: Comparison method violates its general contract!	at org.apache.spark.util.collection.TimSort$SortState.mergeLo(TimSort.java:794)	at org.apache.spark.util.collection.TimSort$SortState.mergeAt(TimSort.java:525)	at org.apache.spark.util.collection.TimSort$SortState.mergeCollapse(TimSort.java:453)	at org.apache.spark.util.collection.TimSort$SortState.access$200(TimSort.java:325)	at org.apache.spark.util.collection.TimSort.sort(TimSort.java:153)	at org.apache.spark.util.collection.Sorter.sort(Sorter.scala:37)	at org.apache.spark.util.collection.unsafe.sort.UnsafeInMemorySorter.getSortedIterator(UnsafeInMemorySorter.java:228)	at org.apache.spark.util.collection.unsafe.sort.UnsafeExternalSorter.spill(UnsafeExternalSorter.java:186)	at org.apache.spark.memory.TaskMemoryManager.acquireExecutionMemory(TaskMemoryManager.java:175)	at org.apache.spark.memory.TaskMemoryManager.allocatePage(TaskMemoryManager.java:249)	at org.apache.spark.memory.MemoryConsumer.allocatePage(MemoryConsumer.java:112)	at org.apache.spark.util.collection.unsafe.sort.UnsafeExternalSorter.acquireNewPageIfNecessary(UnsafeExternalSorter.java:318)	at org.apache.spark.util.collection.unsafe.sort.UnsafeExternalSorter.insertRecord(UnsafeExternalSorter.java:333)	at org.apache.spark.sql.execution.UnsafeExternalRowSorter.insertRow(UnsafeExternalRowSorter.java:91)	at org.apache.spark.sql.execution.UnsafeExternalRowSorter.sort(UnsafeExternalRowSorter.java:168)	at org.apache.spark.sql.execution.Sort$$anonfun$1.apply(Sort.scala:90)	at org.apache.spark.sql.execution.Sort$$anonfun$1.apply(Sort.scala:64)	at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$21.apply(RDD.scala:728)	at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$21.apply(RDD.scala:728)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)	at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)	at org.apache.spark.scheduler.Task.run(Task.scala:89)	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)	at java.lang.Thread.run(Thread.java:745)