Searched on Google with the first line of a JAVA stack trace?

We can recommend more relevant solutions and speed up debugging when you paste your entire stack trace with the exception message. Try a sample exception.

Recommended solutions based on your search

Solutions on the web

via Spring JIRA by Sina Sojoodi, 1 year ago
20000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/192.168.109.61:50010 remote=/192.168.109.11:46349]
via GitHub by kings129
, 5 months ago
20000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/*********** remote=********]
via hadoop-common-user by Tenghuan He, 1 year ago
11000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/192.168.179.1:53765 remote=/192.168.179.135:50010]
via hadoop-hdfs-user by Chris Nauroth, 1 year ago
11000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/192.168.179.1:53765<http://192.168.179.1:53765> remote=/192.168.179.135:50010<http://192.168.179.135:50010>]
via hadoop-mapreduce-user by Chris Nauroth, 1 year ago
11000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/192.168.179.1:53765<http://192.168.179.1:53765> remote=/192.168.179.135:50010<http://192.168.179.135:50010>]
via hadoop-common-user by Chris Nauroth, 1 year ago
11000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/192.168.179.1:53765<http://192.168.179.1:53765> remote=/192.168.179.135:50010<http://192.168.179.135:50010>]
java.net.SocketTimeoutException: 20000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/192.168.109.61:50010 remote=/192.168.109.11:46349]	at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:164)	at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:159)	at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:129)	at java.io.FilterInputStream.read(FilterInputStream.java:116)	at java.io.BufferedInputStream.fill(BufferedInputStream.java:218)	at java.io.BufferedInputStream.read1(BufferedInputStream.java:258)	at java.io.BufferedInputStream.read(BufferedInputStream.java:317)	at java.io.DataInputStream.read(DataInputStream.java:132)	at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:192)	at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213)	at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)	at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)	at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:424)	at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:656)	at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:512)	at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:98)	at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:65)	at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:219)	at java.lang.Thread.run(Thread.java:662)