私は以下を使用してファイルからバイトを読み取っています:
FileSystem fs = config.getHDFS();
try {
Path path = new Path(dirName + '/' + fileName);
byte[] bytes = new byte[(int)fs.getFileStatus(path)
.getLen()];
in = fs.open(path);
in.read(bytes);
result = new DataInputStream(new ByteArrayInputStream(bytes));
} catch (Exception e) {
e.printStackTrace();
if (in != null) {
try {
in.close();
} catch (IOException e1) {
e1.printStackTrace();
}
}
}
私が読んでいるディレクトリには約15,000個のファイルがあります。ある時点の後、in.read(bytes)の行でこの例外が発生します:
2012-05-31 14:11:45,477 [INFO:main] (DFSInputStream.java:414) - Failed to connect to /165.36.80.28:50010, add to deadNodes and continue
java.io.EOFException
at java.io.DataInputStream.readShort(DataInputStream.java:298)
at org.apache.hadoop.hdfs.protocol.DataTransferProtocol$Status.read(DataTransferProtocol.java:115)
at org.apache.hadoop.hdfs.BlockReader.newBlockReader(BlockReader.java:427)
at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:725)
at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:390)
at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:514)
at java.io.DataInputStream.read(DataInputStream.java:83)
スローされる別の例外は次のとおりです。
2012-05-31 15:09:14,849 [INFO:main] (DFSInputStream.java:414) - Failed to connect to /165.36.80.28:50010, add to deadNodes and continue
java.net.SocketException: No buffer space available (maximum connections reached?): connect
at sun.nio.ch.Net.connect(Native Method)
at sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:507)
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:192)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:373)
at org.apache.hadoop.hdfs.DFSInputStream.getBlockReader(DFSInputStream.java:719)
at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:390)
at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:514)
at java.io.DataInputStream.read(DataInputStream.java:83)
何が問題になるのかアドバイスしてください。