Background: Production environment There is a script to read the ST Layer table data when an IO error, view the table directory of files, are compressed files. Detailed information is as follows:
Task with the most failures (4):
-----
Task ID:
task_201408301703_172845_m_003505
Url:
http://master:50030/taskdetails.jsp?jobid=job_201408301703_172845&tipid=task_201408301703_172845_m_003505
-----
Diagnostic Messages for this Task:
Java.io.IOException:IO error in Map input file hdfs://master:9000/user/hive/warehouse/pc.db/dwd_st_pc_list/dt= Startup/startup-m-03653.gz
At Org.apache.hadoop.mapred.maptask$trackedrecordreader.movetonext (maptask.java:242)
At Org.apache.hadoop.mapred.maptask$trackedrecordreader.next (maptask.java:216)
At Org.apache.hadoop.mapred.MapRunner.run (maprunner.java:48)
At Org.apache.hadoop.mapred.MapTask.runOldMapper (maptask.java:436)
At Org.apache.hadoop.mapred.MapTask.run (maptask.java:372)
At Org.apache.hadoop.mapred.child$4.run (child.java:255)
At java.security.AccessController.doPrivileged (Native Method)
At Javax.security.auth.Subject.doAs (subject.java:415)
At Org.apache.hadoop.security.UserGroupInformation.doAs (usergroupinformation.java:1149)
At Org.apache.hadoop.mapred.Child.main (child.java:249)
caused by:java.io.IOException:java.io.EOFException:Unexpected end of input stream
At Org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain.handleRecordReaderNextException ( hiveioexceptionhandlerchain.java:121)
At Org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil.handleRecordReaderNextException ( hiveioexceptionhandlerutil.java:77)
At Org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext (hivecontextawarerecordreader.java:276)
At Org.apache.hadoop.hive.ql.io.HiveRecordReader.doNext (hiverecordreader.java:79)
At Org.apache.hadoop.hive.ql.io.HiveRecordReader.doNext (hiverecordreader.java:33)
At Org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.next (hivecontextawarerecordreader.java:108)
At Org.apache.hadoop.mapred.maptask$trackedrecordreader.movetonext (maptask.java:236)
... 9 more
caused by:java.io.EOFException:Unexpected end of input stream
At Org.apache.hadoop.io.compress.DecompressorStream.decompress (decompressorstream.java:137)
At Org.apache.hadoop.io.compress.DecompressorStream.read (decompressorstream.java:77)
At Java.io.InputStream.read (inputstream.java:101)
At Org.apache.hadoop.util.LineReader.readLine (linereader.java:134)
At Org.apache.hadoop.mapred.LineRecordReader.next (linerecordreader.java:176)
At Org.apache.hadoop.mapred.LineRecordReader.next (linerecordreader.java:43)
At Org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext (hivecontextawarerecordreader.java:274)
... More
Failed:execution Error, return code 2 from Org.apache.hadoop.hive.ql.exec.MapRedTask
MapReduce Jobs Launched:
Job 0:map:12137 reduce:100 Cumulative cpu:23227.51 sec HDFs read:2121148099 HDFs write:0 FAIL
Total MapReduce CPU time spent:0 days 6 hours minutes 7 seconds 510 msec
Check the corresponding job task log, found some files decompression error, find the corresponding file, download to local decompression, or error.
Taking into account the backup of Hadoop several, so the wrong file deleted, delete the script above the retry.
Hadoop platform Read File error