我正在尝试使用Spark阅读镶木地板文件 . Spark版本是1.3.0 . 当我尝试阅读时,我得到以下异常 . 任何帮助都非常感谢 . CDH版本:5.4.3

在parquet.format.converter.ParquetMetadataConverter.readParquetMetadata的parquet.format.converter.ParquetMetadataConverter.fromParquetMetadata(ParquetMetadataConverter.java:543)的parquet.format.converter.ParquetMetadataConverter.fromParquetStatistics(ParquetMetadataConverter.java:249)中的java.lang.NullPointerException (ParquetMetadataConverter.java:520)在parquet.hadoop.ParquetFileReader.readFooter(ParquetFileReader.java:426)的parquet.hadoop.ParquetFileReader.readFooter(ParquetFileReader.java:381)在parquet.hadoop.ParquetRecordReader.initializeInternalReader(ParquetRecordReader.java: 155)在org.apache.spark.sql的org.apache.spark.sql.sources.SqlNewHadoopRDD $$ anon $ 1 . (SqlNewHadoopRDD.scala:153)的parquet.hadoop.ParquetRecordReader.initialize(ParquetRecordReader.java:138) . sources.qlNewHadoopRDD.compute(SqlNewHadoopRDD.scala:124)atg.apache.spark.sql.sources.SqlNewHadoopRDD.compute(SqlNewHadoopRDD.scala:66)at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala: 277)在org.apache.spark.rdd.RDD.iterator(RDD.scala:244)org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)org.apache.spark.rdd.RDD.computeOrReadCheckpoint( RDD.scala:277)org.apache.spark.rdd.RDD.iterator(RDD.scala:244)org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)atg.apache.spark .rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)org.apache.spark.rdd.RDD.iterator(RDD.scala:244)org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:70) )org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)org.apache.spark.scheduler.Task.run(Task.scala:70)at org.apache.spark.executor.Executor $ TaskRunner.run(Executor.scala:213)位于java.util.conte当前的Java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)java.util.ThreadPoolExecutor $ Worker.run(ThreadPoolExecutor.java:615) .Thread.run(Thread.java:745)