在intellij使用spark scala,我正在阅读来自couchbase bucket的500万条记录 . 对于一些记录,它工作正常,但它显示了500万

failed with {"msg":"Error performing bulk get operation - cause: {3 errors, 
starting with read tcp 127.0.0.1:54295->127.0.0.1:11210: i/o timeout}","code":12008}

为了这

N1qlQuery.simple(s"select * from data")

   //and if i use 

N1qlQuery.simple(s"select * from data where date >= somevalue1 and date <= somevalue2")

它会引发以下错误

ERROR executor.Executor:阶段3.0(TID 3)中任务0.0的异常java.lang.RuntimeException:rx.internal中rx.exceptions.Exceptions.propagate(Exceptions.java:58)的java.util.concurrent.TimeoutException . operators.BlockingOperatorToIterator $ SubscriberIterator.hasNext(BlockingOperatorToIterator.java:100)在scala.collection.convert.Wrappers $ JIteratorWrapper.hasNext(Wrappers.scala:42)在com.couchbase.spark.internal.LazyIterator $$匿名$ 1.hasNext( LazyIterator.scala:27)scala.collection.Iterator $$ anon $ 11.hasNext(Iterator.scala:408)at org.apache.spark scala.collection.Iterator $$ anon $ 11.hasNext(Iterator.scala:408)at org.apache.spark .sql.catalyst.expressions.GeneratedClass $ GeneratedIterator.processNext(来源不明)在org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)在org.apache.spark.sql.execution.WholeStageCodegenExec $ $ anonfun $ 8 $$ anon $ 1.hasNext(WholeStageCodegenExec.scala:395)at scala.collection.Iterator $$ anon $ 11.hasNext(Iterator.scala:408)at scala.colle ction.Iterator $$ anon $ 12.hasNext(Iterator.scala:439)at scala.collection.Iterator $ class.foreach(Iterator.scala:893)at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)at scala .collection.TraversableOnce $ class.foldLeft(TraversableOnce.scala:157)在scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)在scala.collection.TraversableOnce $ class.fold(TraversableOnce.scala:212)在阶 . collection.AbstractIterator.fold(Iterator.scala:1336)atg.apache.spark.rdd.RDD $$ anonfun $ fold $ 1 $$ anonfun $ 20.apply(RDD.scala:1087)at org.apache.spark.rdd . RDD $$ anonfun $ fold $ 1 $$ anonfun $ 20.apply(RDD.scala:1087)org.apache.spark.SparkContext $$ anonfun $ 33.apply(SparkContext.scala:2125)at org.apache.spark.SparkContext $ $ anonfun $ 33.apply(SparkContext.scala:2125)org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)at org.apache.spark.scheduler.Task.run(Task.scala:108) at org.apache.spark.executor.Executor $ TaskRunner.run(Executor.scala:338)at java.util.concurrent.ThreadP oolExecutor.runWorker(ThreadPoolExecutor.java:1149)java.util.concurrent.ThreadPoolExecutor $ Worker.run(ThreadPoolExecutor.java:624)at java.lang.Thread.run(Thread.java:748)引起:java.util .concurrent.TimeoutException在rx.internal.operators.OperatorTimeoutBase $ TimeoutSubscriber.onTimeout(OperatorTimeoutBase.java:177)在rx.internal.operators.OperatorTimeout $ 1 $ 1.call(OperatorTimeout.java:41)在rx.internal.schedulers.EventLoopsScheduler $ EventLoopWorker $ 2.call(EventLoopsScheduler.java:189)at java.util.concurrent.Executors上的rx.internal.schedulers.ScheduledAction.run(ScheduledAction.java:55)java上的$ RunnableAdapter.call(Executors.java:511) .util.concurrent.FutureTask.run(FutureTask.java:266)在java.util.concurrent.ScheduledThreadPoolExecutor中$ $ ScheduledFutureTask.access 201(ScheduledThreadPoolExecutor.java:180)在java.util.concurrent.ScheduledThreadPoolExecutor中$ ScheduledFutureTask.run(的ScheduledThreadPoolExecutor . java:293)... 3更多18/12/03 13:12:39错误scheduler.TaskSetManag呃:阶段3.0中的任务0失败了1次;中止作业线程“main”中的异常org.apache.spark.SparkException:作业因阶段失败而中止:阶段3.0中的任务0失败1次,最近失败:阶段3.0中丢失任务0.0(TID 3,localhost, Actuator 驱动程序) ):了java.lang.RuntimeException:java.util.concurrent.TimeoutException在rx.exceptions.Exceptions.propagate(Exceptions.java:58)在rx.internal.operators.BlockingOperatorToIterator $ SubscriberIterator.hasNext(BlockingOperatorToIterator.java:100)在scala.collection.convert.Wrappers $ JIteratorWrapper.hasNext(Wrappers.scala:42)at com.couchbase.spark.internal.LazyIterator $$ anon $ 1.hasNext(LazyIterator.scala:27)at scala.collection.Iterator $$ anon $ 11.hasNext(Iterator.scala:408)at scala.collection.Iterator $$ anon $ 11.hasNext(Iterator.scala:408)at org.apache.spark.sql.catalyst.expressions.GeneratedClass $ GeneratedIterator.processNext(Unknown Source )org.apache.spark中的org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)org.apache.spark.sql.execution.Whol eStageCodegenExec $$ anonfun $ 8 $$ anon $ 1.hasNext(WholeStageCodegenExec.scala:395)scala.collection.Iterator $$ anon $ 11.hasNext(Iterator.scala:408)at scala.collection.Iterator $$ anon $ 12.hasNext( Iterator.scala:439)scala.collection.Iterator $ class.foreach(Iterator.scala:893)at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)at scala.collection.TraversableOnce $ class.foldLeft(TraversableOnce .scala:157)scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)at scala.collection.TraversableOnce $ class.fold(TraversableOnce.scala:212)at scala.collection.AbstractIterator.fold(Iterator.scala: 1336)在org.apache.spark.rdd.RDD $$ anonfun $ fold $ 1 $$ anonfun $ 20.apply(RDD.scala:1087)at org.apache.spark.rdd.RDD $$ anonfun $ fold $ 1 $$ anonfun $ 20.apply (RDD.scala:1087)org.apache.spark.SparkContext $$ anonfun $ 33.apply(SparkContext.scala:2125)at org.apache.spark.SparkContext $$ anonfun $ 33.apply(SparkContext.scala:2125)at at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)atg.apache.spark.scheduler.Task.run(Task.scala:108)org.apache.spark.executor.Executor $ TaskRunner . 运行(Executor.scala:338)在java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)在java.util.concurrent.ThreadPoolExecutor中的$ Worker.run(ThreadPoolExecutor.java:624)在java.lang.Thread中.RUN(Thread.java:748)产生的原因:在rx.internal.operators.OperatorTimeout $ 1 $ 1.call在rx.internal.operators.OperatorTimeoutBase $ TimeoutSubscriber.onTimeout(OperatorTimeoutBase.java:177)java.util.concurrent.TimeoutException (OperatorTimeout.java:41)at rx.internal.schedulers.EventLoopsScheduler $ EventLoopWorker $ 2 .CALL(EventLoopsScheduler.java:189)在rx.internal.schedulers.ScheduledAction.run(ScheduledAction.java:55)在java.util.concurrent.Executors $ RunnableAdapter.call(Executors.java:511)在java.util中 . concurrent.FutureTask.run(FutureTask.java:266)在java.util.concurrent.ScheduledThreadPoolExecutor中$ $ ScheduledFutureTask.access 201(ScheduledThreadPoolExecutor.java:180)在java.util.concurrent.ScheduledThreadPoolExecutor中$ ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293 )