Last active
March 22, 2016 04:57
-
-
Save ondine/56186677a0414d8bc98b to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| FROM THE COMMAND LINE: | |
| HOME% jupyter notebook | |
| [I 21:42:37.461 NotebookApp] Serving notebooks from local directory: /Users/edlc4sz/workspace/apache-toree-quickstart | |
| [I 21:42:37.461 NotebookApp] 0 active kernels | |
| [I 21:42:37.461 NotebookApp] The Jupyter Notebook is running at: http://localhost:8888/ | |
| [I 21:42:37.462 NotebookApp] Use Control-C to stop this server and shut down all kernels (twice to skip confirmation). | |
| [I 21:42:57.255 NotebookApp] Creating new notebook in | |
| [I 21:42:57.816 NotebookApp] Kernel started: 89665ca7-cfe4-4f20-be77-62609c5d0fb6 | |
| Starting Spark Kernel with SPARK_HOME=/Applications/spark-1.6.1-bin-hadoop2.6 | |
| 16/03/21 21:43:01 [WARN] o.a.t.Main$$anon$1 - No external magics provided to MagicLoader! | |
| 16/03/21 21:43:02 [WARN] o.a.h.u.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable | |
| [W 21:43:07.910 NotebookApp] Timeout waiting for kernel_info reply from 89665ca7-cfe4-4f20-be77-62609c5d0fb6 | |
| 16/03/21 21:43:07 [WARN] o.a.t.k.p.v.k.s.Shell - Parent header is null for message ED6D770C827244258A0F500845941808 of type comm_open | |
| 16/03/21 21:43:08 [WARN] o.a.t.k.a.Kernel - Locked to Scala interpreter with SparkIMain until decoupled! | |
| 16/03/21 21:43:08 [WARN] o.a.t.k.a.Kernel - Unable to control initialization of REPL class server! | |
| 16/03/21 21:43:08 [WARN] o.a.t.k.p.v.s.KernelOutputStream - Suppressing empty output: '' | |
| 16/03/21 21:43:08 [INFO] Remoting - Starting remoting | |
| 16/03/21 21:43:09 [INFO] Remoting - Remoting started; listening on addresses :[akka.tcp://[email protected]:58018] | |
| 16/03/21 21:43:11 [INFO] D.Persistence - Property hive.metastore.integral.jdo.pushdown unknown - will be ignored | |
| 16/03/21 21:43:11 [INFO] D.Persistence - Property datanucleus.cache.level2 unknown - will be ignored | |
| 16/03/21 21:43:14 [INFO] D.Datastore - The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table. | |
| 16/03/21 21:43:14 [INFO] D.Datastore - The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table. | |
| 16/03/21 21:43:15 [INFO] D.Datastore - The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table. | |
| 16/03/21 21:43:15 [INFO] D.Datastore - The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table. | |
| 16/03/21 21:43:15 [WARN] o.a.h.h.m.ObjectStore - Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 1.2.0 | |
| 16/03/21 21:43:15 [WARN] o.a.h.h.m.ObjectStore - Failed to get database default, returning NoSuchObjectException | |
| 16/03/21 21:43:16 [WARN] - Your hostname, OXC02PN690FVH7 resolves to a loopback/non-reachable address: fe80:0:0:0:7877:6ff:fedd:3e7d%awdl0, but we couldn't find any external IP address! | |
| 16/03/21 21:43:22 [INFO] D.Datastore - The class "org.apache.hadoop.hive.metastore.model.MResourceUri" is tagged as "embedded-only" so does not have its own datastore table. | |
| 16/03/21 21:43:24 [INFO] D.Persistence - Property hive.metastore.integral.jdo.pushdown unknown - will be ignored | |
| 16/03/21 21:43:24 [INFO] D.Persistence - Property datanucleus.cache.level2 unknown - will be ignored | |
| 16/03/21 21:43:27 [INFO] D.Datastore - The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table. | |
| 16/03/21 21:43:27 [INFO] D.Datastore - The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table. | |
| 16/03/21 21:43:28 [INFO] D.Datastore - The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table. | |
| 16/03/21 21:43:28 [INFO] D.Datastore - The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table. | |
| 16/03/21 21:43:28 [WARN] o.a.h.h.m.ObjectStore - Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 1.2.0 | |
| 16/03/21 21:43:28 [WARN] o.a.h.h.m.ObjectStore - Failed to get database default, returning NoSuchObjectException | |
| 16/03/21 21:43:28 [INFO] D.Datastore - The class "org.apache.hadoop.hive.metastore.model.MResourceUri" is tagged as "embedded-only" so does not have its own datastore table. | |
| 16/03/21 21:43:32 [WARN] o.a.t.k.p.v.h.CommOpenHandler - Received invalid target for Comm Open: ipython.widget | |
| [I 21:44:57.774 NotebookApp] Saving file at /OK_TestNotebook.ipynb | |
| 16/03/21 21:45:31 [ERROR] o.a.s.e.Executor - Exception in task 0.0 in stage 0.0 (TID 0) | |
| java.net.SocketTimeoutException: connect timed out | |
| at java.net.PlainSocketImpl.socketConnect(Native Method) | |
| at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350) | |
| at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206) | |
| at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188) | |
| at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) | |
| at java.net.Socket.connect(Socket.java:589) | |
| at sun.net.NetworkClient.doConnect(NetworkClient.java:175) | |
| at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) | |
| at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) | |
| at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) | |
| at sun.net.www.http.HttpClient.New(HttpClient.java:308) | |
| at sun.net.www.http.HttpClient.New(HttpClient.java:326) | |
| at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1169) | |
| at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1105) | |
| at sun.net.www.protocol.http.HttpURLConnection$6.run(HttpURLConnection.java:989) | |
| at sun.net.www.protocol.http.HttpURLConnection$6.run(HttpURLConnection.java:987) | |
| at java.security.AccessController.doPrivileged(Native Method) | |
| at java.security.AccessController.doPrivilegedWithCombiner(AccessController.java:782) | |
| at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:986) | |
| at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:933) | |
| at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:588) | |
| at org.apache.spark.util.Utils$.fetchFile(Utils.scala:394) | |
| at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:405) | |
| at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:397) | |
| at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) | |
| at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) | |
| at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) | |
| at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) | |
| at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) | |
| at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) | |
| at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) | |
| at org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:397) | |
| at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:193) | |
| at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
| at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
| at java.lang.Thread.run(Thread.java:745) | |
| 16/03/21 21:45:31 [WARN] o.a.s.s.TaskSetManager - Lost task 0.0 in stage 0.0 (TID 0, localhost): java.net.SocketTimeoutException: connect timed out | |
| at java.net.PlainSocketImpl.socketConnect(Native Method) | |
| at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350) | |
| at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206) | |
| at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188) | |
| at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) | |
| at java.net.Socket.connect(Socket.java:589) | |
| at sun.net.NetworkClient.doConnect(NetworkClient.java:175) | |
| at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) | |
| at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) | |
| at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) | |
| at sun.net.www.http.HttpClient.New(HttpClient.java:308) | |
| at sun.net.www.http.HttpClient.New(HttpClient.java:326) | |
| at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1169) | |
| at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1105) | |
| at sun.net.www.protocol.http.HttpURLConnection$6.run(HttpURLConnection.java:989) | |
| at sun.net.www.protocol.http.HttpURLConnection$6.run(HttpURLConnection.java:987) | |
| at java.security.AccessController.doPrivileged(Native Method) | |
| at java.security.AccessController.doPrivilegedWithCombiner(AccessController.java:782) | |
| at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:986) | |
| at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:933) | |
| at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:588) | |
| at org.apache.spark.util.Utils$.fetchFile(Utils.scala:394) | |
| at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:405) | |
| at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:397) | |
| at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) | |
| at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) | |
| at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) | |
| at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) | |
| at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) | |
| at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) | |
| at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) | |
| at org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:397) | |
| at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:193) | |
| at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
| at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
| at java.lang.Thread.run(Thread.java:745) | |
| 16/03/21 21:45:31 [ERROR] o.a.s.s.TaskSetManager - Task 0 in stage 0.0 failed 1 times; aborting job | |
| org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): java.net.SocketTimeoutException: connect timed out | |
| at java.net.PlainSocketImpl.socketConnect(Native Method) | |
| at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350) | |
| at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206) | |
| at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188) | |
| at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) | |
| at java.net.Socket.connect(Socket.java:589) | |
| at sun.net.NetworkClient.doConnect(NetworkClient.java:175) | |
| at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) | |
| at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) | |
| at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) | |
| at sun.net.www.http.HttpClient.New(HttpClient.java:308) | |
| at sun.net.www.http.HttpClient.New(HttpClient.java:326) | |
| at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1169) | |
| at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1105) | |
| at sun.net.www.protocol.http.HttpURLConnection$6.run(HttpURLConnection.java:989) | |
| at sun.net.www.protocol.http.HttpURLConnection$6.run(HttpURLConnection.java:987) | |
| at java.security.AccessController.doPrivileged(Native Method) | |
| at java.security.AccessController.doPrivilegedWithCombiner(AccessController.java:782) | |
| at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:986) | |
| at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:933) | |
| at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:588) | |
| at org.apache.spark.util.Utils$.fetchFile(Utils.scala:394) | |
| at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:405) | |
| at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:397) | |
| at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) | |
| at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) | |
| at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) | |
| at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) | |
| at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) | |
| at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) | |
| at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) | |
| at org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:397) | |
| at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:193) | |
| at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
| at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
| at java.lang.Thread.run(Thread.java:745) | |
| Driver stacktrace: | |
| at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431) | |
| at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419) | |
| at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418) | |
| at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
| at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) | |
| at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418) | |
| at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) | |
| at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) | |
| at scala.Option.foreach(Option.scala:236) | |
| at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799) | |
| at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640) | |
| at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599) | |
| at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588) | |
| at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) | |
| at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620) | |
| at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832) | |
| at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845) | |
| at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858) | |
| at org.apache.spark.rdd.RDD$$anonfun$take$1.apply(RDD.scala:1328) | |
| at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) | |
| at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) | |
| at org.apache.spark.rdd.RDD.withScope(RDD.scala:316) | |
| at org.apache.spark.rdd.RDD.take(RDD.scala:1302) | |
| at $iwC$$iwC$$iwC$$iwC.<init>(<console>:24) | |
| at $iwC$$iwC$$iwC.<init>(<console>:29) | |
| at $iwC$$iwC.<init>(<console>:31) | |
| at $iwC.<init>(<console>:33) | |
| at <init>(<console>:35) | |
| at .<init>(<console>:39) | |
| at .<clinit>(<console>) | |
| at .<init>(<console>:7) | |
| at .<clinit>(<console>) | |
| at $print(<console>) | |
| at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
| at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
| at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
| at java.lang.reflect.Method.invoke(Method.java:497) | |
| at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065) | |
| at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346) | |
| at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840) | |
| at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871) | |
| at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819) | |
| at org.apache.toree.kernel.interpreter.scala.ScalaInterpreter$$anonfun$interpretAddTask$1$$anonfun$apply$3.apply(ScalaInterpreter.scala:356) | |
| at org.apache.toree.kernel.interpreter.scala.ScalaInterpreter$$anonfun$interpretAddTask$1$$anonfun$apply$3.apply(ScalaInterpreter.scala:351) | |
| at org.apache.toree.global.StreamState$.withStreams(StreamState.scala:81) | |
| at org.apache.toree.kernel.interpreter.scala.ScalaInterpreter$$anonfun$interpretAddTask$1.apply(ScalaInterpreter.scala:350) | |
| at org.apache.toree.kernel.interpreter.scala.ScalaInterpreter$$anonfun$interpretAddTask$1.apply(ScalaInterpreter.scala:350) | |
| at org.apache.toree.utils.TaskManager$$anonfun$add$2$$anon$1.run(TaskManager.scala:140) | |
| at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
| at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
| at java.lang.Thread.run(Thread.java:745) | |
| Caused by: java.net.SocketTimeoutException: connect timed out | |
| at java.net.PlainSocketImpl.socketConnect(Native Method) | |
| at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350) | |
| at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206) | |
| at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188) | |
| at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) | |
| at java.net.Socket.connect(Socket.java:589) | |
| at sun.net.NetworkClient.doConnect(NetworkClient.java:175) | |
| at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) | |
| at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) | |
| at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) | |
| at sun.net.www.http.HttpClient.New(HttpClient.java:308) | |
| at sun.net.www.http.HttpClient.New(HttpClient.java:326) | |
| at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1169) | |
| at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1105) | |
| at sun.net.www.protocol.http.HttpURLConnection$6.run(HttpURLConnection.java:989) | |
| at sun.net.www.protocol.http.HttpURLConnection$6.run(HttpURLConnection.java:987) | |
| at java.security.AccessController.doPrivileged(Native Method) | |
| at java.security.AccessController.doPrivilegedWithCombiner(AccessController.java:782) | |
| at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:986) | |
| at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:933) | |
| at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:588) | |
| at org.apache.spark.util.Utils$.fetchFile(Utils.scala:394) | |
| at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:405) | |
| at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:397) | |
| at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) | |
| at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) | |
| at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) | |
| at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) | |
| at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) | |
| at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) | |
| at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) | |
| at org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:397) | |
| at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:193) | |
| ... 3 more | |
| lastException: Throwable = null | |
| 16/03/21 21:45:32 [WARN] o.a.t.k.p.v.s.KernelOutputStream - Suppressing empty output: '' | |
| FROM THE JUPYTER UI: | |
| Jupyter NotebookOK_TestNotebook Last Checkpoint: 5 minutes ago (autosaved) | |
| CellToolbar | |
| In [2]: | |
| sc.parallelize(1 to 100). | |
| filter(x => x % 2 == 0). | |
| map(x => x * x). | |
| take(10) | |
| Out[2]: | |
| Name: org.apache.spark.SparkException | |
| Message: Job aborted due to stage failure: Task 0 in stage 1.0 failed 1 times, most recent failure: Lost task 0.0 in stage 1.0 (TID 1, localhost): java.net.SocketTimeoutException: connect timed out | |
| at java.net.PlainSocketImpl.socketConnect(Native Method) | |
| at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350) | |
| at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206) | |
| at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188) | |
| at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) | |
| at java.net.Socket.connect(Socket.java:589) | |
| at sun.net.NetworkClient.doConnect(NetworkClient.java:175) | |
| at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) | |
| at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) | |
| at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) | |
| at sun.net.www.http.HttpClient.New(HttpClient.java:308) | |
| at sun.net.www.http.HttpClient.New(HttpClient.java:326) | |
| at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1169) | |
| at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1105) | |
| at sun.net.www.protocol.http.HttpURLConnection$6.run(HttpURLConnection.java:989) | |
| at sun.net.www.protocol.http.HttpURLConnection$6.run(HttpURLConnection.java:987) | |
| at java.security.AccessController.doPrivileged(Native Method) | |
| at java.security.AccessController.doPrivilegedWithCombiner(AccessController.java:782) | |
| at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:986) | |
| at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:933) | |
| at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:588) | |
| at org.apache.spark.util.Utils$.fetchFile(Utils.scala:394) | |
| at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:405) | |
| at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:397) | |
| at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) | |
| at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) | |
| at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) | |
| at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) | |
| at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) | |
| at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) | |
| at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) | |
| at org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:397) | |
| at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:193) | |
| at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
| at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
| at java.lang.Thread.run(Thread.java:745) | |
| Driver stacktrace: | |
| StackTrace: org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431) | |
| org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419) | |
| org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418) | |
| scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) | |
| scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) | |
| org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418) | |
| org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) | |
| org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799) | |
| scala.Option.foreach(Option.scala:236) | |
| org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799) | |
| org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640) | |
| org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599) | |
| org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588) | |
| org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) | |
| org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620) | |
| org.apache.spark.SparkContext.runJob(SparkContext.scala:1832) | |
| org.apache.spark.SparkContext.runJob(SparkContext.scala:1845) | |
| org.apache.spark.SparkContext.runJob(SparkContext.scala:1858) | |
| org.apache.spark.rdd.RDD$$anonfun$take$1.apply(RDD.scala:1328) | |
| org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150) | |
| org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111) | |
| org.apache.spark.rdd.RDD.withScope(RDD.scala:316) | |
| org.apache.spark.rdd.RDD.take(RDD.scala:1302) | |
| $line20.$read$$iwC$$iwC$$iwC$$iwC.<init>(<console>:24) | |
| $line20.$read$$iwC$$iwC$$iwC.<init>(<console>:29) | |
| $line20.$read$$iwC$$iwC.<init>(<console>:31) | |
| $line20.$read$$iwC.<init>(<console>:33) | |
| $line20.$read.<init>(<console>:35) | |
| $line20.$read$.<init>(<console>:39) | |
| $line20.$read$.<clinit>(<console>) | |
| $line20.$eval$.<init>(<console>:7) | |
| $line20.$eval$.<clinit>(<console>) | |
| $line20.$eval.$print(<console>) | |
| sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
| sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
| sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
| java.lang.reflect.Method.invoke(Method.java:497) | |
| org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065) | |
| org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346) | |
| org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840) | |
| org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871) | |
| org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819) | |
| org.apache.toree.kernel.interpreter.scala.ScalaInterpreter$$anonfun$interpretAddTask$1$$anonfun$apply$3.apply(ScalaInterpreter.scala:356) | |
| org.apache.toree.kernel.interpreter.scala.ScalaInterpreter$$anonfun$interpretAddTask$1$$anonfun$apply$3.apply(ScalaInterpreter.scala:351) | |
| org.apache.toree.global.StreamState$.withStreams(StreamState.scala:81) | |
| org.apache.toree.kernel.interpreter.scala.ScalaInterpreter$$anonfun$interpretAddTask$1.apply(ScalaInterpreter.scala:350) | |
| org.apache.toree.kernel.interpreter.scala.ScalaInterpreter$$anonfun$interpretAddTask$1.apply(ScalaInterpreter.scala:350) | |
| org.apache.toree.utils.TaskManager$$anonfun$add$2$$anon$1.run(TaskManager.scala:140) | |
| java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) | |
| java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) | |
| java.lang.Thread.run(Thread.java:745) |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment