Created
August 22, 2016 11:55
-
-
Save HyukjinKwon/4a7e0848173d045fd6faf3e3030f62db to your computer and use it in GitHub Desktop.
[SPARK][R] test output (stdout) on Windows 7 32bit
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Warning message: | |
package 'testthat' was built under R version 3.3.1 | |
Loading required package: methods | |
Attaching package: 'SparkR' | |
The following object is masked from 'package:testthat': | |
describe | |
The following objects are masked from 'package:stats': | |
cov, filter, lag, na.omit, predict, sd, var, window | |
The following objects are masked from 'package:base': | |
as.data.frame, colnames, colnames<-, drop, endsWith, intersect, | |
rank, rbind, sample, startsWith, subset, summary, transform, union | |
binary functions: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
.1...2 | |
functions on binary files: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
3456 | |
broadcast variables: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
.. | |
functions in client.R: ..... | |
test functions in sparkR.R: ........................ | |
include R packages: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
MLlib functions: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
..................................7..........................8......9.....a....b....c.....d......efE...E | |
parallelize() and collect(): Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
............................. | |
basic RDD functions: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
................................................................................................................................................................................................................................................................................................................................................................................E..E.................................................... | |
SerDe functionality: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
................... | |
partitionBy, groupByKey, reduceByKey etc.: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
.................... | |
SparkSQL functions: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
.........................................................S.....................................................................................................................EE.......................................................................................................................................S............................................................................................................................................................................................................................................................S..................................................................................................................................................S | |
tests RDD function take(): Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
................ | |
the textFile() function: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
EEEEEEEEE | |
functions in utils.R: Spark package is found in SPARK_HOME: C:\Users\IEUser\workspace\spark\bin\.. | |
..E............................... | |
Windows-specific tests: . | |
Skipped ------------------------------------------------------------------------ | |
1. create DataFrame from RDD (@test_sparkSQL.R#200) - Hive is not build with SparkSQL, skipped | |
2. test HiveContext (@test_sparkSQL.R#1041) - Hive is not build with SparkSQL, skipped | |
3. read/write ORC files (@test_sparkSQL.R#1748) - Hive is not build with SparkSQL, skipped | |
4. enableHiveSupport on SparkSession (@test_sparkSQL.R#2480) - Hive is not build with SparkSQL, skipped | |
Failed ------------------------------------------------------------------------- | |
1. Error: union on two RDDs (@test_binary_function.R#38) ----------------------- | |
java.net.URISyntaxException: Illegal character in opaque part at index 2: C:\Users\IEUser\AppData\Local\Temp\Rtmp0WxVBi\spark-testee8286a28.tmp | |
at java.net.URI$Parser.fail(URI.java:2848) | |
at java.net.URI$Parser.checkChars(URI.java:3021) | |
at java.net.URI$Parser.parse(URI.java:3058) | |
at java.net.URI.<init>(URI.java:588) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.api.java.JavaSparkContext.textFile(JavaSparkContext.scala:179) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: textFile(sc, fileName) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_binary_function.R:38 | |
2: callJMethod(sc, "textFile", path, getMinPartitions(sc, minPartitions)) | |
3: invokeJava(isStatic = FALSE, objId$id, methodName, ...) | |
4: stop(readString(conn)) | |
2. Error: zipPartitions() on RDDs (@test_binary_function.R#84) ----------------- | |
java.net.URISyntaxException: Illegal character in opaque part at index 2: C:\Users\IEUser\AppData\Local\Temp\Rtmp0WxVBi\spark-testee848d42457.tmp | |
at java.net.URI$Parser.fail(URI.java:2848) | |
at java.net.URI$Parser.checkChars(URI.java:3021) | |
at java.net.URI$Parser.parse(URI.java:3058) | |
at java.net.URI.<init>(URI.java:588) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.api.java.JavaSparkContext.textFile(JavaSparkContext.scala:179) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: textFile(sc, fileName, 1) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_binary_function.R:84 | |
2: callJMethod(sc, "textFile", path, getMinPartitions(sc, minPartitions)) | |
3: invokeJava(isStatic = FALSE, objId$id, methodName, ...) | |
4: stop(readString(conn)) | |
3. Error: saveAsObjectFile()/objectFile() following textFile() works (@test_binaryFile.R#31) | |
java.net.URISyntaxException: Illegal character in opaque part at index 2: C:\Users\IEUser\AppData\Local\Temp\Rtmp0WxVBi\spark-testee879447d80.tmp | |
at java.net.URI$Parser.fail(URI.java:2848) | |
at java.net.URI$Parser.checkChars(URI.java:3021) | |
at java.net.URI$Parser.parse(URI.java:3058) | |
at java.net.URI.<init>(URI.java:588) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.api.java.JavaSparkContext.textFile(JavaSparkContext.scala:179) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: textFile(sc, fileName1, 1) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_binaryFile.R:31 | |
2: callJMethod(sc, "textFile", path, getMinPartitions(sc, minPartitions)) | |
3: invokeJava(isStatic = FALSE, objId$id, methodName, ...) | |
4: stop(readString(conn)) | |
4. Error: saveAsObjectFile()/objectFile() works on a parallelized list (@test_binaryFile.R#46) | |
java.net.URISyntaxException: Illegal character in opaque part at index 2: C:\Users\IEUser\AppData\Local\Temp\Rtmp0WxVBi\spark-testee835f35271.tmp | |
at java.net.URI$Parser.fail(URI.java:2848) | |
at java.net.URI$Parser.checkChars(URI.java:3021) | |
at java.net.URI$Parser.parse(URI.java:3058) | |
at java.net.URI.<init>(URI.java:588) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$sequenceFile$1.apply(SparkContext.scala:1144) | |
at org.apache.spark.SparkContext$$anonfun$sequenceFile$1.apply(SparkContext.scala:1141) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.sequenceFile(SparkContext.scala:1141) | |
at org.apache.spark.SparkContext$$anonfun$objectFile$1.apply(SparkContext.scala:1214) | |
at org.apache.spark.SparkContext$$anonfun$objectFile$1.apply(SparkContext.scala:1212) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.objectFile(SparkContext.scala:1212) | |
at org.apache.spark.api.java.JavaSparkContext.objectFile(JavaSparkContext.scala:340) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: objectFile(sc, fileName) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_binaryFile.R:46 | |
2: callJMethod(sc, "objectFile", path, getMinPartitions(sc, minPartitions)) | |
3: invokeJava(isStatic = FALSE, objId$id, methodName, ...) | |
4: stop(readString(conn)) | |
5. Error: saveAsObjectFile()/objectFile() following RDD transformations works (@test_binaryFile.R#57) | |
java.net.URISyntaxException: Illegal character in opaque part at index 2: C:\Users\IEUser\AppData\Local\Temp\Rtmp0WxVBi\spark-testee828376db9.tmp | |
at java.net.URI$Parser.fail(URI.java:2848) | |
at java.net.URI$Parser.checkChars(URI.java:3021) | |
at java.net.URI$Parser.parse(URI.java:3058) | |
at java.net.URI.<init>(URI.java:588) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.api.java.JavaSparkContext.textFile(JavaSparkContext.scala:179) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: textFile(sc, fileName1) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_binaryFile.R:57 | |
2: callJMethod(sc, "textFile", path, getMinPartitions(sc, minPartitions)) | |
3: invokeJava(isStatic = FALSE, objId$id, methodName, ...) | |
4: stop(readString(conn)) | |
6. Error: saveAsObjectFile()/objectFile() works with multiple paths (@test_binaryFile.R#85) | |
java.net.URISyntaxException: Illegal character in opaque part at index 2: C:\Users\IEUser\AppData\Local\Temp\Rtmp0WxVBi\spark-testee81353a.tmp,C:\Users\IEUser\AppData\Local\Temp\Rtmp0WxVBi\spark-testee8241b4b38.tmp | |
at java.net.URI$Parser.fail(URI.java:2848) | |
at java.net.URI$Parser.checkChars(URI.java:3021) | |
at java.net.URI$Parser.parse(URI.java:3058) | |
at java.net.URI.<init>(URI.java:588) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$sequenceFile$1.apply(SparkContext.scala:1144) | |
at org.apache.spark.SparkContext$$anonfun$sequenceFile$1.apply(SparkContext.scala:1141) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.sequenceFile(SparkContext.scala:1141) | |
at org.apache.spark.SparkContext$$anonfun$objectFile$1.apply(SparkContext.scala:1214) | |
at org.apache.spark.SparkContext$$anonfun$objectFile$1.apply(SparkContext.scala:1212) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.objectFile(SparkContext.scala:1212) | |
at org.apache.spark.api.java.JavaSparkContext.objectFile(JavaSparkContext.scala:340) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: objectFile(sc, c(fileName1, fileName2)) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_binaryFile.R:85 | |
2: callJMethod(sc, "objectFile", path, getMinPartitions(sc, minPartitions)) | |
3: invokeJava(isStatic = FALSE, objId$id, methodName, ...) | |
4: stop(readString(conn)) | |
7. Error: spark.glm save/load (@test_mllib.R#158) ------------------------------ | |
java.io.IOException: No FileSystem for scheme: C | |
at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421) | |
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) | |
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) | |
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) | |
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) | |
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.ml.r.RWrappers$.load(RWrappers.scala:36) | |
at org.apache.spark.ml.r.RWrappers.load(RWrappers.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: read.ml(modelPath) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_mllib.R:158 | |
2: callJStatic("org.apache.spark.ml.r.RWrappers", "load", path) | |
3: invokeJava(isStatic = TRUE, className, methodName, ...) | |
4: stop(readString(conn)) | |
8. Error: glm save/load (@test_mllib.R#288) ------------------------------------ | |
java.io.IOException: No FileSystem for scheme: C | |
at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421) | |
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) | |
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) | |
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) | |
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) | |
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.ml.r.RWrappers$.load(RWrappers.scala:36) | |
at org.apache.spark.ml.r.RWrappers.load(RWrappers.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: read.ml(modelPath) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_mllib.R:288 | |
2: callJStatic("org.apache.spark.ml.r.RWrappers", "load", path) | |
3: invokeJava(isStatic = TRUE, className, methodName, ...) | |
4: stop(readString(conn)) | |
9. Error: spark.kmeans (@test_mllib.R#336) ------------------------------------- | |
java.io.IOException: No FileSystem for scheme: C | |
at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421) | |
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) | |
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) | |
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) | |
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) | |
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.ml.r.RWrappers$.load(RWrappers.scala:36) | |
at org.apache.spark.ml.r.RWrappers.load(RWrappers.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: read.ml(modelPath) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_mllib.R:336 | |
2: callJStatic("org.apache.spark.ml.r.RWrappers", "load", path) | |
3: invokeJava(isStatic = TRUE, className, methodName, ...) | |
4: stop(readString(conn)) | |
10. Error: spark.naiveBayes (@test_mllib.R#403) -------------------------------- | |
java.io.IOException: No FileSystem for scheme: C | |
at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421) | |
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) | |
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) | |
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) | |
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) | |
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.ml.r.RWrappers$.load(RWrappers.scala:36) | |
at org.apache.spark.ml.r.RWrappers.load(RWrappers.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: read.ml(modelPath) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_mllib.R:403 | |
2: callJStatic("org.apache.spark.ml.r.RWrappers", "load", path) | |
3: invokeJava(isStatic = TRUE, className, methodName, ...) | |
4: stop(readString(conn)) | |
11. Error: spark.survreg (@test_mllib.R#460) ----------------------------------- | |
java.io.IOException: No FileSystem for scheme: C | |
at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421) | |
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) | |
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) | |
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) | |
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) | |
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.ml.r.RWrappers$.load(RWrappers.scala:36) | |
at org.apache.spark.ml.r.RWrappers.load(RWrappers.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: read.ml(modelPath) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_mllib.R:460 | |
2: callJStatic("org.apache.spark.ml.r.RWrappers", "load", path) | |
3: invokeJava(isStatic = TRUE, className, methodName, ...) | |
4: stop(readString(conn)) | |
12. Error: spark.isotonicRegression (@test_mllib.R#505) ------------------------ | |
java.io.IOException: No FileSystem for scheme: C | |
at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421) | |
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) | |
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) | |
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) | |
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) | |
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.ml.r.RWrappers$.load(RWrappers.scala:36) | |
at org.apache.spark.ml.r.RWrappers.load(RWrappers.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: read.ml(modelPath) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_mllib.R:505 | |
2: callJStatic("org.apache.spark.ml.r.RWrappers", "load", path) | |
3: invokeJava(isStatic = TRUE, className, methodName, ...) | |
4: stop(readString(conn)) | |
13. Error: spark.gaussianMixture (@test_mllib.R#567) --------------------------- | |
java.io.IOException: No FileSystem for scheme: C | |
at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421) | |
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) | |
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) | |
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) | |
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) | |
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.ml.r.RWrappers$.load(RWrappers.scala:36) | |
at org.apache.spark.ml.r.RWrappers.load(RWrappers.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: read.ml(modelPath) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_mllib.R:567 | |
2: callJStatic("org.apache.spark.ml.r.RWrappers", "load", path) | |
3: invokeJava(isStatic = TRUE, className, methodName, ...) | |
4: stop(readString(conn)) | |
14. Error: spark.lda with libsvm (@test_mllib.R#600) --------------------------- | |
java.io.IOException: No FileSystem for scheme: C | |
at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421) | |
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) | |
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) | |
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) | |
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) | |
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:992) | |
at org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:987) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:987) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:803) | |
at org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:801) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) | |
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) | |
at org.apache.spark.SparkContext.withScope(SparkContext.scala:683) | |
at org.apache.spark.SparkContext.textFile(SparkContext.scala:801) | |
at org.apache.spark.ml.r.RWrappers$.load(RWrappers.scala:36) | |
at org.apache.spark.ml.r.RWrappers.load(RWrappers.scala) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: read.ml(modelPath) at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_mllib.R:600 | |
2: callJStatic("org.apache.spark.ml.r.RWrappers", "load", path) | |
3: invokeJava(isStatic = TRUE, className, methodName, ...) | |
4: stop(readString(conn)) | |
15. Error: spark.lda with text input (@test_mllib.R#613) ----------------------- | |
org.apache.spark.sql.AnalysisException: Path does not exist: file:/C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/data/mllib/sample_lda_data.txt; | |
at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$12.apply(DataSource.scala:375) | |
at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$12.apply(DataSource.scala:364) | |
at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241) | |
at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241) | |
at scala.collection.immutable.List.foreach(List.scala:381) | |
at scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:241) | |
at scala.collection.immutable.List.flatMap(List.scala:344) | |
at org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:364) | |
at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:149) | |
at org.apache.spark.sql.DataFrameReader.text(DataFrameReader.scala:490) | |
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) | |
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) | |
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) | |
at java.lang.reflect.Method.invoke(Method.java:498) | |
at org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:141) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:86) | |
at org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:38) | |
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244) | |
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308) | |
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294) | |
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846) | |
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) | |
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) | |
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) | |
at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111) | |
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) | |
at java.lang.Thread.run(Thread.java:745) | |
1: read.text("data/mllib/sample_lda_data.txt") at C:/Users/IEUser/workspace/spark/R/lib/SparkR/tests/testthat/test_mllib.R:613 | |
2: dispatchFunc("read.text(path)", x, ...) | |
3: f(x, ...) | |
4: callJMethod(read, "text", paths) | |
5: invokeJava(isStatic = FALSE, objId$id, methodName, ...) | |
6: stop(readString(conn)) | |
DONE =========================================================================== | |
Error: Test failures | |
Execution halted |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment