ПФ
Size: a a a
ПФ
ПФ
RI
val df = spark.sqlContext.read.parquet(s"/path/to/parquet")
ПФ
RI
RI
RI
ПФ
ПФ
RI
ПФ
ПФ
RI
RI
val ss = SparkSession.builder.config(rdd.sparkContext.getConf).getOrCreate(), который генерировал ss внутри rdd. Падает с ошибкой org.apache.spark.SparkException: A master URL must be set in your configuration
at org.apache.spark.SparkContext.<init>(SparkContext.scala:367)
at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493)
at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:934)
at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:925)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:925)
at org.apache.spark.ml.util.BaseReadWrite$class.sparkSession(ReadWrite.scala:74)
at org.apache.spark.ml.util.MLReader.sparkSession(ReadWrite.scala:193)
at org.apache.spark.ml.util.BaseReadWrite$class.sc(ReadWrite.scala:85)
at org.apache.spark.ml.util.MLReader.sc(ReadWrite.scala:193)
at org.apache.spark.ml.PipelineModel$PipelineModelReader.load(Pipeline.scala:410)
at org.apache.spark.ml.PipelineModel$PipelineModelReader.load(Pipeline.scala:404)
at ru.model.ModelApplication$$anonfun$6$$anonfun$apply$2.apply(ModelApplication.scala:143)
val model = PipelineModel.read.load(modelPath), перед функцией генерацииt
RI
AC
RI
val sparkSession: SparkSession = SparkSession.builder.config(rdd.context.getConf).getOrCreate()?AC
rdd.contextа из него создать sqlContext и читать
GP
val df = spark.sqlContext.read.parquet(s"/path/to/parquet")