import org.apache.spark.ml.feature.val stringColumns = Array("domain","size", "form_factor")
val index_transformers: Array[org.apache.spark.ml.PipelineStageindexColumns = df_indexed.columns.filter(x => x contains "in
可以将一个经过训练的管道模型从pyspark环境加载到scala中吗?我正在尝试这样做,但是我遇到了这个错误 requirement failed: Error loading metadata: Expected class name org.apache.spark.ml.PipelineModelbut found class name pyspark.ml.pipeline.PipelineModel 更准确地说,我有一个pyspark管道模型: pipe =
现在,我想通过交叉验证使用Pipeline进行模型选择。org.apache.spark.ml.recommendation.ALS.transformSchema(ALS.scala:304)at org.apache.spark.ml.PipelineStage.transformSchema(Pipeline.scala:58)
at org