) at org.apache.spark.sql.DataFrame$$anonfun$collect$1.apply(DataFrame.scala:1385) at org.apache.spark.sql.DataFrame$$anonfun$collect$1.apply(DataFrame.scala:1385) at org.apache.spark.sql.execution.SQLExecution$org.apache.spark.
$1$1.apply(DataFrame.scala:1499) at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame:2086) at org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$execute$1(DataFrame.scala:1498)
我正在使用Kakfa和MongoDB,我的输出(Df1)是一个流数据帧,我想把它保存到Kakfa中。有什么建议吗?非常感谢! val df= lines.selectExpr("CAST(value AS STRING)").as[(String)]
.select("data.*") .format("console")
.option("truncat
我正在尝试设置一个Pandas Dataframe来处理Databricks中的数据。我的数据是从本地计算机上的文件导入的,如下所示:Snip of the data # Import packagesimport numpy as np
#print(ownr.shape)df1 = pd.DataFrame</em