擅长:python、mysql、java
<p>我找到了解决办法:</p>
<pre><code>from pyspark.sql.functions import regexp_extract, col, split
from pyspark.sql import functions as sf
df_test=spark.sql("select * from brand_cleanup")
#Applying the transformations to the data
split_col=split(df_test.item_eng_desc,' ')
df_split=df_test.withColumn('item_desc_clean',sf.concat(split_col.getItem(0),sf.lit(' '),split_col.getItem(1),sf.lit(' '),split_col.getItem(2)))
</code></pre>