'How to create a CSV file with PySpark?
I have a short question about pyspark write.
read_jdbc = spark.read \
.format("jdbc") \
.option("url", "jdbc:postgresql:dbserver") \
.option("dbtable", "schema.tablename") \
.option("user", "username") \
.option("password", "password") \
.load()
read_jdbc.show()
When I command show, it works perfectly,
However, when I try to write,
read_jdbc.write.csv("some/aaa")
it only creates aaa folder, but the CSV is not created.
I also tried
read_jdbc.write.jdbc(mysql_url,table="csv_test",mode="append")
This does not work either. Any help?
Solution 1:[1]
you can write dataframe to csv
df.write.csv("file-path")
or
df.write.format("csv").save("file-path")
Sources
This article follows the attribution requirements of Stack Overflow and is licensed under CC BY-SA 3.0.
Source: Stack Overflow
Solution | Source |
---|---|
Solution 1 | Sudhin |