You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Can someone here explain how the temporary files are created in S3 when calling the driver? It seems:
When calling spark.read, a temporary file folder will be created with a file of length 1
When calling any functions to evaluate the returned Dataframe, another temporary file folder will be created with the actual data
What is the file created in first step?
The temporary files created seems to be in plain text without schema. Is there anyway to make that parquet or export parquet files with schema?
Thanks!
The text was updated successfully, but these errors were encountered:
Can someone here explain how the temporary files are created in S3 when calling the driver? It seems:
When calling any functions to evaluate the returned Dataframe, another temporary file folder will be created with the actual data
What is the file created in first step?
Thanks!
The text was updated successfully, but these errors were encountered: