Replies: 1 comment
-
Hello, for your case where you are creating files of size 10MB its better to use file_cache instead. Please try that and let me know if you still see latency, thanks! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I run my application on a Ubuntu VM, the app uses spark to write about 10M of parquet data to ADLS blob container (mounted by blobfuse2) every 5 minutes. We found that each write took around 4 minutes to complete. I copied the same amount of data directly (without spark session) to the ADLS, it still took over 2 minutes. The Ubuntu box is not under pressure ( lower CPU and memory usage), the upload speed of the internet link is over 1500Mb/sec. My questions are:
Beta Was this translation helpful? Give feedback.
All reactions