https://delta.io logo
s

sabari dass

04/06/2023, 8:16 PM
Hi All, I have a use case where I need to write the data into same delta-lake location Parallely. When I try to do this using pyspark,I got the concurrency error is there any way to handle this issue?
I am not sure whether this I need to post it in #random or here. Apologies if I posted wrongly
l

Lennart Skogmo

04/06/2023, 8:37 PM
I dont have any experience with this but there are some tips in the docs. For example sticking to inserts or using partitions. https://docs.delta.io/latest/concurrency-control.html
s

sabari dass

04/06/2023, 9:00 PM
@Lennart Skogmo I am unable to open this link.
l

Lennart Skogmo

04/06/2023, 9:09 PM
Its because the certificate expired i guess. But its the offical site.
s

sabari dass

04/06/2023, 9:48 PM
Ok, now I am able to access it. Also I think I found the solution in stack overflow for my exact use case. When I read concurrency control page asking to create partition for the delta table which will not help for my usecase 🙂 Thanks for your help!
39 Views