https://delta.io logo
r

Rahul Madnawat

08/14/2023, 5:50 PM
Hey everyone! Does anyone know how to get the DynamoDB Log Store going within Databricks for writes to Delta? I tried doing something simple like
Copy code
spark.conf.set("spark.delta.logStore.s3.impl", "io.delta.storage.S3DynamoDBLogStore")
spark.conf.set("spark.delta.DynamoDBLogStore.ddb.region", "us-west-2")
spark.conf.set("spark.io.delta.storage.S3DynamoDBLogStore.ddb.tableName", "ddb_test")

df = spark.createDataFrame(
    [
        (1, "foo"),
        (2, "bar"),
    ],
    ["id", "label"]
)

df.write.mode("append").format("delta").save(<S3 path>)
and didn't see a DynamoDB table created. Would appreciate any pointers, thanks!
n

Nick Karpov

08/14/2023, 6:56 PM
what version of DBR?
r

Rahul Madnawat

08/14/2023, 8:06 PM
13.2
n

Nick Karpov

08/14/2023, 8:18 PM
hm, i don't think it works yet at session level, can you set these at the cluster level
Copy code
spark.delta.logStore.class io.delta.storage.S3DynamoDBLogStore
spark.databricks.tahoe.logStore.allowOSSLogStores true
spark.io.delta.storage.S3DynamoDBLogStore.ddb.tableName <dynamodb_table_name>
spark.io.delta.storage.S3DynamoDBLogStore.ddb.region <dynamodb_region>
👍 1