I currently have an append table in databricks (spark 3, databricks 7.5)
parsedDf
.select("somefield", "anotherField",'partition', 'offset')
.write
.format("delta")
.mode("append")
.option("mergeSchema", "true")
.save(f"/mnt/defaultDatalake/{append_table_name}")
It was created with a create table command before and I don't use INSERT
commands to write to it (as seen above)
Now I want to be able to use SQL logic to query it without everytime going through createOrReplaceTempView every time. Is is possible to add a table to the current data without removing it? what changes do I need to support this?
UPDATE:
I've tried:
res= spark.sql(f"CREATE TABLE exploration.oplog USING DELTA LOCATION '/mnt/defaultDataLake/{append_table_name}'")
But get an AnalysisException
You are trying to create an external table exploration
.dataitems_oplog
from /mnt/defaultDataLake/specificpathhere
using Databricks Delta, but the schema is not specified when the
input path is empty.
While the path isn't empty.
question from:
https://stackoverflow.com/questions/65652312/delta-lake-in-databricks-creating-a-table-for-existing-storage 与恶龙缠斗过久,自身亦成为恶龙;凝视深渊过久,深渊将回以凝视…