For most common read and write operations on Delta tables, you can use Apache Spark reader and writer APIs (see Table Batch Reads and Writes and Table Streaming Reads and Writes). However, there are some operations that are specific to Delta Lake and you must use Delta Lake programmatic APIs. This topic describes these programmatic APIs.
Some programmatic APIs are still evolving and are indicated with the Evolving qualifier in the API docs.
Databricks ensures binary compatibility between the Delta Lake project and Delta Lake in Databricks Runtime. The following table lists the Delta Lake release packaged in each Databricks Runtime version.
Delta Lake in Databricks Runtime compatibility matrix
|Databricks Runtime version||Delta Lake version||Documentation|