Skip to content

Releases: databricks/databricks-jdbc

v3.3.2

28 Apr 09:58
d54693a

Choose a tag to compare

Added

  • Added CallableStatement support with IN parameters. Connection.prepareCall() now returns a working DatabricksCallableStatement that supports positional parameter binding and execution via {call proc(?)} JDBC escape syntax. OUT/INOUT parameters and named parameters throw SQLFeatureNotSupportedException.
  • Added AI coding agent detection to the User-Agent header. When the driver is invoked by a known AI coding agent (e.g. Claude Code, Cursor, Gemini CLI), agent/<product> is appended to the User-Agent string.

Updated

  • Added support for using SQL SHOW commands for Thrift-mode metadata operations (getTables, getColumns, getSchemas, getFunctions, getPrimaryKeys, getImportedKeys, getCrossReference). Enable by setting UseQueryForMetadata=1. This aligns Thrift metadata behavior with Statement Execution API (SEA) mode.

Fixed

  • Improved error messages for cancelled statements: operations cancelled via Statement.cancel() or closed connections now return SQL state HY008 (operation cancelled) instead of generic error codes, making it easier for applications to detect and handle cancellations.
  • Fixed race condition between chunk download error handling and result set close that could cause invalid state transition warnings (CHUNK_RELEASED -> DOWNLOAD_FAILED) during Arrow Cloud Fetch operations in resource-constrained environments.
  • Fixed EnableBatchedInserts silently falling back to individual execution when table or schema names contain special characters (e.g., hyphens) inside backtick-quoted identifiers. Added a warn log when the fallback occurs.
  • Fixed IntervalConverter crash (IllegalArgumentException: Invalid interval metadata) when INTERVAL columns are returned via CloudFetch. Arrow metadata from CloudFetch uses underscored format (INTERVAL_YEAR_MONTH, INTERVAL_DAY_TIME) which the driver's regex did not accept.
  • Fixed Statement being prematurely closed after queries that return inline results, which prevented re-execution, getResultSet(), and getExecutionResult() from working. Statements now remain open and reusable until explicitly closed by the caller.
  • Fixed primitive types within complex types (ARRAY, MAP, STRUCT) not being correctly parsed when Arrow serialization uses alternate formats: TIMESTAMP/TIMESTAMP_NTZ as epoch microseconds or component arrays, and BINARY as base64-encoded strings.
  • Fixed PARSE_SYNTAX_ERROR for column names containing special characters (e.g., dots) when EnableBatchedInserts is enabled, by re-quoting column names with backticks in reconstructed multi-row INSERT statements.
  • Fixed Volume ingestion for SEA mode, which was broken due to statement being closed prematurely.
  • Fixed unclear error: [null] messages during transient HTTP failures (e.g. 502 Bad Gateway) in Thrift polling. Error messages now include server error details and use SQL state 08S01 (communication link failure) so callers can identify retryable errors. Also fixed DatabricksError (RuntimeException) from SDK client being unhandled in CloudFetch download paths.
  • Fixed escaped pattern characters in catalogName for getSchemas, as returned catalogName should be unescaped.
  • Fixed getColumnClassName() returning null for VARIANT columns in SEA mode by adding VARIANT to the type system.
  • Fixed getColumns() returning DATA_TYPE=0 (NULL) for GEOMETRY/GEOGRAPHY columns in Thrift mode. Now returns Types.VARCHAR (12) when geospatial is disabled and Types.OTHER (1111) when enabled, consistent with SEA mode.
  • Fixed getCrossReference() returning 0 rows when parent args are passed in uppercase. The client-side filter used case-sensitive comparison against server-returned lowercase names.

Release Databricks OSS JDBC driver version 3.3.1

19 Mar 08:12
c6ab4de

Choose a tag to compare

Added

  • Added DatabaseMetaData.getProcedures() and DatabaseMetaData.getProcedureColumns() to discover stored procedures and their parameters. Queries information_schema.routines and information_schema.parameters using parameterized SQL for both SEA and Thrift transports.
  • Added connection property OAuthWebServerTimeout to configure the OAuth browser authentication timeout for U2M (user-to-machine) flows, and also updated hardcoded 1-hour timeout to default 120 seconds timeout.
  • Added connection property UseQueryForMetadata to use SQL SHOW commands instead of Thrift RPCs for metadata operations (getCatalogs, getSchemas, getTables, getColumns, getFunctions). This fixes incorrect wildcard matching where _ was treated as a single-character wildcard in Thrift metadata pattern filters.
  • Added connection property TreatMetadataCatalogNameAsPattern to control whether catalog names are treated as patterns in Thrift metadata RPCs. When disabled (default), unescaped _ in catalog names is escaped to prevent single-character wildcard matching. This aligns with JDBC spec which treats catalogName as identifier and not pattern.

Updated

  • Bumped com.fasterxml.jackson.core:jackson-core from 2.18.3 to 2.18.6.
  • Fat jar now routes SDK and Apache HTTP client logs through Java Util Logging (JUL), removing the need for external logging libraries.
  • Added Apache Arrow on-heap memory management for processing Arrow query results. Previously, Arrow result processing was unusable on JDK 16+ without passing the --add-opens=java.base/java.nio=ALL-UNNAMED JVM argument, due to stricter encapsulation of internal APIs. With this change, there is no JVM argument required - the driver automatically falls back to an on-heap memory path that uses standard JVM heap allocation instead of direct memory access.
  • Log timestamps now explicitly display timezone.
  • [Breaking Change] PreparedStatement.setTimestamp(int, Timestamp, Calendar) now properly applies Calendar timezone conversion using LocalDateTime pattern (inline with getTimestamp). Previously Calendar parameter was ineffective.
  • DatabaseMetaData.getColumns() with null catalog parameter now retrieves columns from all available catalogs when using SQL Execution API.

Fixed

  • Fixed statement timeout when the server returns TIMEDOUT_STATE directly in the ExecuteStatement response (e.g. query queued under load), the driver now throws SQLTimeoutException instead of DatabricksHttpException.
  • Fixed Thrift polling infinite loop when server restarts invalidate operation handles, and added configurable timeout (MetadataOperationTimeout, default 300s) with sleep between polls for metadata operations.
  • Fixed DatabricksParameterMetaData.countParameters and DatabricksStatement.trimCommentsAndWhitespaces with a SqlCommentParser utility class.
  • Fixed rollback() to throw SQLException when called in auto-commit mode (no active transaction), aligning with JDBC spec. Previously it silently sent a ROLLBACK command to the server.
  • Fixed fetchAutoCommitStateFromServer() to accept both "1"/"0" and "true"/"false" responses from SET AUTOCOMMIT query, since different server implementations return different formats.
  • Fixed socket leak in SDK HTTP client that prevented CRaC checkpointing. The SDK's connection pool was not shut down on connection.close(), leaving TCP sockets open.
  • Fixed IdleConnectionEvictor thread leak in long-running services. The feature-flags context shared per host was ref-counted incorrectly and held a stale connection UUID after the owning connection closed; on the next 15-minute refresh it silently recreated an HTTP client (and its evictor thread) that was never cleaned up. Connection UUIDs are now tracked idempotently and the stored connection context is updated when the owning connection closes.
  • Fixed Date fields within complex types (ARRAY, STRUCT, MAP) being returned as epoch day integers instead of proper date values.
  • Fixed DatabaseMetaData.getColumns() returning the column type name in COLUMN_DEF for columns with no default value. COLUMN_DEF now correctly returns null per the JDBC specification.
  • Coalesce concurrent expired cloud fetch link refreshes into a single batch FetchResults RPC to prevent thread pool exhaustion under high concurrency.

v3.2.1

25 Feb 09:30
1638998

Choose a tag to compare

[v3.2.1] - 2026-02-16

Added

  • Added streaming prefetch mode for Thrift inline results (columnar and Arrow) with background batch prefetching and configurable sliding window for improved throughput.
  • Added EnableInlineStreaming connection parameter to enable/disable streaming mode (default: enabled).
  • Added ThriftMaxBatchesInMemory connection parameter to control the sliding window size for streaming (default: 3).
  • Added support for disabling CloudFetch via EnableQueryResultDownload=0 to use inline Arrow results instead.
  • Added EnableMetricViewMetadata connection parameter to enable/disable Metric View table type (default: disabled).
  • Added NonRowcountQueryPrefixes connection parameter to specify comma-separated query prefixes that should return result sets instead of row counts.

Updated

  • Enhanced error logging for token exchange failures.
  • Geospatial column type names now include SRID information (e.g., GEOMETRY(4326) instead of GEOMETRY).
  • Implemented lazy loading for inline Arrow results, fetching arrow batches on demand instead of all at once. This improves memory usage and initial response time for large result sets when using the Thrift protocol with Arrow format.
  • Enhanced enableMultipleCatalogSupport behavior: When this parameter is disabled (enableMultipleCatalogSupport=0), metadata operations (such as getSchemas(), getTables(), getColumns(), etc.) now return results only when the catalog parameter is either null or matches the current catalog. For any other catalog name, an empty result set is returned. This ensures metadata queries are restricted to the current catalog context. When enabled (enableMultipleCatalogSupport=1), metadata operations continue to work across all accessible catalogs.

Fixed

  • Fixed getTypeInfo() and getClientInfoProperties() to return fresh ResultSet instances on each call instead of shared static instances. This resolves issues where calling these methods multiple times would fail due to exhausted cursor state (Issue #1178).
  • Fixed complex data type metadata support when retrieving 0 rows in Arrow format
  • Normalized TIMESTAMP_NTZ to TIMESTAMP in Thrift path for consistency with SEA behavior
  • Fixed complex types not being returned as objects in SEA Inline mode when EnableComplexDatatypeSupport=true.
  • Fixed StringIndexOutOfBoundsException when parsing complex data types in Thrift CloudFetch mode. The issue occurred when metadata contained incomplete type information (e.g., "ARRAY" instead of "ARRAY"). Now retrieves complete type information from Arrow metadata.
  • Fixed timeout exception handling to throw SQLTimeoutException instead of DatabricksHttpException when queries timeout during result fetching phase. This completes the timeout exception fix to handle both query execution polling and result fetching phases.
  • Fixed getResultSet() to return null in case of DML statements to honour JDBC spec.

v3.1.1

13 Jan 11:46
0e8b1ad

Choose a tag to compare

Added

  • Added token caching for all authentication providers to reduce token endpoint calls.
  • We will be rolling out the use of Databricks SQL Execution API by default for queries submitted on DBSQL. To continue using Databricks Thrift Server backend for execution, set UseThriftClient to 1.

Updated

  • Changed default value of IgnoreTransactions from 0 to 1 to disable multi-statement transactions by default. Preview participants can opt-in by setting IgnoreTransactions=0. Also updated supportsTransactions() to respect this flag.

Fixed

  • [PECOBLR-1131] Fix incorrect refetching of expired CloudFetch links when using Thrift protocol.
  • Fixed logging to respect params when the driver is shaded.
  • Fixed isWildcard to return true only when the value is *

Release 3.0.7

18 Dec 19:32
4ed9125

Choose a tag to compare

Updated

  • Log timestamps now explicitly display timezone.
  • [Breaking Change] PreparedStatement.setTimestamp(int, Timestamp, Calendar) now properly applies Calendar timezone conversion using LocalDateTime pattern (inline with getTimestamp). Previously Calendar parameter was ineffective.
  • DatabaseMetaData.getColumns() with null catalog parameter now retrieves columns from all catalogs when using SQL Execution API, aligning the behaviour with thrift.
  • DatabaseMetaData.getFunctions() with null catalog parameter now retrieves columns from the current catalog when using SQL Execution API, aligning the behaviour with thrift.

Fixed

  • Fix timeout exception handling to throw SQLTimeoutException instead of DatabricksSQLException when queries timeout.
  • Removes dangerous global timezone modification that caused race conditions.
  • Fixed Statement.getLargeUpdateCount() to return -1 instead of throwing Exception when there were no more results or result is not an update count.
  • CVE-2025-66566. Updated lz4-java dependency to 1.10.1.
  • Fix INVALID_IDENTIFIER error when using catalog/schema/table names for SQL Exec API with hyphens or special characters in metadata operations (getSchemas(), getTables(), getColumns(), etc.) and connection methods (setCatalog(), setSchema()). Per Databricks identifier rules, special characters are now properly enclosed in backticks.
  • Fix Auth_Scope handling inconsistency in Azure U2M OAuth.

Release Databricks OSS JDBC driver version 3.0.6

12 Dec 09:28
6e1c164

Choose a tag to compare

Added

  • Added the EnableTokenFederation url param to enable or disable Token federation feature. By default it is set to 1
  • Added the ApiRetriableHttpCodes, ApiRetryTimeout url params to enable retries for specific HTTP codes irrespective of Retry-After header. By default the HTTP codes list is empty.

Updated

  • Added validation for positive integer configuration properties (RowsFetchedPerBlock, BatchInsertSize, etc.) to prevent hangs and errors when set to zero or negative values.
  • Updated Circuit breaker to be triggered by 429 errors too.
  • Refactored chunk download to keep a sliding window of chunk links. The window advances as the main thread consumes chunks. These changes can be enabled using the connection property EnableStreamingChunkProvider=1. The changes are expected to make chunk download faster and robust.
  • Added separate circuit breaker to handle 429 from SQL Exec API connection creation calls, and fall back to Thrift.

Fixed

  • Fix driver crash when using INTERVAL types.
  • Fix connection failure in restricted environments when LogLevel.OFF is used.
  • Fix U2M by including SDK OAuth HTML callback resources.
  • Fix microsecond precision loss in PreparedStatement.setTimestamp(int,Timestamp, Calendar) and address thread-safety issues with global timezone modification.
  • Fix metadata methods (getColumns, getFunctions, getPrimaryKeys, getImportedKeys) to return empty ResultSets instead of throwing exceptions when catalog parameter is NULL, for SQL Exec API.

Release 3.0.5

21 Nov 07:48
6317c1e

Choose a tag to compare

Added

  • Added support for high-performance batched writes with parameter interpolation:
    • supportManyParameters=1: Enables parameter interpolation to bypass 256-parameter limit (default: 0)
    • EnableBatchedInserts=1: Enables multi-row INSERT batching (default: 0)
    • BatchInsertSize=<SIZE>: Maximum rows per batch (default: 1000)
    • Note: Large batches are chunked for execution. If a chunk fails, previous chunks remain committed (no transaction rollback). Consider using staging tables for critical workflows.
  • Added Feature-flag integration for SQL Exec API rollout
  • Call statements will return result sets in response
  • Add a gating flag for enabling GeoSpatial support: EnableGeoSpatialSupport. By default, it will be disabled

Updated

  • Minimized OAuth requests by reducing calls in feature flags and telemetry.
  • Geospatial getWKB() now returns OGC-compliant WKB values.

Fixed

  • Fix: SQLInterpolator failing to escape temporal fields and special characters.
  • Fixed: Errors in table creation when using BIGINT, SMALLINT, TINYINT, or VOID types.
  • Fixed: PreparedStatement.getMetaData() now correctly reports TINYINT columns as Types.TINYINT (java.lang.Byte) instead of Types.SMALLINT (java.lang.Integer).
  • Fixed: TINYINT to String conversion to return numeric representation (e.g., "65") instead of character representation (e.g., "A").
  • Fixed: Complex types (Structs, arrays, maps) now show detailed type information in metadata calls in Thrift mode
  • Fixed: incorrect chunk download/processing status codes.
  • Shade SLF4J to avoid conflicts with user applications.

v3.0.4

13 Nov 19:15
5057a1a

Choose a tag to compare

Added

  • Added support for telemetry log levels, which can be controlled via the connection parameter TelemetryLogLevel. This allows users to configure the verbosity of telemetry logging from OFF to TRACE.
  • Added full support for JDBC transaction control methods in Databricks. Transaction support in Databricks is currently available as a Private Preview. The IgnoreTransactions connection parameter can be set to 1 to disable or no-op transaction control methods.
  • Added a new config attribute DisableOauthRefreshToken to control whether refresh tokens are requested in OAuth exchanges. By default, the driver does not include the offline_access scope. If offline_access is explicitly provided by the user, it is preserved and not removed.

Updated

  • Updated sdk version from 0.65.0 to 0.69.0

Fixed

  • Fixed SQL syntax error when LIKE queries contain empty ESCAPE clauses.
  • Fix: driver failing to authenticate on token update in U2M flow.
  • Fix: driver failing to parse complex data types with nullable attributes.
  • Fixed: Resolved SDK token-caching regression causing token refresh on every call. SDK is now configured once to avoid excessive token endpoint hits and rate limiting.
  • Fixed: TimestampConverter.toString() returning ISO8601 format with timezone conversion instead of SQL standard format.
  • Fixed: Driver not loading complete JSON result in the case of SEA Inline without Arrow

Releasing 3.0.3 to Maven

29 Oct 19:33

Choose a tag to compare

Fixing regression in increased traffic for Token endpoint

v1.0.9-oss

19 Aug 10:03
ed7483d

Choose a tag to compare

Added

  • Added support for providing custom HTTP options: HttpMaxConnectionsPerRoute and HttpConnectionRequestTimeout.
  • Add V2 of chunk download using async http client with corresponding implementations of AbstractRemoteChunkProvider and
    AbstractArrowResultChunk
  • Telemetry is being enabled by default, with a configuration to opt-out.

Updated

Fixed

  • Fixed Statement.getUpdateCount to return -1 for non-DML queries.
  • Fixed Statement.setMaxRows(0) to be interepeted as no limit.
  • Fixed retry behaviour to not throw an exception when there is no retry-after header for 503 and 429 status codes.
  • Fixed encoded UserAgent parsing in BI tools.
  • Fixed setting empty schema as the default schema in the spark session.