Limitations and known issues for data sources in Data Virtualization
The following limitations and known issues apply to data sources in Data Virtualization.
For more information about data sources and connections in, see Supported data sources in Data Virtualization.
For additional solutions to problems that you might encounter with data source connections, see the troubleshooting topic Troubleshooting data source connections in Data Virtualization.- VIRTUALIZETABLE procedure fails on Presto if user only has partial column access
- Connections created using the cpd-cli connection create command can't be found or used
- Using the INSERT operation on a Microsoft Azure Data Lake Storage connection causes error
- Runtime Error displays when you try to find data sources
- Collecting statistics for a virtualized Presto table results in an error
- Unable to connect to a SSL-enabled data source in Data Virtualization by using a remote connector if you use custom certificates
- Querying a virtualized table in a Presto catalog with a matching schema from a previous catalog might result in an error
- Special characters are not supported in MongoDB database names
- Limited file types are supported with the Microsoft Azure Data Lake Storage Gen2 data source connection
- Special characters are not preserved in databases or schemas with MongoDB connections after you upgrade to Data Virtualization on IBM Software Hub
- The DECFLOAT data type is not supported in Data Virtualization
- The Data sources page might fail to load data sources when remote connectors are added, edited, or removed
- Unable to add a connection to SAP S/4HANA data source with a SAP OData connection
- You cannot connect to a MongoDB data source with special characters in a database name
- When you virtualize data that contains LOB (CLOB/BLOB) or Long Varchar data types, the preview might show the columns as empty
- Remote data sources - Performance issues when you create data source connection
- Schema map refresh in-progress message appears for reloaded connections that do not require a refresh schema map
Data source issues
- VIRTUALIZETABLE procedure fails on Presto if user only has partial column access
- Connections created using the cpd-cli
connection createcommand can't be found or used -
Applies to: 5.2.1
- Using the
INSERToperation on a Microsoft Azure Data Lake Storage connection causes error -
Applies to: 5.2.1
When you use theINSERToperation to insert data into a table on a Microsoft Azure Data Lake Storage connection, the operation might fail and cause an error resembling this example:2025-08-06T20:08:16,210 WARN com.ibm.biginsights.bigsql.dfsrw.scheduler.DfsTempDirManager [TThreadPoolServer WorkerProcess-4] SId:32.0-6633: Error calling setPermission on abfs://bigdatacontainer2@bigdataqasa2.dfs.core.windows.net/FVT/dv_fvt_cos_data2124/COSBVT/path5571/_TEMP_1754499267690_320574492_20250906080915760 2025-08-06T20:08:26,720 ERROR com.ibm.biginsights.bigsql.dfsrw.scheduler.DfsBaseWriterCommitHandler [TThreadPoolServer WorkerProcess-4] {bigsql.COMMIT} SId:32.0-6633: Failed to commit for cosbvt.table5571 org.apache.hadoop.fs.FileAlreadyExistsException: Operation failed: "The specified path already exists.", 409, PUT, https://bigdataqasa2.dfs.core.windows.net/bigdatacontainer2/FVT/dv_fvt_cos_data2124/COSBVT/path5571/.COMMITTING___TEMP_1754499267690_320574492_20250906080915760?resource=file&timeout=90, rId: dbf44b18-601f-0090-6e0d-05e4c1000000, PathAlreadyExists, "The specified path already exists. RequestId:dbf22b38-603f-0090-7f0d-07e8s1000000 Time:2025-08-06T20:08:26.6748181Z"Workaround:-
- Primary workaround
- Delete the temporary commit marker file manually and retry the insert. From the previous
example, you would delete the
.COMMITTING___TEMP_1754499267690_320574492_20250906080915760file.
-
- Secondary workaround
- Bypass the use of temporary commit markers by setting
bigsql.insert.temporary.storage.modetoBYPASS. For more information, see Bypassing temporary directories to increase the performance of insert operations into object stores.
-
Runtime Errordisplays when you try to find data sources
- Collecting statistics for a virtualized Presto table results in an error
- Unable to connect to an SSL-enabled data source in Data Virtualization by using a remote connector if you use custom certificates
-
Applies to: 5.2.0 and later
- Querying a virtualized table in a Presto catalog with a matching schema from a previous catalog might result in an error
-
Applies to: 5.2.0 and later
- Special characters are not supported in MongoDB database names
- Limited file types are supported with the Microsoft Azure Data Lake Storage Gen2 data source connection
-
Applies to: 5.2.0 and later
- Special characters are not preserved in databases or schemas with MongoDB connections after you upgrade to Data Virtualization on IBM Software Hub
-
Applies to: 5.2.0 and later
- The DECFLOAT data type is not supported in Data Virtualization
- Applies to: 5.2.0 and later
- The Data sources page might fail to load data sources when remote connectors are added, edited, or removed
-
Applies to: 5.2.0 and later
- Unable to add a connection to SAP S/4HANA data source with a SAP OData connection
-
Applies to: 5.2.0 and later
- You cannot connect to a MongoDB data source with special characters in a database name
-
Applies to: 5.2.0 and later
- When you virtualize data that contains LOB (CLOB/BLOB) or Long Varchar data types, the preview might show the columns as empty
-
Applies to: 5.2.0 and later
- Remote data sources - Performance issues when you create data source connection
-
Applies to: 5.2.0 and later
- Query fails due to unexpectedly closed connection to data source
-
Applies to: 5.2.0 and later
- Schema map refresh in-progress message appears for reloaded connections that do not require a refresh schema map
Applies to: 5.2.0 and later
The Schema map refresh in-progress message appears when you reload connections in Data Virtualization, even when the data source does not require a refresh schema map.