The analytics platform provider joins a long list of vendors who already support the open table formats. Credit: Stokkete/Shutterstock Teradata is adding support for two open table formats, Apache Iceberg and Linux Foundation’s Delta Lake, to its multi-cloud analytics platform VantageCloud Lake and its AI and machine learning engine AI Unlimited. Typically, open table formats are architected to generate performance for data lakes using cloud-based object storage. The performance is achieved by creating a layer of abstraction atop a data lake via the use of columnar storage and metadata management that allows enterprises to manage and update data more efficiently. The fundamental advantage of using an open table format is that enterprises can modify their data schema or partitioning strategy without having to reprocess the entire dataset. A number of Teradata’s rivals, including providers of cloud-based analytics and software such as Snowflake, Starburst, Dremio, Cloudera, and Clickhouse, already support Apache Iceberg. The Linux Foundation’s Delta Live tables format is supported by the likes of Google Cloud, AWS, and Databricks. The addition of support for the open table formats will, according to Teradata, result in its customers being able to allow cross-read and cross-write data stored in multiple open table formats. This interoperability extends to AWS Glue, Unity, and Apache Hive catalogs and works in multi-cloud and multi-data lake environments, the company said, adding that support for the open table formats will be available for VantageCloud Lake and AI Unlimited on AWS and Azure in June 2024. AI Unlimited will be available for purchase under public preview on the AWS and Azure Marketplaces in the second quarter of the year. Teradata is also integrating third-party tools such as Airbyte Cloud, Apache Airflow, and dbt. The Airbyte Cloud integration will help streamline data ingestion into VantageCloud with a fully managed and hosted service that eliminates the need for time consuming infrastructure setup and management, while the Apache Airflow integration will allow enterprise teams to programmatically author, schedule, and monitor workflows. The dbt tool integration, on the other hand, helps manage the transform part of the extract, load, and transform (ETL) process. It can be used as a tool for data transformation in databases, data lakes, and data warehouses, the company said, adding that all the integrations have already made generally available. Related content feature Dataframes explained: The modern in-memory data science format Dataframes are a staple element of data science libraries and frameworks. Here's why many developers prefer them for working with in-memory data. By Serdar Yegulalp Nov 06, 2024 6 mins Data Science Data Management analysis Cloud providers make bank with genAI while projects fail Generative AI is causing excitement but not success for most enterprises. This needs to change quickly, but it will take some work that enterprises may not be willing to do. By David Linthicum Nov 05, 2024 5 mins Generative AI Cloud Computing Data Management feature Overcoming data inconsistency with a universal semantic layer Disparate BI, analytics, and data science tools result in discrepancies in data interpretation, business logic, and definitions among user groups. A universal semantic layer resolves those discrepancies. By Artyom Keydunov Nov 01, 2024 7 mins Business Intelligence Data Management feature Bridging the performance gap in data infrastructure for AI A significant chasm exists between most organizations’ current data infrastructure capabilities and those necessary to effectively support AI workloads. By Colleen Tartow Oct 28, 2024 12 mins Generative AI Data Architecture Artificial Intelligence Resources Videos