Skip to main content

Release notes 25.3

· 13 min read

Dataedo version 25.3.0 is out! Take a look at this release's highlights:

  • Dataedo Browser Extension
  • Import Metadata in Portal - Amazon Redshift, Azure SQL Database, MySQL, Oracle, PostgreSQL, Power BI Reports and Dashboards, Snowflake, SQL Server
  • Data Quality - Configurable thresholds, Reusable rules, Rules on views, Bulk assignment of rules, Bulk threshold editing, Expanded rule suggestions, Unified metrics everywhere, New dashboard filters, Custom failed rows count for SQL rules, Operational Dashboard enhancements
  • Scheduler - Delete data sources and connections in Portal, In progress task indicator, Deprecated connectors label
  • Connectors - Improvements for Databricks, AWS Athena, Redshift, Power BI, SSRS, SSIS, SSAS Tabular, Snowflake Enterprise Lineage, PostgreSQL and MySQL.
  • Dataedo Agent - New cross-platform application, New Windows installer
  • Steward Hub - Foreign key suggestions, Interactive Data Quality suggestions
  • UX/UI - Smarter navigation, Consistent layouts, Refreshed components, Filters in URLs
  • Login - Email field synchronization for SAML, Default login method fix
  • Other improvements - Task scheduling via Public API, New repository creation option in Portal

Dataedo Browser Extension

Now you can bring Dataedo right into your BI workflow with our new Browser Extension.
It allows you to see key catalog details (such as descriptions, glossary terms, owners, certification status, or lineage) directly inside your Power BI and Tableau dashboards. Thanks to this, you can better understand and trust your data without leaving the reports you’re working on.

This means less context switching and quicker insights, as all the critical business context is available exactly where you need it most.

Report in browser extension

Read the full guide in our documentation

Import Metadata in Portal

Until now, metadata import was only available through Dataedo Desktop. Starting with this release, you can perform imports directly in the Portal – initially for a selected set of connectors:

Import in Portal connectors list

Import multiple sources from one host

A major improvement is the ability to import multiple sources from the same host in one go. Previously, you had to repeat the entire import process for each source separately, and each one created its own Data Source. Now, you can import multiple sources at once – each will still be created as a separate Data Source, but all of them will belong to the same Connection.

Multiple sources

Connections and shared credentials

This is also the moment to highlight the Connections concept: they group shared settings such as credentials, host, and port, making it much easier to manage configurations across multiple sources. Multiple connections can reuse the same credentials, so when login details change, you only need to update them in one place. With this release, imported sources that share the same host will also share the same credentials within a single connection.

Permissions

Metadata import in the Portal can only be performed by users with the Connection Manager role. To start an import, the user needs to open Connections in the Portal and click Add connection.

Add connection

Dataedo Desktop support

Importing with these connectors is still fully supported in Dataedo Desktop. We are extending functionality to the Portal, but not removing any existing options.

Data Quality

This release brings a number of new features and enhancements that make it easier to manage and monitor data quality:

Configurable thresholds

By default, Data Quality rule instances failed if even a single row did not meet the rule’s condition. Now, you can define a custom threshold (e.g., allow up to 5% invalid rows) to make rules less strict where appropriate. Read more in the documentation

DQ thresholds

Reusable rules

In addition to single-use rule instances or custom SQL rules, you can now create reusable Data Quality rules and later assign them to any object (table, view, or column). This reduces duplication, ensures consistency, and makes it easier to manage your Data Quality setup. Only users with the Power Data Steward role at the repository level can create reusable rules, ensuring that only authorized users can define queries that access real data across connected sources. Read more in the documentation

Reusable DQ rule list

Rules on views

Data Quality rules can now be assigned not only to tables and their columns, but also to views and their columns.

Bulk assignment of rules

Assign the same rule to up to 50 objects at once (e.g., multiple columns). This creates individual rule instances that can be managed separately later.

Bulk assign rule

Bulk threshold editing

Thresholds can now be updated for multiple rule instances at once from any Data Quality list view. This is especially helpful for existing customers who would otherwise need to update instances one by one.

Bulk edit threshold

Expanded rule suggestions

We now suggest a wider set of rules based on column names and context to speed up the manual process of assigning them:

CategoryColumn name containsSuggested rules
EmailmailValid email address
Government IDsssn, nip, pesel, einUS SSN, Polish NIP, Polish PESEL, US EIN
Datescreated_at, updated_at, timestamp, expiry_dateFuture dates exist, Not in future, Is fresh, No suspicious dates
URLs & IPsurl, website, link, ip_addressValid HTTP address, Valid IPv4 address
Postal codeszip_codeValid US zip code
Identifiers & product codesasin, iban, imei, vin, cusip, dunsValid ASIN, Valid IBAN, Valid IMEI, Valid VIN, Valid CUSIP
JSON / Metadatajson, payloadValid JSON

Unified metrics everywhere

Both Data Quality score and Data Quality ratio are now displayed consistently across the Portal, including the Data Quality dashboard. This provides a more complete picture of data quality at every level.

DQ dashboard

New dashboard filters

The Data Quality dashboard now supports filtering by Data Source, Business Domains, and Data Domains for more focused analysis.

DQ filters

Custom failed rows count for SQL rules

When creating a custom SQL rule (for specific use cases), you can now manually provide a query that returns the number of failed rows. This is useful in situations where the default wrapping logic (SELECT COUNT(*) FROM (...)) would fail due to subquery limitations or complex SQL.

Operational Dashboard enhancements

We added an optional column Instance Description in both the Operational Dashboard and Runs History. Once enabled, it remains visible for easier tracking.

Scheduler

Delete data sources and connections in Portal

It is now possible to remove Data Sources and Connections directly in the Portal, without having to use Dataedo Desktop.

Delete connection
Delete source

In progress task indicator

When a task (Metadata Import, Refresh Profiling, or Data Quality check) is currently processing, this status is now visible both in the connection header and in the Scheduler task list.

Task in progress

Deprecated connectors label

Connectors that are no longer supported (e.g., the old dbt integration) are now marked with a Deprecated label. Importing from these sources is no longer possible, but previously imported data remains available.

Depreciated source

Connectors

Databricks

The Databricks connector now supports data profiling and data quality features. We have optimized the lineage extraction process – instead of calling the Databricks API for each column, lineage is now retrieved from system tables.

New requirement – warehouse

The connector now requires providing warehouse information during connection setup.

AWS Athena

The AWS Athena connector now supports data profiling. We also improved the handling of external tablesscripts and locations are now imported, and data lineage to source files is created automatically.

Redshift

The Redshift connector now supports data quality.

Power BI

We improved lineage when Power BI calls stored procedures to fetch data.

Power BI lineage

In addition, the import process is now noticeably faster. Datasets and dataset tables are also linked through a parent–child relationship.

SQL Server Reporting Services (SSRS)

Dataedo now supports creating lineage when a stored procedure is called to fetch data. We’ve also developed a simple app that imports report usage statistics. Please contact support if you would like to use it (requires direct connection to the ReportServer database).

SQL Server Integration Services (SSIS)

Lineage is now created when the source SQL script is stored in a variable (SQL Command from variable option). We also improved lineage when a stored procedure is called to fetch data.

SSIS lineage

SQL Server Analysis Services (SSAS) Tabular

You can now import the IsHidden property. Lineage is built correctly when column names are modified in the M Language script stage. We also improved lineage when a stored procedure is called to fetch data.

SSAS lineage

Snowflake Enterprise Lineage

We introduced behavioral changes to make imported script objects easier to document. Scripts with the same parameterized query hash are now grouped under a single object. This prevents documentation from being cluttered by cyclically executed scripts with different parameters.

First import changes after Dataedo upgrade

If you have already used the Snowflake Enterprise Lineage connector, please note that after upgrading Dataedo, the first import changes will remove duplicate objects (with the same parameterized query hash). Only the most recently imported object will be kept.

PostgreSQL

Added import of SQL Query objects for functions with RETURNS TABLE or RETURNS SETOF. These objects:

  • include the data-returning statement in the Script tab,
  • have a parent–child relationship,
  • have data lineage to the source tables.

MySQL

Added import of SQL Query objects for procedures that contain a SELECT statement returning data. These objects:

  • include the data-returning statement in the Script tab,
  • have a parent–child relationship,
  • have data lineage to the source tables.

Dataedo Agent

New cross-platform application

Dataedo Agent is now a completely new, cross-platform application, no longer tied to Dataedo Desktop. It can be installed with Docker on Linux, macOS, and Windows – just like the Dataedo Portal.

Agent on docker

New Windows installer

For Windows users, the old console-based installer has been replaced with an improved graphical installer that makes it easier to install the Agent as a Windows Service. It also simplifies upgrading from the old Agent to the new application. Upgrading from an existing Agent installation will automatically replace it with the new application and handle all required adjustments.

Steward Hub

The Steward Hub continues to evolve into a powerful assistant for managing and improving your metadata.

Foreign key suggestions

A new section now highlights potential foreign keys for each table. Users can review these suggestions, confirm them through a verification popup, and assign a name to the new relationship before saving.

FK suggestions

Interactive Data Quality suggestions

Suggestions in the Data Quality area are no longer read-only. Instead of re-creating them manually, you can now accept them directly in the interface for a smoother workflow.

DQ suggestions

UX/UI

This quarter we introduced a series of improvements to make the Dataedo Portal more intuitive, consistent, and pleasant to use:

Smarter navigation

The sidebar now expands on hover to show all section names and collapses automatically when you move away. Even in the collapsed view, your current section stays highlighted, saving space without losing orientation.

new nav

Consistent layouts

All pages now follow a unified layout. For example, domains adopt the same look and feel as data products, including counters of linked objects for a clear overview.

Refreshed components

Tags, badges, and checkboxes received a redesign for a cleaner and more consistent look across the app.

Filters in URLs

When copying a link, applied filters, sorting, and pagination are now included in the URL, making it easier to share exact views with others.

Login

We’ve refined authentication to make login more reliable and better aligned with user management:

Email field synchronization for SAML

When logging in with SAML, the email value from the SAML response is now stored not only in the login field but also in the email field in Dataedo. For existing users, the system automatically fills in the missing email field during login if it was previously empty.

Default login method fix

We corrected an issue from the previous release. It is now possible to select only one default login method. In the case of SAML, this means that only a single provider can be set as default, instead of the entire login method.

Other improvements

We also introduced several technical enhancements and options requested by our users:

Task scheduling via Public API

In version 25.2 we removed the ability to schedule tasks (Metadata Import, Refresh Profiling, Data Quality) via command line. This functionality is now restored in a new and more flexible way: through the Public API.

  • GET /public/v1/scheduler/tasks – returns all scheduled tasks with their IDs.
  • POST /public/v1/scheduler/tasks/{taskId}/trigger – triggers a specific task by ID.
    Tasks triggered this way also appear in Schedule → Calendar in the Portal. In the details, they are marked with: This task was triggered manually via Public API.

New repository creation option

Until now, in Portal creating a new repository meant either letting Dataedo generate a fresh one, or connecting to an existing repository. We have now added a third option: you can prepare your own empty database, and Dataedo will create all the necessary tables and structures inside it. This gives you full flexibility to choose the exact database type, pricing tier, and configuration you prefer.