Making Data Relevant Again

Metadata Management

Data Governance

Data Discovery

Data Management

Data Quality

Data Catalog

Swaroop Jagadish

Jan 29, 2024

Metadata Management

Data Governance

Data Discovery

Data Management

Data Quality

Data Catalog

Picture This

You are a risk analyst with a financial services organization. You’re part of a team that creates products and services for small- and medium-sized business (SMB) customers.

But … your team’s efforts are constrained by poor-quality data. You struggle to identify and integrate the data you need to analyze SMB market segmentation and to model product viability and financial risk. The data isn’t where it should be, isn’t what it should be, isn’t in the expected format, and is riddled with anomalies.

This is a “fictional” scenario, but it’s also not-so-fictional, because many organizations can identify with it. Increasingly, decision-makers and stakeholders just don’t trust their data and analytics—usually because what they’re seeing is out-of-date, incomplete, inconsistent, and sometimes flat-out wrong.

Relevant Data Is Difficult to Find

There are four common causes of this. First, data is siloed across a myriad of cloud applications, services, and resources—Databricks, Snowflake, and BigQuery, for example, or cloud storage resources like S3 or Google Cloud Storage. There’s also data in the on-premises environment, which is often siloed in legacy systems. Thanks to all of this siloing, it’s difficult to ask simple questions like “Where is my customer data?”—or complex analytic questions like “Who are my customers?”—because the relevant data lives in so many different places and is structured according to different schemas.

In the best-case scenario, you know where your data is, you know what it is, you know where it came from, you have some idea of what format it’s in, and you also know what’s wrong with it.

Sadly, the reverse is usually the case. The best-case scenario almost never happens.

Relevant Data Lacks Essential Context

In fact, for many organizations, knowledge of all five of these things is too frequently a luxury.

They can’t pinpoint, exactly, the applications or resources in which all of their relevant customer data is siloed. And even if they can, this data isn’t always understandable—because it lacks context.

This is the second reason people don’t trust their data or analytics.

What is it? Where did it come from? What was done to it? By whom? For what purpose? Is it first-party data, created by an internal producer, or is it external, generated by a third party? Is it in any way sensitive? For example, are the values recorded in the `ID` column `device_ID`s or `customer_ID`s? The former could be sensitive, the latter probably aren’t. Without context, you just don’t know.

Too often, people who want to discover and work with data … just don’t know.

Relevant Data Is Nondescript

The third reason has to do with timeliness. When you’re dealing with siloed datasets, even when you know something about the history of a specific dataset, you don’t always know just how “historical” it is.

In other words, how fresh is it? Is it the product of a routine batch process, or is it a one-off thing—i.e., created at a specific time for a specific purpose? If it’s the product of a batch process, when was it last updated? How frequently is it supposed to be updated? If it hasn’t been updated, why not? Timely data isn’t just critical for the accuracy and reliability of operational analytics and executive dashboards: a lack of timely data (because of data outages, or delays in preparing and integrating data) can also lead to missed opportunities, paralyzed decision-making, poor customer service, and other bad outcomes.

Any one of these sound familiar?

The fourth reason has to do with semantics. Let’s say that you are dealing with high-quality datasets. Can you meaningfully compare attributes, entities, or definitions across them? For example, does `last_purchase_date` in Dataset A mean the same thing as `RecentPurchase` in Dataset B? With a little investigative effort, you can usually answer this question. But shouldn’t the answer be obvious?

For data practitioners, it usually isn’t.

Relevant Data and the Challenge of AI

The issues I’ve raised have particular salience for organizations pivoting to take advantage of AI.

Machine intelligence requires access to high-quality, context-rich data. Full stop. In undergraduate computer science classes, we still teach students the truism “Garbage in, garbage out” because we want to emphasize the importance of quality input data. But in ML, “Garbage in, garbage out” isn’t just a truism, it’s also an Iron Law. Train a model on poor-quality data, and you’ll get erroneous, biased, and/or totally unpredictable outputs. Feed a production model low-quality data and you’ll get inconsistent, imprecise, inaccurate results. In AI, everything depends on the quality of input data.

But there’s one another critical factor to consider….

Relevant Metadata for AI Transparency, Reproducibility, and Performance

In the AI Era, having high-quality data and metadata are more important than ever.

For one thing, companies need to collect and track the metadata created during LLM fine-tuning and prompt construction to demonstrate that their models are transparent, reproducible, ethical, and compliant with regulations.

For another, having access to high-quality metadata is essential for improving the accuracy and precision of AI models—including LLMs. Metadata helps identify data biases and gaps—are certain demographic groups over-represented in the dataset? Is the data overwhelmingly derived from one source?—which is critical for training models that are accurate, precise, fair, and aligned with human values. Metadata also provides a baseline for creating new datasets for traditional ML model retraining, and helps accelerate data preprocessing and feature engineering. Finally, metadata allows for experiment tracking and version control, improving the reproducibility and reliability of ML models.

Lack of high-quality data is the main reason organizations struggle to create and operationalize LLM prototypes that perform accurately and reliably in production. And lack of quality metadata is one of the main reasons organizations struggle to maintain and improve the performance of production LLMs.

When Relevant Data Isn’t Relevant, or: The Status Quo

One last thing to keep in mind is that data work is very much a team sport, requiring not only coordination and collaboration but also an irreducible element of trust.

Teammates must trust not only one another, but also the ground on which they’re playing.

Like a soccer pitch, data is the ground truth that defines the boundaries of the game and provides an orienting context for all players. Its condition has a material impact on how the game is played.

Or whether it can be played at all.

The problem is that people too often don’t want to play … because they don’t trust their data, either because it’s of poor quality or because it just isn’t relevant for them. How could it be? In order to discover relevant data, consumers have to learn to use third-party tools that force them to switch outside of their preferred workflows.

Technical and logistical barriers—using unfamiliar tools, learning unfamiliar languages, or jumping through process-based hoops to gain access to useful data—absolutely discourage business analysts and other consumers from tracking down relevant data that they could use to enrich their analyses.

But the thing that people forget is that “relevance” is also a function of comfort, familiarity, and value.

By this standard, we fail to make data relevant for consumers whenever we silo it in a system or fail to describe what it is, when it was created, where it came from, what it’s for, and why it’s important. We fail to make data relevant for consumers whenever we erect process-related hurdles or hoops to accessing and using it. Unfortunately, by these criteria, most data, in most organizations, is irrelevant.

It’s irrelevant because it’s too much trouble for people to access it. The value just isn’t there for them.

Acryl Cloud: Making Data Relevant Again

This is why Acryl Data exists. In Acryl Cloud, we deliver a reliable, trustworthy metadata platform—built on top of the thriving open-source DataHub project—that data practitioners can use to easily search for and discover data, as well as quickly answer questions like:

What kind of data is this?
Where did it come from?
When was it created?
By whom or what?
For what purpose?
What other data does it relate to or depend on?
What other data can it be compared to?

And we’re marrying these capabilities with best-in-class automated features—like metadata-driven governance, with the ability to continuously monitor and enforce policy-based governance standards.

Intrigued? Transform your data operations with Acryl Cloud, the best-in-class metadata platform that makes your data relevant again: discoverable, accessible, usable, understandable—and governable.

Metadata Management

Data Governance

Data Discovery

Data Management

Data Quality

Data Catalog


Governing the Kafka Firehose

Kafka’s schema registry and data portal are great, but without a way to actually enforce schema standards across all your upstream apps and services, data breakages are still going to happen. Just as important, without insight into who or what depends on this data, you can’t contain the damage. And, as data teams know, Kafka data breakages almost always cascade far and wide downstream—wrecking not just data pipelines, and not just business-critical products and services, but also any reports, dashboards, or operational analytics that depend on upstream Kafka data.

When Data Quality Fires Break Out, You're Always First to Know with Acryl Observe

Acryl Observe is a complete observability solution offered by Acryl Cloud. It helps you detect data quality issues as soon as they happen so you can address them proactively, rather than waiting for them to impact your business’ operations and services. And it integrates seamlessly with all data warehouses—including Snowflake, BigQuery, Redshift, and Databricks. But Acryl Observe is more than just detection. When data breakages do inevitably occur, it gives you everything you need to assess impact, debug, and resolve them fast; notifying all the right people with real-time status updates along the way.

John Joyce


Five Signs You Need a Unified Data Observability Solution

A data observability tool is like loss-prevention for your data ecosystem, equipping you with the tools you need to proactively identify and extinguish data quality fires before they can erupt into towering infernos. Damage control is key, because upstream failures almost always have cascading downstream effects—breaking KPIs, reports, and dashboards, along with the business products and services these support and enable. When data quality fires become routine, trust is eroded. Stakeholders no longer trust their reports, dashboards, and analytics, jeopardizing the data-driven culture you’ve worked so hard to nurture

John Joyce


Get started with Acryl today.
Acryl Data delivers an easy to consume DataHub platform for the enterprise
See it in action
Acryl Data Logo
Acryl DataHub
Acryl ObserveCustomer Stories
© 2024 Acryl Data