Last week, I heard a story about a VP who greenlit a $200,000 marketing campaign based on what looked like rock-solid customer data. Three weeks in, they noticed something odd: conversions were 60% lower than projected.

Turns out, the report was counting duplicate customer IDs. The "new customer surge" they thought they had? Just the same people showing up twice in the spreadsheet.

The Real Problem

The campaign wasn't the problem. The data was. And here's the scary part: the numbers looked perfectly fine. Clean columns. Nice charts. Professional dashboard. No red flags.

Bad data doesn't announce itself.

It sits there, quietly, looking legitimate until you make a decision that costs real money, wastes real time, or embarrasses you in front of your team.

So how do you avoid this?

You run checks. Simple ones. Before you trust any data, before you build a strategy around it, and definitely before you spend a dollar on it.

Here are the 10 checks that will save you from making decisions on bad data.

Wake-Up Call

Bad data costs organizations an average of $15 million annually, yet 47% of newly created data records have at least one critical error.

Why Most Data Problems Are Silent

The biggest misconception about data quality is that bad data will look bad.

It won't.

A spreadsheet with missing values doesn't crash. A dashboard with incorrect dates doesn't throw an error. A report with duplicates doesn't come with a warning label.

Everything looks normal until someone asks a question the data can't answer correctly.

That's when you find out your "top-performing region" was actually just reporting data twice. That's when you realize your "customer growth rate" included test accounts. That's when you discover the pricing column has been pulling from last quarter's data for three months.

Data quality issues hide in plain sight.

And the people who spot them early? They're not doing anything magical. They're just running the same checks, every time, before they trust what they're looking at.

Let's walk through them.

The Data Quality Mindset

Before we dive into the checks, remember this: Your job isn't to trust data. Your job is to verify it. Every dataset is guilty until proven innocent. This mindset alone will save you from more bad decisions than any tool or technique.

The 10 Data Quality Checks

1. Check for Missing Values

What it means: Look for blank cells, nulls, or placeholders like "N/A" or "TBD."

Why it matters: Missing data skews averages, throws off counts, and makes analysis unreliable. If 30% of your revenue column is blank, your revenue total is wrong.

Real Example: You're analyzing customer support response times. You filter for "average response time under 2 hours" and feel great about your team's performance. But then you notice 40% of tickets don't have a response time logged at all. Those are probably the ones that took 6 hours. Your "fast response" metric just became meaningless.

2. Check for Uniqueness

What it means: Make sure things that should be unique actually are. Customer IDs, order numbers, email addresses.

Why it matters: Duplicates inflate your counts and make metrics like "total customers" or "total sales" completely unreliable.

Real Example: You pull a list of 5,000 customers who bought in Q4. Looks great until you realize 1,200 of them appear twice because someone merged two systems and didn't deduplicate. Your actual customer count? 3,800. That's a 24% difference, and every decision you make on "5,000 customers" is wrong.

3. Check for Broken Relationships

What it means: Make sure data that's supposed to connect actually does. Orders should have customers. Invoices should have line items. Transactions should have accounts.

Why it matters: Broken links create orphan records that either get ignored (undercounting) or cause confusion (mismatching).

Example: You're reviewing sales by sales rep. You see 300 orders with no rep assigned. Are those self-service? Web orders? Or is it a data entry problem? If you don't know, your "top performer" ranking is useless, and your commission calculations might be wrong.

Quick Tip

When you find broken relationships, don't just fix the symptoms. Ask why they broke in the first place. Often, it reveals a process problem that will keep creating bad data until you address the root cause.

4. Check for Accepted Values

What it means: Fields that should have a limited set of options (like status: "active," "inactive," "pending") shouldn't have random values like "activ" or "Pending" or "TBD."

Why it matters: Typos and inconsistencies break your filters and counts. If you filter for "active" but some records say "Active" or "ACTIVE," you'll miss data.

Real Example: You're counting how many projects are "Complete." The system shows 120. But someone also entered "Completed," "Done," and "Finished" in other rows. Your real count? 180. That's a 50% undercount, and it makes your team look slower than they actually are.

5. Check Business Logic

What it means: Does the data make sense in the real world? Can someone have a purchase date before their account was created? Can a refund be larger than the original sale?

Why it matters: Data can be "technically correct" but logically impossible. These contradictions expose errors earlier in the process.

Real Example: You see a customer who signed up on January 15th but made a purchase on January 10th. That's impossible. Either the signup date is wrong, the purchase date is wrong, or this customer record was merged from another system incorrectly. Until you fix it, any "time to first purchase" analysis is broken.

6. Check for Reasonable Ranges

What it means: Numbers should fall within expected boundaries. Ages between 0-120. Prices above $0. Percentages between 0-100%.

Why it matters: Outliers that break reality signal data entry errors or system glitches. A single bad value can ruin an average or total.

Example: You're calculating average order value. It's coming out to $8,400, which feels high. You dig in and find one order for $500,000. Turns out someone added an extra zero during data entry. That one mistake just made all your AOV analysis useless.

The Single Bad Value Problem

One outlier can destroy an average. One negative number where there should be none can throw off a total. Always check the extremes, not just the middle.

7. Check Data Types

What it means: Make sure numbers are stored as numbers, dates as dates, and text as text. A price stored as text can't be summed. A date stored as text can't be sorted correctly.

Why it matters: The wrong data type breaks calculations, sorting, and filtering. You'll get weird results and not know why.

Technical Trap

Example: You try to calculate total revenue and get an error. Turns out the revenue column has values like "$1,200" (with the dollar sign). That's text, not a number. Your spreadsheet can't add it. You'll need to clean every single row before your totals work.

8. Check Freshness

What it means: Is the data current? When was it last updated? Are you looking at today's numbers or last month's?

Why it matters: Stale data leads to decisions based on outdated reality. The world moved on, but your data didn't.

The Stale Data Problem

A study found that 27% of business decisions are made on data that's more than a week old. In fast-moving industries, that's ancient history.

Example: You pull a dashboard showing "current inventory levels." You decide to order more product based on low stock. But the dashboard hasn't refreshed in three days. The warehouse actually received a shipment yesterday. Now you're about to over-order and tie up cash in inventory you don't need.

9. Check Time Consistency

What it means: Make sure time-based data is measured consistently. Are all dates in the same timezone? Are all timestamps using the same format?

Why it matters: Inconsistent time tracking creates false patterns and makes trend analysis unreliable.

Example: You're tracking daily sign-ups across multiple regions. Your chart shows a massive spike every day at 5 PM. Is it a real pattern? Nope. Turns out the European data is in UTC, but the US data is in EST. The "spike" is just a timezone mismatch creating a phantom trend.

Time Zone Trap

Time inconsistencies are one of the most common and hardest-to-spot data quality issues. Always verify that timestamps across different systems use the same timezone, or clearly document when they don't.

10. Check for Sudden Changes

What it means: Look for spikes, drops, or shifts that don't match reality. Did your traffic triple overnight? Did your conversion rate suddenly drop to zero?

Why it matters: Dramatic changes are usually data errors, not business changes. Trusting them leads to panic or overconfidence.

Example: Your sales dashboard shows a 90% drop in revenue yesterday. You freak out. But then you check and realize the payment processor had an outage, so yesterday's sales didn't sync yet. The drop isn't real. If you'd acted on that number, you might have made a bad call.

Real-World Win

A retail client running these 10 checks discovered that their "declining sales" were actually a reporting bug that double-counted returns. Once fixed, they realized sales were actually up 12%, which completely changed their inventory strategy and saved them from cutting marketing spend at the worst possible time.

The Big Picture: Data Trust is Earned, Not Assumed

Here's what most people miss about data quality: good-looking data isn't the same as trustworthy data.

A dashboard can have beautiful charts and still be built on garbage. A spreadsheet can be perfectly formatted and completely wrong. A report can have executive-level polish and lead you to a terrible decision.

These 10 checks aren't about being paranoid. They're about being smart.

They help you catch the silent errors that cost money, waste time, and destroy credibility. They turn "this looks fine" into "I've verified this is fine."

And the best part? They don't require technical skills. You don't need to write code or learn SQL. You just need to ask the right questions:

Your Data Quality Checklist:

  • Missing values: Are there gaps in critical columns?
  • Uniqueness: Are there duplicates where there shouldn't be?
  • Relationships: Do the connections between data make sense?
  • Valid values: Are all entries using accepted formats and options?
  • Business logic: Does the data reflect reality?
  • Ranges: Are the values within reasonable boundaries?
  • Data types: Is everything stored correctly?
  • Freshness: Is this data current and up-to-date?
  • Time consistency: Are timestamps and timezones aligned?
  • Sudden changes: Do dramatic shifts match real events?

The 5-Minute Rule

Before making any significant decision based on data, spend 5 minutes running through these checks. It's a small investment that can save you from massive mistakes. Think of it as insurance against bad decisions.

The people who run these checks don't have more data than you. They just trust it less until it proves itself.

Don't Mistake Convenience for Accuracy

Data is everywhere now. Dashboards auto-refresh. Reports generate instantly. Analytics tools spit out insights in seconds.

It's easy to assume that because something is automated, it must be accurate.

It's not.

Automation makes bad data faster. It doesn't make it better.

The next time you're about to make a decision based on data, pause. Run these checks. Ask the uncomfortable questions. Look for the things that don't add up.

Because the cost of bad data isn't just a wrong number. It's a wasted campaign. A missed opportunity. A strategy built on sand.

Good decisions deserve good data.

Don't settle for anything less.

Need Help With Your Data Quality?

We help businesses implement automated data quality checks that catch errors before they cost you money. From simple spreadsheet validation to enterprise-wide monitoring systems.

Schedule a Free Data Audit