SHARE
Facebook X Pinterest WhatsApp

Data Engineers Spend Two Days Per Week Fixing Bad Data

thumbnail
Data Engineers Spend Two Days Per Week Fixing Bad Data

Data engineers spend 40% of their workweek dealing with incidents relating to poor data quality, which may cost an organization 20% of its revenue.

Written By
thumbnail
David Curry
David Curry
Sep 12, 2022

Data engineers are spending two days every workweek fixing issues with data quality, according to a new survey conducted by Wakefield Research and published by data reliability company Monte Carlo.

The 2022 data quality report, which surveyed 300 data professionals, found that it takes four hours on average to detect an incident, and a further nine hours on average to resolve it. In a month, the average data professional has to deal with 61 incidents. 

SEE ALSO: Observability and Software Supply Chain Security

Not many of the data engineers have formal tracking for how long it takes to detect an incident. In the report, co-founder and CEO at Monte Carlo, Barr Moses, found that engineers with more experience thought they took longer to detect an incident. 

The amount of incidents has also increased rapidly over the past 12 months, with 58 percent of respondents saying that it has greatly or somewhat increased. Less than 20 percent said data incidents had somewhat or greatly decreased over that same period. 

“The first step to improving data quality and trust is measuring it, starting with the number and type of incidents, and setting baselines on response rates and data downtime,” said Shane Murray, CTO at Monte Carlo. 

Teams that ran tests more frequently reported lower levels of data incidents, although respondents did not see the reduction in downtime having a meaningful effect on revenue. It may be suggested that from this data professionals see any downtime from data incidents as a serious impediment, regardless of length. 

Accordingly, data engineers perceive that 26 percent of their organization’s revenue could be harmed by bad data quality. Only five percent of respondents said bad data had no effect on revenue generation. 

“This may seem like a shockingly high number for some, but there are multiple surveys that repeatedly reveal the high cost of poor data quality,” said Francisco Alberini, product manager at Monte Carlo. “For example, Gartner finds bad data costs an organization about $13 million a year. The value of data as a business driver is increasing every year, which also makes the cost of data downtime get more expensive.” 

When discussing the impact of bad data for decision makers and stakeholders, 47 percent of respondents said it had an impact all of the time or most of the time. Less than 25 percent said it rarely or never had an impact. 

“For companies to become data driven there needs to be trust in the data and in the data team,” said Mei Tao, product manager at Monte Carlo. “A shift needs to take place where the data team can catch and resolve data incidents before bad data is acted upon.”

Data observability and other metrics to evaluate and filter bad data could reduce the amount of incidents that impact an organization’s dataflow, thus reducing the costs that come with providing decision makers and stakeholders with inaccurate data.

thumbnail
David Curry

David is a technology writer with several years experience covering all aspects of IoT, from technology to networks to security.

Recommended for you...

Smart Talk Episode 5: Disaggregation of the Observability Stack
4 Things Every Business Must Know About The Application Generation
Gregg Ostrowski
Feb 22, 2024
Full-Stack Observability Improves Uptime, Lessens Outage Cost
David Curry
Oct 9, 2023
Report: Don’t Neglect Open Source Security

Featured Resources from Cloud Data Insights

The Difficult Reality of Implementing Zero Trust Networking
Misbah Rehman
Jan 6, 2026
Cloud Evolution 2026: Strategic Imperatives for Chief Data Officers
Why Network Services Need Automation
The Shared Responsibility Model and Its Impact on Your Security Posture
RT Insights Logo

Analysis and market insights on real-time analytics including Big Data, the IoT, and cognitive computing. Business use cases and technologies are discussed.

Property of TechnologyAdvice. © 2026 TechnologyAdvice. All Rights Reserved

Advertiser Disclosure: Some of the products that appear on this site are from companies from which TechnologyAdvice receives compensation. This compensation may impact how and where products appear on this site including, for example, the order in which they appear. TechnologyAdvice does not include all companies or all types of products available in the marketplace.