What Is Scientific Debt?
What Is Scientific Debt?
You've probably heard of technical debt, but data scientists have to contend with scientific debt. David Robinson explains what it is and how to contend with it.
Join the DZone community and get the full member experience.Join For Free
How to Simplify Apache Kafka. Get eBook.
A very useful concept in software engineering is technical debt.
Technical debt occurs when engineers choose a quick but suboptimal solution to a problem or don't spend the time to build sustainable infrastructure. Maybe they're using an approach that doesn't scale well as the team and codebase expand (such as hardcoding "magic numbers"), or they're using a tool for reasons of convenience rather than appropriateness ("We'll write the DevOps infrastructure in PHP since that's what our team already knows."). Either way, it's something that seems like it's working at first but causes real challenges in the long-term in the form of delayed feature launches and hard-to-fix bugs.
In my new job as Chief Data Scientist at DataCamp, I've been thinking about the role of data science within a business and discussing this with other professionals in the field. On a panel earlier this year, I realized that data scientists have a rough equivalent to this concept: scientific debt.
Scientific debt is when a team takes shortcuts in data analysis, experimental practices, and monitoring that could have long-term negative consequences.
When you hear a statement like:
- "We don't have enough time to run a randomized test; let's launch it."
- "To a first approximation, this effect is probably linear."
- "This could be a confounding factor, but we'll look into that later."
- "It's directionally accurate, at least."
...you're hearing a little scientific debt being "borrowed."
Most engineers have a sense of what it's like for a company to struggle with technical debt. What would a company struggling with scientific debt look like?
Imagine a small startup, WidgetCorp, is developing a B2B product, and deciding on their sales strategy. One year, they decide to start focusing their sales efforts on larger corporate clients. They notice that as they take on this new strategy, their monthly revenue increases. They're encouraged by this, and in the following years, hire half a dozen salespeople with experience working with large clients and spend marketing and design effort building that as part of their brand.
Years later, the strategy doesn't seem to be paying off: their revenue is struggling and the early successes aren't repeating themselves. They hire an analyst who looks at their sales data and finds that, in fact, it had never been the case that they'd had a higher return-on-investment selling to large companies. In that early year, their revenue had been rising because of a seasonal effect (the demand for widgets goes up in the fall and winter), which was compounded with some random noise and anecdotes (e.g. "SmallCompany.com was a waste of our time, but we just closed a huge deal with Megabiz!")
WidgetCorp took on too much scientific debt.
Some ways this might have happened:
They made irreversible decisions based on flawed analyses. It's reasonable to take a quick look at metrics and be happy that they're going in the right direction. But once the company made product, sales, and marketing changes, it became difficult to reverse them. Before making a major shift in business, it's worth making sure that the data supports it: that they've accounted for seasonal effects and applied proper statistical tests.
Lack of monitoring. Early on, there may not have been enough data to tell whether larger clients were contributing to better sales. But as more data was collected, it would be worth continually testing this assumption in the form of a dashboard or a quarterly report. If this isn't tracked, no one will notice that the hypothesis was falsified, even once they have the data.
Lack of data infrastructure. Maybe early in the company, the leads were locked in a sales CRM, while accounting data was stored in Excel spreadsheets that were emailed around. Even if there were a dedicated analyst within the company, they may not have easy access to the relevant data (linking together sales success and company size). Even if it were theoretically possible to combine the datasets with some effort, schlep blindness might have made everyone avoid the analysis entirely. This is an area where technical debt and scientific debt often appear together since it takes engineering effort to make scientific problems easy to solve.
Spreading inaccurate lore. Suppose that the WidgetCorp CEO had given a series of company-wide talks and public blog posts with the message, "The future of WidgetCorp is serving big companies!" Product teams got into the habit of prioritizing features in this direction, and every failure got blamed on, "I guess we weren't focused enough on big clients." This kind of "cultural inertia" can be very difficult to reverse, even if the executive team is willing to publicly admit their mistake (which isn't guaranteed!)
Just about every experienced data scientist has at least a few of these stories, even from otherwise successful companies. They are to scientific debt what the Daily WTF is to technical debt.
Is Scientific Debt Always Bad?
Not at all!
I often take shortcuts in my own analyses. Running a randomized experiment for a feature launch is sometimes too expensive, especially if the number of users is fairly small or the change pretty uncontroversial (you wouldn't A/B test a typo fix). And while correlation doesn't imply causation, it's usually better than nothing when making business decisions.
The comparison to technical debt is useful here: a small engineering team's first goal is typically to build a minimum viable product quickly, rather than overengineer a system that they think will be robust in the distant future. (The equivalent in scientific debt is typically called overthinking, e.g., "Yes, I suppose we could control for weather when we examine what sales deals succeed, but I'm pretty sure you're overthinking this."). And the comparison to financial debt is meaningful, too: Companies typically take on debt (or, similarly, give up equity) while they're growing. Just like you can't build a company without borrowing money, you can't build a company while being certain every decision is thoroughly supported by data.
What's important in both technical and scientific debt is to keep the long-term cost in mind.
It isn't technical debt if you aren't...
Leveraging it to get something valuable up front.
Paying interest on it regularly.
Treating it as a liability that you may eventually need to pay in full.
Code that doesn't meet these criteria isn't debt; it's just low-quality work.
Wrong decisions are expensive, and not paying attention to data is a risk. We can do a cost-benefit analysis of whether the risk is worth it, but we shouldn't write it off as "data scientists always find something to complain about."
Why Even Call It Debt?
To a data scientist or analyst, this post might sound pretty obvious. Of course, there are downsides to ignoring statistical rigor, so why bother giving it a "buzzword-y" name? Because it puts the concept in terms executives and managers can easily understand.
Again, let's go back to technical debt. There are lots of reasons individual engineers may want to write so-called clean code: they appreciate its elegance, they want to impress their peers, or they're perfectionists procrastinating on other work. These reasons don't generally matter to non-technical employees, who care about product features and reliability. The framing of technical debt helps emphasize what the company loses by not investing in architecture: the idea that even if a product looks like it's working, the flaws have a long-term cost in actual dollars and time.
Engineer: "It bothers me that different internal projects use different naming conventions."
CTO: "Sorry it annoys you, but code is code — I don't see why you should waste time on this."
Engineer: "Our inconsistent naming conventions are technical debt: they make it harder for new developers to learn the system."
CTO: "I've been looking for ways to reduce our onboarding time! Great idea; let me know what you need to fix it."
Similarly, scientists, especially from an academic background, often have a particular interest in discovering truths about reality. So the idea of, "I'd like to analyze whether X is a confounding factor here," can sound like an indulgence rather than an immediate business need. Statisticians, in particular, are often excited by finding flaws in mathematical methods. So, when a data scientist says something like, "We can't use that method; Jones et al. 2012 proved that it is asymptotically inconsistent," non-technical colleagues might assume they're overthinking it or even showing off. Framing it in terms of what we're actually risking helps communicate why it's worth spending time on.
How Can We Manage Scientific Debt Well?
- Let data scientists "pay interest" on it. Just as not every engineering project will lead to a new feature, not every analysis will lead to an exciting discovery or novel algorithm. Some time needs to be spent confirming or invalidating existing assumptions. Jonathan Nolis has a great article about prioritizing data science work, where he describes this quadrant as "providing proof."
- Build data engineering processes. As described earlier, one reason a company might fall into scientific debt is that analysts may not have easy access to the data they need. It could be locked away in a platform that hasn't been ingested, or in Google Sheets that are edited by hand. Ingesting relevant data into a data warehouse or a data lake makes it more likely data scientists can make relevant discoveries.
- Revisit old analyses. One common reason early-stage companies go into scientific debt is that they don't yet have enough data to draw robust conclusions. Even if you don't have enough data yet, that doesn't mean you should forget about the problem. Sometimes I put time on my calendar to run an analysis once I expect enough data to be available, even if it's a few months away. This can also help confirm an important analysis is still relevant: just like you'd keep track of a KPI over time, you want to keep track of whether a conclusion remains true.
- Have data expertise spread throughout the company. Just as someone who can't program may not recognize technical debt, someone who doesn't have experience analyzing and understanding data may not recognize scientific debt. This is yet another reason to democratize data science within your company, as we do at DataCamp.
Published at DZone with permission of David Robinson , DZone MVB. See the original article here.
Opinions expressed by DZone contributors are their own.