{{announcement.body}}
{{announcement.title}}

What's Worse Than Coding Without Tests? Coding With Bad Tests.

DZone 's Guide to

What's Worse Than Coding Without Tests? Coding With Bad Tests.

The focus has been on test quantity/coverage without mentioning test quality. If tests are a developer's safety net, shouldn't they be treated like safety gear?

· Agile Zone ·
Free Resource

safety net cartoon

For a long time we've (and by "we" I mean the industry) harped on test coverage. We've sold each other the idea that tests are the safety net developers need to code without fear. Because if the tests pass, your changes must be good, and ready for production, right? Advocates of Test Driven Development even say you should write the tests before the code. But what happens when the tests are wrong? When they're buggy? When they don't test what you think they do, or worse don't test anything? In short, how do you test the tests?

So far, the industry focus has been on test quantity, i.e. coverage, without really talking about test quality. But the time to shift focus is coming. Because if tests really are a developer's safety net, then shouldn't they be treated like safety gear? And once you think of them as safety gear, it becomes obvious that they should be inspected on a regular basis. After all, firefighters inspect their gear after every fire, and climbers and acrobats inspect their gear before every use. 

So when's the last time you inspected a test? And by "inspected", I don't mean "ran". Running the tests is an inspection of your code, not the tests themselves. If you're like me, it was just before you committed the last change to it. If your commit got peer review, then you were lucky enough to get a second pair of eyes on that test. But after you committed the changes, I'm guessing it was entirely "out of sight, out of mind."

That's a real pity if you consider some of the things that can go wrong with a test. Just as much as developers are prone to copy-paste errors during coding, we're maybe even more so while writing tests.

Copy/paste errors happen in tests too: `assertEquals(exp3, exp3);`

Because it's easy to relax and make absent-minded errors when you're not writing "real" code.

When multiple calls in your `try` can throw the same exception, what does the assertion in the `catch` really test?

And of course, most of us have come to rely on IDEs and compilers to catch the silly, obvious stuff.

`Assert.assertNotNull` shouldn't be passed a primitive value; it will never be `null`.

Fortunately, there's one more open source tool in the toolbox: static analysis. That's why SonarQube recently added ten new Java rules to help you write better unit tests for a new total of 33, with a lot more coming on the subject. In the near term, additional test-related rules are planned for Java, Python and PHP, with more languages targeted long-range. We're also working on changes in the SonarQube UI to help you understand the quality of your tests. Beyond that, the long-requested application of some "normal" rules to tests is also in the pipe.

Because until we figure out how to test the tests, static analysis is the next best thing. 

Topics:
static analysis, static analysis tools, test quality

Published at DZone with permission of G. Ann Campbell . See the original article here.

Opinions expressed by DZone contributors are their own.

{{ parent.title || parent.header.title}}

{{ parent.tldr }}

{{ parent.urlSource.name }}