testthat 2.3.0

  r-lib, devtools, testthat

  Hadley Wickham

We’re pumped to announce that testthat 2.3.0 is now available on CRAN! testthat makes it easy to turn your existing informal tests into formal automated tests that you can rerun quickly and easily. testthat is the most popular unit-testing package for R, and is used by over 4,000 CRAN and Bioconductor packages. You can learn more about unit testing at https://r-pkgs.org/tests.html.

(We didn’t write a blog post about testthat 2.2.0 because it only introduced a single, experimental, new feature: verify_output(). It has now matured to the point we think you should also try it out, so it’s discussed below.)

Install the latest version of testthat with:

install.packages("testthat")

This release features two main improvements:

  • A general overhaul of condition and error handling.

  • A polished verify_output() that is ready for you to use to test your print methods and error messages.

For a complete list of changes, please see the release notes.

library(testthat)

Errors

The main improvements are mostly behind the scenes: we have overhauled the handling of errors and backtraces so that you should get more informative outputs when tests error unexpectedly or fail. It’s hard to authentically demonstrate this in an RMarkdown document, but if you have an error inside a test, like this:

f <- function() g()
g <- function() h()
h <- function() stop("This is an error!")

test_that("f() works as expected", {
  expect_equal(f(), 10)
})

You’ll now get an informative backtrace that should allow you to quickly locate the source of the error:

test-error.R:6: error: f() works as expected
This is an error!
Backtrace:
 1. testthat::expect_equal(f(), 10) tests/testthat/test-error.R:6:2
 4. testthat:::f()
 5. testthat:::g() tests/testthat/test-catch.R:1:5
 6. testthat:::h() tests/testthat/test-catch.R:2:5

The previous version only showed the error message, which wasn’t terribly useful!

verify_output()

verify_output() is designed to test output aimed at a human, like print methods and error messages. Here you want to test that the output is useful to a human, but there’s obviously no way to do that automatically. Instead, the best you can do is to check the results with your eyeballs every time the results change. verify_output() is designed to help you do this, making it a type of visual regression test.

You’ll need to use verify_output() in concert with git. Whenever the output changes, you’ll get a test failure, but to see the change, you’ll need to use git. If the change is correct, commit it with git. If it’s incorrect, fix your code and rerun the tests. Once fixed, the git diff will disappear.

verify_output() works a little like RMarkdown: you give it some R code and it will run it, interleaving the input and output. For example, imagine we were writing some tests to check that tibbles print correctly:

library(tibble)

test_that("tibbles print usefully", {
  verify_output(test_path("test-print-dataframe.txt"), {
    df1 <- tibble(x = 1:1e6)
    print(df1)
  })
})

That test yields a test-print-dataframe.txt containing this output:

> df1 <- tibble(x = 1:1e+06)
> print(df1)
# A tibble: 1,000,000 x 1
       x
   <int>
 1     1
 2     2
 3     3
 4     4
 5     5
 6     6
 7     7
 8     8
 9     9
10    10
# ... with 999,990 more rows

Unfortunately, there’s no way for verify_output() to capture comments, so you can instead use bare strings if you want comments to appear in the output. If you start the comment with # it will be formatted as a heading:

test_that("tibbles print usefully", {
  verify_output(test_path("test-print-dataframe-2.txt"), {
    "# long tibbles"
    df1 <- tibble(x = 1:1e6)
    print(df1)
    
    "# wide tibbles"
    "not yet written"
  })
})

long tibbles
============

> df1 <- tibble(x = 1:1e+06)
> print(df1)
# A tibble: 1,000,000 x 1
       x
   <int>
 1     1
 2     2
 3     3
 4     4
 5     5
 6     6
 7     7
 8     8
 9     9
10    10
# ... with 999,990 more rows


wide tibbles
============

> # not yet written

verify_output() is automatically skipped when run on CRAN. This avoids false positives because it’s very easy to accidentally depend on the code from other packages, and failure does not imply incorrect computation, just a change in presentation. In other words, verify_output() is meant to monitor the evolution of outputs produced by your package rather than checking for regressions. In this way, it is similar to the vdiffr extension to testthat which uses the same approach to monitor the evolution of plots.

Acknoweldgements

A big thanks to everyone who helped make this version happen! @aaronrudkin, @aneudecker, @atheriel, @batpigandme, @billdenney, @cimentadaj, @CorradoLanera, @cosi1, @DavisVaughan, @dtm2451, @EmielSchmeink, @flying-sheep, @gaborcsardi, @gdurif, @hadley, @hughjonesd, @ianmcook, @jameslamb, @jcheng5, @jcubic, @jennybc, @jeroen, @jimhester, @jozefhajnala, @jpritikin, @keesh0, @kenahoo, @KrishanBhasin, @krlmlr, @lionel-, @lorenzwalthert, @maxheld83, @mcol, @MichaelChirico, @mkurdej, @MLopez-Ibanez, @moodymudskipper, @mwmclean, @Nachtfeuer, @nalzok, @nbenn, @oxr463, @patr1ckm, @richierocks, @rillig, @sebastiangonsal, @shrektan, @StevenMaude, @tdhock, @theclue, @tjbell, @topepo, @torbjorn, @trevorld, @wlandau, and @zappingseb