Typically, we test code to ensure its output meets our expectations.
is_plausible <- function(value, minimum = -Inf, maximum = Inf) value >= minimum & value <= maximum
What do you expect the output to be?
is_plausible(c(10, 37, -999), min = 0, max = 112)
is_plausible <- function(value, minimum = -Inf, maximum = Inf) value >= minimum & value <= maximum
What do you expect the output to be?
is_plausible(c(10, 37, -999), min = 0, max = 112)
# [1] TRUE TRUE FALSE
is_plausible <- function(value, minimum = -Inf, maximum = Inf) value >= minimum & value <= maximum
What do you expect the output to be?
is_plausible(c(10, 37, -999), min = 0, max = 112)
# [1] TRUE TRUE FALSE
is_plausible("2", min = 0, max = 10)
is_plausible <- function(value, minimum = -Inf, maximum = Inf) value >= minimum & value <= maximum
What do you expect the output to be?
is_plausible(c(10, 37, -999), min = 0, max = 112)
# [1] TRUE TRUE FALSE
is_plausible("2", min = 0, max = 10)
# [1] FALSE
is_plausible <- function(value, minimum = -Inf, maximum = Inf) value >= minimum & value <= maximum
What do you expect the output to be?
is_plausible(c(10, 37, -999), min = 0, max = 112)
# [1] TRUE TRUE FALSE
is_plausible("2", min = 0, max = 10)
# [1] FALSE
Why? Because "10" < "2"
in lexical order
Should is_plausible.character
:
NA
?FALSE
? (But is it really implausible?)As an exercise, try sorting the vector c(2, 3, "a", 11, "b", "+4")
Unit testing is the process in which the smallest testable parts of source code are tested individually and independently, usually in an automated way.
The testthat framework comprises three parts:
expect_
test_that()
Expectations in testthat compare an object with a reference value or property. If they do not match, an error is thrown.
(If they do match, the tested object is returned, invisibly.)
expect_equal
expect_is
expect_length
expect_true
expect_error
expect_equal
compares a number with a reference value
hyp <- 3^2 + 4^2expect_equal(hyp, 25) # Runs without errorexpect_equal(hyp, 26) # Throws an error:
# Error: `hyp` not equal to 26.# 1/1 mismatches# [1] 25 - 26 == -1
expect_equal
compares a number with a reference value
hyp <- 3^2 + 4^2expect_equal(hyp, 25) # Runs without errorexpect_equal(hyp, 26) # Throws an error:
# Error: `hyp` not equal to 26.# 1/1 mismatches# [1] 25 - 26 == -1
expect_equal(sqrt(2), 1.41) # Throws an error:
# Error: sqrt(2) not equal to 1.41.# 1/1 mismatches# [1] 1.41 - 1.41 == 0.00421
expect_equal(sqrt(2), 1.41, tolerance = .01) # No error thrown
expect_is
checks the data type of an object
val <- 43.7expect_is(val, 'numeric') # Runs without errorexpect_is(val, 'character') # Produces an error:
# Error: `val` inherits from `'numeric'` not `'character'`.
expect_length
checks the number of elements in a vector
expect_length(letters, 26) # Runs without errorexpect_length(letters, 25) # Throws an error:
# Error: `letters` has length 26, not length 25.
Some expectation functions check against a (fixed) condition, rather than a custom reference value.
codes <- c(no = F, yes = T)expect_true(codes['yes']) # Runs without errorexpect_true('yes') # Throws an error:
# Error: "yes" is not TRUE# # `actual` is a character vector ('yes')# `expected` is a logical vector (TRUE)
expect_named(codes) # Runs without errorexpect_false(as.logical(0)) # Runs without error
Other such functions include expect_error
, which checks that a function call throws an error when evaluated.
Analogously, there are expect_warning
, expect_message
, etc.
expect_error(3.14 + 'hello') # Runs without errorexpect_error(3.14 < 'hello') # Throws an error:
# Error: `3.14 < "hello"` did not throw an error.
expect_true( is_plausible(5, min = 0, max = 10) ) # No errorexpect_false( is_plausible(-1, min = 0, max = 1) ) # No errorexpect_true( is_plausible('2', min = 0, max = 10) ) # Error:
# Error: is_plausible("2", min = 0, max = 10) is not TRUE# # `actual`: FALSE# `expected`: TRUE
expect_true( is_plausible(5, min = 0, max = 10) ) # No errorexpect_false( is_plausible(-1, min = 0, max = 1) ) # No errorexpect_true( is_plausible('2', min = 0, max = 10) ) # Error:
# Error: is_plausible("2", min = 0, max = 10) is not TRUE# # `actual`: FALSE# `expected`: TRUE
Now decide:
Fix to pass the test (not recommended here; will cause warning
s):
is_plausible <- function(value, minimum = -Inf, maximum = Inf) { value <- as.numeric(value) value >= minimum & value <= maximum}expect_true( is_plausible('2', min = 0, max = 10) ) # No error
Fix to pass the test (not recommended here; will cause warning
s):
is_plausible <- function(value, minimum = -Inf, maximum = Inf) { value <- as.numeric(value) value >= minimum & value <= maximum}expect_true( is_plausible('2', min = 0, max = 10) ) # No error
Or rethink, by expecting stricter handling of inputs:
is_plausible <- function(value, minimum = -Inf, maximum = Inf) { stopifnot(is.numeric(value)) value >= minimum & value <= maximum}expect_error( is_plausible('2', min = 0, max = 10) ) # No error
test_that('Scalar integers', { expect_true(is_plausible(5, min = 0, max = 10)) expect_false(is_plausible(-1, min = 0, max = 1))})
# Test passed
test_that('Integer vectors', { expect_equal(is_plausible(0:10, min = 0, max = 10), rep(TRUE, 11)) expect_equal(is_plausible(-1:2, min = 0, max = 1), c(F, T, T, F))})
# Test passed
test_that('Handle unusual or missing inputs', { expect_error(is_plausible('2', min = 0, max = 10)) expect_error(is_plausible(min = 0, max = 10)) expect_error(is_plausible(2, min = "0", max = 10)) expect_error(is_plausible(2, min = 0, max = "10"))})
# -- Failure (<text>:4:3): Handle unusual or missing inputs ----------------------# `is_plausible(2, min = "0", max = 10)` did not throw an error.# # -- Failure (<text>:5:3): Handle unusual or missing inputs ----------------------# `is_plausible(2, min = 0, max = "10")` did not throw an error.
Add a tests/
folder to an R package
<Ctrl> + <Shift> + T
R CMD CHECK
Or run tests locally in analysis folder:
test_file()
/ test_dir()
auto_test()
R Markdown 'test chunks' with error=TRUE
or testrmd
Note: most documentation for testthat assumes you are writing a package. Future talk: how to do analysis as an R package...
Write a test before any other code.
Check that the test fails.
Write enough code to make it pass.
Add another test and iterate steps 1–3.
Refactor the code, while ensuring it passes all tests.
In a way, R CMD CHECK
could be considered test-driven development, because the tests your package has to pass were written long before the package.
In software engineering, continuous integration (CI) is the practice of merging all changes to a central repository, and automatically rebuilding & testing the code after every change.
Tools: Travis CI, GitHub Actions, Bitbucket Pipelines, Gitlab CI/CD, Jenkins
People used to recommend Travis CI, but the company sold out and stopped offering a free service. If browsing the web you might still find old articles talking about an R + GitHub + Travis workflow.
Make life easier for yourself:
DESCRIPTION
fileR/
subfoldertests/testthat/
subfoldertest_local()
/ <Ctrl>+<Shift>+T
Minimal working DESCRIPTION
file:
Package: blahVersion: 0.1
Example repo: https://github.com/Selbosh/unittesting
Friday 3 December
Advent of Code discussion
Attempt days 1–2 and share your approach
See also https://selbydavid.com/2020/12/06/advent-2020/
Typically, we test code to ensure its output meets our expectations.
Keyboard shortcuts
↑, ←, Pg Up, k | Go to previous slide |
↓, →, Pg Dn, Space, j | Go to next slide |
Home | Go to first slide |
End | Go to last slide |
Number + Return | Go to specific slide |
b / m / f | Toggle blackout / mirrored / fullscreen mode |
c | Clone slideshow |
p | Toggle presenter mode |
t | Restart the presentation timer |
?, h | Toggle this help |
Esc | Back to slideshow |