The test recorder is the easiest way to create test scripts, but it is not the only way. You can create and edit test scripts manually.
The final product would look something like:
# File: ./tests/testthat/test-shinytest2.R
test_that("App initialization is consistent", {
app <- AppDriver$new()
app$expect_values()
})
A test script has this basic structure: first, there is an initialization, then the expectations are performed.
In the initialization, the script creates a new AppDriver
object.
Next, define some interactions with the application and makes expectations via snapshots.
app$set_inputs(check_group = c("1", "2"))
app$set_inputs(check_group = c("1", "2", "3"))
app$click("action")
app$expect_values()
app$click("action")
app$expect_values()
For customizing a script, the second portion – the interactions and expectations – is the part you will want to modify.
app$set_inputs()
and app$click()
With app$set_inputs()
, you provide the name of one or more inputs and corresponding values to set them to. Consider this set of directives:
app$set_inputs(check_group = c("1", "2"))
app$set_inputs(check_group = c("1", "2", "3"))
app$click("action")
Notice that we set the value of check_group
two times in a row. When we recorded this test script, it started with the value "1"
, and then we checked the "2"
and "3"
boxes. The recorded script set the value to c("1", "2")
, and then c("1", "2", "3")
. The c("1", "2")
value was simply an intermediate step.
It is possible to simplify and speed up the tests by dropping the intermediate step, which leaves us with this:
Multiple calls to app$set_inputs()
should be reduced to a single line unless the side effects of setting inputs multiple times is desired. When app$set_inputs()
is called, it normally returns control and moves on to the next step only after the server sends a response to the client. The reason it waits for a response is so that a subsequent call to app$snapshot()
will be sure to capture the updated output values. If app$set_inputs()
did not wait for a update, then, if the output update did not happen very quickly, a snapshot might capture the state of the application before the outputs are updated.
For an in-depth discussion of expectations, see the Robust testing vignette.
There are a couple standard ways to make expectations from your AppDriver
object. * app$expect_values()
: This is the simplest way to make an expectation. It creates a list of values and compares them to the current values of the application. It also takes a debug screenshot and stores it as a snapshot. However, this snapshot will never fail an expectation test. The extra screenshot file is useful for manually determining if UI changes have happened over time while not producing numerous false-positive test failures. * app$expect_screenshot()
: This method takes a screenshot of the application and compares it to a reference screenshot. If the screenshots are different, it fails the test.
There are few ways to use app$expect_values()
. The simplest way is to call it with no arguments:
The first time this is called in a test script, it will record all input
, output
, and export
values from the application, in a file called 001.json
. The next call will save the values in 002.json
, and so on.
Each time you call app$expect_values()
, it will also save a debug screenshot of the web browser, in a file 001_.png
, 002_.png
, and so on. These screen shots are useful for debugging your tests and inspecting what they’re doing. You can tell it to not take screen shots, to save space and make the tests run slightly faster, in the initialization step, with:
If you want to disable screenshots for a single app$expect_values()
call, you can use:
If you want more targeted tests, you can snapshot specific items by using the input
, output
, and export
parameters. For example, to capture the value of just the outputs named "a"
and "b"
, you would call:
You could also capture specific inputs or exports:
Finally, if you want to snapshot all output
s but no inputs or exports, you can simply set output
to TRUE
as only values that are requested are recorded:
The same can be used to snapshot all input
and/or all export
values. To capture all output
and export
values, but no input
values:
In some cases, it’s useful to snapshot some bits of internal state of an application – state that’s not reflected directly in the inputs or outputs. This can be done by exporting values.
Consider this toy example where an application has an internal reactive expression that would be convenient to capture.
shinyApp(
fluidPage(
numericInput("x", "x", 4),
numericInput("y", "y", 10),
numericInput("z", "z", 100),
verbatimTextOutput("result", placeholder = TRUE)
),
function(input, output, session) {
xy <- reactive(input$x * input$y)
yz <- reactive(input$y + input$z)
output$result <- renderText({
xy() / yz()
})
exportTestValues(
xy = {
xy()
},
yz = {
yz()
}
)
}
)
Notice these lines:
This creates exported values called xy
and yz
. When a test snapshot is taken, it evaluates the reactive expression and inserts their values in the snapshot:
{
"input": {
"x": 4,
"y": 10,
"z": 100
},
"output": {
"result": "0.3636364"
},
"export": {
"xy": 40,
"yz": 110
}
}
Note that the expression (i.e. xy()
and yz()
) is saved when exportTestValues()
is called, not the specific values. Generally, you should keep these expressions as simple as possible so that test failures yield informative outputs. Note that any value here has to be serializable to JSON, so it’s not suitable for testing complex objects.
In some cases, you may need to wait for some amount of time between steps (e.g. waiting for CSS animations). You can do this by adding Sys.sleep()
in your script. For example:
However, if you are waiting for some Shiny reactives to be calculated, please read more in the Waiting for an output
(or input
) value section.
If your application uses randomly-generated values (by calling functions like rnorm
, runif
, sample
directly or indirectly), then, in normal operation, it will produce different results on each run. Because {shinytest2}
works by comparing the current state of an application to a previous state, these random changes will cause test failures.
To make such applications exactly repeatable, you can set the random seed. This can be done by specifying a random seed in the recorder. The seed can be any integer value.
If your test script has already been created, you can set the random seed by editing the test script so that the AppDriver$new()
call has a value for seed
. Moreover, if you need these tests to pass on multiple versions of R with different RNGkind()
defaults (e.g., 3.5 and 3.6), you may want to set a fixed RNGkind()
across your tests (perhaps via RNGversion("3.5.0")
). However, using a AppDriver$new(variant=)
value that contains the R version can also remedy this situation.
In many cases, instead of setting the random seed, it is more elegant to use an output preprocessor.
With tabbed views, such as tabsetPanel()
and navbarPage()
, in order for {shinytest2}
to keep track of which tab is currently selected, they must have an id
. For example:
or
If you record a file upload event to a fileInput
, the test script will have a line like this:
Notice that the filename "mtcars.csv"
appears in the script without a full file path. This is because the test recorder does not know where the file came from; it only knows the name of the file.
Before you run the test script, you must copy the file to the tests/testthat/
directory. See the tests/testthat/
subdirectory of this app for an example.
After copying the file to that directory, run test_app()
as usual.
As you step through the script, you can inspect the state of the application in a few different ways.
The recommended way is to call app$view()
. This will open a browser window in your default Chrome browser.
Another way is to take a screenshot of your application. You can call app$get_screenshot()
to take a screenshot of the application. This will display the screenshot as if it were a plot. (In RStudio, it will show in the Viewer pane.)
Using both methods, you can inspect the screenshot to see the state of the application.
It can also be useful to get the current input, output, and export values.
To fetch all values without incrementing the expectation counter, you can call app$get_values()
. This returns a list, which you can inspect with the str()
function. It may look something like this:
vals <- app$get_values()
str(vals)
#> List of 3
#> $ input :List of 4
#> ..$ action :Classes 'integer', 'shinyActionButtonValue' int 0
#> ..$ checkbox : logi TRUE
#> ..$ check_group: chr "1"
#> ..$ text : chr "Enter text..."
#> $ output:List of 12
#> ..$ action_out : chr "[1] 0\nattr(,\"class\")\n[1] \"integer\" #> \"shinyActionButtonValue\""
#> ..$ checkbox_out : chr "[1] TRUE"
#> ..$ check_group_out: chr "[1] \"1\""
#> ..$ text_out : chr "[1] \"Enter text...\""
#> $ export: Named list()
The values retrieved this way can be used for expectation-based testing. For example, we could retrieve the checkbox_out
value and compare it to a known string:
input
(or output
) valueIn most situations, when an input
value is set, app$set_inputs()
will wait up to 3 seconds (default) for the next output value to be sent from the server. If, however, you have an application that does not immediately send the value you are interested in, you will need to insert at app$wait_for_value()
function call. This is readily useful for htmlwidgets or anything that involves javascript code to set input
values. This method can also be used to determine when input
(or output
) values have been set after the initial output values have been set or for when reactivity is done in two or more stages.
For example, when checking a checkbox that adds dynamic UI, such as the old faithful Shiny example, three round trips from the browser to the Shiny server will need to occur.
If we were to only use app$set_inputs("chkbx", TRUE)
and immediately take a snapshot, we would most likely miss the correct plot value.
To dig into this further, we can look at this situation’s example code:
app.R
:
library(shiny)
ui <- fluidPage(
checkboxInput("chkbx", "Display Graph?", FALSE),
uiOutput("dynamic_output")
)
server <- function(input, output) {
output$dynamic_output <- renderUI({
if(!input$chkbx) {
return(NULL)
}
tagList(
sliderInput(inputId = "bins", label = "Number of bins:", min = 1, max = 50, value = 30),
plotOutput(outputId = "dist_plot")
)
})
# Will not execute until `input$bins` is available
# `input$bins` is not available until `chkbx` is checked
output$dist_plot <- renderPlot({
# artificially slow the plot
Sys.sleep(5)
x <- faithful$waiting
bins <- seq(min(x), max(x), length.out = input$bins + 1)
hist(x, breaks = bins, col = "#75AADB", border = "white",
xlab = "Waiting time to next eruption (in mins)",
main = "Histogram of waiting times")
})
}
shinyApp(ui = ui, server = server)
Naively, we could add Sys.sleep()
function calls to slow down {shinytest2}
testing to hope that the Shiny application is in a stable state.
./tests/testthat/test-use-sleep-bad.R
:
test_that("Sleeping can work to wait for an app to stabilize", {
app <- AppDriver$new()
# Check the box to display the slider and plot
app$set_inputs('chkbx' = TRUE)
# Hope that the plot appears within 7 seconds
Sys.sleep(7)
app$snapshot()
app$set_inputs(bins = 40)
# Hope the newer plot appears within 7 seconds
Sys.sleep(7)
app$expect_values()
})
The above pattern will probably fail if executed on a slower machine. This pattern of sleeping and hoping should be replaced with app$wait_for_idle()
or app$wait_for_value()
. app$wait_for_idle()
will wait for the whole application to be idle for 500ms (default). This method is more intuitive, generic, and is easier to code. app$wait_for_value()
will wait for a specific input value to not be an ignore
ed value. This method is very specific and requires more logic to code.
./tests/testthat/test-wait-good.R
:
test_that("Use $wait_for_idle() for app to stabilize", {
app <- AppDriver$new()
# Check the box to display the slider and plot
app$set_inputs(chkbx = TRUE)
# Wait until Shiny is not busy for 500ms
app$wait_for_idle(500)
# Make expectation
app$expect_values()
# Change slider value to 40
app$set_inputs(bins = 40, wait_ = FALSE)
# Wait until Shiny is not busy for 500ms
app$wait_for_idle(500)
# Make expectation
app$expect_values()
})
test_that("Use $wait_for_value() for app to stabilize", {
app <- AppDriver$new()
# Check the box to display the slider and plot
app$set_inputs(chkbx = TRUE)
# Wait until `output$dist_plot` is not `NULL`
# (Store the retrieved plot value for later comparison)
first_plot_value <- app$wait_for_value(output = "dist_plot", ignore = list(NULL))
# Make expectation
app$expect_values()
# Change slider value to 40
app$set_inputs(bins = 40, wait_ = FALSE)
# Wait until plot does not equal the plot with 30 bins
app$wait_for_value(output = "dist_plot", ignore = list(first_plot_value))
app$expect_values()
})
If your application uses a data source that changes over time, then a snapshot taken yesterday may not match a snapshot taken today, even if the app itself hasn’t changed. Dynamic data inherently poses a challenge for snapshot-based testing.
This problem can be avoided by detecting when the application is being tested, and in that case use a static data set instead. To do the detection, you can do something like the following:
variant
to expect different snapshotsYou can variants to deal with cases where the snapshot output varies and you want to capture and test the variations. Common use cases include variations for operating system, R version, or version of key dependency. Variants are an advanced feature.
When you use them, you’ll need to carefully think about your testing strategy to ensure that all important variants are covered by automated tests, and ensure that you have a way to get snapshot changes out of your CI system and back into the repo.
It is recommended to use AppDriver$new(variant= platform_variant())
to capture the operating system and R version where the test is running. Variants are required when calling $expect_screenshot()
. A NULL
value may be used, but is not recommended for testing reproducibility.
All {shinytest2}
expectation methods eventually call testthat::expect_snapshot_file()
to save the snapshot files. Snapshot expectations will only compare to their variant’s version of the snapshot file.
Most input components in Shiny set their values by communicating through something called an input binding. {shinytest2}
works well with input
values that are set via input bindings.
However, there are some components that set input values without using an input binding. These include some htmlwidgets, such as DT and plotly, as well as Shiny’s built-in plot interactions with renderPlot()
.
To set these unbound inputs, the resulting test script code will look something like this:
app$set_inputs(table_rows_selected = 1, allow_no_input_binding_ = TRUE)
app$set_inputs(table_row_last_clicked = 1, allow_no_input_binding_ = TRUE)
When the test script is replayed, the headless browser will set these inputs to the specified values and send them to the server running Shiny in R. However, the browser will not be able to tell the input object to do the exact same behaviors. For example, with DT, when a row is selected, the mouse click event in the browser triggers the DataTable to highlight the row in the browser and also set the input value for Shiny. When {shinytest2}
plays the script, it can only do the latter part, setting the input value. The result is that when a screenshot is taken, it will not have highlighted rows. For components that have internal state that is updated in response to user interaction, that internal state will not be updated when the test script is played. In some cases, this may mean that when the script is played, the behavior when the script is played will not be the same as when a user actually interacts with the application.
If the input component sets multiple input values (without bindings) in response to a single user event, it may make sense to coalesce them into a single set_inputs
call, such as this:
Learn about using shinytest2 with continuous integration.