This vignette is current as of furniture
1.9.7.
furniture
We will first make a fictitious data set:
df <- data.frame(a = rnorm(1000, 1.5, 2),
b = seq(1, 1000, 1),
c = c(rep("control", 400), rep("Other", 70), rep("treatment", 500), rep("None", 30)),
d = c(sample(1:1000, 900, replace=TRUE), rep(-99, 100)))
There are four functions that we’ll demonstrate here:
washer
table1
tableC
tableF
washer
is a great function for quick data cleaning. In situations where there are placeholders, extra levels in a factor, or several values need to be changed to another.
Now that the data is “washed” we can start exploring and reporting.
##
## ────────────────────────────────
## Mean/Count (SD/%)
## n = 900
## a
## 1.6 (2.0)
## b
## 450.5 (260.0)
## factor(c)
## control 470 (52.2%)
## treatment 430 (47.8%)
## d
## 496.2 (291.4)
## ────────────────────────────────
The variables must be numeric or factor. Since we use a special type of evaluation (i.e. Non-Standard Evaluation) we can change the variables in the function (e.g., factor(c)
). This can be extended to making a whole new variable in the function as well.
##
## ───────────────────────────────────────
## Mean/Count (SD/%)
## n = 900
## a
## 1.6 (2.0)
## b
## 450.5 (260.0)
## d
## 496.2 (291.4)
## ifelse(a > 1, 1, 0)
## 0.6 (0.5)
## ───────────────────────────────────────
This is just the beginning though. Two powerful things the function can do are shown below:
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: b
##
## ─────────────────────────────────────────────────────────
## factor(c)
## control treatment P-Value
## n = 470 n = 430
## a 0.967
## 1.6 (2.0) 1.6 (2.0)
## b <.001
## 235.5 (135.8) 685.5 (124.3)
## d 0.961
## 496.7 (286.9) 495.7 (296.6)
## ifelse(a > 1, 1, 0) 0.753
## 0.6 (0.5) 0.6 (0.5)
## ─────────────────────────────────────────────────────────
The splitby = ~factor(c)
stratifies the means and counts by a factor variable (in this case either control or treatment). When we use this we can also automatically compute tests of significance using test=TRUE
.
We can also use it intuitively within the pipe (for more about this, see the “Table 1” vignette):
##
## ─────────────────────────────────────────────────────────
## c
## control treatment P-Value
## n = 470 n = 430
## a 0.967
## 1.6 (2.0) 1.6 (2.0)
## b <.001
## 235.5 (135.8) 685.5 (124.3)
## d 0.961
## 496.7 (286.9) 495.7 (296.6)
## ifelse(a > 1, 1, 0) 0.753
## 0.6 (0.5) 0.6 (0.5)
## ─────────────────────────────────────────────────────────
In this case, we used the group_by()
function from dplyr
(within the tidyverse
) and table1()
knows to use that as the grouping variable in place of the splitby
argument.
If the parametric tests (default) are not appropriate, you can set param = FALSE
.
##
## ─────────────────────────────────────────────────────────
## c
## control treatment P-Value
## n = 470 n = 430
## a 0.713
## 1.6 (2.0) 1.6 (2.0)
## b <.001
## 235.5 (135.8) 685.5 (124.3)
## d 0.981
## 496.7 (286.9) 495.7 (296.6)
## ifelse(a > 1, 1, 0) 0.753
## 0.6 (0.5) 0.6 (0.5)
## ─────────────────────────────────────────────────────────
Finally, you can polish it quite a bit using a few other options. For example, you can do the following:
table1(df, a, b, d, ifelse(a > 1, 1, 0),
splitby=~factor(c),
test=TRUE,
var_names = c("A", "B", "D", "New Var"),
type = c("simple", "condensed"))
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: B
##
## ─────────────────────────────────────────────
## factor(c)
## control treatment P-Value
## n = 470 n = 430
## A 1.6 (2.0) 1.6 (2.0) 0.967
## B 235.5 (135.8) 685.5 (124.3) <.001
## D 496.7 (286.9) 495.7 (296.6) 0.961
## New Var 0.6 (0.5) 0.6 (0.5) 0.753
## ─────────────────────────────────────────────
Note that var_names
can be used for more complex naming (e.g., with spaces, brackets) that otherwise cannot be used with data frames. Alternatively, for more simple naming, we can name them directly.
table1(df, A = a, B = b, D = d, A2 = ifelse(a > 1, 1, 0),
splitby=~factor(c),
test=TRUE,
type = c("simple", "condensed"))
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: B
##
## ────────────────────────────────────────
## factor(c)
## control treatment P-Value
## n = 470 n = 430
## A 1.6 (2.0) 1.6 (2.0) 0.967
## B 235.5 (135.8) 685.5 (124.3) <.001
## D 496.7 (286.9) 495.7 (296.6) 0.961
## A2 0.6 (0.5) 0.6 (0.5) 0.753
## ────────────────────────────────────────
You can also format the numbers (adding a comma for big numbers such as in 20,000 instead of 20000):
table1(df, a, b, d, ifelse(a > 1, 1, 0),
splitby=~factor(c),
test=TRUE,
var_names = c("A", "B", "D", "New Var"),
format_number = TRUE)
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: B
##
## ─────────────────────────────────────────────
## factor(c)
## control treatment P-Value
## n = 470 n = 430
## A 0.967
## 1.6 (2.0) 1.6 (2.0)
## B <.001
## 235.5 (135.8) 685.5 (124.3)
## D 0.961
## 496.7 (286.9) 495.7 (296.6)
## New Var 0.753
## 0.6 (0.5) 0.6 (0.5)
## ─────────────────────────────────────────────
The table can be exported directly to a folder in the working directory called “Table1”. Using export
, we provide it with a string that will be the name of the CSV containing the formatted table.
table1(df, a, b, d, ifelse(a > 1, 1, 0),
splitby=~factor(c),
test=TRUE,
var_names = c("A", "B", "D", "New Var"),
format_number = TRUE,
export = "example_table1")
This can also be outputted as a latex, markdown, or pandoc table (matching all the output types of knitr::kable
). Below shows how to do a latex table (not using kable
however, but a built-in function that provides the variable name at the top of the table):
table1(df, a, b, d, "new var" = ifelse(a > 1, 1, 0),
splitby = ~factor(c),
test = TRUE,
output = "latex2")
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: b
## \begin{table}[ ht ]
## \centering
## \caption{}\label{}
## \begin{tabular}{ l c c c }
## \toprule
## & \multicolumn{ 2 }{c}{ factor(c) }\\
## & control & treatment & P-Value \\
## & n = 470 & n = 430 & \\
## \midrule
## a & & & 0.967\\
## \hspace{6pt} & 1.6 (2.0) & 1.6 (2.0) & \\
## b & & & <.001\\
## \hspace{6pt} & 235.5 (135.8) & 685.5 (124.3) & \\
## d & & & 0.961\\
## \hspace{6pt} & 496.7 (286.9) & 495.7 (296.6) & \\
## new var & & & 0.753\\
## \hspace{6pt} & 0.6 (0.5) & 0.6 (0.5) & \\
## \bottomrule
##
## \end{tabular}
## \end{table}
Last item to show you regarding table1()
is that it can be printed in a simplified and condensed form. This instead of reporting counts and percentages for categorical variables, it reports only percentages and the table has much less white space.
table1(df, a, b, d, "new var" = ifelse(a > 1, 1, 0),
splitby = ~factor(c),
test = TRUE,
type = c("simple", "condensed"))
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: b
##
## ─────────────────────────────────────────────
## factor(c)
## control treatment P-Value
## n = 470 n = 430
## a 1.6 (2.0) 1.6 (2.0) 0.967
## b 235.5 (135.8) 685.5 (124.3) <.001
## d 496.7 (286.9) 495.7 (296.6) 0.961
## new var 0.6 (0.5) 0.6 (0.5) 0.753
## ─────────────────────────────────────────────
This function is to create simple, beautiful correlation tables. The syntax is just like table1()
in most respects. Below we include all the numeric variables to see their correlations. Since there are missing values in d
we will use the natural na.rm=TRUE
.
## N = 900
## Note: pearson correlation (p-value).
##
## ─────────────────────────────────────────
## [1] [2] [3]
## [1]a 1.00
## [2]b 0.013 (0.699) 1.00
## [3]d 0.017 (0.616) -0.008 (0.819) 1.00
## ─────────────────────────────────────────
All the adjustments that you can make in table1()
can be done here as well. For example,
## N = 900
## Note: pearson correlation (p-value).
[1] | [2] | [3] | |
---|---|---|---|
[1]A | 1.00 | ||
[2]B | 0.013 (0.699) | 1.00 | |
[3]D | 0.017 (0.616) | -0.008 (0.819) | 1.00 |
This function is to create simple frequency tables. The syntax is just like table1()
and tableC()
in most respects, except that it uses only one variable instead of many.
##
## ────────────────────────────────────────────────
## a Freq CumFreq Percent CumPerc
## -4.75268974093533 1 1 0.10% 0.10%
## -4.2907267597833 1 2 0.10% 0.20%
## -4.08692254442744 1 3 0.10% 0.30%
## -3.76517947534703 1 4 0.10% 0.40%
## -3.30715514097551 1 5 0.10% 0.50%
## -3.27591116895658 1 6 0.10% 0.60%
## -3.15455103272007 1 7 0.10% 0.70%
## -3.04062506698273 1 8 0.10% 0.80%
## -2.92689538594695 1 9 0.10% 0.90%
## -2.8776832147635 1 10 0.10% 1.00%
## ... ... ... ... ...
## 6.72790266665327 1 990 0.10% 99.00%
## 6.77840419538758 1 991 0.10% 99.10%
## 6.88796042421174 1 992 0.10% 99.20%
## 6.92113758878904 1 993 0.10% 99.30%
## 7.06565585404149 1 994 0.10% 99.40%
## 7.09422135438498 1 995 0.10% 99.50%
## 7.16403289035171 1 996 0.10% 99.60%
## 7.27382322581555 1 997 0.10% 99.70%
## 7.5369142155617 1 998 0.10% 99.80%
## 7.67204955283881 1 999 0.10% 99.90%
## 8.84741860259068 1 1000 0.10% 100.00%
## ────────────────────────────────────────────────
Similarly to table1()
we can use a splitby
argument (or group_by()
).
## Variable:d
##
## ─────────────────────────────────────────────────────
## control Freq CumFreq Percent CumPerc Valid CumValid
## 1 1 1 0.20% 0.20% 0.21% 0.21%
## 8 1 2 0.20% 0.40% 0.21% 0.43%
## 9 1 3 0.20% 0.60% 0.21% 0.64%
## 10 1 4 0.20% 0.80% 0.21% 0.85%
## 14 2 6 0.40% 1.20% 0.43% 1.28%
## 16 1 7 0.20% 1.40% 0.21% 1.49%
## 18 1 8 0.20% 1.60% 0.21% 1.70%
## 20 1 9 0.20% 1.80% 0.21% 1.91%
## 21 1 10 0.20% 2.00% 0.21% 2.13%
## 23 1 11 0.20% 2.20% 0.21% 2.34%
## ... ... ... ... ... ... ...
## 972 1 460 0.20% 92.00% 0.21% 97.87%
## 974 2 462 0.40% 92.40% 0.43% 98.30%
## 975 1 463 0.20% 92.60% 0.21% 98.51%
## 977 1 464 0.20% 92.80% 0.21% 98.72%
## 979 1 465 0.20% 93.00% 0.21% 98.94%
## 981 1 466 0.20% 93.20% 0.21% 99.15%
## 988 1 467 0.20% 93.40% 0.21% 99.36%
## 993 1 468 0.20% 93.60% 0.21% 99.57%
## 998 1 469 0.20% 93.80% 0.21% 99.79%
## 1000 1 470 0.20% 94.00% 0.21% 100.00%
## Missing 30 500 6.00% 100.00%
## ─────────────────────────────────────────────────────
##
## ───────────────────────────────────────────────────────
## treatment Freq CumFreq Percent CumPerc Valid CumValid
## 3 1 1 0.20% 0.20% 0.23% 0.23%
## 5 1 2 0.20% 0.40% 0.23% 0.47%
## 8 1 3 0.20% 0.60% 0.23% 0.70%
## 10 1 4 0.20% 0.80% 0.23% 0.93%
## 11 1 5 0.20% 1.00% 0.23% 1.16%
## 13 1 6 0.20% 1.20% 0.23% 1.40%
## 14 2 8 0.40% 1.60% 0.47% 1.86%
## 16 1 9 0.20% 1.80% 0.23% 2.09%
## 19 1 10 0.20% 2.00% 0.23% 2.33%
## 25 4 14 0.80% 2.80% 0.93% 3.26%
## ... ... ... ... ... ... ...
## 966 3 418 0.60% 83.60% 0.70% 97.21%
## 969 1 419 0.20% 83.80% 0.23% 97.44%
## 973 1 420 0.20% 84.00% 0.23% 97.67%
## 975 1 421 0.20% 84.20% 0.23% 97.91%
## 982 1 422 0.20% 84.40% 0.23% 98.14%
## 984 1 423 0.20% 84.60% 0.23% 98.37%
## 989 2 425 0.40% 85.00% 0.47% 98.84%
## 990 2 427 0.40% 85.40% 0.47% 99.30%
## 992 2 429 0.40% 85.80% 0.47% 99.77%
## 994 1 430 0.20% 86.00% 0.23% 100.00%
## Missing 70 500 14.00% 100.00%
## ───────────────────────────────────────────────────────
## Using dplyr::group_by() groups: c
## Variable:d
##
## ─────────────────────────────────────────────────────
## control Freq CumFreq Percent CumPerc Valid CumValid
## 1 1 1 0.20% 0.20% 0.21% 0.21%
## 8 1 2 0.20% 0.40% 0.21% 0.43%
## 9 1 3 0.20% 0.60% 0.21% 0.64%
## 10 1 4 0.20% 0.80% 0.21% 0.85%
## 14 2 6 0.40% 1.20% 0.43% 1.28%
## 16 1 7 0.20% 1.40% 0.21% 1.49%
## 18 1 8 0.20% 1.60% 0.21% 1.70%
## 20 1 9 0.20% 1.80% 0.21% 1.91%
## 21 1 10 0.20% 2.00% 0.21% 2.13%
## 23 1 11 0.20% 2.20% 0.21% 2.34%
## ... ... ... ... ... ... ...
## 972 1 460 0.20% 92.00% 0.21% 97.87%
## 974 2 462 0.40% 92.40% 0.43% 98.30%
## 975 1 463 0.20% 92.60% 0.21% 98.51%
## 977 1 464 0.20% 92.80% 0.21% 98.72%
## 979 1 465 0.20% 93.00% 0.21% 98.94%
## 981 1 466 0.20% 93.20% 0.21% 99.15%
## 988 1 467 0.20% 93.40% 0.21% 99.36%
## 993 1 468 0.20% 93.60% 0.21% 99.57%
## 998 1 469 0.20% 93.80% 0.21% 99.79%
## 1000 1 470 0.20% 94.00% 0.21% 100.00%
## Missing 30 500 6.00% 100.00%
## ─────────────────────────────────────────────────────
##
## ───────────────────────────────────────────────────────
## treatment Freq CumFreq Percent CumPerc Valid CumValid
## 3 1 1 0.20% 0.20% 0.23% 0.23%
## 5 1 2 0.20% 0.40% 0.23% 0.47%
## 8 1 3 0.20% 0.60% 0.23% 0.70%
## 10 1 4 0.20% 0.80% 0.23% 0.93%
## 11 1 5 0.20% 1.00% 0.23% 1.16%
## 13 1 6 0.20% 1.20% 0.23% 1.40%
## 14 2 8 0.40% 1.60% 0.47% 1.86%
## 16 1 9 0.20% 1.80% 0.23% 2.09%
## 19 1 10 0.20% 2.00% 0.23% 2.33%
## 25 4 14 0.80% 2.80% 0.93% 3.26%
## ... ... ... ... ... ... ...
## 966 3 418 0.60% 83.60% 0.70% 97.21%
## 969 1 419 0.20% 83.80% 0.23% 97.44%
## 973 1 420 0.20% 84.00% 0.23% 97.67%
## 975 1 421 0.20% 84.20% 0.23% 97.91%
## 982 1 422 0.20% 84.40% 0.23% 98.14%
## 984 1 423 0.20% 84.60% 0.23% 98.37%
## 989 2 425 0.40% 85.00% 0.47% 98.84%
## 990 2 427 0.40% 85.40% 0.47% 99.30%
## 992 2 429 0.40% 85.80% 0.47% 99.77%
## 994 1 430 0.20% 86.00% 0.23% 100.00%
## Missing 70 500 14.00% 100.00%
## ───────────────────────────────────────────────────────
Lastly, tableX()
is a pipe-able two-way version of table()
with a similar syntax to that of the rest of the furniture
functions.
## ifelse(d > 500, 1, 0)
## c 0 1 Missing Total
## control 241 229 30 500
## treatment 220 210 70 500
## Total 461 439 100 1000
By default, it provides the total counts for the rows and columns with flexibility as to what is displayed and where.
The four functions: table1()
, tableC()
, tableF()
, and washer()
add simplicity to cleaning up and understanding your data. Use these pieces of furniture to make your quantitative life a bit easier.