Welcome to the hands-on workshop “Tidy Work in Tidyverse”. Most of the things necessary to complete the tutorials and challenges were covered in the lecture. However, sometimes the tasks require that you check the docs or search online. Not all our solutions are optimal. Let us know if you can do better or solve things in a different way. If stuck, look at hints, next google and if still stuck, turn to TA. It is a lot of material, do not fee bad if you do not solve all tasks. If you completed Challenge 3 you are good and have used the most important features of tidyverse! Good luck!
Datasets are available here.
Rewrite the following code chunks as pipes (Load package magrittr
because tidyverse
supports only the %>%
pipe!):
my_cars <- mtcars[, c(1:4, 10)]
my_cars <- my_cars[my_cars$disp > mean(my_cars$disp), ]
my_cars <- colMeans(my_cars)
mtcars %>%
select(c(1:4, 10)) %>%
filter(disp > mean(disp)) %>%
colMeans() -> my_cars
What is wrong with our solution?
# It is better to have the result assigned on the left hand side: result <- expression. In this case the expression is the **whole** pipe.
# Our 'expression -> result' is correct but can easily be missed when reading the code.
The summary(x)
function is a bit special: when you type summary(x)
in the console, print
is called in an implicit way. Pipe call does not do, so you will have to invoke print
in an explicit way. But the %T>%
does unbranch for one call only, you will have to make printing of the summary
a one single composed call using {}
.
summary(cars)
colSums(cars)
cars %T>% {print(summary(.))} %>% colSums()
Rewrite correlations to pipes.
cor(mtcars$gear, mtcars$mpg)
mtcars %$% cor(gear, mpg)
cor(mtcars)
mtcars %>% cor()
Given is the dim_summary(nrows, ncols)
function which takes matrix nrows
and ncols
as arguments and prints this info about its dimensions:
dim_summary <- function(nrows, ncols) {
print(
paste0('Matrix M has: ', nrows, ' rows and ', ncols, ' columns.')
)
}
Rewrite the code chunks below as pipes:
distr1 <- rnorm(16)
M <- matrix(distr1, ncol = 4)
plot(M)
M <- M + sample(M)
dim_summary(nrows = nrow(M), ncols = ncol(M))
distr2 <- rnorm(16)
N <- matrix(distr2, ncol = 4)
colnames(N) <- (letters[1:4])
summary(N)
P <- M %x% t(N)
heatmap(P)
colnames(P) <- letters[1:dim(P)[2]]
cor(P[ ,'a'], P[ ,'i'])
A class of functions, called the replacement functions are of the form function(arguments)<-value
and rownames(x) <- c('a', 'b', 'c')
is a good example of a replacement function. When writing pipes, we have bear in mind that whole function<-
is the name of the replacement function and thus we have to use it as such in the pipe.
Sometimes, it may not be possible to put everything into one single pipe, the results of running two or more pipes have to be used in the final pipe.
dim_summary <- function(nrows, ncols) {
print(paste0('Matrix M has: ', nrows, ' rows and ', ncols, ' columns.'))
}
M <- rnorm(16) %>%
matrix(ncol = 4) %T>%
plot() %>%
`+`(., sample(.)) %T>%
{dim_summary(nrow(.), ncol(.))}
N <- rnorm(16) %>%
matrix(ncol = 4) %>%
`colnames<-`(letters[1:4]) %>%
summary()
P <- M %>%
`%x%`(., N) %T>%
heatmap() %>%
`colnames<-`(letters[1:dim(.)[2]]) %>%
as_data_frame() %$%
cor(a, i)
mtcars
dataset to a tibble vehicles
.cyl
) variable using:
[[index]]
accessor,[[string]]
accessor,$
accessor.tibble.print_max
and dplyr.print_min
? Is there any? Test it.vehicles
back to a data.frame
called automobiles
.
# 1
vehicles <- mtcars %>% as_tibble()
# 2
vehicles[['cyl']]
vehicles[[2]]
vehicles$cyl
# 3
vehicles %T>%
{print(.[['cyl']])} %T>%
{print(.[[2]])} %>%
.$cyl
# 4
vehicles
# 5
vehicles %>% head(n = 30)
# 6
options(tibble.print_min = 15, tibble.print_max = 30)
# 7
# In theory there should be no difference. dplyr imports tibble from the tibble package
# and dplyr.width, dplyr.print_min and dplyr.print_min are passed down to the tibble.
# But test both behaviours. First with only the tibble package loaded, later with dplyr # loaded.
# 8
automobiles <- as.data.frame(vehicles)
Create the following tibble using tribble()
:
tab <- tribble(
~id, ~event, ~date,
1, 'success', '24-04-2017',
2, 'failed', '25-04-2017',
3, 'failed', '25-04-2017',
4, 'success', '27-04-2017'
)
Compare the performance of as.data.frame()
, as_data_frame()
and as_tibble()
on a
100 x 30 matrix filled with random integers. Use package microbenchmark
. Fill in your result here in the Tidyverse Lab sheet, Tibbles – performance.
tst <- replicate(30, sample(100), simplify = TRUE)
colnames(tst) = paste0(rep('col', times = dim(tst)[2]), 1:dim(tst)[2])
microbenchmark::microbenchmark(
as.data.frame(tst),
as_data_frame(tst),
as_tibble(tst)
)
Do you think tibbles are lazy? Try to create a tibble that tests whether lazy evaluation applies to tibbles too.
tibble(x = sample(1:10, size = 10, replace = T), y = log10(x))
Parse the following vectors using parse_
functions:
vec1 <- c(1, 7.2, 3.84, -5.23)
– parse it as double (any problems? why?).c(1, 7.2, 3.84, -5.23)
as integer. What happens?vec2 <- c('2', '3,45', '?', '-7,28')
vec3 <- c('2', '3,45', '?', '-7.28')
vec4 <- c('barrel: 432.7$', 'liter: 15.42PLN', 'gallon costs approx 32.1SEK', 'sunny, wind gusts up till 55m/s')
vec4
as number? Do it if you can.vec5 <- "25 Dec 2015"
as date (hint: ?parse_date()
).10_Jul_1410
as date.
vec1 <- c(1, 7.2, 3.84, -5.23)
vec2 <- c('2', '3,45', '?', '-7,28')
vec3 <- c('2', '3,45', '?', '-7.28')
vec4 <- c('barrel: 432.7$', 'liter: 15.42PLN', 'gallon costs approx 32.1SEK', 'sunny, wind gusts up till 55m/s')
vec5 <- "25 Dec 2015"
parse_double(vec1)
parse_integer(vec1)
parse_integer(as.integer(vec1)) # Is it the best way? Hint: rounding.
parse_double(vec2, na = '?', locale = locale(decimal_mark = ','))
parse_number(vec2, na = '?', locale = locale(decimal_mark = ','))
guess_parser(vec4)
parse_guess(vec4)
# Yes, you can:
parse_number(vec4)
parse_date(vec5, format="%d %b %Y")
parse_date("10_Jul_1410", format="%d%.%b%.%Y")
The nycflights13
package contains information about all flights that departed from NYC (i.e., EWR, JFK and LGA) in 2013: 336,776 flights with 16 variables. To help understand what causes delays, it also includes a number of other useful datasets: weather, planes, airports, airlines. We will use it to train working with tibbles and dplyr
.
nycflights13
package (install if necessary),flights
tibble.carrier
and arr_time
,carrier
, tailnum
and origin
,day
through carrier
,arr
ival (hint: ?tidyselect
),v <- c("arr_time", "sched_arr_time", "arr_delay")
,dest
to destination
using:
select()
andrename()
What is the difference between the two approaches?
install.packages('nycflights13')
library('nycflights13')
?nycflights13
flights
flights %>% select(-carrier, -arr_time)
flights %>% select(carrier, tailnum, origin)
flights %>% select(-(day:carrier))
flights %>% select(contains('arr_')) # or
v <- c("arr_time", "sched_arr_time", "arr_delay")
flights %>% select(v) # ambiguous, or better
flights %>% select(all_of(v))
flights %>% select(destination = dest)
flights %>% rename(destination = dest)
# select keeps only the renamed column while rename returns the whole dataset
# with the column renamed
?slice
),?sample_n()
) 3 random flights per day in March,unique()
routes and sort them by destination,distinct()
routes and sort them by destination,unique()
more efficient than distinct()
?
flights %>% filter(arr_delay < 0)
flights %>% filter(dep_delay >= 10, dep_delay <= 33) # or
flights %>% filter(between(dep_delay, 10, 33))
flights %>% filter(is.na(arr_time))
flights %>% slice(1234:1258)
nycflights13::flights %>% filter(month == 3) %>%
group_by(day) %>%
slice_sample(n = 3)
nycflights13::flights %>%
filter(month == 1) %>%
group_by(carrier) %>%
slice_max(dep_delay, n = 5)
nycflights13::flights %>%
select(origin, dest) %>%
unique() %>%
arrange(dest)
nycflights13::flights %>%
mutate(route = paste(origin, dest, sep="-")) %>%
select(route) %>%
unique()
nycflights13::flights %>%
select(origin, dest) %>%
distinct() %>%
arrange(dest)
# or
flights %>%
mutate(route = paste(origin, dest, sep="-")) %>%
distinct(route)
microbenchmark::microbenchmark(
unique = nycflights13::flights %>%
select(origin, dest) %>%
unique() %>%
arrange(dest),
distinct = nycflights13::flights %>%
distinct(origin, dest) %>%
arrange(dest),
times = 10L
)
Distinct is faster.
air_time
is the amount of time in minutes spent in the air. Add a new column air_spd
that will contain aircraft’s airspeed in mph,
as above, but keep only the new air_spd
variable,
use rownames_to_column()
on mtcars
to add car model as an extra column,
flights %>% mutate(air_spd = distance/(air_time / 60))
flights %>% transmute(air_spd = distance/(air_time / 60))
mtcars %>% rownames_to_column('model')
group_by()
, summarise()
and n()
to see how many planes were delayed (departure) every month,
flights %>%
filter(dep_delay > 0) %>%
group_by(month) %>%
summarise(num_dep_delayed = n())
tally()
and count()
,
flights %>%
filter(dep_delay > 0) %>%
group_by(month) %>%
tally()
flights %>%
filter(dep_delay > 0) %>%
count(month)
dep_delay
per month?
flights %>%
group_by(month) %>%
summarise(mean_dep_delay = mean(dep_delay, na.rm = T))
flights %>%
filter(arr_delay > 0) %>%
group_by(origin) %>%
summarise(cnt = n()) %>%
arrange(desc(cnt))
tally()
flights %>%
filter(arr_delay > 0) %>%
group_by(origin) %>%
tally(sort = T)
summarise()
to sum total dep_delay
per month in hours,
flights %>%
group_by(month) %>%
summarize(tot_dep_delay = sum(dep_delay/60, na.rm = T))
wt
parameter of count()
(works with tally()
too) to achieve the same,
flights %>%
group_by(month) %>%
count(wt = dep_delay/60)
group_size()
on carrier
what does it return?
flights %>%
group_by(carrier) %>%
group_size()
n_groups()
to check the number of unique origin-carrier pairs,
flights %>%
group_by(carrier) %>%
n_groups()
Note on ungroup
: Depending on the version of dplyr
, you may or may need to use the ungroup()
if you want to group your data on some other variables:
flights %>%
group_by(origin) %>%
mutate(mean_delay_orig = (mean(dep_delay, na.rm = T) + mean(arr_delay, na.rm = T)) / 2) %>%
ungroup() %>%
group_by(carrier) %>%
mutate(mean_delay_carr = (mean(dep_delay, na.rm = T) + mean(arr_delay, na.rm = T)) / 2) %>%
select(origin, carrier, mean_delay_orig, mean_delay_carr)
may or may need ungroup depending on your dplyr
version. In the newer versions, summarise
and mutate
drop one aggregation level.
Given the following tibbles set1
and set2
:
set1 <- tribble(
~id, ~color,
'id1', 'grey',
'id1', 'red',
'id2', 'green',
'id3', 'blue'
)
set2 <- tribble(
~id, ~size,
'id2', 'XL',
'id3', 'M',
'id4', 'M'
)
set1
set2
Perform joins on id
that result in the grey area from the Venn diagrams below. We have not talked about all possible joins, so read the docs if you do not know which join to use.
left_join(set1, set2, by = 'id')
right_join(set1, set2, by = 'id')
inner_join(set1, set2, by = 'id') # or
semi_join(set1, set2, by = 'id') # semi_join removes duplicates in x
# and also returns only columns from x.
full_join(set1, set2, by = 'id') # or
anti_join(set1, set2, by = 'id') # or
Now time to do some data tidying. First install a package with some untidy data:
#renv::install("rstudio/EDAWR")
library(EDAWR)
cases
so that years are not in separate columns, but in the column called year
containing a value per each year.
tidy_cases <- cases %>%
pivot_longer(-country, names_to = "year", values_to = "count")
pollution
dataset. Tidy it so that there separate columns for large
and small
pollution values.
tidy_pollution <- pollution %>%
pivot_wider(city, names_from = size, values_from = amount)
storms
dataset contains the date
column. Make it into 3 columns: year
, month
and day
. Store the result as tidy_storms
tidy_storms <- storms %>%
separate(col = date,
into = c("year", "month", "day"),
sep = "-")
year
, month
and day
in tidy_storms
into a date
column again but in the “DD/MM/YYYY” format.
tidy_storms %>% unite(col = "date", 4:6, sep = "/")
You will be given a fastq
file coming from MinION sequencer (Oxford Nanopore). This file contains test reads from the chicken genome. The flow-cell used here has 512 channels, each channel consists of 4 pores and only one pore is active at a time. Once your sequence gets stuck for some reason, the device will attempt to remove it from the pore by playing with reversing polarity on that pore. If this was successful, the pore will be re-used. Your task will be to visualise reading events from the meta-data in the fastq
dataset and to see how each and every channel behaved. Also, you will plot the distribution of reading times.
Datasets are available here.
First, we will need to load the necessary libraries. I will give you a hint – you need the following libraries:
here
– not necessary, but it is an elegant way of reading the data locally from the project folder,tidyverse
– well, quite obvious why,ShortRead
from Bioconductor – to deal with short reads in fastq
,lubridate
– to figure out reading times.library(here)
library(tidyverse)
library(ShortRead)
library(lubridate)
Now, let’s read the fastq data. Check ShortRead
documentation to see how to read our fastq
file. Also, try to use package here
. If you write: data <- here::here('data/my.fastq')
, the my.fastq
file will be read from the data
folder which is a sub folder of your project root, i.e. the folder where your r script is. It is a good practice and also prevents Jenny Bryan from coming to your office and setting your computer on fire.
Now think a bit, to plot reading events, do we need all the data in the file or only some specific part? You may want to see some few first lines of the fastq
to learn about the data structure.
raw_data <- here::here('docs/tidyverse_Marcin/lab/lab_assets/FUL1_fastqs_GRID2.fastq')
f <- ShortRead::FastqFile(raw_data)
rfq <- ShortRead::readFastq(f)
headers <- rfq@id
close(f)
In this step, we are extracting data from fastq headers of each and every read in the fastq file. Not super efficient and perhaps the slowest step of the whole analyses. Can you do it better than our example solution?
Desired output: a table (tibble/data.frame) with reads as rows and meta-data as columns.
Use strsplit()
to explode string data into columns and str_remove_all()
to get rid of unwanted characters. Since we did not talk much about regular expressions and the stringr
package:
strsplit
. Here we split on comma.text <- "This text is long, or not?"
strsplit(text, ',')
str_remove_all(text, ",.*")
data <- dplyr::as_tibble(matrix(NA_character_, ncol = 6, nrow = length(headers)), .name_repair = 'minimal')
colnames(data) <- c('id', 'run_id', 'sample_id', 'read', 'channel', 'start_time')
for (i in 1:length(headers)) {
data[i,] <- toString(headers[[i]]) %>%
strsplit(' ') %>%
unlist() %>%
str_remove_all(".*=") %>% t()
}
Now, the fun part begins:
start_dttm
that represents start time for a given read as proper datetime
object (read lubridate
docs) andchan
that is the proper numeric representation of the channel, Read about lead()
data2 <- data %>%
mutate(start_dttm = as_datetime(start_time)) %>%
mutate(chan=as.numeric(channel)) %>%
group_by(chan) %>%
arrange(start_dttm) %>%
mutate(time_diff = lead(start_dttm) - start_dttm) %>%
arrange(chan)
Here, we want to see what was happening in each channel over time. Plot the data you have just prepared so that:
Can you visualise this in a better way? Different geometry?
ggplot(data2, mapping = aes(x = start_dttm,
y = as.factor(chan),
col = as.numeric(time_diff)
)
) +
geom_point(size = .5) +
theme_bw()
Now, we want to see how time-to-next-read is distributed. Since it has a veeeeery long right tail, I am cutting off everything above 200 seconds (just by eyeballing).
histogram
),
# Show time-to-next read distribution
# thr <- mean(data2$time_diff, na.rm = T) + 3 * sd(data2$time_diff, na.rm = T)
tmp <- data2 %>%
ungroup() %>%
filter(time_diff < 200) %>%
select(time_diff)
hist(as.numeric(tmp$time_diff), breaks = 1000, las=1)
In this challenge, your task will be to analyse species composition of some samples. The samples, were actual products containing parts of plants. DNA has been isolated form the samples and an amplicon metabarcoding was performed using two sets of primers: for the ITS1 and the ITS2 region. Each sample had 3 technical replicates. Your task will be to transform BLAST output to a tidy form suitable for further analyses or visualisation.
We will obviously need tidyverse
, we will also do some string manipulations with stringr
also here
package is good to have.
library(tidyverse)
library(stringr)
library(here)
Here, we will define our input variables. We need:
file
that contains the path to the dataset,sample_name
is a string, the name of the sample you want to analyse,threshold
is an integer saying what is a the minimal number of replicates that have to contain an OTU in order to call it a true positive (TP),strict
a logical. If set to TRUE, only the OTUs deemed TP will be shown.Below we set some example values:
# Change the path to your project path, where your data is
file <- here::here("docs/tidyverse_Marcin/lab/lab_assets/blast_result.csv")
sample_name <- 'SAMPLE12'
threshold <- 1
strict <- F
Now, you should read the data:
species_orig <- read_csv(file, col_names = c("sample","its","replicate","OTU","size","hit","perc_ident","score","family","species")) %>%
select(-score)
head(species_orig,n = 10)
As you see, the following information are included in the data:
sample
is simply the name of the sample,its
is either ITS1 or ITS2 and tells which set of PCR primers has been used,replicate
contains information on which replicate the sequences come from,OTU
is a unique identifier of the so-called Operational Taxonomic Unit, an OTU often corresponds to one species but not always. Sometimes 2 OTUs represent the same species, sometimes 1 OTU consists of more than one species,size
is the number of reads that support that particular OTU,hit
is the BLAST hit identifier. The 4 top BLAST hits are reported per OTU,perc_identity
is the percentage identity of the sequence to the BLAST hit,family
is the identified plant family,species
is the identified plant species.Create a new dataset species
that contains an extra column n_replicates
. The column contains number of replicates this particular species is present in. Do it per sample and its.
species <- species_orig %>%
group_by(sample, its, species) %>%
mutate(n_replicates = n_distinct(replicate)) %>%
ungroup()
head(species,n = 10)
Now, your task is to filter out all but your sample_name
samples from the dataset.
Call the resulting dataset my_sample
.
my_sample <- species %>%
filter(sample == sample_name)
What happens if a set of primers failed to amplify or if one replicate was lost?
Use complete()
to make sure you have NA
values in such cases.
my_sample <- my_sample %>%
complete(its = c("ITS1", "ITS2"),
replicate = c("R1","R2","R3"))
Look, the first sample in the table is SAMPLE10
. Why not SAMPLE1
?
That’s a sorting issue: if sorted as character, 10 will come before 1. WE have to fix this by adding trailing zero to the values in OTU
. We do not expect more than 99 OTUs in a sample, so it is ok with only one trailing 0 (otherwise the 100-th sample will spoil our sorting and come out like: SAMPLE100, SAMPLE01, SAMPLE10). We will need to use regular expression:
OTU
column that follow pattern “OTUdigit” we need to change to “OTU0digit”. Regular expression that matches this is OTU([0-9]$)
and it should be replaced by: OTU0\\1
. Ask your TAs to explain this if you do not know much about regular expressions and pattern matching.my_sample <- my_sample %>%
mutate(OTU = str_replace(OTU,pattern = "OTU([0-9]$)",
replacement = "OTU0\\1"))
head(my_sample,n = 10)
Sometimes, an OTU generates two or more top BLAST hits that come from the same species. We have decided to sum reads in such cases. Do it!
my_sample <- my_sample %>%
ungroup() %>%
group_by(sample, its, replicate, OTU, species, n_replicates) %>%
summarise(n_reads = sum(size)) %>%
ungroup() %>%
group_by(its, species, OTU)
head(my_sample,n=10)
Now, we want to see how many identifications an OTU got. Implement this. Store the result in a new tibble diversity
.
diversity <- my_sample %>%
ungroup() %>%
group_by(its, replicate, OTU) %>%
summarise(n_species = n())
head(diversity,n=10)
Add the diversity
data to my_sample
using appropriate join
function. Also, remove the column with sample names since we are dealing with only one sample.
my_sample <- my_sample %>%
left_join(diversity) %>%
select(-sample)
head(my_sample,n=10)
Can you think of a good way of visualising the data? Think of:
ITS
, replicate
, n_reads
, n_replicates
, OTU
, threshold
, n_species
and specie
. Well, pretty much all of them :-)Our example solution involves some ggplot2
magics. But would base-R be good enough for this type of plot?
Use the FAA report and tidyverse
to learn more about aircraft incidents with wildlife. Use your imagination and NYC data science blog for inspiration!
## R version 4.0.2 (2020-06-22)
## Platform: x86_64-conda_cos6-linux-gnu (64-bit)
## Running under: Ubuntu 20.04.2 LTS
##
## Matrix products: default
## BLAS/LAPACK: /home/roy/miniconda3/envs/r-4.0/lib/libopenblasp-r0.3.10.so
##
## locale:
## [1] LC_CTYPE=en_GB.UTF-8 LC_NUMERIC=C
## [3] LC_TIME=en_GB.UTF-8 LC_COLLATE=en_GB.UTF-8
## [5] LC_MONETARY=en_GB.UTF-8 LC_MESSAGES=en_GB.UTF-8
## [7] LC_PAPER=en_GB.UTF-8 LC_NAME=C
## [9] LC_ADDRESS=C LC_TELEPHONE=C
## [11] LC_MEASUREMENT=en_GB.UTF-8 LC_IDENTIFICATION=C
##
## attached base packages:
## [1] stats4 parallel stats graphics grDevices
## [6] utils datasets methods base
##
## other attached packages:
## [1] lubridate_1.7.10
## [2] ShortRead_1.48.0
## [3] GenomicAlignments_1.26.0
## [4] SummarizedExperiment_1.20.0
## [5] Biobase_2.50.0
## [6] MatrixGenerics_1.2.1
## [7] matrixStats_0.59.0
## [8] Rsamtools_2.6.0
## [9] GenomicRanges_1.42.0
## [10] GenomeInfoDb_1.26.7
## [11] Biostrings_2.58.0
## [12] XVector_0.30.0
## [13] IRanges_2.24.1
## [14] S4Vectors_0.28.1
## [15] BiocParallel_1.24.1
## [16] BiocGenerics_0.36.1
## [17] here_1.0.1
## [18] eulerr_6.1.0
## [19] bsplus_0.1.2
## [20] EDAWR_0.1
## [21] magrittr_2.0.1
## [22] gganimate_1.0.7
## [23] plotly_4.9.3
## [24] ggiraph_0.7.10
## [25] randomcoloR_1.1.0.1
## [26] ggpointdensity_0.1.0
## [27] patchwork_1.1.1
## [28] forcats_0.5.1
## [29] stringr_1.4.0
## [30] dplyr_1.0.6
## [31] purrr_0.3.4
## [32] readr_1.4.0
## [33] tidyr_1.1.3
## [34] tibble_3.1.2
## [35] tidyverse_1.3.1
## [36] fontawesome_0.2.1
## [37] captioner_2.2.3
## [38] ggplot2_3.3.3
## [39] reticulate_1.20
## [40] bookdown_0.22
## [41] knitr_1.33
## [42] renv_0.13.2
##
## loaded via a namespace (and not attached):
## [1] utf8_1.2.1 tidyselect_1.1.1
## [3] htmlwidgets_1.5.3 grid_4.0.2
## [5] Rtsne_0.15 munsell_0.5.0
## [7] codetools_0.2-18 ica_1.0-2
## [9] future_1.21.0 gifski_1.4.3-1
## [11] miniUI_0.1.1.1 withr_2.4.2
## [13] colorspace_2.0-1 highr_0.9
## [15] uuid_0.1-4 rstudioapi_0.13
## [17] Seurat_4.0.2 ROCR_1.0-11
## [19] xaringan_0.21 tensor_1.5
## [21] listenv_0.8.0 labeling_0.4.2
## [23] GenomeInfoDbData_1.2.4 hwriter_1.3.2
## [25] polyclip_1.10-0 farver_2.1.0
## [27] rprojroot_2.0.2 parallelly_1.25.0
## [29] vctrs_0.3.8 generics_0.1.0
## [31] xfun_0.23 pagedown_0.14
## [33] R6_2.5.0 DelayedArray_0.16.3
## [35] bitops_1.0-7 spatstat.utils_2.1-0
## [37] assertthat_0.2.1 promises_1.2.0.1
## [39] scales_1.1.1 gtable_0.3.0
## [41] globals_0.14.0 processx_3.5.2
## [43] goftest_1.2-2 rlang_0.4.11
## [45] systemfonts_1.0.2 splines_4.0.2
## [47] lazyeval_0.2.2 spatstat.geom_2.1-0
## [49] broom_0.7.6 BiocManager_1.30.15
## [51] yaml_2.2.1 reshape2_1.4.4
## [53] abind_1.4-5 modelr_0.1.8
## [55] backports_1.2.1 httpuv_1.6.1
## [57] rsconnect_0.8.18 tools_4.0.2
## [59] ellipsis_0.3.2 spatstat.core_2.1-2
## [61] jquerylib_0.1.4 RColorBrewer_1.1-2
## [63] servr_0.22 ggridges_0.5.3
## [65] Rcpp_1.0.6 plyr_1.8.6
## [67] progress_1.2.2 zlibbioc_1.36.0
## [69] RCurl_1.98-1.3 ps_1.6.0
## [71] prettyunits_1.1.1 rpart_4.1-15
## [73] deldir_0.2-10 pbapply_1.4-3
## [75] cowplot_1.1.1 zoo_1.8-9
## [77] SeuratObject_4.0.1 haven_2.4.1
## [79] ggrepel_0.9.1 cluster_2.1.2
## [81] fs_1.5.0 data.table_1.14.0
## [83] scattermore_0.7 lmtest_0.9-38
## [85] reprex_2.0.0 RANN_2.6.1
## [87] fitdistrplus_1.1-5 hms_1.1.0
## [89] mime_0.10 evaluate_0.14
## [91] xtable_1.8-4 jpeg_0.1-8.1
## [93] readxl_1.3.1 gridExtra_2.3
## [95] compiler_4.0.2 KernSmooth_2.23-20
## [97] V8_3.4.2 crayon_1.4.1
## [99] websocket_1.4.0 htmltools_0.5.1.1
## [101] mgcv_1.8-36 later_1.2.0
## [103] DBI_1.1.1 tweenr_1.0.2
## [105] dbplyr_2.1.1 MASS_7.3-54
## [107] rappdirs_0.3.3 Matrix_1.3-4
## [109] cli_2.5.0 igraph_1.2.6
## [111] pkgconfig_2.0.3 spatstat.sparse_2.0-0
## [113] xml2_1.3.2 bslib_0.2.5.1
## [115] rvest_1.0.0 callr_3.7.0
## [117] digest_0.6.27 sctransform_0.3.2
## [119] RcppAnnoy_0.0.18 spatstat.data_2.1-0
## [121] polylabelr_0.2.0 rmarkdown_2.8
## [123] cellranger_1.1.0 leiden_0.3.8
## [125] uwot_0.1.10 curl_4.3.1
## [127] shiny_1.6.0 lifecycle_1.0.0
## [129] nlme_3.1-152 jsonlite_1.7.2
## [131] viridisLite_0.4.0 fansi_0.5.0
## [133] pillar_1.6.1 lattice_0.20-44
## [135] fastmap_1.1.0 httr_1.4.2
## [137] pkgbuild_1.2.0 survival_3.2-11
## [139] glue_1.4.2 remotes_2.4.0
## [141] png_0.1-7 stringi_1.6.2
## [143] sass_0.4.0 latticeExtra_0.6-29
## [145] irlba_2.3.3 future.apply_1.7.0
Built on: 17-Jun-2021 at 09:58:31.
2021 • SciLifeLab • NBIS • RaukR