We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from.
# random normal dataset.seed(666)r_two <-data.frame(dat =c(rnorm(n =20, mean =4, sd =1),rnorm(n =20, mean =5, sd =1)),sample =c(rep("A", 20), rep("B", 20)))# perform t-test# note how we set the `var.equal` argument to TRUE because we know # our data has the same SD (they are simulated as such!)t.test(dat ~ sample, data = r_two, var.equal =TRUE)
Two Sample t-test
data: dat by sample
t = -1.9544, df = 38, p-value = 0.05805
alternative hypothesis: true difference in means between group A and group B is not equal to 0
95 percent confidence interval:
-1.51699175 0.02670136
sample estimates:
mean in group A mean in group B
4.001438 4.746584
Repeat this analyses using the Welch’s t.test(). (/5)
Repeat your analysis, above, using the even more old-fashioned Equation 4 in the lecture. Show the code and talk us through the step you followed to read the p-values off the table of t-statistics. (/10)
Answer
We simply change the argument var.equal to FALSE in the t.test() function. This is because we are using a Welch’s t-test, which does not assume equal variances between the two groups.
# random normal dataset.seed(666)r_two <-data.frame(dat =c(rnorm(n =20, mean =4, sd =1),rnorm(n =20, mean =5, sd =1)),sample =c(rep("A", 20), rep("B", 20)))# perform t-test# note how we set the `var.equal` argument to FALSE because we want# to do a Welch's t-testt.test(dat ~ sample, data = r_two, var.equal =FALSE)
Welch Two Sample t-test
data: dat by sample
t = -1.9544, df = 36.501, p-value = 0.05835
alternative hypothesis: true difference in means between group A and group B is not equal to 0
95 percent confidence interval:
-1.51803431 0.02774392
sample estimates:
mean in group A mean in group B
4.001438 4.746584
✓ (x 5) for this simple calculation.
To do the Welch’s t-test manually (as one would do 30 years ago with a hand-held calculator), we need to calculate the t-statistic and the degrees of freedom. The t-statistic is calculated by:
Here, and are the variances of groups and , respectively (see Section X). The d.f. to use with Welch’s t-test is obtained using the Welch–Satterthwaite equation:
where and are the sample sizes of groups and , respectively. The p-value is then obtained from the t-distribution with the degrees of freedom calculated above (see below).
✓ (x 10) What do we do with the d.f. and the t-statistic? We look them up in the t-distribution table (find one!) to find the critical value of t for a given significance level (e.g., 0.05). If our calculated t-statistic (take the absolute) is greater than the critical value at the d.f. we are using, we reject the null hypothesis. If it is less than the critical value, we fail to reject the null hypothesis.
Question 3
Please report back on Task F.3 presented in the lecture. Write up formal Methods and Results sections. (/15)
Answer
✓
Question 4
Please report back the analysis and results for Task F.4. in the lecture. Write up formal Methods and Results sections. (/15)
---title: "BCB744 Task F"format: html: fig-format: svg fig_retina: 2 fig-dpi: 400params: hide_answers: false---```{r, echo=FALSE}knitr::opts_chunk$set(echo =TRUE,eval =TRUE,warning =FALSE,message =FALSE,fig.width =6,fig.asp =0.65,out.width ="100%",fig.align ="center")library(tidyverse)library(ggpubr)library(ggthemes)```# [[Assessment Sheet](BCB744_Task_F_Surname.xlsx)]{.my-highlight} {#sec-assessment}# 7. *t*-Tests## Question 1Please report back on Task F.1 presented in the lecture. Write up formal Methods and Results sections. **(/15)**`r if (params$hide_answers) "::: {.content-hidden}"`**Answer**```{r}```- ✓ `r if (params$hide_answers) ":::"`## Question 2Please refer to the [two-sided two-sample *t*-test in the lecture.](https://tangledbank.netlify.app/BCB744/basic_stats/07-t_tests.html#two-sided-two-sample-t-test). It is recreated here:```{r}# random normal dataset.seed(666)r_two <-data.frame(dat =c(rnorm(n =20, mean =4, sd =1),rnorm(n =20, mean =5, sd =1)),sample =c(rep("A", 20), rep("B", 20)))# perform t-test# note how we set the `var.equal` argument to TRUE because we know # our data has the same SD (they are simulated as such!)t.test(dat ~ sample, data = r_two, var.equal =TRUE)```a. Repeat this analyses using the Welch's `t.test()`. **(/5)**b. Repeat your analysis, above, using the even more old-fashioned Equation 4 in the lecture. Show the code and talk us through the step you followed to read the *p*-values off the table of *t*-statistics. **(/10)**`r if (params$hide_answers) "::: {.content-hidden}"`**Answer**a. We simply change the argument `var.equal` to `FALSE` in the `t.test()` function. This is because we are using a Welch's *t*-test, which does not assume equal variances between the two groups.```{r}# random normal dataset.seed(666)r_two <-data.frame(dat =c(rnorm(n =20, mean =4, sd =1),rnorm(n =20, mean =5, sd =1)),sample =c(rep("A", 20), rep("B", 20)))# perform t-test# note how we set the `var.equal` argument to FALSE because we want# to do a Welch's t-testt.test(dat ~ sample, data = r_two, var.equal =FALSE)```- ✓ (x 5) for this simple calculation.b. To do the Welch's *t*-test manually (as one would do 30 years ago with a hand-held calculator), we need to calculate the *t*-statistic and the degrees of freedom. The *t*-statistic is calculated by:$$t=\frac{\bar{A}-\bar{B}}{\sqrt{\frac{S^{2}_{A}}{n}+\frac{S^{2}_{B}}{m}}}$$Here, $S_{A}$ and $S_{B}$ are the variances of groups $A$ and $B$, respectively (see Section X). The *d.f.* to use with Welch's *t*-test is obtained using the Welch--Satterthwaite equation:$$d.f. = \frac{\left( \frac{S^{2}_{A}}{n}+\frac{S^{2}_{B}}{m} \right)^{2}}{\left( \frac{S^{4}_{A}}{n-1} + \frac{S^{4}_{B}}{m-1} \right)}$$where $n$ and $m$ are the sample sizes of groups $A$ and $B$, respectively. The *p*-value is then obtained from the *t*-distribution with the degrees of freedom calculated above (see below).```{r}# calculate the t-statistict_stat <- (mean(r_two$dat[r_two$sample =="A"]) -mean(r_two$dat[r_two$sample =="B"])) /sqrt((var(r_two$dat[r_two$sample =="A"]) /length(r_two$dat[r_two$sample =="A"])) + (var(r_two$dat[r_two$sample =="B"]) /length(r_two$dat[r_two$sample =="B"])))# calculate the degrees of freedomdf <- ((var(r_two$dat[r_two$sample =="A"]) /length(r_two$dat[r_two$sample =="A"])) + (var(r_two$dat[r_two$sample =="B"]) /length(r_two$dat[r_two$sample =="B"])))^2/ (((var(r_two$dat[r_two$sample =="A"]) /length(r_two$dat[r_two$sample =="A"]))^2/ (length(r_two$dat[r_two$sample =="A"]) -1)) + ((var(r_two$dat[r_two$sample =="B"]) /length(r_two$dat[r_two$sample =="B"]))^2/ (length(r_two$dat[r_two$sample =="B"]) -1)))print(t_stat)print(df)```- ✓ (x 10) What do we do with the d.f. and the *t*-statistic? We look them up in the *t*-distribution table (find one!) to find the critical value of *t* for a given significance level (e.g., 0.05). If our calculated *t*-statistic (take the absolute) is greater than the critical value at the d.f. we are using, we reject the null hypothesis. If it is less than the critical value, we fail to reject the null hypothesis.`r if (params$hide_answers) ":::"`## Question 3Please report back on Task F.3 presented in the lecture. Write up formal Methods and Results sections. **(/15)**`r if (params$hide_answers) "::: {.content-hidden}"`**Answer**```{r}```- ✓`r if (params$hide_answers) ":::"`## Question 4Please report back the analysis and results for Task F.4. in the lecture. Write up formal Methods and Results sections. **(/15)**`r if (params$hide_answers) "::: {.content-hidden}"`**Answer**```{r}```- ✓`r if (params$hide_answers) ":::"`