<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en-GB">
	<id>https://training-course-material.com/index.php?action=history&amp;feed=atom&amp;title=Inferential_Statistics_for_b_and_r</id>
	<title>Inferential Statistics for b and r - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://training-course-material.com/index.php?action=history&amp;feed=atom&amp;title=Inferential_Statistics_for_b_and_r"/>
	<link rel="alternate" type="text/html" href="https://training-course-material.com/index.php?title=Inferential_Statistics_for_b_and_r&amp;action=history"/>
	<updated>2026-05-13T10:25:28Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.45.1</generator>
	<entry>
		<id>https://training-course-material.com/index.php?title=Inferential_Statistics_for_b_and_r&amp;diff=17986&amp;oldid=prev</id>
		<title>Ahnboyoung: /* Assumptions */</title>
		<link rel="alternate" type="text/html" href="https://training-course-material.com/index.php?title=Inferential_Statistics_for_b_and_r&amp;diff=17986&amp;oldid=prev"/>
		<updated>2014-06-03T23:14:30Z</updated>

		<summary type="html">&lt;p&gt;&lt;span class=&quot;autocomment&quot;&gt;Assumptions&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{{Cat|Regression| 05}}&lt;br /&gt;
=Assumptions=&lt;br /&gt;
* Although no assumptions were needed to determine the best-fitting straight line, assumptions are made in the calculation of inferential statistics. &lt;br /&gt;
* Naturally, these assumptions refer to the population, not the sample.&lt;br /&gt;
# Linearity: The relationship between the two variables is linear.&lt;br /&gt;
# Homoscedasticity: The variance around the regression line is the same for all values of X. A clear violation of this assumption is shown in below. (Notice that the predictions for students with high high-school GPAs are very good, whereas the predictions for students with low high-school GPAs are not very good. In other words, the points for students with high high-school GPAs are close to the regression line, whereas the points for low high-school GPA students are not.)&lt;br /&gt;
# The errors of prediction are distributed normally. This means that the deviations from the regression line are normally distributed. It does not mean that X or Y is normally distributed.&lt;br /&gt;
&lt;br /&gt;
[[File:ClipCapIt-140603-221400.PNG]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Significance Test for the Slope (b)=&lt;br /&gt;
;The general formula for a t test:&lt;br /&gt;
:[[File:ClipCapIt-140603-235845.PNG]]&lt;br /&gt;
&lt;br /&gt;
As applied here, the statistic is the sample value of the slope (b) and the hypothesized value is 0. &lt;br /&gt;
&lt;br /&gt;
;The number of degrees of freedom for this test is:&lt;br /&gt;
 df = N-2&lt;br /&gt;
 where N is the number of pairs of scores.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
;The estimated standard error of b is computed using the following formula:&lt;br /&gt;
 [[File:ClipCapIt-140603-235928.PNG]]&lt;br /&gt;
 s&amp;lt;sub&amp;gt;b&amp;lt;/sub&amp;gt; is the estimated standard error of b, &lt;br /&gt;
 s&amp;lt;sub&amp;gt;est&amp;lt;/sub&amp;gt; is the standard error of the estimate&lt;br /&gt;
 SSX is the sum of squared deviations of X from the mean of X&lt;br /&gt;
&lt;br /&gt;
; SSX is calculated as&lt;br /&gt;
 [[File:ClipCapIt-140604-000043.PNG]]&lt;br /&gt;
 where Mx is the mean of X&lt;br /&gt;
&lt;br /&gt;
;The standard error of the estimate can be calculated as&lt;br /&gt;
 [[File:ClipCapIt-140604-000058.PNG]]&lt;br /&gt;
&lt;br /&gt;
==Example==&lt;br /&gt;
[[File:ClipCapIt-140604-000213.PNG]]&lt;br /&gt;
* The column X has the values of the predictor variable &lt;br /&gt;
* The column Y has the values of the criterion variable&lt;br /&gt;
* The column x has the differences between the values of column X and the mean of X&lt;br /&gt;
* The column x&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; is the square of the x column&lt;br /&gt;
* The column y has the differences between the values of column Y and the mean of Y. &lt;br /&gt;
* The column y&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; is simply square of the y column&lt;br /&gt;
&lt;br /&gt;
;The standard error of the estimate&lt;br /&gt;
The computation of the standard error of the estimate (s&amp;lt;sub&amp;gt;est&amp;lt;/sub&amp;gt;) for these data is shown in the section on the standard error of the estimate. It is equal to 0.964.&lt;br /&gt;
 s&amp;lt;sub&amp;gt;est&amp;lt;/sub&amp;gt; = 0.964&lt;br /&gt;
&lt;br /&gt;
;SSX&lt;br /&gt;
SSX is the sum of squared deviations from the mean of X. i.e. it is equal to the sum of the x2 column and is equal to 10.&lt;br /&gt;
 SSX = 10.00&lt;br /&gt;
&lt;br /&gt;
We now have all the information to compute the standard error of b:&lt;br /&gt;
&lt;br /&gt;
;the slope (b) is &lt;br /&gt;
 b= 0.425. &lt;br /&gt;
 df = N-2 = 5-2 = 3.&lt;br /&gt;
&lt;br /&gt;
*The p value for a two-tailed t test is 0.26. &lt;br /&gt;
*Therefore, the slope is not significantly different from 0.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Confidence Interval for the Slope=&lt;br /&gt;
* The method for computing a confidence interval for the population slope is very similar to methods for computing other confidence intervals. &lt;br /&gt;
* For the 95% confidence interval, the formula is:&lt;br /&gt;
 lower limit: b - (t.95)(sb)&lt;br /&gt;
 upper limit: b + (t.95)(sb)&lt;br /&gt;
 where t.95 is the value of t to use for the 95% confidence interval&lt;br /&gt;
&lt;br /&gt;
==Example==&lt;br /&gt;
[[File:ClipCapIt-140604-000620.PNG]]&lt;br /&gt;
* The values of t to be used in a confidence interval can be looked up in a table of the t distribution. &lt;br /&gt;
* A small version of such a table is shown above. &lt;br /&gt;
* The first column, df, stands for degrees of freedom.&lt;br /&gt;
* You can also use the &amp;quot;inverse t distribution&amp;quot; calculator to find the t values to use in a confidence interval.&lt;br /&gt;
* Applying these formulas to the example data,&lt;br /&gt;
 lower limit: 0.425 - (3.182)(0.305) = -0.55&lt;br /&gt;
 upper limit: 0.425 + (3.182)(0.305) = 1.40&lt;br /&gt;
&lt;br /&gt;
=Significance Test for the Correlation=&lt;br /&gt;
The formula for a significance test of Pearson&amp;#039;s correlation is shown below:&lt;br /&gt;
 [[File:ClipCapIt-140604-000727.PNG]]&lt;br /&gt;
 where N is the number of pairs of scores. &lt;br /&gt;
&lt;br /&gt;
For the example data,&lt;br /&gt;
 [[File:ClipCapIt-140604-000806.PNG]]&lt;br /&gt;
&lt;br /&gt;
Notice that this is the same t value obtained in the t test of b. &lt;br /&gt;
As in that test, the degrees of freedom is &lt;br /&gt;
 N - 2 = 5 -2 = 3.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Quiz=&lt;br /&gt;
&amp;lt;quiz display=simple &amp;gt;&lt;br /&gt;
{ Which of the following are assumptions made in the calculation of regression inferential statistics? &lt;br /&gt;
&lt;br /&gt;
|type=&amp;quot;[]&amp;quot;}&lt;br /&gt;
+A:The errors of prediction are normally distributed.&lt;br /&gt;
-B:X is normally distributed.&lt;br /&gt;
-C:Y is normally distributed.&lt;br /&gt;
+D:The variance around the regression line is the same for all values of X.&lt;br /&gt;
+E:The relationship between X and Y is linear.&lt;br /&gt;
&lt;br /&gt;
{&lt;br /&gt;
{{Show Answer|&lt;br /&gt;
A,D,E&lt;br /&gt;
&lt;br /&gt;
The assumptions are linearity, homoscedasticity, and normally distributed errors. See the text for more information.&lt;br /&gt;
}}&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{The slope of a regression line is 0.8, and the standard error of the slope is 0.3. The sample used to compute this regression line consisted of 12 participants. Compute the 95% confidence interval for the slope. Type the upper limit of the confidence interval in the box below.  &lt;br /&gt;
&lt;br /&gt;
|type=&amp;quot;{}&amp;quot;}&lt;br /&gt;
{ 1.47 }&lt;br /&gt;
&lt;br /&gt;
{&lt;br /&gt;
{{Show Answer|&lt;br /&gt;
1.47&lt;br /&gt;
&lt;br /&gt;
Use the table in this section or the inverse t distribution calculator to find that the critical value is t(N-2).&lt;br /&gt;
&lt;br /&gt;
t(10) s 2.23. &lt;br /&gt;
&lt;br /&gt;
The upper limit of the 95% CI is b + (t)(sb)&lt;br /&gt;
&lt;br /&gt;
.8 + 2.23(.3) equals to 1.47.&lt;br /&gt;
}}&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
{In a sample of 20, the correlation between two variables is .5. Determine if this correlation is significant at the .05 level by calculating the t value. &lt;br /&gt;
&lt;br /&gt;
|type=&amp;quot;{}&amp;quot;}&lt;br /&gt;
{ 1.47 }&lt;br /&gt;
&lt;br /&gt;
{&lt;br /&gt;
{{Show Answer|&lt;br /&gt;
2.45&lt;br /&gt;
&lt;br /&gt;
t is (r) sqrt(N-2)/sqrt(1-r2) equals to (0.5) sqrt(18)/sqrt(1-.25) is 2.45 (This is significant at the .05 level.)&lt;br /&gt;
}}&lt;br /&gt;
}&lt;br /&gt;
&lt;br /&gt;
{Calculate the lower limit of the 95% confidence interval for the correlation of .75 (N = 25). &lt;br /&gt;
&lt;br /&gt;
|type=&amp;quot;{}&amp;quot;}&lt;br /&gt;
{ 0.505 }&lt;br /&gt;
&lt;br /&gt;
{&lt;br /&gt;
{{Show Answer|&lt;br /&gt;
0.505&lt;br /&gt;
&lt;br /&gt;
First, convert r to z&amp;#039; (so .75 -&amp;gt; .973). The standard error of z&amp;#039; is 1/sqrt(N-3)is .213. &lt;br /&gt;
&lt;br /&gt;
Lower limit of CI is .973 - 1.96(.213) equals to 0.556. Now convert back from z&amp;#039; to r. r is .505&lt;br /&gt;
}}&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/quiz&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ahnboyoung</name></author>
	</entry>
</feed>