SBI-listserv participants
See below for an invitation to participate in an assessment project.
----------------------------------------------------------------------------
As you may know, we recently received NSF funding (DUE-1323210) to
facilitate assessment of (algebra-based) introductory statistics courses,
with a focus on gaining a better understanding of potential differences in
student learning between “traditional” and simulation/randomization-based
introductory statistics courses. As such, we are asking you to consider
having your students participate in the assessment project *regardless of
how much (if any) simulation- and randomization-based inference methods you
use in your course*. As a thank you for your participation, we are happy to
offer a $100 stipend and a customized report on your students’ performance
in your class. If you are interested in participating, please fill out this
short survey, as soon as possible, but early enough to allow time to set up
individualized links for your class before your term starts:
https://www.surveymonkey.com/s/9SYS8H3.
Some brief details follow, with answers to some commonly asked questions
here <http://homepages.dordt.edu/ntintle/faqs.pdf>.
1. Students in your introductory statistics course (undergraduate or
high school level) will take a pre-test (preferably before the course
starts, but no later than the first week of classes). The “test” contains
multiple choice questions that assess conceptual understanding and student
attitudes toward statistics. Most students take approximately 45 minutes
to complete the test. The test is administered completely online (we will
provide you the link). Students can complete the test either inside or
outside of class. See the FAQ for more information on encouraging student
participation and IRB including opt-out options. After the closing date
you specify, we will send you the student names and individual performance
data. At the end of your course, students will take a single multiple
choice, online post-test about attitudes and concepts (or, for the
post-test only, separated concepts and attitudes tests). Again, we’ll
provide you the individualized link.
2. Assuming your sections attain at least 75% participation rates, and
you fill out a brief (<30 minute) survey about your course (e.g., size,
pedagogy, classroom technology, etc.) at the conclusion of the course, you
will receive the stipend and customized report.
We anticipate publishing a series of articles on the data gathered as part
of this project. Neither student-nor instructor-level information will be
reported individually (only in aggregate) in these articles. If you are
interested in participating but have questions about your institution’s
IRB, please contact us as well.
Again, if you are interested in participating, please fill out this short
survey, as soon as possible, but early enough to allow time to set up links
for your class before your semester starts:
https://www.surveymonkey.com/s/9SYS8H3. After we receive your survey
responses, you will be contacted directly by us with more
information/details.
Please direct additional questions to either Cindy Nederhoff (assessment
administrator: cindy.nederhoff(a)dordt.edu) or Nathan Tintle (project
director: nathan.tintle(a)dordt.edu).
Thanks for considering this,
Nathan Tintle (on behalf of the PIs: Nathan Tintle, Beth Chance, Dennis
Pearl, Soma Roy and Todd Swanson)
--
Nathan Tintle, Ph.D.
Associate Professor of Statistics and Dept. Chair
Director for Research and Scholarship
Dordt College
Sioux Center, IA 51250
nathan.tintle(a)dordt.edu
Phone: (712) 722-6264
Office: SB1612
Sorry for cross-posting, but just in case any of you have not seen this
announcement ...
-- Allan Rossman
-------- Forwarded Message --------
Subject: [CAUSE] Fwd: Teaching and Learning Webinar:Nathan Tintle,
Dordt College and Camille Fairbourn, Utah State University 12:00 to
12:30p.m. Eastern time, September 8th, 2015 (note special time)
Date: Thu, 6 Aug 2015 14:08:19 -0400
From: LAURA BURGHARD <lfb109(a)psu.edu>
To: cause(a)causeweb.org
------------------------------------------------------------------------
*Subject: *Teaching and Learning Webinar:Nathan Tintle, Dordt College
and Camille Fairbourn, Utah State University 12:00 to 12:30p.m. Eastern
time, September 8th, 2015 (note special time)
*Teaching and Learning Webinar Series*
"Reflections on making the switch to a simulation-based inference
curriculum"
with Nathan Tintle, Dordt College and Camille Fairbourn, Utah State
University
12:00 to 12:30p.m. Eastern time, Tuesday,September 8^th , 2015
/(note special time)/
*Abstract*: In this webinar some recent new adopters of simulation-based
inference (SBI) curricula will share their responses to questions such
as: What made you switch to SBI from a traditional curriculum? What have
you enjoyed most about the switch? What were some of the challenges in
switching? What would you do different next time?
To register for this webinar:
https://attendee.gotowebinar.com/register/6796765635866240257
*Logistics*: The webinar will be conducted using the GoToWebinar
software platform. A computer with internet access is all you need.
GoToWebinar offers audio participation through your computer
microphone. For participants in the US and Canada, if you prefer the
telephone for audio participation, this feature is also available.
All registered webinar attendees will receive a confirmation email
generated by the GoToWebinar system upon registering. This email
includes a link to enter the webinar. Keep the confirmation email as you
will use this link to enter the webinar â you will also be sent a
reminder with the link two hours before the webinar begins. Once you
leave the webinar, you cannot re-enter. If you have not used GoToWebinar
before, please review the information below. The webinar will be
recorded and the archived version will be available on-line within a few
days following the presentation, if you are unable to attend.
*/New to GoToWebinar?
/*
You will see the live presentation on your computer screen and the sound
of the presentation will come through your computer. If you can listen
to music or hear videos on your computer, your computer has the
capability for you to hear the presentation. For your voice to be heard
by the presenters, if you wish to ask a question, GTW gives viewers the
option of VoIP (voice over the internet) or a telephone option for U.S.
and Canadian participants. VoIP will allow everyone from around the
world to participate without additional cost. The option to change from
VoIP to phone will appear in the screen of options after you enter the
webinar. If you wish to change to phone, follow the instructions and
enter the provided pin number. If you are using VoIP, you will need a
microphone attached to your computer so that your question can be heard
by the presenter/audience.
GoToWebinar offers a short video, âAttendee Quick Start (5:09)
<http://support.citrixonline.com/en_US/GoToMeeting/video/GTMV00012>â,
which you may find helpful:
http://support.citrixonline.com/en_US/GoToMeeting/video/GTMV00012
*For PC-based participants:
*
* Internet Explorer 7.0 or newer, Mozilla Firefox 4.0 or newer or
Google Chrome 5.0 or newer. JavaScript must be enabled.
* Windows 8, 7, Vista, XP or 2003 Server.
* Cable modem, DSL, or better Internet connection.
* Dual-core 2.4GHz CPU or faster with 2GB of RAM or more.
* Participants wishing to connect to audio using VoIP will need a
fast Internet connection, a microphone and speakers (or USB headset).
*For Mac-based participants:
*
* Safari 3.0 or newer, Firefox 4.0 or newer or Google Chrome 5.0 or
newer. JavaScript must be enabled.
* Mac OS X 10.6 â Snow Leopard or newer.
* Intel processor with 1GB of RAM or more.
* Cable modem, DSL, or better Internet connection.
* Macs have built-in speakers and a microphone with ambient noise
reduction that will work well for VoIP.
*For participants with GoToMeeting app for iPad, iPhone, or Android:
*
* Free GoToMeeting app from the App Store or Google Play.
* WiFi connection recommended for VoIP audio.
*For attendees with GoToMeeting app for Windows RT tablet:
*
* Free GoToMeeting app from the Windows Store.
* x86, x64 or ARM processor.
Hello SBI listserv participants and SBI blog readers!
Happy 4th of July!
I want to let you know, in case you don't already know, that we have three new posts on the Simulation-based Inference blog (https://www.causeweb.org/sbi/):
1) We have posts by Ann Cannon and Erin Blankenship addressing the topic "How do you convince/train others to teach SBI?"
Ann's post is titled, "Dragged kicking and screaming by an algebraist" and Erin's post is titled "There's no convincing necessary if you're the boss: implementing the simulation-based approach with TA instructors." The links for these posts can be found here https://www.causeweb.org/sbi/?page_id=923
2) We have also added a post by Randall Pruim titled "How I teach SBI using R." This post can be found under the discussion topic "How I utilize technology."
https://www.causeweb.org/sbi/?page_id=402
We would love to hear your feedback and suggestions!
On behalf of the ISI team, I'd like to thank our blog contributors for writing these pieces for us, and you, our readers for reading these. Please let us know your thoughts and ideas on the various posts by leaving comments, suggestions, etc. on our blog.
Wish you a very nice weekend!
- Soma
-----------------------
Soma Roy
Associate Professor
Statistics
California Polytechnic State University
San Luis Obispo CA 93407
Phone no.: (805)-756-5250<tel:%28805%29-756-5250>
"... for whenever you learn something new, the whole world becomes that much richer." - Norton Juster, The Phantom Tollbooth
Hello SBI listserv participants and SBI blog readers,
Hope you are enjoying your Saturday morning!
First, Thank you for your discussions on/contributions to the listserv - it is great to hear about all the things that statistics teachers are doing in their classes!
Second, we have several new articles on the Simulation-based Inference blog (https://www.causeweb.org/sbi/) that have been recently posted:
1) We have two new posts on "How to use real data" by Kevin Ross and Nathan Tintle.
2) Erin Blankenship, Karen McGaughey, and Kathryn Dobeck have written about their experiences and what they thought was "The hardest thing about getting started with simulation-based curricula."
3) For readers interested in "How to implement simulation-based methods in high school classrooms/AP Statistics classes" - we have articles from Bob Peterson, Catherine Case, and Josh Tabor, all AP Statistics teachers, writing about their experiences.
On behalf of the ISI team, I'd like to thank all our blog contributors for writing these pieces for us.
I hope you enjoy reading these articles, and others posted on the blog, as much as I do!
Have a nice weekend!
- Soma
-----------------------
Soma Roy
Associate Professor
Statistics
California Polytechnic State University
San Luis Obispo CA 93407
Phone no.: (805)-756-5250
"… for whenever you learn something new, the whole world becomes that much richer." - Norton Juster, The Phantom Tollbooth
Happy summer everyone,
I thought I would send out a message I received yesterday. A local
journalist reached out to me (I taught his daughters) with the following
question, and I thought it was a nice problem to send out to everyone.
Even though our school year is over, I've sent it along to my students to
see if anyone is interested to try and answer this question for him, or
even to possibly present to the local town council.
Enjoy!
In 2007, Ross Valley residents voted on a flood tax. You had to sign the
ballot (which was unusual) and 21% of the ballots were not signed and
thrown out. The "valid" votes were split essentially 50-50 (50.1% "no" vs.
49.9% "yes") but the 1,672 tossed out votes were 56.33% "no" and 43.67%
"yes." Assuming there is no reason why the unsigned ballots would be more
likely "yes" or "no, " can you calculate the odds of this 56.33-43.67 split
for a 50-50 event? Many of us suspect foul play, and the matter is of
urgency now as the tax money is about to be used to tear up San Anselmo's
Memorial Park. Your names will NOT be used; it's purely my mathematical
curiosity. Thanks. -- Barry
Kevin
--
Kevin Rees
Math Department Chair
Marin Academy
www.ma.org
415-482-3260
I believe that Scott Rifkin has it exactly right with the bootstrap. I have used the approach that he described with my students in a first college course in Statistical Science. In the last week of the course just completed, my students worked in teams of three using resampling and bootstrapping. The focus, of course, is on understanding the variability of as estimate derived from a sample.
I believe that, at a basic level, students can understand and appreciate the bootstrap. Of course there are subtleties, but then that's true in most real statistical problems! For those who want to learn more about the subtleties and the performance of various procedures that grow out of bootstrapping, see the excellent recent paper, written for teachers, by Tim Hesterberg. I learned a lot from it, and my students verified some of the points that Tim makes for bootstrap confidence intervals.
John Emerson
Middlebury, VT
Date: Sat, 16 May 2015 09:56:55 -0700
From: Scott Rifkin <sarifkin(a)ucsd.edu>
To: sbi(a)causeweb.org
Subject: [SBI] How to estimate parameters: To bootstrap or not?
Message-ID: <555776D7.7020209(a)ucsd.edu>
Content-Type: text/plain; charset=utf-8; format=flowed
My approach to the conceptual hurdle of using a single sample to mimic the population runs something along the following lines:
First get them comfortable with the idea that a statistic measured on the sample (usually we talk about mean) is our best estimate of the corresponding population parameter. The goal is to make the link between sample properties and population properties.
Then present the problem:
Problem: This sample statistic can move around depending on the sample.
How variable is it? How big is this sample variability? Since the population parameter is fixed, we want to get an idea of how close our 'best estimate' is.
Ideal solution: keep taking samples from the population and get a collection of statistics. Then we'd know because we'd have a sampling distribution (a distribution of sample statistics).
Problem: This is impractical. Usually we can only afford to take one sample.
Question: is there a way we can mimic/simulate this ideal solution?
Problem: We don't know what the population actually looks like.
But we do know something about it. In fact, everything we know about it is encapsulated in the sample. The sample is our best estimate of the population in a particular way: we expect the population to be much bigger in size, but the frequencies of each value in the sample are representative (in a statistical way) of the frequencies of each value in the actual population. And we already have it in hand.
So instead of doing the impractical - continually using scarce resources to sample from the population and calculate our statistic on each sample
- we use what we already have and sample from our best estimate of the population - our sample itself - and calculate our statistic from each bootstrap sample.
For my students, the key is when they understand that the sample itself plays the role of an estimate of the population. And that we use bootstrapping to study the variability (not the location) of our statistic of interest.
- Scott Rifkin
------------
EBE, Division of Biological Sciences
UCSD
>
> Hello All,
>
> As person who spends most of her summer working with high school
> teachers on stats and probability content and creating lesson plans,
> which are used in the next school year, I've followed this discussion
> eagerly.
>
> High school teachers are relatively easily convinced that a large
> enough, random sample is usually representative of the population.
> Convincing teachers that one of these samples could be used to mimic
> the entire population and then be utilized to generate more random
> samples is quite a different thing. I am convinced of the
> bootstrapping process, but to leap there immediately with teachers
> versus the more cumbersome routes discussed in this chain of responses
> might cause serious distress.
>
> Are there resources to help educate high school teachers (and myself
> further) in regard to bootstrapping? Research and experience shows
> that teachers with either omit or superficially enact contact that
> they feel is beyond their current knowledge base.
>
> Simulation, in general, has been daunting for high school teachers.
> Of 23 we worked with last summer, only 25% took the plunge with
> re-randomization. However, the ones that did, thoroughly enjoyed the
> experience, as did their students.
>
> Best,
>
> Maryann
>
> /----------------------------------------/
>
> /Maryann E. Huey/
>
> /Mathematics and Computer Science/
>
> /Drake University/
>
> /515/271-2839///
>
> <46A78617-9B87-4C74-8821-5F91750143B1[6].png>
>
------------------------------
Message: 2
Date: Sat, 16 May 2015 19:27:17 -0400
From: Daren Starnes <dstarnes(a)lawrenceville.org>
To: Simulation-Based Inference <sbi(a)causeweb.org>
Subject: Re: [SBI] How to estimate parameters: To bootstrap or not?
Message-ID:
<CAMo0yhrLr_9y6JKxO6s9cv9jAM-T-pGVH0Kw3LmOT5c9bihHkg(a)mail.gmail.com>
Content-Type: text/plain; charset="utf-8"
Hi, Maryann. From my own work with high school teachers, I have found that the best entry point for simulation-based inference is to introduce them to two cases that are pretty accessible:
1. Using simulation to test a claim about a population proportion based on a random sample from that population. Just simulate many, many samples of that size under the assumption that the claim is true and record the value of the sample proportion for each one in a dotplot. Then look where the observed result falls in the simulated sampling distribution, and ask whether the sample result is sufficiently surprising (far out in the tails of the distribution) to provide convincing evidence against the claim.
Ideally, we'd have learners do this with a spinner or some other physical device first before proceeding to technology, which would necessitate using a fairly small sample size for practical reasons.
2. Using simulation to determine whether the difference between two proportions is statistically significant in a randomized experiment.
Assume that there is no difference in the effects of the two treatments on the subjects in the study (null hypothesis). Simulate re-doing the random assignment of subjects to treatments many, many times, keeping each subject's response (success or failure) the same as it was in the original experiment. Each time, record the difference in proportions of successes for the two groups on a dotplot. Then look where the observed result falls in the simulated randomization distribution, and ask whether the observed difference in proportions is sufficiently surprising (far out in the tails of the distribution) to provide convincing evidence against the null hypothesis. Ideally, we'd have learners do this with by shuffling and dealing cards or some other physical device first before proceeding to technology, which would necessitate using fairly small group sizes for practical reasons.
There are great resources available from several members of this list that could be used as the basis for these two distinct activities that would introduce teachers to the different scope of inference for random sampling and randomized experiments.
Daren Starnes
**********************************************************************************************
Hello All,
As person who spends most of her summer working with high school teachers on stats and probability content and creating lesson plans, which are used in the next school year, I've followed this discussion eagerly.
High school teachers are relatively easily convinced that a large enough,
random sample is usually representative of the population. Convincing
teachers that one of these samples could be used to mimic the entire population and then be utilized to generate more random samples is quite a different thing. I am convinced of the bootstrapping process, but to leap there immediately with teachers versus the more cumbersome routes discussed in this chain of responses might cause serious distress.
Are there resources to help educate high school teachers (and myself
further) in regard to bootstrapping? Research and experience shows that teachers with either omit or superficially enact contact that they feel is beyond their current knowledge base.
Simulation, in general, has been daunting for high school teachers. Of 23
we worked with last summer, only 25% took the plunge with re-randomization.
However, the ones that did, thoroughly enjoyed the experience, as did their students.
Best,
Maryann
*----------------------------------------*
*Maryann E. Huey*
*Mathematics and Computer Science*
*Drake University*
*515/271-2839 <515%2F271-2839>*
Hi, Maryann. From my own work with high school teachers, I have found that
the best entry point for simulation-based inference is to introduce them to
two cases that are pretty accessible:
1. Using simulation to test a claim about a population proportion based on
a random sample from that population. Just simulate many, many samples of
that size under the assumption that the claim is true and record the value
of the sample proportion for each one in a dotplot. Then look where the
observed result falls in the simulated sampling distribution, and ask
whether the sample result is sufficiently surprising (far out in the tails
of the distribution) to provide convincing evidence against the claim.
Ideally, we'd have learners do this with a spinner or some other physical
device first before proceeding to technology, which would necessitate using
a fairly small sample size for practical reasons.
2. Using simulation to determine whether the difference between two
proportions is statistically significant in a randomized experiment.
Assume that there is no difference in the effects of the two treatments on
the subjects in the study (null hypothesis). Simulate re-doing the random
assignment of subjects to treatments many, many times, keeping each
subject's response (success or failure) the same as it was in the original
experiment. Each time, record the difference in proportions of successes
for the two groups on a dotplot. Then look where the observed result falls
in the simulated randomization distribution, and ask whether the observed
difference in proportions is sufficiently surprising (far out in the tails
of the distribution) to provide convincing evidence against the null
hypothesis. Ideally, we'd have learners do this with by shuffling and
dealing cards or some other physical device first before proceeding to
technology, which would necessitate using fairly small group sizes for
practical reasons.
There are great resources available from several members of this list that
could be used as the basis for these two distinct activities that would
introduce teachers to the different scope of inference for random sampling
and randomized experiments.
Daren Starnes
**********************************************************************************************
Hello All,
As person who spends most of her summer working with high school teachers
on stats and probability content and creating lesson plans, which are used
in the next school year, I've followed this discussion eagerly.
High school teachers are relatively easily convinced that a large enough,
random sample is usually representative of the population. Convincing
teachers that one of these samples could be used to mimic the entire
population and then be utilized to generate more random samples is quite a
different thing. I am convinced of the bootstrapping process, but to leap
there immediately with teachers versus the more cumbersome routes discussed
in this chain of responses might cause serious distress.
Are there resources to help educate high school teachers (and myself
further) in regard to bootstrapping? Research and experience shows that
teachers with either omit or superficially enact contact that they feel is
beyond their current knowledge base.
Simulation, in general, has been daunting for high school teachers. Of 23
we worked with last summer, only 25% took the plunge with re-randomization.
However, the ones that did, thoroughly enjoyed the experience, as did
their students.
Best,
Maryann
*----------------------------------------*
*Maryann E. Huey*
*Mathematics and Computer Science*
*Drake University*
*515/271-2839 <515%2F271-2839>*
My approach to the conceptual hurdle of using a single sample to mimic
the population runs something along the following lines:
First get them comfortable with the idea that a statistic measured on
the sample (usually we talk about mean) is our best estimate of the
corresponding population parameter. The goal is to make the link
between sample properties and population properties.
Then present the problem:
Problem: This sample statistic can move around depending on the sample.
How variable is it? How big is this sample variability? Since the
population parameter is fixed, we want to get an idea of how close our
'best estimate' is.
Ideal solution: keep taking samples from the population and get a
collection of statistics. Then we'd know because we'd have a sampling
distribution (a distribution of sample statistics).
Problem: This is impractical. Usually we can only afford to take one
sample.
Question: is there a way we can mimic/simulate this ideal solution?
Problem: We don't know what the population actually looks like.
But we do know something about it. In fact, everything we know about it
is encapsulated in the sample. The sample is our best estimate of the
population in a particular way: we expect the population to be much
bigger in size, but the frequencies of each value in the sample are
representative (in a statistical way) of the frequencies of each value
in the actual population. And we already have it in hand.
So instead of doing the impractical - continually using scarce resources
to sample from the population and calculate our statistic on each sample
- we use what we already have and sample from our best estimate of the
population - our sample itself - and calculate our statistic from each
bootstrap sample.
For my students, the key is when they understand that the sample itself
plays the role of an estimate of the population. And that we use
bootstrapping to study the variability (not the location) of our
statistic of interest.
- Scott Rifkin
------------
EBE, Division of Biological Sciences
UCSD
>
> Hello All,
>
> As person who spends most of her summer working with high school
> teachers on stats and probability content and creating lesson plans,
> which are used in the next school year, I've followed this discussion
> eagerly.
>
> High school teachers are relatively easily convinced that a large
> enough, random sample is usually representative of the population.
> Convincing teachers that one of these samples could be used to mimic
> the entire population and then be utilized to generate more random
> samples is quite a different thing. I am convinced of the
> bootstrapping process, but to leap there immediately with teachers
> versus the more cumbersome routes discussed in this chain of responses
> might cause serious distress.
>
> Are there resources to help educate high school teachers (and myself
> further) in regard to bootstrapping? Research and experience shows
> that teachers with either omit or superficially enact contact that
> they feel is beyond their current knowledge base.
>
> Simulation, in general, has been daunting for high school teachers.
> Of 23 we worked with last summer, only 25% took the plunge with
> re-randomization. However, the ones that did, thoroughly enjoyed the
> experience, as did their students.
>
> Best,
>
> Maryann
>
> /----------------------------------------/
>
> /Maryann E. Huey/
>
> /Mathematics and Computer Science/
>
> /Drake University/
>
> /515/271-2839///
>
> <46A78617-9B87-4C74-8821-5F91750143B1[6].png>
>
Beth, Robin, et al.
While the method of finding plausible values (i.e. a confidence interval)
for a single proportion by testing many different null hypothesis values is
inefficient, I personally find it valuable, at least as a starting point,
because (a) It reinforces the idea of how to do tests of significance and
(b) It reinforces the language of 'null is plausible' vs. the common
student mistake of 'null is true' when the p-value is large. While I don't
spend a lot of time with this approach (as others have mentioned it is time
consuming from the students perspective and limited in terms of cases when
it is applicable) it seems to act as a nice 'bridge' to other techniques,
like estimating the SE from the simulated null distribution and taking 2*SE
as a rough 95% CI, and/or theory-based approaches, without needing to
introduce another type of simulation (e.g., bootstrap).
Nathan
On Tue, May 5, 2015 at 6:01 PM, Beth Chance <bchance(a)calpoly.edu> wrote:
> Hi,
>
>
>
> I of course have to argue with Robin J But not on all points
>
>
>
> In the one sample, quantitative variable case, instead of bootstrapping, I
> have students sample from a made-up population. So this is still a bit ad
> hoc, but I think helps them better see the sampling from population
> connection we are emphasizing at this point in the course.
>
>
>
> With proportions, I agree that you have to decide whether you want to use
> a hypothesized value or the sample proportion to estimate the SD. In my
> class I want students to think about both methods, partly to see it often
> doesn’t make a difference, especially with a large sample size, which I
> assume is what the CCSSM will focus on. And of course, “traditional”
> methods make the same “arbitrary” decision – use hypothesized if you have
> one, use sample if you don’t. We do have students try lots of different
> null values the first time we are creating a confidence interval of
> plausible values (and that’s when we introduce the idea of level of
> significance), but we have added a feature to the technology to make this a
> little more efficient once they get the idea (think slider).
>
>
>
> My hope, though I don’t have a lot of data and what I do have isn’t
> “great,” is that students will be better able to focus on the interval
> being for the *parameter* rather than the common misconception that it’s an
> interval for sample proportions that I worry bootstrapping might
> reinforce. Basically I want students to think about a confidence interval
> as estimate +- 2SD, which they seem to get pretty easily, and then we can
> worry about the details of how to estimate the (right) SD in different
> cases/use technology. This is what we carry over to other statistics later
> in the course. I think CCSSM is focused on having students understand
> sampling variability and the idea of margin-of-error, of the proportion
> being close to the parameter and that the “plus or minus part” depends on
> sample size. Lots of good ways to get those ideas across. Like them, I’ve
> been starting with proportion, I guess they thought mean would be too tough
> as they didn’t want to have them get into bootstrapping.
>
>
>
> There is more discussion on exactly this issue on the SBI blog:
> https://www.causeweb.org/sbi/?cat=14
>
>
>
> Beth
>
>
>
>
>
> *From:* sbi-bounces(a)causeweb.org [mailto:sbi-bounces@causeweb.org] *On
> Behalf Of *Robin Lock
> *Sent:* Tuesday, May 05, 2015 1:17 PM
> *To:* Simulation-Based Inference
> *Subject:* Re: [SBI] How to estimate parameters: To bootstrap or not?
>
>
>
> Daren -
> I'm a firm believer in using the bootstrap as the way to get a margin
> of error via simulation. The ideal way would be to form a sampling
> distribution but, in the real world, taking 1000's of new samples from the
> actual population is not a feasible way to assess the accuracy of the one
> estimate you have from an original sample!
>
> I think that the logic of hypothesis testing is already tricky for
> students to get a handle on, to have intervals depend on inverting that
> logic seems even trickier. The example you provided in the CCSS
> Progressions document is even more confusing. Here's what I gather is its
> "logic"
>
> Start with a sample of size 50 with a sample proportion of 0.40. You
> want to estimate its "margin of error".
> 1. Suppose the population proportion is really p=0.50. Simulate a
> sampling distribution using that p.
> 2. Observe that the sample phat=0.40 is not far in the tail of that
> distribution, so 0.50 is a "plausible" value for the population
> proportion. - Not bad so far.
> 3. Estimate the standard error by finding the std. dev. of the sample
> proportions in the distribution generated around p=0.50 (SE=0.07).
> 4. Use 0.4 +/- 2*(0.07) = 0.4 +/-0.14 = 0.26 to 0.54 to get a CI of
> similar "plausible values.
>
> Of course the SE for p=0.5 and p=0.4 are not a lot different, but I don't
> see the logic of picking some random "other" proportion, when you can do
> the simulation just as well aound p=0.40 (which is what the bootstrap would
> do in the first place!). I wonder what advice the document would give
> for finding a CI when phat=0.12?
>
> I think it is possible to do an interval more coherently by doing lots of
> tests for lots of null parameters and seeing which would be rejected for
> the sample data, but
> (a) That sort of guess/check process is not very efficient.
> (b) I'd like to downplay the hard 5% reject Ho decision, and would
> rather have a test p-value be interpreted as "strength of evidence"
> (c) Creating the simulations to test lots of nulls is more problematic
> (especially via simulation) for other parameter situations like a
> difference in proportions, difference in means, or correlation.
>
> The bootstrap procedure is pretty straightforward: take lots of samples
> (with replacement) from the original sample, calculate the statistics of
> interest, estimate SE as the std. dev. of all those bootstrap statistics.
> A rough margin of error as 2*SE is easy to find and the same process works
> for lots of different parameters.
>
> Robin
>
> On 5/3/2015 4:03 PM, Daren Starnes wrote:
>
> Happy May, everyone. There is an interesting thread on the AP Statistics
> Teacher Community about two distinct views on estimating parameters via
> simulation. This came up because the Common Core State Standards includes
> this Statistics and Probability standard
>
>
>
> S-IC.4 Use data from a sample survey to estimate a population mean or
> proportion; develop a margin of error through the use of simulation models
> for random sampling.
>
>
>
> View #1: Use bootstrapping.
>
>
>
> View #2: Determine whether "nearby" values of the parameter are plausible
> by simulating a "null distribution" with that parameter value and seeing if
> the observed statistic is a believable outcome from such a null
> distribution. Keep doing this for other nearby values until you have an
> interval of plausible values for the parameter.
>
>
>
> The attached CCSS Progression document seems to suggest View #2, at least
> as far as estimating a proportion is concerned. There is no discussion of
> how to estimate the margin of error for a mean in this way (I wonder why!).
>
>
>
> This seems like an issue that this experienced group of SBI folks would
> already have grappled with--both philosophically and pedagogically. So I
> thought I would ask what the prevailing wisdom is.
>
>
>
> Daren Starnes
>
>
>
>
>
> _______________________________________________
>
> SBI mailing list
>
> SBI(a)causeweb.org
>
> https://www.causeweb.org/mailman/listinfo/sbi
>
>
>
> --
>
> Robin Lock
>
> Burry Professor of Statistics
>
> St. Lawrence University
>
>
--
Nathan Tintle, Ph.D.
Associate Professor of Statistics and Dept. Chair
Director for Research and Scholarship
Dordt College
Sioux Center, IA 51250
nathan.tintle(a)dordt.edu
Phone: (712) 722-6264
Office: SB1612
Happy May, everyone. There is an interesting thread on the AP Statistics
Teacher Community about two distinct views on estimating parameters via
simulation. This came up because the Common Core State Standards includes
this Statistics and Probability standard
S-IC.4 Use data from a sample survey to estimate a population mean or
proportion; develop a margin of error through the use of simulation models
for random sampling.
View #1: Use bootstrapping.
View #2: Determine whether "nearby" values of the parameter are plausible
by simulating a "null distribution" with that parameter value and seeing if
the observed statistic is a believable outcome from such a null
distribution. Keep doing this for other nearby values until you have an
interval of plausible values for the parameter.
The attached CCSS Progression document seems to suggest View #2, at least
as far as estimating a proportion is concerned. There is no discussion of
how to estimate the margin of error for a mean in this way (I wonder why!).
This seems like an issue that this experienced group of SBI folks would
already have grappled with--both philosophically and pedagogically. So I
thought I would ask what the prevailing wisdom is.
Daren Starnes