36-402, Spring 2012: Self-Assessment and Lessons Learned (Advanced Data Analysis from an Elementary Point of View)

November 14, 2012
By

(This article was originally published at Three-Toed Sloth , and syndicated at StatsBlogs.)

Attention conservation notice: Complacent navel-gazing about teaching statistics to undergraduates at a very un-representative university. Written back in May, not posted due to laziness.

Overall, I think this went tolerably but not perfectly. The content was improved and is near (but not at) a local optimum. I've also got a better sense than last year of how much they're learning, and I think the homework is helping them learn. But the course could definitely do more to move them towards doing data analysis, as opposed to answering homework questions.

It was a substantially bigger class than last time (88 students vs. 63), and this led to some real issues. The lecture room was fine, but office hours were over-crowded, the stream of e-mail from student seemed unending, and, worst of all, there simply wasn't enough TA time available for grading. (88 weekly assignments of serious data analysis is a lot to grade.) Everything did get graded eventually, but not as fast as it should have been. From the number of students registered for the fall course in regression which is ADA's pre-requisite, I can expect about as many again in 2013, or a few less. This is about twice as many students as I'd like to see in it.

Speaking of grading, I had to completely re-do the assignments, since solutions from last year were circulating1 --- and I did not put the new solutions on-line. This was a big chunk of work I was not anticipating, but at least now I have two sets of problems for the future.

In terms of actual content, I cut nothing major, but, by eliminating exam review sessions, etc., squeezed in some extra stuff: a reminder lecture on multivariate distributions, smooth tests and relative densities, time series, and more on estimating causal effects. I am relatively happy with the over-all content, though less so with the organization. In particular I wonder about devoting the very first part of the course to regression with one input variable, as a really elementary way to introduce fundamental concepts (smoothing, kernels, splines, generalization, cross-validation, bootstrap, specification checking), and only then go to regression with multiple inputs. I worry, however, that this would be coddling them.

The "quality control samples" were eye-opening. Some students found them nerve-wracking, which was counterproductive; but lots of them also gave every sign of being candid. What I found most surprising was how many of them (a third? more?) were taking ADA because they had been funneled into it by major other than statistics, often one of CMU's several larval-quant programs. (I'm not sure how those took my plugging Red Plenty in class while explaining Lagrange multipliers.) Their self-de-motivating question was not "Is this going to be on the exam?", but rather, "Is this going to come up in interviews?". If I were a better teacher I'd work more at reaching these students.

Week-to-week, the quality control samples were very helpful for figuring out what students actually grasped, and, even more, how their thinking went astray (when it did). In particular, there turned out to be much more variation than I'd anticipated in readiness to apply material they had already learned in a new context — the "but that's from chapter 7 and this is chapter 12" problem, or even the "but that's from introduction to statistical inference, or matrices and linear transformations, which I took a year ago" problem. I am not sure if I should provide more cues, or if, again, that would be coddling.

These facts about motivation and preparation don't please me, exactly, but at least for right now these seem like constraints I have to (and can) work within.

I will certainly use the quality-control samples again in 2013, but I may take fewer each week, and make it clearer to the students that they're not being graded.

Some specifics:

  • I need to do more to set expectations at the beginning, especially in light of why the students are taking the class. Namely: ADA is a course in statistical methodology; it will cover models (e.g., kernel regression), methods (e.g., bootstrap), and knowing when to use which tools. It is not a research-projects course. (That's 36-490.) It is not, as one (good) student said he'd expected, "linear models plus programming". It is not a "get a job in a data startup" or "get a job as a quant" course. (It may be a data science course, since I continue to think that "data science" means "computational applied statistics", pace Cathy.)
  • There's a definite trade-off between having the students gain understanding by building their own tools, and letting them do more data analysis by using pre-built tools. I think most of them now get bootstrapping intuitvely, because I didn't let them use boot, and kept drilling them on it. I am less confident as to whether they grasp, say, factor analysis, or even additive models, at the same level.
  • Relatedly, and very importantly, all the assignments probably give the students too much "scaffolding", that is, guide them far too much as to exactly what analysis to do, with what steps in what order. But deciding on these things is a key part of doing good data analysis. As time allows, re-do the assignments so that they leave more and more to student initiative as the class goes on.
  • Making The R Cookbook required was a good idea. Things would have been better still if everyone had taken some programming, if not necessarily statistical computing. (I suppress here my usual rant about how, if you are taking a class called "Advanced Data Analysis" in 2012, it is really not unreasonable to expect you to write some code.) Look, down the road, into making this a pre-requisite.
  • pcalg proved too hard for many students to install, or rather Rgraphviz did. If I had unlimited time I'd write my own R interface to graphviz, which wouldn't choke if you installed the latter with a different version number.
    (If I had really unlimited time, I'd re-write pcalg so that, as Peter and Clark suggested originally, you could give it a partial ordering of the variables, and force the graph to be compatible with that. [Currently you can say that some undirected edges must be present, and others must be absent, which is not at all the same.] This might make a suitable project for one group of The Kids in statistical computing in the fall, whereas hacking at Rgraphviz is probably too much. [In the event, I did not give them that project.])
  • Having a required textbook is no longer useful. Make it clear that the notes take the place of a textbook, and that they're expected to study them — and not just the figure captions. (But "police the reading" homework wastes time they could spend doing more important stuff.) Consider keeping Faraway as a recommended but optional text, or perhaps replacing it with Wood's Generalized Additive Models.
  • Introduce some simple simulations very early on, before the main chapters on simulation and, especially, before the chapter on the bootstrap. Move the main simulation chapter back before bootstrapping?
  • Do not call a test statistic t unless it is really a t-test. (This led to a lot of confusion when doing specification testing.)
  • More practice in finding confidence intervals for differences directly. ("The difference between 'significant' and 'not significant' is not statistically significant.")
  • More practice in answering "what if?" questions from non-linear regression models, including error bars on the what-ifs, before getting to causal inference.
  • Say more about measurement models in the context of graphical models and especially causality.
  • Enough students have taken econometrics that more explicit de-programming about instrumental variables is called for.
  • I gave a smaller proportion of A's than last year, but also failed absolutely fewer students. Some who had struggled during the semester managed to pull it together for the final, which was very gratifying.
  • Shorten the final and make it due earlier; the rush to get in grades in time for graduating seniors was Not Good.
  • Coursekit proved to be vastly less annoying to use than Blackboard. That is, it was not actively painful to perform such obscure operations as "enter grades" or "send e-mail to all students" or "link to lecture notes". It's a free service, so I am sure that it's all part of some nefarious plan ("if you're not paying for something, you're not the customer; you're the product being sold"), but at this point I really don't care, and intend to keep using it. [Update, November 2012: Coursekit, now Lore, got "upgraded" over the summer to a level of dysfunctionality that equals or even exceeds Blackboard's. I will thus switch back to the user-hostile software which doesn't obviously have exploitative intellectual-property or social-marketing designs on my students.]

Finally, if students again complain that the class makes it harder for them to take seriously what they hear in economics or psychology or the like, remember not to laugh maniacally, or use expressions like "Victory is mine!" Rather, maintain a sober, thoughtful, benevolent, and above all sane expression, and talk solemnly about the great prospects for using good statistical methods to solve real-world problems.

[1]: One advantage to putting bad jokes in your solutions is that it becomes pretty obvious when someone is parroting them back to you.

Advanced Data Analysis from an Elementary Point of View



Please comment on the article here: Three-Toed Sloth


Subscribe

Email:

  Subscribe