Tag Archives: cathy o’neil

Should we fire people who pick the wrong Final Four?

A thought experiment touched off by Cathy’s latest post on value-added modeling.

Suppose I’m in charge of a big financial firm and I made every trader who worked for me fill out an NCAA tournament bracket.  Then, every year, I fired the people whose brackets ended up in the lowest quintile.

This makes sense, right?  Successful prediction of college basketball games involves a lot of the same skills you want traders to have:  an ability to aggregate information about uncertain outcomes, fluency in quantitative reasoning, a certain degree of strategic thinking (what choices do you make if your objective is to minimize the probability that your bracket is in the bottom 20%?  What if your fellow traders are also following the same strategy…?)  You might even do a study that finds that firms whose traders did better at bracket prediction actually ended up with better returns 5 years later.  Even if the effect is small, that might add up to a lot of money.  Yes, the measure isn’t perfect, but why wouldn’t I want to fire the people who, on average, are likely to make less money for my firm?

And yet we wouldn’t do this, right?  Just because we think it would be obnoxious to fire people based on a measure predominantly not under their control.  At least we think this when it comes to high-paid financial professionals.  Somehow, when it comes to schoolteachers, we think about it differently.

Tagged , , ,

OKCathy

This week’s Aunt Pythia column features Cathy O’Neil’s take on what questions online daters ought to have to answer in their profiles:

How sexual are you? (super important question)
How much fun are you? (people are surprisingly honest when asked this)
How awesome do you smell? (might need to invent technology for this one)
What bothers you more: the big bank bailout or the idea of increasing the minimum wage?
Do you like strong personalities or would you rather things stay polite?
What do you love arguing about more: politics or aesthetics?
Where would you love to visit if you could go anywhere?
Do you want kids?
Dog person or cat person?
Do you sometimes wish the girl could be the hero, and not always fall for the hapless dude at the end?

I gotta say, thinking back to when I was single, during the second Clinton administration, I don’t think these are the questions I personally would most want to ask of my prospective dates.

On the other hand, I think the questions provide a near-perfect portrait of Cathy!  So let me offer my own suggestion:  maybe profiles shouldn’t have any answers.  Maybe they should just have questions.  And you contact the person whose questions you’d like to answer.

What would your questions be?

 

 

Tagged , , ,

My other daughter is a girl

I like Cathy’s take on this famous probability puzzle.  Why does this problem give one’s intuition such a vicious noogie?

It is relevant that the two questions below have two different answers.

  • I have two children.  One of my children is a girl who was born on Friday.  What’s the probability I have two girls?
  • I have two children.  One of my children is a girl.  Before you came in, I selected a daughter at random from the set of all my daughters, and this daughter was born on Friday.  What’s the probability I have two girls?

 

Tagged , , ,

Learn to be a crappy programmer

“If a thing’s worth doing, it’s worth doing well” is a nice old saying, but is it true?  Cathy’s advice column today reminded me of this question, as regards coding.  I think learning to write good code is quite hard.  On the other hand, learning to write fairly crappy yet functional code is drastically less hard.  Drastically less hard and incredibly useful!  For many people, it’s probably the optimal point on the reward/expenditure curve.

It feels somehow wrong to give advice like “Learn to be a crappy programmer” but I think it might actually be good advice.

 

Tagged , , ,

How much is the stacks project graph like a random graph?

Cathy posted some cool data yesterday coming from the new visualization features of the magnificent Stacks Project.  Summary:  you can make a directed graph whose vertices are the 10,445 tagged assertions in the Stacks Project, and whose edges are logical dependency.  So this graph (hopefully!) doesn’t have any directed cycles.  (Actually, Cathy tells me that the Stacks Project autovomits out any contribution that would create a logical cycle!  I wish LaTeX could do that.)

Given any assertion v, you can construct the subgraph G_v of vertices which are the terminus of a directed path starting at v.  And Cathy finds that if you plot the number of vertices and number of edges of each of these graphs, you get something that looks really, really close to a line.

Why is this so?  Does it suggest some underlying structure?  I tend to say no, or at least not much — my guess is that in some sense it is “expected” for graphs like this to have this sort of property.

Because I am trying to get strong at sage I coded some of this up this morning. One way to make a random directed graph with no cycles is as follows:  start with N edges, and a function f on natural numbers k that decays with k, and then connect vertex N to vertex N-k (if there is such a vertex) with probability f(k).  The decaying function f is supposed to mimic the fact that an assertion is presumably more likely to refer to something just before it than something “far away” (though of course the stack project is not a strictly linear thing like a book.)

Here’s how Cathy’s plot looks for a graph generated by N= 1000 and f(k) = (2/3)^k, which makes the mean out-degree 2 as suggested in Cathy’s post.

stacksgraph_expmean2

Pretty linear — though if you look closely you can see that there are really (at least) a couple of close-to-linear “strands” superimposed! At first I thought this was because I forgot to clear the plot before running the program, but no, this is the kind of thing that happens.

Is this because the distribution decays so fast, so that there are very few long-range edges? Here’s how the plot looks with f(k) = 1/k^2, a nice fat tail yielding many more long edges:

stacksgraph_inversesquare

My guess: a random graph aficionado could prove that the plot stays very close to a line with high probability under a broad range of random graph models. But I don’t really know!

Update:  Although you know what must be happening here?  It’s not hard to check that in the models I’ve presented here, there’s a huge amount of overlap between the descendant graphs; in fact, a vertex is very likely to be connected all but c of the vertices below it for a suitable constant c.

I would guess the Stacks Project graph doesn’t have this property (though it would be interesting to hear from Cathy to what extent this is the case) and that in her scatterplot we are not measuring the same graph again and again.

It might be fun to consider a model where vertices are pairs of natural numbers and (m,n) is connected to (m-k,n-l) with probability f(k,l) for some suitable decay.  Under those circumstances, you’d have substantially less overlap between the descendant trees; do you still get the approximately linear relationship between edges and nodes?

Tagged , , , ,

Why “I don’t know” doesn’t read as macho

Great post from Cathy about the need to be able to assert uncertainty in a, well, assertive way.  Why is this so hard?  Why do we find people trustworthy when they say, with complete confidence, “Here’s your answer?”

One reason must be that in some contexts, confidence really does correlate with knowledge; people who truly know nothing about a subject are (hopefully) more willing to express uncertainty about it.  So when we hear someone answer a question with “I’m not sure,” we have to carry out some inferential computation:  do we think they’re saying that because they’ve never thought about the question, or because they’ve thought enough about the question to understand that it’s actually difficult?

I don’t have an answer to this conundrum, but I do have an extremely scientific infographic that I hope will illustrate the issue.

Image

 

Tagged , , ,

Guest post: Stephanie Tai on deference to experts

My colleague Steph Tai at the law school wrote a long, amazing Facebook message to me about the question Cathy and I have been pawing at:  when and in what spirit should we be listening to experts?  It was too good to be limited to Facebook, so, with her permission, I’m reprinting it below.

Steph deals with these issues because her academic specialty is the legal status of scientific knowledge and scientific evidence.  So yes:  in a discussion on whether we should listen to experts I am asking you to listen to the opinions of an expert on expertise.

Also, Steph very modestly doesn’t link to her own paper on this stuff until the very bottom of this post.  I know you guys don’t always read to the bottom, so I’ve got your link to “Comparing Approaches Toward Governing Scientific Advisory Bodies on Food Safety in the United States and the European Union” right here!

And now, Steph:

*****

Some quick thoughts on this very interesting exchange. What might be helpful, to take everyone out of our own political contexts, perhaps, is to contrast this discussion you’re both having regarding experts and financial models with discussions about experts and climate models, where, it seems, the political dynamics are fairly opposite. There, you have people on the far right making similar claims to Cathy: that climate scientists are to be distrusted because they’re just coming up with scare models because these allegedly biased models are useful to those climate scientists–i.e., to bring money to left-wing causes, to generate grants for more research, etc.

 

So when you apply the claim that Cathy makes at the end of her post–“If you see someone using a model to make predictions that directly benefit them or lose them money – like a day trader, or a chess player, or someone who literally places a bet on an outcome (unless they place another hidden bet on the opposite outcome) – then you can be sure they are optimizing their model for accuracy as best they can. . . . But if you are witnessing someone creating a model which predicts outcomes that are irrelevant to their immediate bottom-line, then you might want to look into the model yourself.”–I’m not sure you can totally put climate scientists in that former category (of those that directly benefit from the accuracy of their predictions). This is due to the nature of most climate work: most researchers in the area only contribute to one tiny part of the models, rather than produce the entire model themselves (thus, the incentives to avoid inaccuracies are diffuse rather than direct); the “test time” for the models are often relatively far into the future (again, making the incentives more indirect); and the sorts of diffuse reputational gains that an individual climate scientist gets from being part of a team that might partly contribute to an accurate climate model is far less direct than the examples given of day traders and chess players or “someone who literally places a bet on an outcome.”

 

What that in turn seems to mean is that under Cathy’s approach, climate scientists would be viewed as in the latter category—those creating models that “predict outcomes that are irrelevant to their immediate bottom-line,” and thus deserve people looking “into the model [themselves].” But at least from what I’ve seen, there is *so* much out there in terms of inaccurate and misleading information about climate models (by folks with stakes in the *perception* of those models) that chances are, a lay person’s inquiry into climate models has high chance to being shaped by similar forces with which Cathy is (in my view appropriately) concerned. Which in turn makes me concerned about applying this approach.
Disclaimer: I used to fall under this larger umbrella of climate scientists, though I didn’t work on the climate models themselves, just one small input to them—the global warming potentials of chlorofluorocarbon substitutes. So this contrast is not entirely unemotional for me. That said, now that I’m an academic who studies the *use* of science in legal decisionmaking (and no longer really an academic who studies the impact of chlorofluorocarbon substitutes on climate), I don’t want to be driven by these past personal ties, but they’re still there, so I feel like I should lay them out.

 

So what’s to be done? I absolutely agree with Cathy’s statement that “when independent people like myself step up to denounce a given statement or theory, it’s not clear to the public who is the expert and who isn’t.” It would seem, from what she says at the end of her essay, that her answer to this “expertise ambiguity” is to get people to look into the model when expertise is unclear.[*] But that in turn raises a whole bunch of questions:

 

(1) What does it take to “look into the model yourself”? That is, how much understanding does it take? Some sociologists of science suggest that translational “experts”–that is, “experts” who aren’t necessarily producing new information and research, but instead are “expert” enough to communicate stuff to those not trained in the area–can help bridge this divide without requiring everyone to become “experts” themselves. But that can also raise the question of whether these translational experts have hidden agendas in some way. Moreover, one can also raise questions of whether a partial understanding of the model might in some instances be more misleading than not looking into the model at all–examples of that could be the various challenges to evolution based on fairly minor examples that when fully contextualized seem minor but may pop out to someone who is doing a less systematic inquiry.

 

(2) How does a layperson avoid, in attempting to understand the underlying model, the same manipulations by those with financial stakes in the matter–the same stakes that Cathy recognizes might shape the model itself? Because that happens as well, so that even if one were to try to look into a model themselves, the educational materials it would take to look into that model can be also argued to be developed by those with stakes in the matter. (I think Cathy sort of raises this in a subsequent post about how entire subfields can be regarded as “captured” by particular interests.)

 

(3) (and to me this is one of the most important questions) Given the high degree of training it takes to understand any of these individual areas of expertise, and given that we encounter so many areas in which this sort of deeper understanding is needed to resolve policy questions, how can any individual actually apply that initial exhortation–to look into the model yourself–in every instance where expertise ambiguity is raised? To me that’s one of the most compelling arguments in favor of deferring to experts to some extent–that lay people (as citizens, as judges, as whatever) simply don’t have time to do the kind of thing that Cathy suggests in every situation where she argues it’s called for. Expert reliance isn’t perfect, sure–but it’s a potentially pragmatic response to an imperfect world with limited time and resources.

 

Do my thoughts on (3) mean that I think we should blindly defer to experts? Absolutely not. I’m just pointing it out as something that weighs in favor of listening to experts a little more. But that also doesn’t mean that the concerns Cathy raises are unwarranted. My friend Wendy Wagner writes about this in her papers on the production of FDA reports and toxic materials testing, and I find her inquiries quite compelling. P.s. I should also plug a work of hers that seems especially relevant to this conversation. It suggests that the part of Nate Silver’s book that might raise the most concerns (I dunno, because I haven’t read it) is its potential contribution to the vision of models as “truth machines,” rather than understanding that models are just one tool to aid in making decisions, and a tool which must be contextualized (for bias, for meaningfulness, for uncertainty) at that.

 

So how to address this balance between skepticism and lack of time to do full inquiries into everything? I totally don’t have the answers, though the kind of stuff I explore are procedural ways to address these issues, at least when legal decisions are raised–for example,
* public participation processes (with questions as to both the timing and scope of those processes, the ability and likelihood that these processes are even used, the accessibility of these processes, the susceptibility of “abuse,” the weight of those processes in ultimate decisionmaking)
* scientific ombudsman mechanisms (which questions of how ombudsman are to be selected, the resources they can use to work with citizen groups, the training of such ombudsmen)
* the formation of independent advisory committees (with questions of the selection of committee members, conflict of interest provisions, the authority accorded to such committees)
* legal case law requiring certain decisionmaking heuristics in the face of scientific uncertainty to avoid too much susceptibility to data manipulation (with questions of the incentives those heuristics create for actual potential funders of scientific research, the ability of judges to apply such heuristics in a consistent manner)
–as well as legal requirements that exacerbate these problems. Anyway, thanks for an interesting back and forth!

 

[*] I’m not getting into the question of “what makes someone an expert?” here, and instead focus on “how do we make decisions given the ambiguousness of who should be considered experts?” because that’s more relevant to what I study, although I should also point out that philosophers and sociologists of science have been studying this in what’s starting to be called the “third wave” of science, technology, and society studies. There’s a lot of debate about this, and I have a teensy summary of it here (since Jordan says it’s okay for me to plug myself :) (Note: the EFSA advisory committee structure, if anyone cares, has changed since I published this article so that the article characterizations are no longer accurate.)

 

 

 

Tagged , , , , , ,

Distrusters of experts all around

The Wall Street Journal op-ed page is always good for a full-throated demand that we distrust the experts:

The general public is not privy to the IPCC debate. But I have been speaking to somebody who understands the issues: Nic Lewis. A semiretired successful financier from Bath, England, with a strong mathematics and physics background, Mr. Lewis has made significant contributions to the subject of climate change.

…Will the lead authors of the relevant chapter of the forthcoming IPCC scientific report acknowledge that the best observational evidence no longer supports the IPCC’s existing 2°-4.5°C “likely” range for climate sensitivity? Unfortunately, this seems unlikely—given the organization’s record of replacing evidence-based policy-making with policy-based evidence-making, as well as the reluctance of academic scientists to accept that what they have been maintaining for many years is wrong.

Domain knowledge, phooey — this dude is successful!

“Distrust the experts,” as a principle, does as much harm as good.  A better principle would be “Distrust people who are bad and trust people who are not bad.”  Of course, it can be hard to tell the difference — but that distinction is one we have to make anyway, in all kinds of contexts, so why not this one?

 

Tagged , ,

In defense of Nate Silver and experts

Cathy goes off on Nate Silver today, calling naive his account of well-meaning people saying false things because they’ve made math mistakes.  In Cathy’s view, people say false things because they’re not well-meaning and are trying to screw you — or, sometimes, because they’re well-meaning but their incentives are pointed at something other than accuracy.  Read the whole thing, it’s more complicated than this paraphrase suggests.

Cathy, a fan of and participant in mass movements, takes special exception to Silver saying:

This is neither the time nor the place for mass movements — this is the time for expert opinion. Once the experts (and I’m not one of them) have reached some kind of a consensus about what the best course of action is (and they haven’t yet), then figure out who is impeding that action for political or other disingenuous reasons and tackle them — do whatever you can to remove them from the playing field. But we’re not at that stage yet.

Cathy’s take:

…I have less faith in the experts than Nate Silver: I don’t want to trust the very people who got us into this mess, while benefitting from it, to also be in charge of cleaning it up. And, being part of the Occupy movement, I obviously think that this is the time for mass movements.

From my experience working first in finance at the hedge fund D.E. Shaw during the credit crisis and afterwards at the risk firm Riskmetrics, and my subsequent experience working in the internet advertising space (a wild west of unregulated personal information warehousing and sales) my conclusion is simple: Distrust the experts.

I think Cathy’s distrust is warranted, but I think Silver shares it.  The central concern of his chapter on weather prediction is the vast difference in accuracy between federal hurricane forecasters, whose only job is to get the hurricane track right, and TV meteorologists, whose very different incentive structure leads them to get the weather wrong on purpose.  He’s just as hard on political pundits and their terrible, terrible predictions, which are designed to be interesting, not correct.

Cathy wishes Silver would put more weight on this stuff, and she may be right, but it’s not fair to paint him as a naif who doesn’t know there’s more to life than math.  (For my full take on Silver’s book, see my review in the Globe.)

As for experts:  I think in many or even most cases deferring to people with extensive domain knowledge is a pretty good default.  Maybe this comes from seeing so many preprints by mathematicians, physicists, and economists flushed with confidence that they can do biology, sociology, and literary study (!) better than the biologists, sociologists, or scholars of literature.  Domain knowledge matters.  Marilyn vos Savant’s opinion about Wiles’s proof of Fermat doesn’t matter.

But what do you do with cases like finance, where the only people with deep domain knowledge are the ones whose incentive structure is socially suboptimal?  (Cathy would use saltier language here.)  I guess you have to count on mavericks like Cathy, who’ve developed the domain knowledge by working in the financial industry, but who are now separated from the incentives that bind the insiders.

But why do I trust what Cathy says about finance?

Because she’s an expert.

Is Cathy OK with this?

Tagged , , , , ,

Startup culture, VC culture, and Mazurblogging

Those of us outside Silicon Valley tend to think of it as a single entity — but venture capitalists and developers are not the same people and don’t have the same goals.  I learned about this from David Carlton’s blog post.  Cathy O’Neil reposted it this morning.  It’s kind of cool that the three of us, who started grad school together and worked with Barry Mazur, are all actively blogging!  We just need to get Matt Emerton in on it and then we’ll have the complete set.  Maybe we could even launch a new blogging platform and call it mazr.  You want startup culture, I’ll give you startup culture!

 

Tagged , , ,
<span>%d</span> bloggers like this: