| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354 | 
							- October 2015This will come as a surprise to a lot of people, but in some cases
 
- it's possible to detect bias in a selection process without knowing
 
- anything about the applicant pool.  Which is exciting because among
 
- other things it means third parties can use this technique to detect
 
- bias whether those doing the selecting want them to or not.You can use this technique whenever (a) you have at least
 
- a random sample of the applicants that were selected, (b) their
 
- subsequent performance is measured, and (c) the groups of
 
- applicants you're comparing have roughly equal distribution of ability.How does it work?  Think about what it means to be biased.  What
 
- it means for a selection process to be biased against applicants
 
- of type x is that it's harder for them to make it through.  Which
 
- means applicants of type x have to be better to get selected than
 
- applicants not of type x.
 
- [1]
 
- Which means applicants of type x
 
- who do make it through the selection process will outperform other
 
- successful applicants.  And if the performance of all the successful
 
- applicants is measured, you'll know if they do.Of course, the test you use to measure performance must be a valid
 
- one.  And in particular it must not be invalidated by the bias you're
 
- trying to measure.
 
- But there are some domains where performance can be measured, and
 
- in those detecting bias is straightforward. Want to know if the
 
- selection process was biased against some type of applicant?  Check
 
- whether they outperform the others.  This is not just a heuristic
 
- for detecting bias.  It's what bias means.For example, many suspect that venture capital firms are biased
 
- against female founders. This would be easy to detect: among their
 
- portfolio companies, do startups with female founders outperform
 
- those without?  A couple months ago, one VC firm (almost certainly
 
- unintentionally) published a study showing bias of this type. First
 
- Round Capital found that among its portfolio companies, startups
 
- with female founders outperformed
 
- those without by 63%. 
 
- [2]The reason I began by saying that this technique would come as a
 
- surprise to many people is that we so rarely see analyses of this
 
- type.  I'm sure it will come as a surprise to First Round that they
 
- performed one. I doubt anyone there realized that by limiting their
 
- sample to their own portfolio, they were producing a study not of
 
- startup trends but of their own biases when selecting companies.I predict we'll see this technique used more in the future.  The
 
- information needed to conduct such studies is increasingly available.
 
- Data about who applies for things is usually closely guarded by the
 
- organizations selecting them, but nowadays data about who gets
 
- selected is often publicly available to anyone who takes the trouble
 
- to aggregate it.
 
- Notes[1]
 
- This technique wouldn't work if the selection process looked
 
- for different things from different types of applicants—for
 
- example, if an employer hired men based on their ability but women
 
- based on their appearance.[2]
 
- As Paul Buchheit points out, First Round excluded their most 
 
- successful investment, Uber, from the study.  And while it 
 
- makes sense to exclude outliers from some types of studies, 
 
- studies of returns from startup investing, which is all about 
 
- hitting outliers, are not one of them.
 
- Thanks to Sam Altman, Jessica Livingston, and Geoff Ralston for reading
 
- drafts of this.
 
 
  |