Skip to content ↓

Using technology to combat bias in hiring

Stephanie Lampkin MBA ’13 says unconscious bias creates unfair hiring practices — and she developed an app for that.
Press Inquiries

Press Contact:

Nancy DuVergne Smith
Phone: 617-253-8217
MIT Alumni Association
Close
Stephanie Lampkin MBA ’13 created an app that takes unconscious bias out of the online hiring process.
Caption:
Stephanie Lampkin MBA ’13 created an app that takes unconscious bias out of the online hiring process.
Credits:
Photo courtesy of Blendoor.

Top tech companies have a diversity problem. Reports show that Facebook and YouTube — and many companies like them — struggle to build a workforce that is reflective of the U.S. population. Why? Some say that the lack of diversity comes from a lack of qualified candidates. Stephanie Lampkin MBA ’13 argues that’s just not true. That’s why she launched Blendoor, a blind job app combatting what she says is one root of the diversity problem — unconscious bias.

Unconscious bias refers to the stereotypes, personal experiences, and cultural exposure that people unknowingly rely on when making a decision. Lampkin says that this is why hiring managers may be more impressed with candidates that resemble themselves, without even knowing it.

“The bias is something innate, but there are ways we can use tech to eliminate it,” says Lampkin. “We’re not relying on the same traditional ways that tend to bring in homogenous teams.”

With Blendoor, hiring managers use the app to sort through a diverse candidate pool without identifiers that can engage unconscious bias — like names, photos, and job dates. Hiring managers see candidate profiles based on how well they match their needs and nothing more.

Lampkin says the app matches companies with the most skilled candidates regardless of gender, age, or ethnicity. “It’s a much bigger value proposition when you say diversity is a great byproduct of the app,” says Lampkin.

Because identities of candidates can only be hidden for so long, Blendoor also tracks how candidates move through the interview process — noting when a candidate is eliminated or gets hired. The app then uses this information to better match candidates in the future and identify at what stage bias may have come into play. 

In addition to the app, Blendoor also offers BlendScore, a metric that ranks top companies based on diversity data, pay equity, and benefits like maternal and paternal leave. The metric serves as a tool for job seekers looking for diverse companies, but also informs companies when they need to make changes. “Shortly after we released a BlendScore for Facebook, they reached out looking to improve it,” says Lampkin. “They’re a customer now.” The BlendScore relies on data shared by companies, packaging it in a way that is accessible. “We’re hoping to be the like U.S. News and World Report pertaining to ethnicity, equality,” she says. “The BlendScore shines the mirror back on the companies — it’s all about transparency.”

Lampkin has personal experience with bias in her own startup journey. She notes that only a handful of black women have raised $1 million from investors — something she wants to change. “I want to pave the way so it’s not so rare for a venture capitalist to be pitched by a black woman as it is now,” says Lampkin. “That legacy is really important to me.”

This article originally appeared on the Slice of MIT blog.

Related Links

Related Topics

Related Articles

More MIT News