AI with Michal

AI bias audit

A structured review of an AI tool's inputs, logic, and outputs to detect patterns that disadvantage protected groups in hiring decisions before those patterns cause legal or reputational harm.

Michal Juhas · Last reviewed May 3, 2026

What is an AI bias audit?

An AI bias audit is a structured review of a hiring tool's inputs, logic, and outputs to detect patterns that produce worse outcomes for protected groups: women, racial minorities, older candidates, people with disabilities, or others covered by employment law. Unlike a one-time sign-off before launch, a useful bias audit is a repeating process tied to the hiring funnel the tool touches.

The audit does not require a data scientist to start. A spreadsheet showing pass rates by protected group at one funnel stage is already an audit. What it requires is a clear owner, a regular cadence, and vendor cooperation when the problem lives in training data rather than the outputs your team controls.

Illustration: AI bias audit showing candidate funnel scores compared by protected group, with a gap flag and a findings report card

In practice

  • A TA ops lead exports 90 days of AI-scored resume decisions, splits them by gender using EEO self-report fields, and finds that profiles from women pass at 71% versus 89% for men. That table is the beginning of a bias audit and a legal conversation.
  • A vendor sends over a model card when asked and it shows training data from 2018 to 2022, a period when fewer women were promoted in the industry the model was calibrated on. The training gap is the audit finding, not just the output gap.
  • An HR partner at a Tuesday debrief asks "when did we last run a bias check on the screening tool?" and nobody knows. That is the most common audit failure in practice.

Quick read, then how hiring teams use it

This is for recruiters, sourcers, TA, and HR partners who need the same vocabulary in debriefs, vendor calls, and policy reviews. Skim the first section when you need a fast shared picture. Use the second when you are deciding how it shows up in your stack, contracts, or quarterly compliance calendar.

Plain-language summary

  • What it means for you: Someone needs to look at whether the AI your team uses treats applicant groups differently. If no one is doing that, you are the first person to notice when legal or press asks.
  • How you would use it: Pull your last quarter of AI-scored or AI-filtered candidates, split by protected group using EEO fields, and check whether any group consistently scores lower or gets rejected at a higher rate.
  • How to get started: Ask your ATS vendor for a decision log export. Ask the AI vendor for their most recent bias audit summary or model card. Name who will review both documents before you go live with a new tool.
  • When it is a good time: Before signing a new vendor contract, before each model update reaches production, and at least once a quarter on any tool that touches the shortlist or offer stage.

When you are running live reqs and tools

  • What it means for you: Every AI ranker, filter, or score that moves candidates in or out of the funnel is a selection procedure under EEOC guidance. Group-rate monitoring is the minimum; tracing bias back to training data or proxy features is the full audit.
  • When it is a good time: Before vendor contract signature, after any model update, and quarterly on production tools. A spike in rejection rates for one req type or one demographic segment is a signal to run the check immediately.
  • How to use it: Log candidate IDs, stage decisions, model version, and timestamp together. Run pass-rate ratios by protected group quarterly. Keep findings with owner names and review dates. Cross-link to adverse impact thresholds so your reporting connects to the legal standard.
  • How to get started: Ask current vendors for their bias audit results and which demographic groups they tested. If no audit exists, that is your first compliance conversation. Add an audit clause to new vendor contracts before you are legally required to.
  • What to watch for: Proxy features (zip code, school prestige, employment gaps) that correlate with protected class. Small-sample results that look clean but are statistically meaningless. Model updates that change scoring without notifying your compliance team. Vendors who conflate "we tested for gender" with "we tested all protected classes."

Where we talk about this

On AI with Michal live sessions, AI bias audits come up in the AI in recruiting track alongside adverse impact because both require the same funnel data and similar calculation skills. We walk through a simplified group-rate table on anonymized data, review a sample model card from a real vendor category, and practice the one-page audit memo that compliance teams need before you sign. If you want the full conversation with peers who bring real vendor contracts, start at Workshops.

Around the web (opinions and rabbit holes)

Third-party creators move fast. Treat these as starting points, not endorsements, and double-check anything before you wire candidate data.

YouTube

Reddit

  • r/humanresources threads on "AI screening bias" and "bias audit" mix practitioner questions with compliance team experience from real debrief rooms.
  • r/recruiting has recurring threads on AI resume screeners where bias risk comes up alongside vendor comparisons and candidate experience concerns.
  • r/AIethics covers audit frameworks, fairness metrics, and research papers for practitioners who want to go deeper than the pass-rate table.

Quora

  • Searching "AI bias audit hiring" on Quora surfaces HR practitioners and researchers explaining the gap between technical fairness metrics and legal adverse impact standards (read critically; quality varies by contributor).

AI bias audit versus adverse impact analysis

AI bias auditAdverse impact analysis
ScopeInputs, logic, training data, and outputsObserved outcome gap at one funnel stage
MethodFeature review, model card review, group-rate analysis4/5ths rule calculation on selection decisions
When to runBefore launch, after each model update, quarterlyQuarterly, plus any time rejection rates spike
Legal driverNYC LL 144, EEOC AI guidance, EU AI ActEEOC Uniform Guidelines, Title VII, ADA
OwnerTA ops and legal or compliance partnerTA ops and legal or compliance partner

Related on this site

Frequently asked questions

What is an AI bias audit in plain language for a recruiter?
An AI bias audit is a structured check that asks: does this tool produce worse outcomes for candidates in a protected group? A recruiter-facing version usually means pulling decision logs from one funnel stage, splitting results by gender, race, or age band, and checking whether any group consistently scores lower, gets rejected faster, or never reaches the shortlist. The goal is not to accuse the vendor of deliberate intent but to catch inherited patterns from historical data before they compound into an adverse impact claim. A quarterly pass-rate comparison by protected group is the minimal version. A full audit adds fairness metrics, model card review, and a written owner.
How is an AI bias audit different from adverse impact analysis?
Adverse impact analysis measures an observed outcome gap at one funnel stage using the 4/5ths rule from EEOC Uniform Guidelines. An AI bias audit goes one level deeper: it investigates why the gap exists, which features the model weighted, which proxy variables correlate with protected class, and whether the training data was representative in the first place. An adverse impact finding can trigger an audit; an audit can prevent an adverse impact finding by surfacing the mechanism before scale. Teams often run both in sequence: adverse impact check flags a stage, bias audit traces the cause. See adverse impact for the legal trigger threshold.
What does a practical AI bias audit cover?
A working bias audit has four parts: data review (who is in training and labeled data, and who is missing), feature review (do inputs include zip code, school name, or employment gap as proxies for protected class), output review (pass rates, score distributions, and rejection reasons split by protected group), and vendor accountability (does the contract require the vendor to share model cards, retrain on drift, and publish independent audit results). Teams in live recruiter workshops often start with part three because the log data is already in the ATS. Parts one and two require vendor cooperation and matter most before signing new contracts.
Who should own the AI bias audit process in a TA team?
In most mid-size teams the best pairing is a TA operations lead who owns the data export and a legal or HR compliance partner who signs off on findings. Someone with basic spreadsheet skills can run the pass-rate table; the legal partner decides whether a ratio below 0.80 needs remediation or documentation. If a People Analytics function exists, bring them in for statistical significance checks on small samples. What breaks down in workshops is no clear owner: the recruiter assumes legal is watching, legal assumes HR built the dashboard, and neither looks until a complaint lands. Assign names, a quarterly calendar slot, and a one-page template before adding any new AI tool to the stack.
What legal requirements mention AI bias audits?
New York City Local Law 144 (effective July 2023) is the most cited: it requires annual independent bias audits for automated employment decision tools used in NYC, public disclosure of summary results, and candidate notification before use. Illinois, Maryland, and California have related transparency or notice laws. EEOC technical guidance from 2023 confirms that AI tools using proxy features can trigger Title VII and ADA liability. In the EU, GDPR Article 22 covers purely automated decisions and adds a human review right. EU AI Act provisions for high-risk hiring tools add conformity assessments. Build the audit into vendor contracts now; retrofitting after a complaint is far more expensive than a quarterly spreadsheet.
How do AI in recruiting workshops cover AI bias audits?
Sessions walk participants through a simplified audit on anonymized funnel data: export the log, split by group, calculate pass-rate ratios, and write the one-paragraph finding memo that legal and People need. The practical goal is operational literacy, not certification: recruiters who can read a group-rate table ask better questions in vendor calls and can push back when a vendor says "our algorithm is fair" without showing the numbers. We also practice reading model cards and vendor bias-audit summaries so participants know what to ask for before signing. Pair the audit framing with human-in-the-loop design so review points are built in before the tool goes live. Start at Workshops.
What tools and methods do teams use to run AI bias audits?
For the output layer, most teams start with what they already have: an ATS export to a spreadsheet, pivot tables by EEO field, and the 4/5ths calculation from adverse impact. Python libraries like Fairlearn, IBM AI Fairness 360, and Google What-If Tool add statistical fairness metrics when you have clean data and a data scientist available. For vendor-side audits, look for ISO/IEC 42001 conformity claims, model cards with demographic performance breakdowns, and independent third-party audit summaries. The NIST AI Risk Management Framework provides a governance checklist useful for internal policy. Pair technical findings with structured output logging so every model decision carries a version tag and a timestamp for the audit trail.

← Back to AI glossary in practice