Our lab was testing constrained LLM architectures when we realized something surprising: by redefining how the AI “thinks,” we could structurally prevent bias — not just detect it.
So we googled and realized it could be useful to HR, and we had Claude AI make some fake JD’s and resumes that it said was loaded with biases and….it worked.
Instead of filtering or training the AI, we have managed to erase demographic and proxy bias (names, schools, ZIP codes) at the cognitive level. It’s pretty cool. Testing shows near-100% blocking of known bias vectors, with full audit trails.
We tried to make a demo video comparing the base model with one with this...firewall...we are calling it. It’s here, if you’re interested.
But...now what? We’re not an HR company. We weren’t even really focusing on this. But it strips away all prestige/race/class/gender biases and evaluates strictly on merit, and that seems pretty important
Is this valuable to practitioners? Should we be talking to compliance teams, product leads, DEI groups? Supposedly it can help an EU AI Safety Act?
We’d love to hear from anyone at Personio (or anyone using Personio) who can point us in the right direction.
Thanks!