The 3 laws of using a recruitment bot

The results from a survey commissioned by MYOB recently found that candidates do not trust artificial intelligence (AI) to make the correct recruitment decisions. Perhaps. But they will soon.
We just need it to be a bit better than it is. Every time a piece of robotics gets it right, we trust it a bit more.
In his article, Trevor Vas outlined a set of principles for the use of AI in recruitment. I think this is a great first step to building trust, but let me go a step further by sitting on the shoulders of Asimov to take a view on this and boil the ethical standards down to three laws:

  1. A recruitment bot may not disadvantage a candidate, or through inaction, allow a candidate to be disadvantaged.
  2. A recruitment bot must select a candidate except where such selection would conflict with the First Law.
  3. A recruitment bot must protect its own data and logic as long as such protection does not conflict with the First or Second Law.
Law 1 – A recruitment bot may not disadvantage a candidate, or through inaction, allow a candidate to be disadvantaged.

Throughout the process a candidate may not be disadvantaged sounds simple and sensible, but it contains a lot of important details that should not be skipped over.
Every candidate needs to be assessed on their own merits against a pre-determined and defensible criterion. The criterion needs to be the same for all candidates applying for or being applied to a role.
When assessing candidates, it should not matter at all whether they are passive or active, have been sourced or applied, are currently employed or not. All candidates need to go through the same process.
Candidates who are not assessed by the recruitment bot may be disadvantaged, or, more likely, those being assessed by humans may put the others at a disadvantage. As such, a recruitment bot must not let candidates fall outside the process.
And the candidate needs to be at the heart of the process. The candidate is the one paying with time and data – they need to be first in design of the recruitment bot. Not doing so would disadvantage the candidates and potentially lead to a poor candidate experience.

Law 2 – A recruitment bot must select a candidate except where such selection would conflict with the First Law.

The purpose of the recruitment exercise is to select a candidate. Therefore, there needs to be an outcome.
More interestingly, by requiring an outcome, except where doing such would conflict with the First Law, ensures that there can be no bias in the system.
Every candidate needs to be assessed on their own merit against the pre-determined and defensible criterion, then measured directly against each other until there is a single victor.
If no victor can be ascertained, making a decision would disadvantage one of the candidates, breaking the First Law.
Therefore, all the candidates at this stage must be selected and put forward; nothing wrong with interviewing three candidates, or five. If you have more than that the criteria are not tight enough to truly be used to assess candidates in a non-biased way.

Law 3 – A recruitment bot must protect its own data and logic as long as such protection does not conflict with the First or Second Laws.

The programmer deserves a commercial advantage, unless their product contains bias or incorrect algorithms.
By enabling the recruitment bot to maintain its own privacy, commercial advantage can be maintained, but the Third Law forces an audit to ensure that the first two laws are being followed.
The problem is that, as with Asimov’s original three laws, there is no way to actually implement them into a piece of machine learning.
My point is that we don’t need to make this complicated, any agreed rules should be simple enough for developers to work within while ensuring the best results for the candidates, hiring managers, and recruiters.
But we do need to do it now.
Recently I was driving home with my daughter and she asked me to play an Adele song – she sang a little bit of it.
Simon: “Hey Google, please play Photograph by Adele.”
Google: “Sure thing, playing When We Were Young by Adele.”
I got the name wrong, but Google knew what I meant and played me the song I wanted. That is helpful; that is building trust. That is where we are going, fast.
The world is changing and, before we know it, it may be that a robot is looking through your entire history of interactions with Siri, Alexa, and Google to assess how polite you are when dealing with your coworkers…
…you do remember to say please and thank you, right?
Cover image: Shutterstock


The upcoming Future of Talent retreat will explore these laws in detail and come up with a set of ethical standards for the use of AI in Talent Acquisition in Australia. If you are an internal TA executive for a corporate organisation, we would love for you to be part of this trailblazing event. Register your interest here.

Article By

Get more articles direct to your inbox

Upcoming Events

Thursday, 15 May 2025

28 & 29 October 2025

You may also enjoy reading...

This Week in Talent: Want to attract, grow, and keep top talent? Internal talent mapping is the secret weapon you’re probably not using enough. It uncovers skills gaps, fuels career growth, and turns high performers into future leaders—while keeping your best people engaged. Ready to make talent mapping your ultimate game-changer? Let’s dive in!
This Week in Talent: Position Descriptions (PDs) are often outdated, uninspiring, and forgotten—but they don’t have to be. Athanasia explores how to breathe new life into PDs, turning them into dynamic, strategic tools that drive performance, career growth, and organisational success. Learn how to simplify, standardise, and transform PDs into something truly impactful.
This Week in Talent: Discover key strategies to shake off the holiday haze, boost collaboration, and set yourself up for success. From fostering teamwork to managing energy wisely, this guide will help you maximize your potential and make this year count. Read more now.