• Sectors we work in banner(2)

    Quick Reads

AI and HR - How can employers reduce the risks associated with using artificial intelligence to help manage their workforce?

The term "Artificial Intelligence" may seem intrinsically linked to the world of the future, conjuring images of evil robots, intent to destroy humanity, or humanoid beings, eerily similar to you or me.  In fact, AI is here today, playing a huge part in the world around us.  AI is any electronic device or system which can solve the kind of complex problems we would usually associate with human intelligence, through machine-learning algorithms (an algorithm is essentially a set of instructions for a computer to follow).  This includes anything from Siri and Alexa to a self-driving Tesla, Netflix's recommendations and personalised thumbnails to Google's sense of direction.  

Like most new technologies, AI can make our lives easier and allow us to achieve more.  It's therefore no wonder that organisations are beginning to utilise AI in all aspects of employment and people management.  People Management recently highlighted the growing use of AI in recruitment, where AI can help by automating candidate screening, scheduling interviews and maintaining candidate databases.  It can perform administrative tasks, including managing holiday entitlement, absences and performance data and can even support talent management, by predicting when an employee might leave.  

It's easy to see why employers are keen to utilise AI.  However, the first adopters of any new technologies can face risks and challenges, particularly if they do not understand all of the implications of the systems they are using.  So, what risks should employers be alive to?

I remember reading about an Uber Eats driver who was dismissed by the company's algorithm last year.  Uber Eats was using facial recognition software (a type of AI) in an attempt to stop people other than drivers using an account.  The software failed to recognise one of its drivers and he was dismissed as a result.  Aside from the issues this raises about the gig economy (which we will have to leave for another day), it highlights some of the potential pitfalls in using AI

Facial recognition software is notoriously worse at recognising black people than white people and worse at recognising women than men.  The dismissed Uber Eats driver said the software used by his employer was "racially discriminatory and should be abolished until perfected".   Use of such software, without careful controls, could not only lead to discrimination claims but could also cause enormous reputational damage.  Employers should tread carefully and ensure that they put proper controls in place to double check any decisions made by AI.

AI software runs on algorithms which are trained on large data sets.  The algorithm uses the data to "learn" and improve.  This means that, if you feed AI a biased data set, you get a biased outcome.  Amazon learnt this lesson the hard way in 2014 when it rolled out a new experimental hiring tool, designed to sift through applicants and narrow down the list of potential new recruits.  The AI was trained on existing recruitment data, looking at the patterns contained in CVs submitted by successful and unsuccessful candidates over the previous 10 years .  Therefore, the AI was taught to recruit exactly the sort of people that Amazon had recruited previously, which might not have been a bad thing were it not for the unfortunate fact that Amazon had mainly hired men over those 10 years. The algorithm learned that men are more desirable candidates than women, penalising any applicant who included words such as "women's" in their CV. 

One may think that Amazon could have dealt with the problem by removing gendered language from the CVs before the AI got involved.  However, to date, this has not proved possible. Research has shown that it's essentially impossible to completely hide gender during the recruitment process; AI can identify when an applicant is male or female even where a human being or, indeed, other AI has supposedly de-gendered the CV.  

Amazon scrapped the tech in 2018. Clearly, the use of recruitment AI which exacerbates existing biases is detrimental to diversity and could open employers up to claims for discrimination.

Aside from the discrimination risks, employers are also very often under an obligation to provide explanations to employees for certain decisions, as part of the implied term of trust and confidence between an employer and employee.  How can an employer, who has blindly relied on the results of an algorithm that it does not understand, ever hope to provide such an explanation?

It is therefore vital that employers conduct their due diligence and have some understanding of the tools they are using - What is the purpose of the tool? What characteristics is it assessing?  How is the algorithm being used to make decisions?

Alongside the discrimination risks, there are potential data protection pitfalls.  UK GDPR limits the circumstances in which employers can make solely automated decisions and requires transparency where such decisions are made.  A "solely automated decision" is a decision where there has been no human influence on the outcome.  For example, if an employer's clocking-in system automatically sends a warning to an employee about punctuality, that would be a decision taken solely by automated means.  However, if the system instead sends a flag to an HR manager, who takes the decision to issue a warning, that decision is not solely automated.  

GDPR says that automated decisions are unlawful if they have a "legal effect" or a "similarly significant effect" on the data subject.  Hiring and firing would definitely fall under this umbrella, meaning employers are limited from relying on fully automated decisions in this regard.  There is an exception where automated decision making is "necessary" to enter into a contract, but this is poorly defined.  It is somewhat tricky to see where some level of human intervention would not be possible for employment purposes.  

Even where decisions are not fully automated, use of AI could potentially lead to other data protection issues.  Have a think about the incredibly broad definition of personal data - "any information relating to an identified or identifiable natural person".  Imagine that my employer is training new AI software on its workforce data.   Say that I am the only person named Briony who has ever worked for my employer but one day another Briony applies and the AI software makes certain assumptions about that other Briony, based on what it learnt about me.  Are those assumptions my personal data?  If the AI says "don't hire this new Briony, she is probably a wrongun'", what does that tell you about me and does that amount to my personal data?  

This might be an overly simplistic example - I doubt that anyone would code AI software to make recommendations based on a single data point - but there really are potential issues here and the answers are not always clear cut.  Employers will need to think carefully about how and why they use employee personal data, ensuring they have a legal basis for each processing operation and are fair and transparent about their actions. 

Ultimately, this is an exciting area and not one to fear.  AI can help to make our lives easier and allow us to make better, more informed decisions.  However, it can only do that if we understand how it works and its limitations and, vitally, consider when it is necessary for an actual human being to step in.  

In 2019, Unilever reported that its use of AI had saved its human recruiters approximately 100,000 hours in interviewing time and nearly £1m per year. However, an overreliance on AI when making recruitment decisions case can see employers easily being wrong-footed and inadvertently breaching UK data protection and anti-discrimination laws.

Our thinking

  • Updates from the Building Safety Regulator - Unblocking the Gateways for Higher Risk Buildings

    Tegan Johnson

    Quick Reads

  • The 1975 Act Turns Fifty: Why Reform was Needed and What Changed

    Tamasin Perkins

    Insights

  • ECCTA for Charities: Maintaining Registers

    Giverny McAndry

    Insights

  • ECCTA 2023 - Failure to prevent fraud offence- what charities need to know and do

    Penelope Byatt

    Insights

  • What do agricultural landlords and workers need to know about the Renters’ Rights Act?

    Emma Preece

    Insights

  • An introduction to Economic Crime and Corporate Transparency Act 2023 for charities: key changes from 18 November 2025

    Liz Gifford

    Insights

  • Succession Stumbling Blocks: Lessons from Thomas v Countryside Solutions Ltd

    Maddie Dunn

    Quick Reads

  • Morning Star UK quotes Julia Cox on the impact of potential inheritance tax rises in the UK Autumn Budget

    Julia Cox

    In the Press

  • What legal developments can the Living Sector expect as we approach the end of 2025 and look ahead to 2026?

    Mark White

    Insights

  • Autumn Budget 2025: Sifting the Rumours on Tax Rises and Reforms

    Charlotte Inglis

    Quick Reads

  • Adjudication under the Construction Act – a case on the residential occupier exception and contesting the validity of a payless notice

    Tegan Johnson

    Insights

  • VAT on Developer’s Biodiversity net gain (BNG) costs

    Elizabeth Hughes

    Insights

  • Understanding the Fire Safety (Residential Evacuation Plans) (England) Regulations 2025: The Living Sector

    David Savage

    Insights

  • Cross-border estates and the new “non-dom” regime: UK IHT reporting on death

    Harriet Betteridge

    Insights

  • What role can construction lawyers play in helping UK construction sector clients achieve greater success and how?

    David Savage

    Insights

  • Amendments to the Non-Contentious Probate Rules in force from today

    Jessie Davies

    Quick Reads

  • Swiss Employment Law: Your Essential Guide to Contracts, Rights, and Regulations

    Remo Wagner

    Quick Reads

  • Installing Chinese Turbines in European Wind Projects – what do you need to know?

    Jue Jun Lu

    Insights

  • Charles Russell Speechlys advises Battery Ventures on its strategic investment round in Signal AI

    Jonathan Morley

    News

  • The AI Advantage: Transforming International Arbitration

    Katy Ackroyd

    Insights

Back to top