By now, everyone is likely to have heard of ChatGPT: the generative AI module which has taken the world by storm, with its ability to review and analyse a vast amount of content to produce a coherent stream of consciousness that is not too dissimilar from a human’s.
While there are some benefits, employers must take care with how such a solution is incorporated into their business practices.
What does this mean for employers?
Legal exposure and reputational concerns
ChatGPT is built on information available from the internet, not all of which is true or accurate. So, what might happen if a company’s client asked for some advice, and in doing so, the company used ChatGPT to generate it but the AI solution relied upon incorrect and/or inaccurate information.
The client may act on this advice and be placed in a disadvantaged position and, consequently, sue the company for negligence, irreparably damaging their reputation.
Inherent bias of AI
Effectively, ChatGPT’s functionality is based on the information provided to it by human decision-makers.
However, assume the human inputting the information was, in one way or another, prejudiced against certain groupings of persons. Might this lead to cases of discrimination?
Take, for example, a HR team relying on ChatGPT to implement a new ‘rewards’ policy in a company which has historically favoured employees who work beyond their contractual hours.
Potentially, such a policy may favour young, single, men who do not have any familial responsibilities as compared to their relevant comparators. Arguably, if this policy was enacted, it may lead to claims of indirect discrimination.
Data security and confidential information
A major concern for employers is ensuring that their employees keep confidential any information, contacts, intellectual property, trade secrets and ‘know-how’ both during and after their employment, as, for most companies, this is where their value resides.
The use of ChatGPT in the workplace may undermine this value if employees offer up such information onto the internet. There may be no guarantee that that information is secure. So, consider what would happen if a data leak occurred on ChatGPT.
Obviously, the information would be at risk but the impact goes wider. It could devalue the company and, because of it, investor confidence could be shaken. Additionally, the company may be hit with breach of data claims of their own, the fines for which can go into the millions of pounds.
Potential Benefits
Despite the risk averse nature of this article, ChatGPT does have some benefits. For example:
That said, employers will need to assess whether the AI solution does more harm than good as compared to when a human employee does the same task.
If you are considering using ChatGPT or other similar AI solutions but need advice with navigating the legal issues, SMB are happy to help.
Please contact Joe Hennessy at joe.hennessy@smb.london
If you are considering separating from your partner, the process of divorce can seem daunting. You are not alone in feeling like this and the family team at SMB are regularly asked about the process of divorce and mediation. Below we have compiled some of the most popular questions that clients ask us about mediation to help sign post you to your next steps.
Read moreAs the Horizon IT Post Office Inquiry Phase 6 comes to an end, and ever more shocking evidence appears about how sub-postmasters were unlawfully convicted as part of a massive criminal conspiracy, when considering all of the bad actors who contributed to the biggest miscarriage of justice in English legal history, it’s worth remembering that there are also people who fall into the opposite category.
Read moreIn order to thwart threatened (or successfully defend actual) defamation claims, publishing lawyers regularly advise journalist/publisher clients on the proper fulfilment of their journalistic obligations.
Read more