Skip to main content

Artificial Intelligence (AI) – Your legal experts

AI has been all over the headlines in recent months. Generative AI can already produce written content, images and music which is often extremely impressive, if not yet perfect – and it is increasingly obvious that in the very near future, AI’s capabilities will revolutionise the way we work and live our lives.

What are the legal implications of AI?

The rapid development and uptake of AI has taken place against a backdrop of legal uncertainty, with many potential legal and regulatory issues arising from its use.

In relation to generative AI, questions include the extent to which the creation and use of AI models is consistent with the rights of those whose creative works have been used to ‘train’ the models. Litigation has commenced in the US and in the UK which may provide some clarity on the legality of AI image generation tools. But the question to consider at the outset is this: how exactly do these tools work and how does the law apply to the content they create.

Our specialist knowledge

Creativity, technology and innovation are at the heart of everything we do at Lewis Silkin. Our team of experienced lawyers is well-equipped to navigate the legal complexities and implications of the rapidly evolving AI landscape.

From startups to multinational corporations, we provide tailored legal solutions to the legal challenges posed by AI technology. We advise on a wide range of matters, including workplace usage policies, algorithmic discrimination, ethics, privacy issues surrounding data sources.

If you would like to arrange a bespoke event on AI please get in touch; this can be organised via LS Presents.

AI 101- our podcast series

AI 101: An introduction to AI

  • In the first introductory episode of our podcast series AI 101, Oliver Fairhurst and JJ Shaw explain some the key concepts and terminology behind this technology.

AI 101: Using AI in employment - explaining decisions and addressing bias 

  • In the second episode of our podcast series AI 101, Tarun Tawakley and Becky Jobling consider some of the implications of using AI in employment related decisions.

Our recent insights:


Related items


EU AI Act:101 – An In-depth Analysis of Europe’s AI Regulatory Framework

28 March 2024

In this article, our Data, Privacy & Cyber team provide an in-depth analysis of Europe’s AI Regulatory Framework.

Introduction to AI podcast

AI 101: An Introduction

12 March 2024

Listen here for the first in our series of podcasts on all things AI in which we’re exploring the key legal issues arising from this hot topic.


White Paper consultation response: regulating AI in the workplace

20 February 2024

In March 2023 the government published its White Paper on AI regulation. In less than a year, development in the field has been rapid, with regulators and legislators chasing to keep up. The government has now published its response to the consultation on the White Paper. We focus on what this tells us about future regulation for the use of AI in the workplace.

ai contracts

A battle is brewing in No10 over AI regulation - Roch Glowacki comments for CITY.A.M.

19 February 2024

An increasing number of MPs are getting frustrated as the government bides its time on creating checks and balances over what it has identified as a potential “existential threat”. Roch Glowacki suggests the UK government may be “deliberately stalling” to see how the EU’s AI Act pans out first.

AI chatbot usage and concepts

NYT v OpenAI: The Publishing Sector’s AI Content-Scraping Conundrum

19 January 2024

Text and Data Mining (TDM) – and “web scraping” more generally – has been thrust into the spotlight with the recent explosion of Generative AI. At the heart of the issue lies a tension between AI companies on the one hand and publishers on the other.

Artificial Intelligence digital concept

Explainability and AI: the silver bullet?

17 January 2024

Lawmakers faced with rapid advances in AI are turning to the safeguards needed to protect individuals whose interests are affected by AI systems and to build trust in automated decision-making. Building trust in automated outcomes will be much more of a challenge for many years following the revelations which are emerging from the Post Office scandal. Common among these safeguards are auditing, human oversight, effective contestability, transparency and explainability. But what is most likely to be effective?

Back To Top