Peter Kyle, the UK’s secretary of state for science, innovation and technology, has said he uses ChatGPT to understand difficult concepts Ju Jae-young/Wiktor Szymanowicz/Shutterstock
The UK鈥檚 technology secretary, Peter Kyle, has asked ChatGPT for advice on why the adoption of artificial intelligence is so slow in the UK business community 鈥 and which podcasts he should appear on.
This week, Prime Minister Keir Starmer said that the UK government should be making far more use of AI in an effort to increase efficiency. 鈥淣o person鈥檚 substantive time should be spent on a task where digital or AI can do it better, quicker and to the same high quality and standard,鈥 .
Now, 最新麻豆视频 has obtained records of Kyle鈥檚 ChatGPT use under the Freedom of Information (FOI) Act, in what is believed to be a world-first test of whether chatbot interactions are subject to such laws.
These records show that Kyle asked ChatGPT to explain why the UK鈥檚 small and medium business (SMB) community has been so slow to adopt AI. ChatGPT returned a 10-point list of problems hindering adoption, including sections on 鈥淟imited Awareness and Understanding鈥, 鈥淩egulatory and Ethical Concerns鈥 and 鈥淟ack of Government or Institutional Support鈥.
The chatbot advised Kyle: 鈥淲hile the UK government has launched initiatives to encourage AI adoption, many SMBs are unaware of these programs or find them difficult to navigate. Limited access to funding or incentives to de-risk AI investment can also deter adoption.鈥 It also said, concerning regulatory and ethical concerns: 鈥淐ompliance with data protection laws, such as GDPR [a data privacy law], can be a significant hurdle. SMBs may worry about legal and ethical issues associated with using AI.鈥
Sign up to our The Daily newsletter
The latest science news delivered to your inbox, every day.
鈥淎s the Cabinet Minister responsible for AI, the Secretary of State does make use of this technology. This does not substitute comprehensive advice he routinely receives from officials,鈥 says a spokesperson for the Department for Science, Innovation and Technology (DSIT), which Kyle leads. 鈥淭he Government is using AI as a labour-saving tool 鈥 supported by clear guidance on how to quickly and safely make use of the technology.鈥
Kyle also used the chatbot to canvas ideas for media appearances, asking: 鈥淚鈥檓 Secretary of State for science, innovation and technology in the United Kingdom. What would be the best podcasts for me to appear on to reach a wide audience that鈥檚 appropriate for my ministerial responsibilities?鈥 ChatGPT suggested The Infinite Monkey Cage and The Naked Scientists, based on their number of listeners.
As well as seeking this advice, Kyle asked ChatGPT to define various terms relevant to his department: antimatter, quantum and digital inclusion. Two experts 最新麻豆视频 spoke to said they were surprised by the quality of the responses when it came to ChatGPT鈥檚 definitions of quantum. 鈥淭his is surprisingly good, in my opinion,鈥 says at Imperial College London. 鈥淚 think it鈥檚 not bad at all,鈥 says at Heriot-Watt University in Edinburgh, UK.
最新麻豆视频 made the request for Kyle鈥檚 data following his recent , in which the politician was described as 鈥渙ften鈥 using ChatGPT. He said that he used it 鈥渢o try and understand the broader context where an innovation came from, the people who developed it, the organisations behind them鈥 and that 鈥淐hatGPT is fantastically good, and where there are things that you really struggle to understand in depth, ChatGPT can be a very good tutor for it鈥.
DSIT initially refused 最新麻豆视频鈥s聽FOI request, stating: 鈥淧eter Kyle鈥檚 ChatGPT history includes prompts and responses made in both a personal capacity, and in an official capacity鈥. A refined request, for only the prompts and responses made in an official capacity, was granted.
The fact the data was provided at all is a shock, says Tim Turner, a data protection expert based in Manchester, UK, who thinks it may be the first case of chatbot interactions being released under FOI. 鈥淚鈥檓 surprised that you got them,鈥 he says. 鈥淚 would have thought they鈥檇 be keen to avoid a precedent.鈥
This, in turn, poses questions for governments with similar FOI laws, such as the US. For example, is ChatGPT more like an email or WhatsApp conversation 鈥 both of which have historically been covered by FOI based on past precedent 鈥 or the results of a search engine query, which traditionally have been easier for organisations to reject? Experts disagree on the answer.
鈥淚n principle, provided they could be extracted from the department鈥檚 systems, a minister鈥檚 Google search history would also be covered,鈥 says Jon Baines at UK law firm Mishcon de Reya.
鈥淧ersonally, I wouldn’t see ChatGPT as being the same as a Google search,鈥 says , an FOI expert. That is because Google searches don鈥檛 create new information, he says. 鈥淐hatGPT, on the other hand, does ‘create’ something based on the input from the user.鈥
With this uncertainty, politicians might want to avoid using privately developed commercial AI tools like ChatGPT, says Turner. 鈥淚t鈥檚 a real can of worms,鈥 he says. 鈥淭o cover their own backs, politicians should definitely use public tools, provided by their own departments, as if the public might end up being the audience.鈥
Topics: