As AI chatbots powered by large language models (LLMs), such as ChatGPT and Bard, become more ubiquitous and people grow increasingly comfortable with them, the way we interact with them is changing. Generative AI is no longer the preserve of data scientists on the cutting edge of technology – deftly proven just a few weeks ago when US Senator Richard Blumenthal opened a hearing with OpenAI CEO Sam Altman with a recording of his own voice. The Senator hadn’t written the words or recorded the audio: both had been generated by AI.
Sam Altman’s appearance before Congress was timely. Current AI regulations are alarmingly inefficient, not least when you consider the slate of globally-significant elections coming up in the next 18 months, with the potential to be exploited by AI-supercharged malevolent actors. We have already seen cases of AI generating mis- and disinformation, and while this is often easy to identify as false – such as the instance of ChatGPT declaring that Elon Musk had died in a Tesla accident – it is not always so obvious. Musk is high-profile, and constantly reminding us of his presence, so people quickly realised he had not died. But less high-profile people are more susceptible to people believing false statements about them that are generated by AI chatbots, such as in the case of Australian Mayor Brian Hood, who ChatGPT falsely named as being implicated in a foreign bribery scandal – when he was actually the whistleblower.
These LLM-generated false statements are, perhaps generously, being called “hallucinations”, but the reality is that the information always originates from one or more sources. There is a very real risk that LLM tools can mistakenly mass-produce, and spread, libellous information. Language proficiency is key to these tools, so it follows that at some point there will be a language-related offence, such as libel, defamation, bribery, or coercion, committed by an AI chatbot. Indeed, the first defamation case has recently been brought against OpenAI, with a man in the US alleging that ChatGPT generated a fake legal summary accusing him of fraud and embezzlement.
Identifying the information sources used by LLMs
To understand where the hallucinations are coming from, we need to take a look at how LLMs work. They are trained by collecting vast amounts of data, which is then organised and processed using a neural network architecture called a transformer, which identifies patterns in the way words are structured, the way phrases relate to one another, and the way language is constructed. Most importantly, transformers can then make (usually) accurate predictions about what words should come next. This is similar to the autocomplete function we are familiar with, but on a far greater scale.
However, AI does not understand at its core what it is saying, or think like a human. An LLM will always attempt to give an answer, and lacks the ability to admit that it does not know. The models are also unable to differentiate clearly between the need to be factually accurate and the desire to conjure something creative. This is partly why, and how, hallucinations occur.
The datasets behind LLMs
The companies creating these tools have not been overly forthcoming in revealing what datasets they are using, but we do have some information. ChatGPT partly uses a dataset called Common Crawl, a publicly-available corpus of billions of web pages. Bard, which is based on a language model called LaMDA, was trained on a dataset called Infiniset, which is comprised of about 12.5% C4 dataset (a specially-filtered version of Common Crawl) and 12.5% Wikipedia. The remaining 75% consists of words scraped from the internet that Google does not provide specifics on, using vague terms such as “non-English web documents” and “dialogs data from public forums” (which makes up 50%) to describe them.
A reasonable guess is that Reddit and Stack Overflow could be the main “public forums” (both have announced their intention to start charging for access to their data, which hints at their importance). Crucially, however, the precise source of information is not known, nor is it even necessarily consistent – LLM tools will frequently present different answers to the same question. Because answers can be generated from multiple blended sources, conflicts and contradictions can arise.
Minimising damage and reducing risk
Hallucinations will continue to be a major headache for both the developers and the users of LLM tools, not least because of the jeopardy they could create. Recently, a lawyer was forced to admit that he had used ChatGPT to generate a legal brief – it turned out that the cases cited by the bot did not actually exist.
The solution remains unclear, with Alphabet CEO Sundar Pichai saying that the question of whether hallucinations can be solved “is a matter of intense debate”. One possible remedy is greater process supervision, known as reinforcement learning with human feedback – providing human approval for each step in an LLM’s chain of thought – but the efficacy and scale of this is not clear.
Hallucinations underline the current limitations of the technology and the reality that they are not yet intelligent in the same way as humans. However, with the threats they pose to governments, companies and individuals around the world, greater understanding of how LLM-powered tools generate information can only be of benefit to those seeking to monitor the content they produce and minimise the reputational damage they could inflict.
Privacy Policy.
Revoke consent.
© Digitalis Media Ltd. Privacy Policy.
Digitalis
We firmly believe that the internet should be available and accessible to anyone, and are committed to providing a website that is accessible to the widest possible audience, regardless of circumstance and ability.
To fulfill this, we aim to adhere as strictly as possible to the World Wide Web Consortium’s (W3C) Web Content Accessibility Guidelines 2.1 (WCAG 2.1) at the AA level. These guidelines explain how to make web content accessible to people with a wide array of disabilities. Complying with those guidelines helps us ensure that the website is accessible to all people: blind people, people with motor impairments, visual impairment, cognitive disabilities, and more.
This website utilizes various technologies that are meant to make it as accessible as possible at all times. We utilize an accessibility interface that allows persons with specific disabilities to adjust the website’s UI (user interface) and design it to their personal needs.
Additionally, the website utilizes an AI-based application that runs in the background and optimizes its accessibility level constantly. This application remediates the website’s HTML, adapts Its functionality and behavior for screen-readers used by the blind users, and for keyboard functions used by individuals with motor impairments.
If you’ve found a malfunction or have ideas for improvement, we’ll be happy to hear from you. You can reach out to the website’s operators by using the following email webrequests@digitalis.com
Our website implements the ARIA attributes (Accessible Rich Internet Applications) technique, alongside various different behavioral changes, to ensure blind users visiting with screen-readers are able to read, comprehend, and enjoy the website’s functions. As soon as a user with a screen-reader enters your site, they immediately receive a prompt to enter the Screen-Reader Profile so they can browse and operate your site effectively. Here’s how our website covers some of the most important screen-reader requirements, alongside console screenshots of code examples:
Screen-reader optimization: we run a background process that learns the website’s components from top to bottom, to ensure ongoing compliance even when updating the website. In this process, we provide screen-readers with meaningful data using the ARIA set of attributes. For example, we provide accurate form labels; descriptions for actionable icons (social media icons, search icons, cart icons, etc.); validation guidance for form inputs; element roles such as buttons, menus, modal dialogues (popups), and others. Additionally, the background process scans all of the website’s images and provides an accurate and meaningful image-object-recognition-based description as an ALT (alternate text) tag for images that are not described. It will also extract texts that are embedded within the image, using an OCR (optical character recognition) technology. To turn on screen-reader adjustments at any time, users need only to press the Alt+1 keyboard combination. Screen-reader users also get automatic announcements to turn the Screen-reader mode on as soon as they enter the website.
These adjustments are compatible with all popular screen readers, including JAWS and NVDA.
Keyboard navigation optimization: The background process also adjusts the website’s HTML, and adds various behaviors using JavaScript code to make the website operable by the keyboard. This includes the ability to navigate the website using the Tab and Shift+Tab keys, operate dropdowns with the arrow keys, close them with Esc, trigger buttons and links using the Enter key, navigate between radio and checkbox elements using the arrow keys, and fill them in with the Spacebar or Enter key.Additionally, keyboard users will find quick-navigation and content-skip menus, available at any time by clicking Alt+1, or as the first elements of the site while navigating with the keyboard. The background process also handles triggered popups by moving the keyboard focus towards them as soon as they appear, and not allow the focus drift outside of it.
Users can also use shortcuts such as “M” (menus), “H” (headings), “F” (forms), “B” (buttons), and “G” (graphics) to jump to specific elements.
We aim to support the widest array of browsers and assistive technologies as possible, so our users can choose the best fitting tools for them, with as few limitations as possible. Therefore, we have worked very hard to be able to support all major systems that comprise over 95% of the user market share including Google Chrome, Mozilla Firefox, Apple Safari, Opera and Microsoft Edge, JAWS and NVDA (screen readers), both for Windows and for MAC users.
Despite our very best efforts to allow anybody to adjust the website to their needs, there may still be pages or sections that are not fully accessible, are in the process of becoming accessible, or are lacking an adequate technological solution to make them accessible. Still, we are continually improving our accessibility, adding, updating and improving its options and features, and developing and adopting new technologies. All this is meant to reach the optimal level of accessibility, following technological advancements. For any assistance, please reach out to webrequests@digitalis.com