Artificial Intelligence chatbots are everywhere. They have captured the public imagination and that of countless Silicon Valley inventors and investors since the arrival of ChatGPT about a year ago.

The stunning human-like abilities of conversational AI – a form of artificial intelligence that enables computers to process and generate human language – have sparked widespread optimism about their potential to transform workplaces and increase productivity.

In what may be a world first, a UK school has appointed an AI chatbot as a “principal headteacher” to support its headmaster. While little is known about the nature of the AI behind it, the chatbot is meant to advise staff on issues such as helping pupils with ADHD and writing school policies.

But, before deploying chatbots in the workplace, it is crucial to understand what they are, how they work and how to use them responsibly.

How chatbots work

Much has been written about the astounding abilities of generative AI, and its sometimes surprising ability to get things wrong.

For example, an AI chatbot can craft a persuasive scholarly argument but, to use chatbot terminology, can “hallucinate” the list of references or get simple facts wrong. To understand why hallucinations occur it is important to understand how these chatbots work.

At their core, AI chatbots are powered by large language models (LLMs), large neural networks trained on massive datasets of text (what we lovingly call “the Internet”).

Importantly, LLMs do not store any data or knowledge in any traditional sense. Rather, when they are built (or “trained”) they encode, in large statistical structures, sophisticated content or language patterns contained in the training data.

Simply speaking, text is turned into numbers, or probabilities.

When in use, LLMs no longer have access to this training data. So when we ask it a question, the response is generated from scratch, every time. Technically, everything is “hallucinated”. When AI chatbots get things right, it is because much of human knowledge is patterned and embedded in language, not because “it knows”.

By design, AI chatbots cannot produce definitive factual answers. They are probabilistic, not deterministic systems, and therefore cannot be relied on as authoritative sources of knowledge. But their ability to recognise linguistic patterns makes them excel at helping humans with tasks that involve text generation or enhancement.

Writing persuasive arguments follows certain patterned regularities, whereas factual answers cannot reliably be generated from probabilistic patterns.

The new workplace assistant

Don’t think of your AI chatbot as an omniscient artificial brain, but as a gifted graduate student assigned to be your personal work assistant.

Like an eager grad student, they work tirelessly and mostly competently on assigned tasks. However, they are also a little bit cocky. Always overconfident, they might take risky shortcuts and provide answers that sound good but lack any factual grounding.

It is wise always to verify chatbot outputs, much as you would double-check a grad student’s work.

Because of their probabilistic foundation they don’t comprehend your question with any human understanding. But in the right roles, when used appropriately, chatbots greatly augment productivity on language-related tasks.

Working with AI chatbots

The three levels of chatbot capability can be summed up using the acronym ACE – assisting, creating, exploring.

  • assisting – chatbots can assist with many writing tasks, such as summarising, analysing and refining text, or extracting key points and themes. They can express arguments in academic text in more accessible ways.
  • creating – chatbots can generate original text, turning dot points into business reports or ideas. They can mimic different genres and write in different styles. As they encode countless bodies of text from different domains, they can be told to take a perspective, impersonating business strategists, scholars, marketers or journalists, to create content useful across many professions.
  • exploring – chatbots make intriguing “discussion partners” about hypothetical ideas (“what would happen if…”). When exploring new issues, let the chatbot set you questions then get it to answer them. If you want to explore what makes a good project report, or social media post, ask the chatbot to write one and then reflect on why it wrote what it did.

The business of chatbots

What do we know about the use of AI chatbots in the workplace so far? Some initial studies point to significant productivity gains.

A pilot project at Westpac found a 46% productivity gain in software coding tasks, with no drop in quality. The experiment compared groups of developers using AI chatbots for a range of programming tasks with a control group that did not.

A study by global management company Boston Consulting Group also reported significant improvements.

In a controlled experiment, consultants used AI chatbots to problem-solve and develop new product ideas, which involved both analytical work and persuasive writing. Those who worked with the chatbot finished 12.2% more tasks, 25.1% more quickly, and at 40% higher quality, than those who didn’t.

In yet another case, an AI chatbot is reportedly being used by a US software company to help write proposals for clients. It scours thousands of internal files for relevant information to generate a suitable response, saving the company time.

These cases give glimpses into the future of AI chatbots, where companies fine-tune generative AI models with their own data or documents, using them for specialist roles such as coders, consultants or call-centre workers.

Many workers are worried AI will be used to automate their work. But given the probabilistic nature of the technology, and its inherent lack of reliability, we do not see automation as the most likely area of application.

AI chatbots might not be coming for your job after all, but they are certainly coming for your job description. AI fluency, the skill to understand and work with AI will soon become essential, similar to working with PCs.

Finally, you might ask, did we let a chatbot write this article? Of course we didn’t. Did we use one in writing it? Of course we did, much like we used a computer, and not a typewriter.

The Conversation requires authors to disclose if they have used AI in the preparation of an article. Articles that use AI for fact-finding or idea generation will generally not be accepted.

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Image: Google DeepMind

Kai Riemer is Professor of Information Technology and Organisation, and Director of Sydney Executive Plus at the University of Sydney Business School. Kai's research interest is in Disruptive Technologies, Enterprise Social Media, Virtual Work, Collaborative Technologies and the Philosophy of Technology.

Dr Sandra Peter is the Director of Sydney Executive Plus at the University of Sydney Business School. Her research and practice focuses on engaging with the future in productive ways, and the impact of emerging technologies on business and society.


We believe in open and honest access to knowledge. We use a Creative Commons Attribution NoDerivatives licence for our articles and podcasts, so you can republish them for free, online or in print.

Related content