Generative AI is transforming much of the daily activities of knowledge workers. Large Language Model (LLM) assistants, like ChatGPT and Gemini, are becoming essential tools for boosting productivity. These assistants manage a wide array of tasks, including but certainly not limited to:
- generating content and ideas
- proofreading text
- getting personalized recommendations
- summarization
- coding
LLM assistants are developed in a way that allows them to understand and respond to user questions and instructions. In that sense, they show great promise to become first-class information sources. Nevertheless, they come with several inconveniences. Some LLMs are trained on data with a cut-off date, so they are unaware of recent events. They may also lack traceability, making it difficult to verify the reliability of their responses. LLMs produce responses based on next word prediction without having a real notion of what is right or wrong. Because of that, responses may sometimes be factually inaccurate, which we have come to call "hallucinations". These drawbacks pose problems in any context, but their impact is amplified in an enterprise setting. Moreover, enterprises are more vulnerable to reputation damage, which is one of the key risks of extensive and public reliance on LLM applications.