Organizations will increasingly be confronted by these questions and others as GenAI comes closer to employees’ livelihoods and identities. For example, there is likely consensus that using software tools of any kind to make final hiring or termination decisions is not an ethical practice. But would there be this same consensus about using GenAI to write the job description, develop the interview questions, or summarize candidate interviews for review by the hiring manager?
These issues pose thorny questions about both real and perceived bias and fairness. For this reason, leaders and their teams need to communicate to establish norms that align with policies, culture, and regulations, especially in the areas of human resources and other areas that have direct impacts on people’s lives and livelihoods.
An Imperfect Solution
These concerns about GenAI have their basis in how the models are trained and how they operate. Large language models (LLMs) are a key component of GenAI solutions. LLMs are trained on massive quantities of data that come from a variety of sources. Where exactly this data comes from is considered proprietary by many GenAI software vendors, but it is safe to assume much of it is gathered from the internet. Thus, much of the data will contain biases, incorrect information, and inappropriate content. While GenAI developers generally work to inhibit overtly offensive, inaccurate, or disturbing content, the system is not perfect.
The risk of bias is not limited to the training data of the models. One common approach to using GenAI at work is to use a technique called retrieval augmented generation (RAG). RAG approaches use some form of knowledge repository to allow users to interact with information in the repository. The knowledge repository is often a collection of documents, such as policies, processes, or regulations.
RAG solutions are a useful way to create common GenAI applications. For instance, an organization may create a knowledge assistant that can help employees with corporate travel and expense policies. However, these solutions are susceptible to biases, as well. If the documents in the knowledge repository have biased or inaccurate information, the GenAI solution will produce biased or inaccurate results.
These biases can be introduced when:
- Documents are inadvertently included (such as an out-of-date version of the policy) or excluded.
- Bad actors include documents (or messages inside documents) that contain inaccurate information or undesirable language, an attack so common that it has a name — data poisoning.
Organizations should have a thoughtful and well-governed approach to how knowledge repositories are constructed, as well as how they are protected from unauthorized access to effectively manage the risks.
Of course, GenAI is not just subject to providing biased responses. It sometimes entirely fabricates statements — known as “hallucinations.” LLMs do not have a database of facts in the model. Rather, they identify patterns in language using complex mathematical inference techniques. They use all the patterns learned from enormous quantities of training data to statistically predict what will come next in a sentence based on the available context, typically the previous prompts. So, statements that have been seen repeatedly in the training data will have a strong statistical correlation.
When the user asks about ideas or concepts where there is not much repetition in the training data, the LLM does its best to find something that sounds plausible. Indeed, the fact that these responses often appear plausible to nonexperts is part of what makes hallucinations so challenging.
The existence of hallucinations should be covered in organizational training. The more employees understand the model’s behavior, the better positioned they are to identify problems. Employees on the lookout for factual errors can fact check GenAI responses before using the output.
Basic prompt engineering can help reduce the occurrence of hallucinations. As the old saying goes, “ask better questions, get better answers.” There also are some technical ways to reduce hallucinations. For example, many GenAI vendors provide settings users can adjust to change the behavior of the model.
“Temperature,” for instance, allows users to increase or decrease the model’s creativity. Turning the temperature up will cause the GenAI to be more creative in its responses. This may be desirable when using it to create ideas for marketing copy. Lower temperatures will prompt less creativity, which may be desirable when using GenAI to summarize facts from a complex document.
Out of Bounds
Unfortunately, when considering how to manage GenAI risks, organizations cannot ignore the risk of insider threats and employee abuse. Beyond the gray area that exists because of the lack of established social norms, there are uses of GenAI that are clearly out of bounds.
So-called prompt injection or jailbreak attacks occur when users attempt to trick the GenAI into saying or doing things that it should not. The goal of such attacks can range from mischief, such as trying to get the GenAI to say something offensive, to fraud, such as trying to access information that the user should not be permitted to access. GenAI also may assist employees that intend to commit fraud through the creation of deepfakes or other forgeries.
Enterprise-class GenAI solutions can implement guardrails and other configurations that prevent or at least reduce the risk of misuse. Guardrails come in many forms and can be configured or given as plain language directives by administrators that are inserted into the processing of the prompt. The guardrails can, for example, prevent the LLM from discussing topics such as violence or crime.
They also can direct the LLM to not perform certain types of tasks. For example, a magazine editor concerned about journalists using GenAI to create articles could create guardrails that would prohibit the LLM from processing prompts that request the creation of a magazine article.
While guardrails can be an important part of managing GenAI, administrators also must have a system to periodically review how employees are using and misusing organizational GenAI resources. Such a system can help identify employee misuse and types of misuse so the organization can implement protections. Reviewing log files also can help identify desirable use cases.
Here to Stay
GenAI is a permanent part of the business landscape. It will fuel tremendous efficiency for the organizations that adopt it. Although there are numerous risks inherent to GenAI’s use, these risks are manageable and should not deter organizations from using it. Early adopters of GenAI have followed identifiable patterns that help get the most out of the technology while effectively managing the biggest risks. Ultimately, organizations that take a strategic and considered approach to GenAI will be well-positioned to enjoy the benefits of this important technology for years to come.