academicAI

Responsible AI in Academia

Author: Kieran Lindsay (ORCID: 0000-0002-1624-7679) | First published: March 31, 2024 | Last update: May 13, 2024 (see file history for versions)

This is a living document. It is open source, and anyone is welcome to make suggestions. You can find the source and document history here or reach out to Kieran at kieran@academicid.net


It is an exciting time for AI research and the use of AI in research. Generative AI has the potential to accelerate and enrich many aspects of the research process, from hypothesis generation to data analysis to dissemination. However, to ensure the reproducibility and integrity of research that involves or utilises generative AI, it is vital to be transparent about its use.

We are still learning about this technology’s impact and are faced with an ever-growing array of AI models, all with different strengths and a wide array of usage parameters. As such, it is essential that generative AI users are transparent and specific in reporting on its use. Providing specific details on AI use will help the community build a clearer picture of AI capabilities and move the field forward.

Below are some preliminary guidance principles that academics, researchers and students can follow when using generative AI to assist with their research or in situations where generative AI forms the subject of their research.

These guidelines are preliminary and are being incrementally expanded and improved. They are designed as a starting point from which the risks, mitigations, and reporting on specific AI use cases can be developed. As such, these should not be relied on solely but used in conjunction with established research procedures, best practices, methodologies and ethics guidelines.

How Generative AI Works: A Primer

Generative AI is a subset of artificial intelligence that focuses on creating new content, such as text, images, audio, or video, based on learned patterns from existing data. The most common type of generative AI is the language model, which can produce human-like text after being trained on vast amounts of written material.

Training Process

There are many variations used for training generative AI models. Typically, training involves feeding the model a large dataset, such as a corpus of books, articles, or websites. The AI system analyses this data to identify patterns and relationships between words, phrases, and concepts. For example, a language model might learn that the word “king” is often associated with words like “queen,” “castle,” and “throne,” and that it frequently appears in sentences about royalty, power, and medieval times.

This first stage of the learning process is based on a technique called “unsupervised learning,” where the AI is not given explicit labels or categories for the data, but rather discovers patterns on its own. The model adjusts its internal parameters to capture these patterns better, essentially building a statistical representation of the language.

Most models then undergo some form of supervised fine-tuning on a smaller dataset for a specific task, such as question answering, text summarisation, or sentiment analysis. This involves providing the model with labelled examples of the desired behaviour and updating its parameters to minimise the difference between its predictions and the correct labels.

Some models incorporate human feedback into the training process to better align the model’s outputs with human preferences and values. This can involve having humans rate the quality or appropriateness of the model’s outputs and using this feedback as a reward signal to update the model’s parameters via reinforcement learning. For example, if the model generates a text that humans deem offensive or biased, it would receive a negative reward, encouraging it to avoid such outputs in the future.

The training process is often iterative, with model creators fine-tuning and adjusting the model based on its performance and user feedback. This can involve updating the training data, modifying the model architecture, or tweaking hyperparameters to improve the model’s outputs.

It is important to note that the specific training process varies widely depending on the type of generative AI model, the intended application, and the researchers’ goals and resources. Some models may rely primarily on unsupervised pre-training, while others may place greater emphasis on human feedback and alignment. Some models replace the human feedback step with feedback from more powerful models such as GPT-4.

Training is the most significant part of a model’s development in influencing performance and behaviour. The key takeaway here is that different models will have vastly different capabilities and behaviours and, for lack of a better phrase, statistically influenced worldviews. Model creators train their models to meet the specifications the creators desire, and, as a result, when assessing a generative AI model for use, it is vital to consider the goals and values of the model’s creator.

Neural Networks and Transformers

Most modern generative AI systems are based on artificial neural networks designed to loosely mimic the human brain’s structure. These networks consist of interconnected nodes (or “neurons”) organised into layers. Each node performs a simple computation based on the inputs it receives from nodes in the previous layer and passes the result to nodes in the next layer. Through this process, the network can learn to recognise complex patterns and relationships in the data.

A particularly influential neural network architecture for generative AI is the “transformer,” introduced in a 2017 paper by Ashish Vaswani et al. The transformer uses a mechanism called “attention” to weigh the relevance of each part of the input data to the current prediction task.

This attention mechanism allows the model to capture long-range dependencies in the data more effectively than previous architectures. It has been the basis for many state-of-the-art language models, such as GPT (Generative Pre-trained Transformer) and BERT (Bidirectional Encoder Representations from Transformers).

Generating New Content

Once a generative AI model has been trained, it can be used to generate new content by providing it with a prompt or seed input. For a language model, this might be the beginning of a sentence or paragraph, which the model then tries to complete based on the patterns it has learned.

The generation process is probabilistic - at each step, the model predicts a distribution over possible following words or tokens and then samples from that distribution to choose the next word or token to generate. This sampling process can be adjusted to make the model’s output more or less “creative” or unpredictable.

For example, suppose we prompt a language model with the phrase “The old king lived in a…”. The model might predict a high probability for words like “castle,” or “palace,” based on the patterns it has seen in its training data. It would then likely generate one of these words to complete the sentence. However, if we adjust the sampling parameters to be more adventurous, it might generate a less expected word like “spaceship” or “treehouse,” leading to a more surprising or imaginative output.

Usually, the user can control some of these parameters and adjust them to fine-tune the model’s outputs for a specific task or application. For example, in a creative writing context, a higher sampling temperature might be used to encourage more diverse and unexpected outputs. In contrast, a lower temperature might be preferred in a summarisation task to ensure the model sticks closely to the input data and, thus, outputs more deterministic content. This is important as the parameters set by the user can influence the model’s outputs, and as such, the user must be aware of these parameters and their impact on the model’s outputs.

When discussing the size or complexity of a generative AI model, it is common to refer to the number of tokens it can process or generate. A token is a basic unit of text that is used as an input or output for a model. Tokens can be individual words, subwords (parts of words), or even single characters, depending on the specific tokenisation method used.

For example, consider the sentence: “The quick brown fox jumps over the lazy dog.” A word-level tokeniser would split this into the following tokens: [“The”, “quick”, “brown”, “fox”, “jumps”, “over”, “the”, “lazy”, “dog”]. However, a subword tokeniser might break down less common words into smaller parts, like this: [“The”, “quick”, “brown”, “fox”, “jump”, “##s”, “over”, “the”, “lazy”, “dog”], where “##s” is a special token indicating that it should be attached to the previous token.

End users have little control over the tokenisation process; however, when assessing the capabilities of a model, the number of tokens a model can process or generate is a necessary metric to consider. For example, it will be necessary to utilise a model with a larger token window if the user is processing long documents. Most models will provide a rough token-to-word conversion estimate. For example, OpenAI provides a general rule of thumb: one token equals roughly four characters, or 100 tokens equate to roughly 75 words.

Risks of Generative AI

Addressing the risks associated with generative AI use requires awareness. Fostering a culture of critical engagement with generative AI is the best method to help mitigate the risks.

The use of generative AI – either as part of day-to-day activities, or within a specific research methodology, or as the object of research – introduces a range of risks that require consideration. A selection of the most obvious risks are highlighted below. Users of generative AI should undertake further research into the risks for specific use cases, especially when using generative AI as part of a research methodology.

Ethical Risks

  1. Bias and Fairness: Generative AI systems, reflecting biases in their training data, can perpetuate or exacerbate biases. This risk is particularly concerning in applications that may impact decision-making in critical areas such as healthcare, justice, and employment or the use of generative AI in qualitative or research methodologies involving any degree of subjective assessment. For instance, a generative AI model trained on historical medical data might inherit biases against certain demographic groups, leading to skewed research outcomes reinforcing health disparities.

  2. Alignment and Control: As noted above, it is vital to remember that companies train large language models for performance, safety and alignment, not necessarily balance and that it is up to these companies to determine what alignment is. Alignment is usually a metric that considers how the model’s behaviour matches what the intended human user expects. For general-purpose or consumer-focused large language models, this alignment is usually to the average user (or whatever the most number of paying customers may prefer in a language model). However, the influence of the creator’s personal views in training the models cannot be discounted. As such, it is essential to consider the alignment of the model in the context of the intended user’s and model creator’s goals and values. This approach to alignment and safety can lead to the model refusing to respond on a topic, producing outputs that are not in line with the research’s objectives or outputs biased towards a particular ideological or political view. Moreover, the alignment of generative AI models can change over time as companies update their training data, algorithms, or content policies. This can create challenges for longitudinal research projects that rely on consistent and predictable model behaviour.

  3. Privacy Concerns: The use of generative AI in processing and synthesising data can inadvertently lead to privacy breaches, especially when dealing with sensitive or personal information. The ability of these systems to generate detailed synthetic data based on real-world datasets poses a risk of re-identification of individuals whose data has inadvertently entered the training set. Further, even when using models where data is not used for retraining, sending data to third-party companies may breach contractual or legislative privacy provisions.

  4. Intellectual Property and Plagiarism: Generative AI’s capacity to produce content that closely mimics existing copyrighted material raises concerns about unintentional plagiarism and the violation of intellectual property rights. While this remains a legal grey area, the concern remains particularly problematic in academic settings where the originality of research outputs and the proper attribution of knowledge are paramount.

Operational Risks

  1. Reliability and Reproducibility: The stochastic nature of generative AI models means that they can produce inconsistent results, challenging the reproducibility of research findings. This variability, coupled with the ‘black box’ nature of many AI systems, can make it difficult to understand and explain how specific results were generated, undermining research’s reproducibility, credibility and integrity. For instance, two researchers using the same generative model but with slightly different parameters will likely get divergent results, making it impossible to replicate findings.

  2. Overreliance and Skill Degradation: The convenience and efficiency of generative AI tools can lead to an overreliance on automated systems, potentially atrophying critical research skills such as literature review, data analysis, and even the formulation of research questions. This overreliance could also lead to researchers accepting AI-generated insights without sufficient scrutiny, increasing the risk of propagating errors or flawed analyses.

  3. Writing Style and Quality This risk of overreliance extends to the use of generative AI for writing published academic material. Whilst anecdotal, most proficient in English will readily question text that uses overly verbose, archaic or unneedingly positive language. In addition to the positive style that models are trained to mimic, most generative AI systems have some form of frequency/presence penalty parameter that discourages the repeated use of the same tokens within close proximity. While this helps prevent the model from getting stuck in loops outputting repeated phrases and words, it can result in an expanded vocabulary that is not always warranted. While the impact on style, readability and substance is obvious, these may be forgiven, to an extent, for the time-saving benefit. However, when content is purported to be human-written, drops in quality may result in a feedback loop amplifying these deficiencies as models are retrained on this “human-generated” content.

  4. Regulatory and Legal Risks: The rapid improvement in capabilities of generative AI and its applications significantly outpace existing legal and regulatory frameworks, leading to uncertainties and potential non-compliance. Further, and more generally, while it may be tempting to copy any and all data into ChatGPT to have it summarised, sending this data to a third-party company (potentially out of your jurisdiction) may be a breach of contract or privacy-focused legislation. With the massive amount of computing resources required to power AI models, without specific assurances as to where data is processed by an AI model, it may be sent to multiple jurisdictions that you are unaware of, potentially breaching data sovereignty laws.

Guidelines for Using Generative AI to Assist Academic Research

Now that we are aware of some of the risks associated with generative AI use, it is important to consider how we can mitigate these risks. Using generative AI in a manner consistent with ethical and legal obligations is not complicated; however, it does require some thought. Below are some guidelines for best practices that will serve as a starting point for the ethical and responsible use of AI in academic settings. These guidelines are not exhaustive and should be applied with a level of discretion aligning with the user’s comfort and familiarity with the technology; additional considerations should be incorporated based on the identified risks and the specific use case.

It can be hard to determine the best practices for using generative AI in academic research, as the technology is still evolving. This is more challenging for those without a technical background, as the technology is often opaque and difficult to understand. The best way to overcome this is to use the technology frequently and in lower-stakes situations. The more you use the technology, the greater your intuitive understanding of its strengths and weaknesses will be. This will allow you to understand better the technology and its limitations and, thus, the risks it may present in your research.

It should also be noted up front that general consumer-facing generative AI services, such as the ChatGPT, Gemini or Claude user interfaces, may not be suitable in situations where generative AI forms part of a research methodology and needs to be reproduced. These services are subject to change, and the parameters used to generate responses are not always public. As such, it is recommended that programmatic solutions or customisable playgrounds be used where possible. You can usually find these playgrounds by signing up for developer accounts or searching something along the lines of ‘OpenAI API playground’. This will allow for the control and reporting of parameters used in the research, which is essential for any research that others may want to reproduce. However, with the risks noted above in mind, consumer-facing services are likely suitable for exploratory research when testing generative AI or just day-to-day use of the tool.

Transparency and disclosure

We used OpenAI’s GPT-4-Turbo model to generate potential interview questions that were reviewed and validated by the research team. The specific generative AI parameters used were as follows:

Human oversight and editorial control

Respect intellectual property

Consider research ethics

Data protection and privacy

Reproducibility and data availability

Mitigate bias

Maintain academic standards

Guidelines for the Scientific Testing of AI Capabilities

Compare to human baselines

Test for robustness & reliability

Interrogate decision-making processes

Acknowledge limitations

Foster open science

Examine social and ethical implications

Considerations for Using Generative AI in Qualitative Research Methodology

The integration of generative AI into qualitative research methodologies offers possibilities for data analysis, theme generation, and insight discovery. The capability of generative AI to carry out “human-level” tasks presents opportunities to speed up the processing of large quantities of data or undertake novel approaches to qualitative research. However, the unique nature of qualitative research, with its emphasis on depth, context, and subjective interpretation, requires careful consideration of several factors to ensure the integrity of research outcomes. Below are several key considerations for researchers aiming to incorporate generative AI into their qualitative research methodologies.

Alignment with research goals and ethics

  1. Purposeful Integration: Clearly define the role of generative AI in qualitative research. Whether it is used for data analysis, generating interview questions, or synthesising themes, ensure that its application directly supports the research objectives without compromising the depth and nuance of the research’s proposed outcomes.

  2. Ethical Considerations: Reflect on the ethical implications of using generative AI, especially in relation to participant consent, confidentiality, and the potential for bias. Be transparent with participants about the use of generative AI in the research process and consider the impact of AI-generated insights on the research subjects and broader societal contexts.

Human-AI collaboration and researcher competency

  1. Human-AI Collaboration: Emphasize the importance of human-AI collaboration in qualitative research. While generative AI can assist in various tasks, it should not replace human judgment, expertise, and contextual understanding. Researchers should work in tandem with AI tools, leveraging their complementary strengths to enhance the quality and depth of the research.

  2. Training and Competency: Researchers should acquire the necessary technical skills and knowledge to integrate AI into their research workflows effectively. This may involve learning about AI fundamentals, familiarising themselves with specific AI tools and platforms, and staying updated on best practices and ethical guidelines for AI use in research.

Data integrity and authenticity

  1. Data Handling: Exercise caution in feeding sensitive or identifiable data into generative AI tools. Anonymise data where possible to protect participant privacy and comply with data protection regulations.

  2. Authenticity of Insights: While generative AI can help identify patterns and themes in data, critically evaluating these insights for authenticity and relevance is crucial. Researchers should not rely solely on AI for data interpretation but use it as a tool to complement their analytical skills and subject matter expertise.

  3. Verify Capabilities: Before undertaking large-scale data processing using generative AI, have it process a representative subset of the data and have a human verify that the AI model is performing as desired. This verification stage should form a key step in the reported methodology and results to demonstrate that the generative AI and the parameters set are verified as performing pursuant to the objectives of its use.

  4. Iterative Analysis: Employ an iterative approach to data analysis, alternating between generative AI-assisted and traditional manual methods. This hybrid strategy can aid in speeding up the research while ensuring the rigour and depth of the qualitative research are maintained.

Methodological transparency and reproducibility

  1. Critical Engagement and Reflexivity: Maintain a reflexive stance towards the use of generative AI, continually questioning how it influences the research process, the construction of knowledge, and the interpretation of data. Acknowledge the limitations of generative AI and the potential for the technology to shape research outcomes in unforeseen ways.

Mitigating bias and acknowledging limitations

  1. Bias Detection and Correction: Like humans, generative AI is not free from subjective viewpoints. Actively seek to identify, measure and mitigate biases in the generative AI’s outputs. This involves critically assessing the data sources, training materials, intended users, and inherent model biases. Employ diverse datasets and consider multiple perspectives to challenge and refine the AI-generated insights. Prompting can be effective in altering, to a degree, the viewpoints expressed by generative AI.

  2. Contextual Sensitivity: Stress the significance of contextual sensitivity when applying generative AI. Qualitative data often involves nuanced meanings, cultural specificities, and personal narratives that may not be fully captured by AI algorithms. Researchers should be cautious not to over-rely on AI-generated insights and should always interpret the results within the broader context of the research setting.

  3. Diversity of Data and Perspectives: Ensure that the data fed into generative AI tools and the resulting analysis reflect the diversity of the research population. Be wary of over-generalisation and the erasure of minority experiences and perspectives. At the same time, be aware that the alignment training of generative AI systems may lead to models overrepresenting minority perspectives.

Responsible Publication and Dissemination

Consider dual use

Engage in public dialogue

Continuously update practices


Disclosure of the use of Generative AI in the Creation of this Document

This document was created with the assistance of the following AI tools.