How to Use AI in Research

đź’ˇ Learning Outcomes

  • Practise designing effective prompts for policy‑oriented research questions and document the results.
  • Critically assess and validate AI‑generated summaries against authoritative policy documents and identify inaccuracies or omissions.
  • Explore how role prompting influences Copilot responses and appreciate the limitations of AI advice.
  • Reflect on responsible AI use through group exercises and develop plans for integrating AI into your research practice.
  • Learn when repeated prompting can become a documented workflow or a simple agent.


âť“ Questions

  1. How can you leverage Copilot to explore policy‑related questions effectively?
  2. How do you verify and critique AI outputs against official sources?
  3. How does the tone and content of Copilot’s response change when you specify different roles?
  4. How will you integrate AI responsibly into your research workflow after this workshop?

Structure & Agenda

  1. Orientation and Prompting – short introduction, a reusable prompt pattern and a prompt practice task.
  2. Group Research with Copilot – small‑group policy exploration and plenary.
  3. Critique and Validation – source-grounded checking, group critique and plenary.
  4. Prompt Framing and Role Effects – role‑prompting experiment and plenary.
  5. Using Agents for Repeated Tasks – workflow framing, agent design and plenary.
  6. Going Further with AI – using AI to help design an agent.

đź”§ The emphasis is still on practical exercises; short teaching segments provide prompt structure, validation habits and workflow ideas to support each activity.

Orientation and Prompting

Introduction

This workshop builds on the principles introduced in the last session. Responsible AI use requires transparency, verification, documentation, equity and continuous monitoring. Participants will practise constructing clear and specific prompts and learn to record prompts and outputs for transparency. Remember not to include confidential or proprietary information in prompts.

A reusable prompt pattern

A good policy when writing research prompt is to usually include:

  • the task: what you want Copilot to do
  • the context: who the answer is for and why
  • the source boundary: what document, website or evidence it should rely on
  • the output format: bullets, table, checklist or short briefing

âť“ A useful extra instruction is to flag uncertainty rather than guessing.

Group Formation

Participants are welcome to form up into groups or work alone

Task 1: Prompt practice

https://m365.cloud.microsoft/chat/

Formulate a simple prompt such as:

📝 “Summarise the University of Nottingham guidance on generative AI for researchers in six bullet points. Use only the linked guidance page. Flag any area that remains ambiguous or context-dependent.”

  • Which part of your prompt defined the task, source boundary and output format?
  • Did Copilot cite sources or mention limitations?
  • What uncertainties remain about the policy?

Building Research with Copilot

Policy Domains

Each group or individual will investigate one AI policy area:

  1. Publisher – Nature / Springer Nature
  2. Academic Society – PSR / SAGE Publications
  3. Research Funder – Cancer Research UK
  4. National Guidelines – UKRI
  5. Professional Body – COPE

👤 We will examine different policy domains to compare how organisations are responding to AI.

Task 2: Research with Copilot

Using Copilot, collect information about your assigned policy area. Document the outputs in a shared sheet or board.

Groups should prepare to:

  • share how they formulated their prompts used
  • identify gaps, hallucinations or incorrect claims
  • evaluate cited sources (if any)

Groups present initial findings and discuss prompt strategies and observed weaknesses in AI responses.

Critique and Validation

Source first, summary second

A safer workflow is:

  1. gather the official policy page or PDF
  2. ask Copilot to summarise or compare only that material
  3. ask it to flag uncertainty, assumptions or missing context
  4. verify the answer against the source yourself
  5. keep only the claims you can defend

🔎 Use AI after you have the source in hand: let it help interpret official documents, not replace reading, checking, and defending the evidence yourself.

Comparing AI Outputs to Official Documents

These are the official policy documents:

⚖️ Basically dont trust the AI summary at face value, test it against the official policy text and see where it is accurate, incomplete, or misleading.

A practical validation checklist

Output What to check
summary or explanation source accuracy, date, missing caveats
quoted rule or citation exact wording, existence, relevance
interpretation scope, ambiguity, whether it follows from the source
agent output consistency, gaps and whether checks were actually applied

âś… Validation should be built into the workflow, not bolted on afterwards.

A worked prompt chain often help

flowchart LR
    A["Define the task"] --> B["Get a first structured answer"]
    B --> C["Interrogate assumptions or gaps"]
    C --> D["Refine format for use"]
    D --> E["Verify against source or data"]

đź§­ A staged prompt chain is less about cleverness than control: it helps separate drafting, critique, refinement, and verification so the final output is more reliable and easier to defend.

Task 3: Group critique

Use Copilot to compair its initial claims with the actual policy text.

For at least two AI-generated statements, verify:

  • whether they appear in the official document
  • whether Copilot added, omitted or distorted information

🔍 Hint Note possible reasons for inaccuracies: outdated training data, vague prompts, or ambiguous policy language.

  • How accurate were Copilot’s outputs overall?
  • What patterns of error or bias emerged?
  • What verification processes should researchers apply?

Prompt Framing and Role Effects

Role Prompting

The tone and content of Copilot’s output can vary depending on the role specified. Asking Copilot to answer “as a researcher”, “as a policy adviser” or “as a science journalist” can change the depth, formality and focus of the response. Prompt engineering strategies emphasise that the quality of outputs depends on the clarity and context provided.

  • useful for adapting tone to a research, policy or public-facing audience
  • risky when style is mistaken for authority
  • best used after the evidence boundary is clear

🎭 Role prompts can be useful for shaping tone and audience, but it should still not be confused with expertise, evidence, or authority.

Task 4: Role effects experiment

Re-ask the same policy question used in Task 1 with three different role instructions.

  1. policy adviser
  2. researcher
  3. journalist

đź§ľ Record differences in tone, structure and certainty.

  • Compare outputs across groups
  • Discuss how role prompting affected perceived authority
  • Reflect on risks of mistaking stylistic confidence for accuracy

Using Agents for Repeatable Tasks

When a repeated task should become a workflow

Good candidates for a reusable workflow have:

  • a stable input pattern
  • a clear output structure
  • public or approved source material
  • an obvious validation step

đź§­ If you are doing the same low-risk task repeatedly, ad hoc prompting becomes inefficient.

Automating Policy Review with AI Agents

AI agents can automate a repeatable, multi-step workflow for policy review.
Instead of prompting for each step separately, the agent can be configured to:

  • retrieve the relevant policy source
  • extract and summarise key points
  • compare positions across organisations
  • flag uncertainty or missing information

âś… Once this workflow is defined, you only need to provide the topic or organisation and the agent should cary out the process consistently each time.

Automate the structure, not the responsibility

Good agent candidates Poor agent candidates
policy lookup from public websites final interpretation of research findings
turning repeated notes into a consistent template authorship or disclosure decisions
monitoring public guidance updates any task based on unclear or sensitive evidence

🤖 Agents work best when the task is narrow and easy to check.

A lightweight AI use note

A short record after each substantial use can include:

  • date and tool
  • task or agent purpose
  • prompt or template used
  • input material
  • checks performed
  • what you kept

If you are using an agent you can ask it to include this information!

Task 5: Build an agent to automate your AI-policy review

Create an AI agent configured to:

  • Accept a single input: the name of the organisation (e.g. Nature, UKRI, COPE)
  • Retrieve and summarise the organisation’s generative-AI policy
  • Identify disclosure and authorship rules
  • Flag uncertainties or possible hallucinations
  • Present information in a consistent structure
  • Include a short validation note explaining how the answer should be checked
  • Also report: date and tool, task or agent purpose, prompt or template used, input material, checks performed, what you kept

Test it on a few organisations.
Note where it succeeds, where it struggles and where human oversight is required. Keep a brief note of the agent instructions and the checks you applied.

  • Discuss benefits of automation vs. risks (e.g. propagating errors)
  • Reflect on where agents can fit into research workflows responsibly

Further Information

📚 Key points

  • Major publishers (e.g. Nature) prohibit AI authorship but allow limited use of AI tools with disclosure.
  • National and international bodies (UKRI, European Commission) provide explicit guidelines on responsible use of generative AI.
  • Universities issue their own integrity guidance — always check local policy.
  • Prompt engineering skills help improve clarity, define source boundaries and reduce ambiguity in AI responses.
  • Ethical use requires validation, transparency, documentation and acknowledgement of limitations.
  • Repeatable workflows and agents are most useful when the task is narrow, low-risk and easy to check.

Hints