The use of large language models and generative AI in evidence synthesis and literature review is a rapidly evolving research area and comprehensive and widely accepted guidelines are still under development. Below are a few useful references to learn more about the application of LLMs in evidence synthesis, such as systematic review.
Sysrev has a built-in generative AI auto-label feature that uses the OpenAI's GPT-4o model. This can be used to automate the labeling process. Sysrev generates an auto-label report that allows users to compare auto-labeling results to human labeling, such that labels can be improved, assessed and optimized to maximize accuracy and provide a transparent assessment process.
There are a few important things to know about the auto-label feature:
There are some useful settings that you can apply to control how the auto-labeler runs, including:
These features and settings are covered in more detail below.
In general, before running the auto-labeler across the entirety of your project documents, you should test, assess and optimize your auto-labels on a small random sample of records. Here is a recommended workflow for optimizing and then using the auto-labeler for a literature or document review project.
[content coming soon]
There are a number of techniques you can use to improve the accuracy of your auto-labels. Here are a few things to try:
Answer true or false for the following questions about this article.
1. Is this a systematic review or meta-analysis?
2. Does this focus on the impacts of wildfires?
3. Does this include at least one health, environmental or economic impact of wildfires?
If all of the answers are true, include this article. If any of the answers are false, exclude the article.