question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Documentation on how to add previous labeling decisions to your project

See original GitHub issue

Feature Request

I request more documentation on how to include previous decisions to your ASReview project. So actually, the feature is already there but lacks documentation. This has been requested (and used) several times now, most often when a user wants to:

  • Switch from screening in another tool to screening with ASReview
  • Updating an existing systematic review with more recent publications

The idea is that you want to include decisions you’ve already made prior to setting up your project.

Currently, this can be done by merging your dataset with labeled and unlabeled publications via Excel for example. Your dataset should contain a column, called label_included which is filled with 1’s or 0’s for the publications that you have already screened, and is empty for the publications that you still need to screen using ASReview.

Describe the solution you’d like We can create an extra section called Add previous labeling decisions to your project under the Datasets page on RTD.

The section should consist of two parts. The first part explains that (and how) ASReview can use previous labeling decisions. The second part explains how one should construct such a dataset. I think a screenshot or two that demonstrate how this should look in Excel would be helpful.

We can use the information that is currently available:

  • In the FAQ, here and here:
    • I already started labeling before I came across ASReview. How can I keep my former screening decisions when starting a new project in ASReview? You can keep your former labeling decisions by adding an extra column in your dataset called ‘included’ or ‘label_included’. In this column, you can indicate previous screening decisions on records with 0s (irrelevant) and 1s (relevant). ASReview will use this information to train the model.
    • Related question: How can I add more publications while I are already started screening in ASReview? You can export the results of ASReview you have so far and add your new publications to this file. Make sure that for your new publications the column included is empty and rename or delete the column of record_id. This latter will save you from running into errors. Then simply import this updated file to ASReview and you can continue the screening process.
  • In issue #360. Note that this issue is still open and that this piece of documentation should elaborate on how to avoid this issue!

Describe alternatives you’ve considered Another option is to write a short blogpost on this. Then we are able to give a bit more detail on possible use cases and implementations. But for now I think a short and technical explanation is enough.

Ideally, this feature is implemented in the frontend in which you can simply upload a labeled and an unlabeled dataset, which are then automatically merged by ASReview.

Issue Analytics

  • State:closed
  • Created 3 years ago
  • Comments:5 (5 by maintainers)

github_iconTop GitHub Comments

1reaction
J535D165commented, Nov 14, 2020

I’m fine with closing this issue. I will close it via the upcoming PR on the documentation.

1reaction
Rensvandeschootcommented, Nov 14, 2020

A description of this feature has been added to the new documentation which will be released with the new version. Shall we close this issue and open a new issue for the option to upload multiple datasets (because this would be a very nice feature to have!)

Read more comments on GitHub >

github_iconTop Results From Across the Web

Creating and managing labels - Google Cloud
Create labels · Open the Labels page in the Google Cloud console. Open the Labels page · Select your project from the Select...
Read more >
Label Studio Documentation — Set up your labeling project
Within a project on the Label Studio UI, click Settings. Click Instructions, or in Label Studio Enterprise, click Annotation Settings.
Read more >
Labels - GitLab Docs
On the left sidebar, select Project information > Labels. Select New label. In the Title field, enter a short, descriptive name for the...
Read more >
Verify and Adjust Labels - Amazon SageMaker
Label verification — Workers indicate if the existing labels are correct, or rate their quality, and can add comments to explain their reasoning....
Read more >
Labeling images and text documents - Azure Machine Learning
In the table of data labeling projects, select the Label data link for your project. You see instructions that are specific to your...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found