deepset-roberta-base-squad2
Version: 17
Last updated April 2025
This is the roberta-base model, fine-tuned using the SQuAD2.0 dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.

Training Details

Hyperparameters

batch_size = 96
n_epochs = 2
base_LM_model = "roberta-base"
max_seq_len = 386
learning_rate = 3e-5
lr_schedule = LinearWarmup
warmup_proportion = 0.2
doc_stride=128
max_query_length=64

Evaluation Results

Evaluated on the SQuAD 2.0 dev set with the official eval script .
"exact": 79.87029394424324,
"f1": 82.91251169582613,

"total": 11873,
"HasAns_exact": 77.93522267206478,
"HasAns_f1": 84.02838248389763,
"HasAns_total": 5928,
"NoAns_exact": 81.79983179142137,
"NoAns_f1": 81.79983179142137,
"NoAns_total": 5945

Model Evaluation samples

TaskUse caseDatasetPython sample (Notebook)CLI with YAML
Question AnsweringExtractive Q&ASquad v2 evaluate-model-question-answering.ipynb evaluate-model-question-answering.yml

Inference samples

Inference typePython sample (Notebook)
Real timesdk-example.ipynb
Real timequestion-answering-online-endpoint.ipynb

Sample inputs and outputs

Sample input

{
    "input_data": {
        "question": "What's my name?",
        "context": "My name is John and I live in Seattle"
    }
}

Sample output

[
  "John"
]
Model Specifications
LicenseCc-by-4.0
Last UpdatedApril 2025
Provider
Languages1 Language