microsoft-dialogrpt-human-vs-rand
Version: 6
Demo
Please try this ➤➤➤ Colab Notebook Demo (click me!)| Context | Response | human_vs_rand score |
|---|---|---|
| I love NLP! | He is a great basketball player. | 0.027 |
| I love NLP! | Can you tell me how it works? | 0.754 |
| I love NLP! | Me too! | 0.631 |
human_vs_rand score predicts how likely the response is corresponding to the given context, rather than a random response.
DialogRPT-human-vs-rand
Dialog Ranking Pretrained Transformers
How likely a dialog response is upvoted 👍 and/or gets replied 💬?This is what DialogRPT is learned to predict.
It is a set of dialog response ranking models proposed by Microsoft Research NLP Group trained on 100 + millions of human feedback data.
It can be used to improve existing dialog generation model (e.g., DialoGPT ) by re-ranking the generated response candidates. Quick Links: We considered the following tasks and provided corresponding pretrained models.
| Task | Description | Pretrained model |
|---|---|---|
| Human feedback | given a context and its two human responses, predict... | |
updown | ... which gets more upvotes? | model card |
width | ... which gets more direct replies? | model card |
depth | ... which gets longer follow-up thread? | model card |
| Human-like (human vs fake) | given a context and one human response, distinguish it with... | |
human_vs_rand | ... a random human response | this model |
human_vs_machine | ... a machine generated response | model card |
Contact:
Please create an issue on our repoCitation:
@inproceedings{gao2020dialogrpt,
title={Dialogue Response RankingTraining with Large-Scale Human Feedback Data},
author={Xiang Gao and Yizhe Zhang and Michel Galley and Chris Brockett and Bill Dolan},
year={2020},
booktitle={EMNLP}
}
microsoft/DialogRPT-human-vs-rand powered by Text Generation Inference (TGI)
Send Request
You can use cURL or any REST Client to send a request to the AzureML endpoint with your AzureML token.curl <AZUREML_ENDPOINT_URL> \
-X POST \
-d '{"inputs":"What is Deep Learning?"}' \
-H "Authorization: Bearer <AZUREML_TOKEN>" \
-H "Content-Type: application/json"
Supported Parameters
- inputs (string): Input prompt.
- parameters (object):
- best_of (integer): Generate best_of sequences and return the one if the highest token logprobs.
- decoder_input_details (boolean): Whether to return decoder input token logprobs and ids.
- details (boolean): Whether to return generation details.
- do_sample (boolean): Activate logits sampling.
- frequency_penalty (float): The parameter for frequency penalty. 1.0 means no penalty Penalize new tokens based on their existing frequency in the text so far, decreasing the model’s likelihood to repeat the same line verbatim.
- grammar (object): One of the following
- #1 (object):
- type (enum): Possible values: json.
- value (string): A string that represents a JSON Schema. JSON Schema is a declarative language that allows to annotate JSON documents with types and descriptions.
- #2 (object):
- type (enum): Possible values: regex.
- value (string): The regular expression.
- #3 (object):
- type (enum): Possible values: json_schema.
- value (object):
- name (string): Optional name identifier for the schema
- schema (object): The actual JSON schema definition
- #1 (object):
- max_new_tokens (integer): Maximum number of tokens to generate.
- repetition_penalty (float): The parameter for repetition penalty. 1.0 means no penalty. See this paper for more details.
- return_full_text (boolean): Whether to prepend the prompt to the generated text
- seed (integer): Random sampling seed.
- stop (string[]): Stop generating tokens if a member of stop is generated.
- temperature (float): The value used to module the logits distribution.
- top_k (integer): The number of highest probability vocabulary tokens to keep for top-k-filtering.
- top_n_tokens (integer): The number of highest probability vocabulary tokens to keep for top-n-filtering.
- top_p (float): Top-p value for nucleus sampling.
- truncate (integer): Truncate inputs tokens to the given size.
- typical_p (float): Typical Decoding mass See Typical Decoding for Natural Language Generation for more information.
- watermark (boolean): Watermarking with A Watermark for Large Language Models.
- stream (boolean): Whether to stream the output tokens or not. Defaults to false.
Example payload
{
"inputs": "What is Deep Learning?",
"parameters": {
"do_sample": true,
"top_p": 0.95,
"temperature": 0.2,
"top_k": 50,
"max_new_tokens": 256,
"repetition_penalty": 1.03,
"stop": ["\nUser:", "<|endoftext|>", "</s>"]
}
}
OpenAI Chat Completion API compatibility
Additionally, Text Generation Inference (TGI) offers an OpenAI Chat Completion API compatible layer under the endpoint/v1/chat/completions,check the full specification in the OpenAI Chat Completion Create Documentation .
Model Specifications
LicenseUnknown
Last UpdatedJuly 2025
ProviderHuggingFace