HH-BTRewardModel-roberta
Bradley-Terry reward model trained on Anthropic/hh-rlhf dataset, using log-sigmoid loss:
Usage
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
tokenizer = AutoTokenizer.from_pretrained("JERRYPAN617/HH-BTRewardModel-roberta")
model = AutoModelForSequenceClassification.from_pretrained("JERRYPAN617/HH-BTRewardModel-roberta")
text = "Your text here"
inputs = tokenizer(text, return_tensors="pt", truncation=True, max_length=512)
with torch.no_grad():
reward = model(**inputs).logits.item()
- Downloads last month
- 1
Model tree for JERRYPAN617/HH-BTRewardModel-roberta
Base model
FacebookAI/roberta-base