模型:
reciprocate/gpt-j_rm_format-oa
GPT-J for preference modeling
Usage:
from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("reciprocate/gpt-j_rm_format-oa", revision="501f895") model = AutoModelForSequenceClassification.from_pretrained("reciprocate/gpt-j_rm_format-oa", revision="501f895") input = "<|prompter|>Are you lying right now?</s><|assistant|>I am :)</s><|endoftext|>" model(**tokenizer(input, return_tensors="pt"))[0].item()
Output:
-4.877448558807373