This model is a fie-tued model based o mistralai/Mistral-7B-v0.1 o the ope source dataset Ope-Orca/SlimOrca. The we alig it with DPO algorithm. For more details, you ca refer our blog: The Practice of Supervised Fie-tuig ad Direct Preferece Optimizatio o Itel Gaudi2. Neural-chat-7b-v3-1 was traied betwee September ad October, 2023. We submit our model to opellmleaderboard, ad the model performace has bee The followig hyperparameters were used durig traiig: Here is the sample code to reproduce the model: Sample Code. eural-chat-7b-v3-1 ca produce factually icorrect output, ad should ot be relied o to produce factually accurate iformatio. eural-chat-7b-v3-1 was traied o Ope-Orca/SlimOrca based o mistralai/Mistral-7B-v0.1. Because of the limitatios of the pretraied model ad the fietuig datasets, it is possible that this model could geerate lewd, biased or otherwise offesive outputs. Therefore, before deployig ay applicatios of eural-chat-7b-v3-1, developers should perform safety testig. The licese o this model does ot costitute legal advice. We are ot resposible for the actios of third parties who use this model. Please cosult a attorey before usig this model for commercial purposes. The NeuralChat team with members from Itel/DCAI/AISE/AIPT. Core team members: Kaokao Lv, Liag Lv, Chag Wag, Wexi Zhag, Xuhui Re, ad Haihao She. Detailed results ca be foud hereFie-tuig o Itel Gaudi2
Model date
Evaluatio
Model
Average ⬆️
ARC (25-s) ⬆️
HellaSwag (10-s) ⬆️
MMLU (5-s) ⬆️
TruthfulQA (MC) (0-s) ⬆️
Wiograde (5-s)
GSM8K (5-s)
DROP (3-s)
mistralai/Mistral-7B-v0.1
50.32
59.58
83.31
64.16
42.15
78.37
18.12
6.14
Itel/eural-chat-7b-v3
67.15
83.29
62.26
58.77
78.06
1.21
50.43
Itel/eural-chat-7b-v3-1
66.21
83.64
62.37
59.65
78.14
19.56
43.84
Traiig procedure
Traiig hyperparameters
Traiig sample code
Prompt Template
### System:
{system}
### User:
{usr}
### Assistat:
Iferece with trasformers
import trasformers
model_ame = 'Itel/eural-chat-7b-v3-1'
model = trasformers.AutoModelForCausalLM.from_pretraied(model_ame)
tokeizer = trasformers.AutoTokeizer.from_pretraied(model_ame)
def geerate_respose(system_iput, user_iput):
# Format the iput usig the provided template
prompt = f"### System:\{system_iput}\### User:\{user_iput}\### Assistat:\"
# Tokeize ad ecode the prompt
iputs = tokeizer.ecode(prompt, retur_tesors="pt", add_special_tokes=False)
# Geerate a respose
outputs = model.geerate(iputs, max_legth=1000, um_retur_sequeces=1)
respose = tokeizer.decode(outputs[0], skip_special_tokes=True)
# Extract oly the assistat's respose
retur respose.split("### Assistat:\")[-1]
# Example usage
system_iput = "You are a math expert assistat. Your missio is to help users uderstad ad solve various math problems. You should provide step-by-step solutios, explai reasoigs ad give the correct aswer."
user_iput = "calculate 100 + 520 + 60"
respose = geerate_respose(system_iput, user_iput)
prit(respose)
# expected respose
"""
To calculate the sum of 100, 520, ad 60, we will follow these steps:
1. Add the first two umbers: 100 + 520
2. Add the result from step 1 to the third umber: (100 + 520) + 60
Step 1: Add 100 ad 520
100 + 520 = 620
Step 2: Add the result from step 1 to the third umber (60)
(620) + 60 = 680
So, the sum of 100, 520, ad 60 is 680.
"""
Ethical Cosideratios ad Limitatios
Disclaimer
Orgaizatios developig the model
Useful liks
Ope LLM Leaderboard Evaluatio Results
Metric
Value
Avg.
59.06
ARC (25-shot)
66.21
HellaSwag (10-shot)
83.64
MMLU (5-shot)
62.37
TruthfulQA (0-shot)
59.65
Wiograde (5-shot)
78.14
GSM8K (5-shot)
19.56
DROP (3-shot)
43.84
点击空白处退出提示
评论