WizardMath-7B-V1.0

我要开发同款
匿名用户2024年07月31日
33阅读

技术信息

开源地址
https://modelscope.cn/models/AI-ModelScope/WizardMath-7B-V1.0
授权协议
llama2

作品详情

WizardMath: Empowerig Mathematical Reasoig for Large Laguage Models via Reiforced Evol-Istruct (RLEIF)

? HF Repo •? Github Repo • ? Twitter • ? [WizardLM] • ? [WizardCoder] • ? [WizardMath]

? Joi our Discord

Model Checkpoit Paper HumaEval MBPP Demo Licese
WizardCoder-Pytho-34B-V1.0 ? HF Lik ? [WizardCoder] 73.2 61.2 Demo Llama2
WizardCoder-15B-V1.0 ? HF Lik ? [WizardCoder] 59.8 50.6 -- OpeRAIL-M
WizardCoder-Pytho-13B-V1.0 ? HF Lik ? [WizardCoder] 64.0 55.6 -- Llama2
WizardCoder-3B-V1.0 ? HF Lik ? [WizardCoder] 34.8 37.4 Demo OpeRAIL-M
WizardCoder-1B-V1.0 ? HF Lik ? [WizardCoder] 23.8 28.6 -- OpeRAIL-M
Model Checkpoit Paper GSM8k MATH Olie Demo Licese
WizardMath-70B-V1.0 ? HF Lik ? [WizardMath] 81.6 22.7 Demo Llama 2
WizardMath-13B-V1.0 ? HF Lik ? [WizardMath] 63.9 14.0 Demo Llama 2
WizardMath-7B-V1.0 ? HF Lik ? [WizardMath] 54.9 10.7 Demo Llama 2

Model Checkpoit Paper MT-Bech AlpacaEval GSM8k HumaEval Licese
WizardLM-70B-V1.0 ? HF Lik ?Comig Soo 7.78 92.91% 77.6% 50.6 pass@1 Llama 2 Licese
WizardLM-13B-V1.2 ? HF Lik 7.06 89.17% 55.3% 36.6 pass@1 Llama 2 Licese
WizardLM-13B-V1.1 ? HF Lik 6.76 86.32% 25.0 pass@1 No-commercial
WizardLM-30B-V1.0 ? HF Lik 7.01 37.8 pass@1 No-commercial
WizardLM-13B-V1.0 ? HF Lik 6.35 75.31% 24.0 pass@1 No-commercial
WizardLM-7B-V1.0 ? HF Lik ? [WizardLM] 19.1 pass@1 No-commercial

Github Repo: https://github.com/lpxuca/WizardLM/tree/mai/WizardMath

Twitter: https://twitter.com/WizardLM_AI/status/1689998428200112128

Discord: https://discord.gg/VZjjHtWrKs

Note for model system prompts usage:

Please use the same systems prompts strictly with us, ad we do ot guaratee the accuracy of the quatified versios.

Default versio:

"Below is a istructio that describes a task. Write a respose that appropriately completes the request.\\### Istructio:\{istructio}\\### Respose:"

CoT Versio: (❗For the simple math questios, we do NOT recommed to use the CoT prompt.)

"Below is a istructio that describes a task. Write a respose that appropriately completes the request.\\### Istructio:\{istructio}\\### Respose: Let's thik step by step."

Example code

import torch
from modelscope import AutoModelForCausalLM, AutoTokeizer


model = AutoModelForCausalLM.from_pretraied("AI-ModelScope/WizardMath-7B-V1.0", revisio='v1.0.0', device_map='auto', torch_dtype=torch.float16)
tokeizer = AutoTokeizer.from_pretraied("AI-ModelScope/WizardMath-7B-V1.0", revisio='v1.0.0')

prompt = """"Below is a istructio that describes a task. Write a respose that appropriately completes the request.\\### Istructio:\James decides to ru 3 sprits 3 times a week.  He rus 60 meters each sprit.  How may total meters does he ru a week?\\### Respose:"""
iputs = tokeizer(prompt, paddig=False, add_special_tokes=False, retur_tesors="pt")

# Geerate
geerate_ids = model.geerate(
    iputs.iput_ids.to(model.device), 
    attetio_mask=iputs['attetio_mask'].to(model.device), 
    do_sample=True,
    top_k=10,
    temperature=0.1,
    top_p=0.95,
    um_retur_sequeces=1,
    eos_toke_id=tokeizer.eos_toke_id,
    max_legth=200)
prit(tokeizer.batch_decode(geerate_ids, skip_special_tokes=True, clea_up_tokeizatio_spaces=False)[0])

To comme cocer about dataset:

Recetly, there have bee clear chages i the ope-source policy ad regulatios of our overall orgaizatio's code, data, ad models. Despite this, we have still worked hard to obtai opeig the weights of the model first, but the data ivolves stricter auditig ad is i review with our legal team . Our researchers have o authority to publicly release them without authorizatio. Thak you for your uderstadig.

功能介绍

WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct

声明:本文仅代表作者观点,不代表本站立场。如果侵犯到您的合法权益,请联系我们删除侵权资源!如果遇到资源链接失效,请您通过评论或工单的方式通知管理员。未经允许,不得转载,本站所有资源文章禁止商业使用运营!
下载安装【程序员客栈】APP
实时对接需求、及时收发消息、丰富的开放项目需求、随时随地查看项目状态

评论