匿名用户2024年07月31日
30阅读
所属分类ai、gpt_bigcode、pytorch、code
开源地址https://modelscope.cn/models/AI-ModelScope/sqlcoder2
授权协议other

作品详情

Defog SQLCoder

Defog's SQLCoder is a state-of-the-art LLM for converting natural language questions to SQL queries.

Interactive Demo | ? HF Repo | ♾️ Colab | ? Twitter

TL;DR

SQLCoder is a 15B parameter model that outperforms gpt-3.5-turbo for natural language to SQL generation tasks on our sql-eval framework, and significantly outperforms all popular open-source models. When fine-tuned on a given schema, it also outperforms gpt-4

SQLCoder is fine-tuned on a base StarCoder model.

Results on novel datasets not seen in training

| model | perccorrect | |-|-| | gpt4-2023-10-04 | 82.0 | | defog-sqlcoder2 | 77.5 | | gpt4-2023-08-28 | 74.0 | | defog-sqlcoder-7b | 71.0 | | gpt-3.5-2023-10-04 | 66.0 | | claude-2 | 64.5 | | gpt-3.5-2023-08-28 | 61.0 | | claudeinstant_1 | 61.0 | | text-davinci-003 | 52.5 |

License

The code in this repo (what little there is of it) is Apache-2 licensed. The model weights have a CC BY-SA 4.0 license, with additional responsible use restrictions added. The TL;DR is that you can use and modify the model for any purpose – including commercial use. However, if you modify the weights (for example, by fine-tuning), you must open-source your modified weights under the same license terms.

Training

Defog was trained on more than 20,000 human-curated questions. These questions were based on 10 different schemas. None of the schemas in the training data were included in our evaluation framework.

You can read more about our training approach and evaluation framework.

Results by question category

We classified each generated question into one of 5 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.

query_category gpt-4 sqlcoder2-15b sqlcoder-7b gpt-3.5 claude-2 claude-instant gpt-3
date 72 80 64 68 52 48 32
group_by 91.4 82.9 82.9 77.1 71.4 71.4 71.4
order_by 82.9 77.1 74.3 68.6 74.3 74.3 68.6
ratio 80 74.3 54.3 37.1 57.1 45.7 25.7
join 82.9 74.3 74.3 71.4 65.7 62.9 57.1
where 80 77.1 74.3 74.3 62.9 60 54.3

Using SQLCoder

You can use SQLCoder via the transformers library by downloading our model weights from the Hugging Face repo. We have added sample code for inference on a sample database schema.

python inference.py -q "Question about the sample database goes here"

# Sample question:
# Do we get more revenue from customers in New York compared to customers in San Francisco? Give me the total revenue for each city, and the difference between the two.

You can also use a demo on our website here, or run SQLCoder in Colab here

Hardware Requirements

SQLCoder has been tested on an A100 40GB GPU with bfloat16 weights. You can also load an 8-bit and 4-bit quantized version of the model on consumer GPUs with 20GB or more of memory – like RTX 4090, RTX 3090, and Apple M2 Pro, M2 Max, or M2 Ultra Chips with 20GB or more of memory.

Todo

  • [x] Open-source the v1 model weights
  • [x] Train the model on more data, with higher data variance
  • [ ] Tune the model further with Reward Modelling and RLHF
  • [ ] Pretrain a model from scratch that specializes in SQL analysis
声明:本文仅代表作者观点,不代表本站立场。如果侵犯到您的合法权益,请联系我们删除侵权资源!如果遇到资源链接失效,请您通过评论或工单的方式通知管理员。未经允许,不得转载,本站所有资源文章禁止商业使用运营!
下载安装【程序员客栈】APP
实时对接需求、及时收发消息、丰富的开放项目需求、随时随地查看项目状态

评论