sqlcoder/README.md

64 lines
4.3 KiB
Markdown
Raw Normal View History

2023-08-18 06:10:53 +08:00
# Defog SQLCoder
2023-10-05 00:46:19 +08:00
**Updated on Oct 4 to reflect benchmarks for SQLCoder2 and SQLCoder-7B**
2023-08-19 04:47:22 +08:00
Defog's SQLCoder is a state-of-the-art LLM for converting natural language questions to SQL queries.
2023-08-18 06:10:53 +08:00
2023-10-04 19:45:05 +08:00
[Interactive Demo](https://defog.ai/sqlcoder-demo/) | [🤗 HF Repo](https://huggingface.co/defog/sqlcoder2) | [♾️ Colab](https://colab.research.google.com/drive/1z4rmOEiFkxkMiecAWeTUlPl0OmKgfEu7?usp=sharing) | [🐦 Twitter](https://twitter.com/defogdata)
2023-08-18 06:10:53 +08:00
## TL;DR
2023-10-04 19:45:05 +08:00
SQLCoder is a 15B parameter model that outperforms `gpt-3.5-turbo` for natural language to SQL generation tasks on our [sql-eval](https://github.com/defog-ai/sql-eval) framework, and significantly outperforms all popular open-source models. When fine-tuned on a given schema, it also outperforms `gpt-4`
2023-08-18 14:42:58 +08:00
SQLCoder is fine-tuned on a base StarCoder model.
2023-08-18 06:10:53 +08:00
2023-09-16 01:47:47 +08:00
## Results on novel datasets not seen in training
2023-08-18 06:10:53 +08:00
| model | perc_correct |
|-|-|
2023-10-04 19:45:05 +08:00
| gpt4-2023-10-04 | 82.0 |
| defog-sqlcoder2 | 77.5 |
2023-10-04 19:45:05 +08:00
| gpt4-2023-08-28 | 74.0 |
| defog-sqlcoder-7b | 71.0 |
| gpt-3.5-2023-10-04 | 66.0 |
| claude-2 | 64.5 |
| gpt-3.5-2023-08-28 | 61.0 |
| claude_instant_1 | 61.0 |
| text-davinci-003 | 52.5 |
2023-08-18 06:10:53 +08:00
2023-08-21 16:31:33 +08:00
## License
2023-08-21 16:32:13 +08:00
The code in this repo (what little there is of it) is Apache-2 licensed. The model weights have a `CC BY-SA 4.0` license. The TL;DR is that you can use and modify the model for any purpose including commercial use. However, if you modify the weights (for example, by fine-tuning), you must open-source your modified weights under the same license terms.
2023-08-21 16:31:33 +08:00
2023-08-18 06:10:53 +08:00
## Training
2023-10-04 19:45:05 +08:00
Defog was trained on more than 20,000 human-curated questions. These questions were based on 10 different schemas. None of the schemas in the training data were included in our evaluation framework.
2023-08-18 06:10:53 +08:00
2023-10-04 19:45:05 +08:00
You can read more about our [training approach](https://defog.ai/blog/open-sourcing-sqlcoder2-7b/) and [evaluation framework](https://defog.ai/blog/open-sourcing-sqleval/).
2023-08-18 06:10:53 +08:00
## Results by question category
2023-08-19 04:47:22 +08:00
We classified each generated question into one of 5 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.
2023-10-04 19:45:05 +08:00
| query_category | gpt-4 | sqlcoder2-15b | sqlcoder-7b | gpt-3.5 | claude-2 | claude-instant | gpt-3 |
|:-----------------|--------:|----------------:|--------------:|----------:|-----------:|-----------------:|--------:|
2023-10-13 01:23:14 +08:00
| date | 72 | 80 | 64 | 68 | 52 | 48 | 32 |
| group_by | 91.4 | 82.9 | 82.9 | 77.1 | 71.4 | 71.4 | 71.4 |
2023-10-04 19:45:05 +08:00
| order_by | 82.9 | 77.1 | 74.3 | 68.6 | 74.3 | 74.3 | 68.6 |
2023-10-13 01:23:14 +08:00
| ratio | 80 | 74.3 | 54.3 | 37.1 | 57.1 | 45.7 | 25.7 |
| join | 82.9 | 74.3 | 74.3 | 71.4 | 65.7 | 62.9 | 57.1 |
2023-10-04 19:45:05 +08:00
| where | 80 | 77.1 | 74.3 | 74.3 | 62.9 | 60 | 54.3 |
2023-08-18 06:10:53 +08:00
## Using SQLCoder
You can use SQLCoder via the `transformers` library by downloading our model weights from the Hugging Face repo. We have added sample code for [inference](./inference.py) on a [sample database schema](./metadata.sql).
```bash
python inference.py -q "Question about the sample database goes here"
# Sample question:
# Do we get more revenue from customers in New York compared to customers in San Francisco? Give me the total revenue for each city, and the difference between the two.
```
2023-08-21 00:08:26 +08:00
You can also use a demo on our website [here](https://defog.ai/sqlcoder-demo), or run SQLCoder in Colab [here](https://colab.research.google.com/drive/13BIKsqHnPOBcQ-ba2p77L5saiepTIwu0#scrollTo=ZpbVgVHMkJvC)
2023-08-18 06:10:53 +08:00
2023-08-18 19:03:43 +08:00
## Hardware Requirements
2023-08-20 14:45:05 +08:00
SQLCoder has been tested on an A100 40GB GPU with `bfloat16` weights. You can also load an 8-bit and 4-bit quantized version of the model on consumer GPUs with 20GB or more of memory  like RTX 4090, RTX 3090, and Apple M2 Pro, M2 Max, or M2 Ultra Chips with 20GB or more of memory.
2023-08-18 19:03:43 +08:00
2023-08-18 06:10:53 +08:00
## Todo
2023-08-18 06:20:36 +08:00
- [x] Open-source the v1 model weights
2023-10-04 19:45:05 +08:00
- [x] Train the model on more data, with higher data variance
2023-08-18 06:20:36 +08:00
- [ ] Tune the model further with Reward Modelling and RLHF
- [ ] Pretrain a model from scratch that specializes in SQL analysis