| | --- |
| | license: afl-3.0 |
| | --- |
| | |
| |
|
| | ## Model description |
| | MathGLM-10B is finetuned from GLM-10B on a dataset with additional multi-step arithmetic operations and math problems described in text, achieves similar performance to GPT-4 on a 5,000-samples Chinese math problem |
| | test set. |
| |
|
| |
|
| | ## How to use |
| | First, you shoud run the following command to pip sat. |
| | ``` |
| | pip install SwissArmyTransformer |
| | ``` |
| | Second, you run the inference code to evaluate our MathGLM-10B. |
| | ``` |
| | bash inference.sh |
| | ``` |
| |
|
| |
|
| | ## Citation |
| | Please cite our paper if you find this code useful for your research: |
| | ``` |
| | @article{yang2023gpt, |
| | title={GPT Can Solve Mathematical Problems Without a Calculator}, |
| | author={Yang, Zhen and Ding, Ming and Lv, Qingsong and Jiang, Zhihuan and He, Zehai and Guo, Yuyi and Bai, Jinfeng and Tang, Jie}, |
| | journal={arXiv preprint arXiv:2309.03241}, |
| | year={2023} |
| | } |
| | ``` |
| |
|
| |
|