Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/THUDM/P-tuning
A novel method to tune language models. Codes and datasets for paper ``GPT understands, too''.
https://github.com/THUDM/P-tuning
few-shot-learning natural-language-processing p-tuning parameter-efficient-learning pre-trained-language-models prompt-tuning
Last synced: 3 days ago
JSON representation
A novel method to tune language models. Codes and datasets for paper ``GPT understands, too''.
- Host: GitHub
- URL: https://github.com/THUDM/P-tuning
- Owner: THUDM
- License: mit
- Created: 2021-03-18T05:33:07.000Z (almost 4 years ago)
- Default Branch: main
- Last Pushed: 2022-10-06T12:36:12.000Z (about 2 years ago)
- Last Synced: 2024-12-18T12:02:34.442Z (4 days ago)
- Topics: few-shot-learning, natural-language-processing, p-tuning, parameter-efficient-learning, pre-trained-language-models, prompt-tuning
- Language: Python
- Homepage:
- Size: 5.98 MB
- Stars: 926
- Watchers: 23
- Forks: 111
- Open Issues: 16
-
Metadata Files:
- Readme: README.md
- License: LICENSE.md
Awesome Lists containing this project
README
# P-tuning
## ❗ News🌟 [2022-10-06] Thrilled to present [GLM-130B: An Open Bilingual Pre-trained Model](https://arxiv.org/abs/2210.02414). It is an open-sourced LLM outperforming GPT-3 175B over various benchmarks. Get model weights and do inference and P-Tuning with only **4 * RTX 3090 or 8 * RTX 2080 Ti** [FOR FREE](https://github.com/THUDM/GLM-130B)!
🌟 [2022-07-14] [Parameter-Efficient Prompt Tuning Makes Generalized and Calibrated Neural Text Retrievers](https://arxiv.org/pdf/2207.07087.pdf) is out! Check our [code](https://github.com/THUDM/P-tuning-v2/tree/main/PT-Retrieval).
🌟 [2021-10-15] [P-tuning v2](https://arxiv.org/abs/2110.07602) is out! Check our [Github repo](https://github.com/THUDM/P-tuning-v2).
A novel method to tune language models. Codes and datasets for paper [``GPT understands, too''](https://arxiv.org/abs/2103.10385).
[Xiao Liu*](https://scholar.google.com.hk/citations?user=VKI8EhUAAAAJ&hl=zh-CN), [Yanan Zheng*](zheng-yanan.github.io), [Zhengxiao Du](https://scholar.google.com/citations?user=A8x07E0AAAAJ&hl=en), [Ming Ding](https://scholar.google.com/citations?user=Va50YzkAAAAJ&hl=en), [Yujie Qian](https://scholar.google.com/citations?user=93a-9kkAAAAJ&hl=en), [Zhilin Yang](https://scholar.google.com.hk/citations?user=7qXxyJkAAAAJ&hl=en), [Jie Tang](http://keg.cs.tsinghua.edu.cn/jietang/)
![](img/PT.png)
You may be also interested in our another work GLM: [All NLP Tasks Are Generation Tasks: A General Pretraining Framework](https://github.com/THUDM/GLM)
## How to use our code
We have released the code and datasets for LAMA and few-shot SuperGLUE (32-dev) experiments. Please check **README.md** and **requirement.txt** in the corresponding subdirectories for details.The [LAMA](https://cloud.tsinghua.edu.cn/f/21b9dcf05cc44adfad25/?dl=1) and [FewGLUE_32dev](https://github.com/THUDM/P-tuning/tree/main/FewGLUE_32dev) datasets are available. The LAMA dataset should be placed in ./data directory, and the SuperGLUE dataset should be placed in the ./ (project root) directory.
## Citation
If you find our work useful, please cite the following paper:
```
@article{liu2021gpt,
title={GPT Understands, Too},
author={Liu, Xiao and Zheng, Yanan and Du, Zhengxiao and Ding, Ming and Qian, Yujie and Yang, Zhilin and Tang, Jie},
journal={arXiv:2103.10385},
year={2021}
}
```