https://github.com/omkarborhade98/text_summarization
Text Summarization using NLP
https://github.com/omkarborhade98/text_summarization
nlp tokenizers transformers
Last synced: 6 months ago
JSON representation
Text Summarization using NLP
- Host: GitHub
- URL: https://github.com/omkarborhade98/text_summarization
- Owner: OmkarBorhade98
- License: apache-2.0
- Created: 2024-01-01T13:00:03.000Z (almost 2 years ago)
- Default Branch: main
- Last Pushed: 2024-01-20T10:35:29.000Z (over 1 year ago)
- Last Synced: 2025-02-09T16:10:01.745Z (8 months ago)
- Topics: nlp, tokenizers, transformers
- Language: Jupyter Notebook
- Homepage:
- Size: 116 KB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# Text Summarization Project
------- Project Work in Progress -------
### Create Conda Environment
```
conda env create -f environment.yml
```### Model Information
- Dataset : [samsum](https://huggingface.co/datasets/samsum)
- Pre-Trained Tokenizer : [google/pegasus-cnn_dailymail](https://huggingface.co/google/pegasus-cnn_dailymail)
- Pre-Trained Model : [google/pegasus-cnn_dailymail](https://huggingface.co/google/pegasus-cnn_dailymail)### Citation
```
@inproceedings{gliwa-etal-2019-samsum,
title = "{SAMS}um Corpus: A Human-annotated Dialogue Dataset for Abstractive Summarization",
author = "Gliwa, Bogdan and
Mochol, Iwona and
Biesek, Maciej and
Wawer, Aleksander",
booktitle = "Proceedings of the 2nd Workshop on New Frontiers in Summarization",
month = nov,
year = "2019",
address = "Hong Kong, China",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/D19-5409",
doi = "10.18653/v1/D19-5409",
pages = "70--79"
}
``````
@misc{zhang2019pegasus,
title={PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization},
author={Jingqing Zhang and Yao Zhao and Mohammad Saleh and Peter J. Liu},
year={2019},
eprint={1912.08777},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```