Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/croqaz/clean-mark
Convert an article into clean text
https://github.com/croqaz/clean-mark
article markdown readability text
Last synced: 1 day ago
JSON representation
Convert an article into clean text
- Host: GitHub
- URL: https://github.com/croqaz/clean-mark
- Owner: croqaz
- License: mit
- Created: 2017-05-05T08:24:37.000Z (over 7 years ago)
- Default Branch: master
- Last Pushed: 2022-09-12T10:32:49.000Z (about 2 years ago)
- Last Synced: 2024-11-24T18:05:24.553Z (18 days ago)
- Topics: article, markdown, readability, text
- Language: JavaScript
- Homepage:
- Size: 1.03 MB
- Stars: 603
- Watchers: 14
- Forks: 51
- Open Issues: 3
-
Metadata Files:
- Readme: README.md
- Contributing: docs/CONTRIBUTING.md
- License: LICENSE
- Code of conduct: docs/CODE-OF-CONDUCT.md
Awesome Lists containing this project
- project-awesome - croqaz/clean-mark - Convert an article into clean text (JavaScript)
README
# ➹ Clean-mark
> Convert a blog article into a clean Markdown text file.
[![NPM Version][npm-image]][npm-url]
[![NPM Downloads][downloads-image]][downloads-url]
[![Build Status][build-image]][build-url]
[![Standard Style Guide][style-image]][style-url]## Example
For example, this article:
[![Original article](article-screen.png)](http://www.techradar.com/news/amd-vs-intel-the-16-core-showdown)
Is converted into this text file:
![Clean text](clean-screen.png)
## Usage
> $ clean-mark "http://some-website.com/fancy-article"
The article will be automatically named using the URL path name. In the case, above, the name will be `fancy-article.md`.
The file type can be specified:
> $ clean-mark "http://some-website.com/fancy-article" -t html
The available types are: HTML, TEXT and Markdown.
The output file and path can be also specified:
> $ clean-mark "http://some-website.com/fancy-article" -o /tmp/article
In that case the output will be `/tmp/article.md`. The extension is added automatically.
## Installation
Simply install with npm:
> $ npm install clean-mark --global
## Why ?
* to save interesting articles offline, in a highly readable text format
* it's easy to read on a tablet, or a Kindle (as it is, or exported to PDF)
* Markdown is easy to export into different formats
* for offline text analysis of multiple articles, using machine learning / AI## How ?
Implementation steps:
1. Downloads the content of a web page
1. Meta-scrape page details (title, author, date, etc)
1. Sanitizes the ugly HTML
1. Minifies the disinfected HTML
1. Converts the result into clean Markdown textThis project depends on the [A-Extractor project](https://github.com/croqaz/a-extractor), a database of expressions used for extracting content from blogs and articles.
## Vision
The goals of the project are are:
1. Good text extraction
1. More useless text is preferred, instead of cutting from the actual article by mistake
1. Extracting media (images, videos, audio) is not that important
1. Extraction speed is not that important## Contributing
Clean-mark was tested on all major news sites. On some websites, the text, or links are cut from the article.
In this case, you have to manually edit the resulted text,AND
please raise an [issue on A-Extractor](https://github.com/croqaz/a-extractor/issues) with the link that doesn't work and we'll add it in the database, so that next time, the text will be extracted correctly.
Also, see [how to contribute](CONTRIBUTING.md).
## Other readability tools
* http://ejucovy.github.io/readability :: Original readability ☠
* https://github.com/luin/readability :: Next readability ☠
* https://github.com/olragon/node-readability :: Node.js readability ☠
* https://github.com/bndr/node-read : Node read ☠
* https://github.com/mozilla/readability :: Mozilla's readability, up to date
* https://github.com/codelucas/newspaper :: Lucas Ou-Yang's Newspaper3k
* https://github.com/postlight/mercury-parser :: Postlight's Mercury Parser
* https://github.com/ageitgey/node-unfluff :: Automatically extract body content (and other cool stuff) from HTML document## Web archiving tools
* https://github.com/croqaz/web-snap :: create "perfect" snapshots of web pages
* https://github.com/Y2Z/monolith :: Rust CLI app
* https://github.com/go-shiori/obelisk :: Go CLI app, inspired by Monolith
* https://github.com/wabarc/cairn :: TypeScript implementation of Obelisk
* https://github.com/gildas-lormeau/SingleFile :: Web Extension for Firefox/Chrome/Edge and CLI tool to save a web page in a single HTML fileMore archiving tools:
* https://crlf.link/mem/web-archiving
* https://github.com/iipc/awesome-web-archiving-----
## License
[MIT](LICENSE) © Cristi Constantin.
[npm-image]: https://img.shields.io/npm/v/clean-mark.svg
[npm-url]: https://www.npmjs.com/package/clean-mark
[david-image]: https://david-dm.org/croqaz/clean-mark/status.svg
[david-url]: https://david-dm.org/croqaz/clean-mark
[downloads-image]: https://img.shields.io/npm/dm/clean-mark.svg
[downloads-url]: https://npmjs.org/package/clean-mark
[build-image]: https://travis-ci.org/croqaz/clean-mark.svg?branch=master
[build-url]: https://travis-ci.org/croqaz/clean-mark
[style-image]: https://img.shields.io/badge/code_style-standard-brightgreen.svg
[style-url]: https://standardjs.com