Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/eggplants/wbsv-cli
CLI for archiving pages and its all links to Wayback Machine
https://github.com/eggplants/wbsv-cli
cli wayback-machine webpage
Last synced: 3 months ago
JSON representation
CLI for archiving pages and its all links to Wayback Machine
- Host: GitHub
- URL: https://github.com/eggplants/wbsv-cli
- Owner: eggplants
- License: mit
- Archived: true
- Created: 2020-01-23T20:23:00.000Z (almost 5 years ago)
- Default Branch: master
- Last Pushed: 2022-03-10T16:27:03.000Z (over 2 years ago)
- Last Synced: 2024-07-23T11:31:05.621Z (4 months ago)
- Topics: cli, wayback-machine, webpage
- Language: Python
- Homepage: https://pypi.org/project/wbsv
- Size: 1.95 MB
- Stars: 14
- Watchers: 2
- Forks: 5
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# `wbsv`
[![PyPI version](https://badge.fury.io/py/wbsv.svg)](https://badge.fury.io/py/wbsv) [![Maintainability](https://api.codeclimate.com/v1/badges/ce84fc17ef2b182eda26/maintainability)](https://codeclimate.com/github/eggplants/wbsv-cli/maintainability)
## `wbsv`(stands for "WayBack machine SavepageNow") is…
CLI tool for saving webpage on Wayback Machine forever.
Enables you to **save all URIs** in a webpage forever on [Wayback Machine](https://archive.org/web/).## Install
```bash
pip install wbsv
```## DEMO
![demo.gif](https://raw.githubusercontent.com/wiki/eggplants/wbsv-cli/demo.gif)
## Run & Examples
### Help
```shellsession
$ wbsv -h
usage: wbsv [-h] [-r times] [-t] [-l level] [-O] [-V] [url ...]CLI tool for save webpage on Wayback Machine forever.
Save webpage and one 's all URI(s) on Wayback Machine.positional arguments:
url Saving pages in order. (default: None)optional arguments:
-h, --help show this help message and exit
-r times, --retry times Set a retry limit on failed save. (default: 3)
-t, --only_target Save just same domain of target (default: False)
-l level, --level level Set maximum recursion depth (default: 1)
-O, --own Only URLs with the same domain as target (default: False)
-V, --version show program's version number and exitIf you don't give the URL,
interactive mode will be launched.
(To quit interactive mode,
type "end", "exit", "exit()",
"break", "bye", ":q" or "finish".)
```### Interactive mode
```shellsession
$ wbsv
[[Input a target url (ex: https://google.com)]]
>>> https://www.u.tsukuba.ac.jp
[+]Target: ['https://www.u.tsukuba.ac.jp']
[+]61 URI(s) found.
[01/60]: https://web.archive.org/web/20200412020015/https://www.u.tsukuba.ac.jp/password/
[02/60]: https://www.u.tsukuba.ac.jp/info_lit/tebiki.html
[03/60]: https://web.archive.org/web/20200412020026/https://www.u.tsukuba.ac.jp/account/
...
[58/60]: https://web.archive.org/web/20200412022608/https://www.u.tsukuba.ac.jp/phishing/
[59/60]: https://www.u.tsukuba.ac.jp/wordpress/wp-content/uploads/note_usingcomputerrooms.png
[60/60]: https://web.archive.org/web/20200412022640/https://www.u.tsukuba.ac.jp/
[+]FIN!: ['https://www.u.tsukuba.ac.jp']
[+]ALL: 60, SAVE: 57, PAST: 0, FAIL: 3
>>>
```### From stdin
```shellsession
$ wbsv https://tsumanne.net
[+]Target: ['https://tsumanne.net']
[+]4 URI(s) found.
[1/4]: https://web.archive.org/web/20200412022931/https://tsumanne.net/si/
[2/4]: https://web.archive.org/web/20200412022935/https://tsumanne.net/
[3/4]: https://web.archive.org/web/20200412022938/https://tsumanne.net/my/
[4/4]: https://web.archive.org/web/20200412022949/https://tsumanne.net/ct/
[+]FIN!: ['https://tsumanne.net']
[+]ALL: 4, SAVE: 4, PAST: 0, FAIL: 0
$
```### Search links recurcively
```bash
wbsv https://programming-place.net/ppp/contents/c/index.html -l 2
```### Increase limit of retry
```bash
wbsv https://tsumanne.net -r 10
```## LISENCE
MIT
## Author
eggplants (haruna)