Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/jeddyhhh/totally-real-news-bot
Takes article headlines from the NYT and creates AI audio/video based on that headline.
https://github.com/jeddyhhh/totally-real-news-bot
ai alltalk artificial-intelligence facebook fake llamacpp llm news sadtalker stable-diffusion text-generation-webui
Last synced: about 15 hours ago
JSON representation
Takes article headlines from the NYT and creates AI audio/video based on that headline.
- Host: GitHub
- URL: https://github.com/jeddyhhh/totally-real-news-bot
- Owner: jeddyhhh
- Created: 2024-07-06T08:35:59.000Z (6 months ago)
- Default Branch: main
- Last Pushed: 2024-07-18T04:17:42.000Z (5 months ago)
- Last Synced: 2024-12-03T17:08:59.702Z (19 days ago)
- Topics: ai, alltalk, artificial-intelligence, facebook, fake, llamacpp, llm, news, sadtalker, stable-diffusion, text-generation-webui
- Language: Python
- Homepage:
- Size: 647 KB
- Stars: 0
- Watchers: 2
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# totally-real-news-bot
Update 6/7/24 - I need to rewrite this, check code annotations for more info.
Generates AI text/images/video based on a NYT news headline, can be modified to create other content.Update - 16/6/24 - Project uses a standalone version of AllTalk TTS v2 BETA, it works pretty much the same as the TWGUI extension but has its own install location, read more here:
https://github.com/erew123/alltalk_tts/tree/alltalkbetaHere's some examples of output: https://www.facebook.com/profile.php?id=61560732713412&sk=videos
Requires:
- Python 3
- A working install of text-generation-webui with the --api flag enabled
(https://github.com/oobabooga/text-generation-webui)
- A working AllTalk TTS v2 BETA installation with api access enabled.
(I'm using the standalone version but the TWGUI extension should work)
(https://github.com/erew123/alltalk_tts/tree/alltalkbeta)
- A working install of stable-diffusion-webui with --api enabled in COMMANDLINE_ARGS
(https://github.com/AUTOMATIC1111/stable-diffusion-webui)
- A New York Times API access key
(https://developer.nytimes.com/get-started)
- (optional) A working install of my fork of sadtalker-api (https://github.com/jeddyhhh/sadtalker-api)
- (optional) A Facebook page access tokenUsage:
1. Clone this repository
2. Run `pip install -r requirements.txt`
3. In run_bot.py, under "API Details", double check that your text-generation-webui, stable-diffusion-webui and alltalk_tts API paths are correct as well as your NYT API key is set
4. In run_bot.py, confirm bot settings under "Config", use True and False to set options.
5. Make sure both text-generation-webui, alltalk_tts v2 BETA and stable-diffusion-webui are running with api enabled
6. Go to alltalk's settings page and enable RVC models under "Global Settings", then tick "Enable RVC" and hit "Update RVC Settings", it will download base models and create a "rvc_voices" folder in the "models" folder.
7. If using random RVC selection mode, edit "rvc_voices.txt" in the root to include your RVC model paths, the ones in there at the moment are examples of how it should be formatted.
RVC models go in alltalk_tts/models/rvc_voices/*folder*/*model*.pth
8. Run `python run_bot.py` in consoleBasic process overview:
1. Bot grabs a NYT headline and short description
2. Bot uses text-generation-webui to analyse the headline for tone and stores it.
3. Bot trys to make up hashtags related to the headline and article summary
4. Bot generates a news article text based on the headline and short description, asks bot to write in a randomly selected tone from emotions.txt and from a random perpective from descriptive.txt
5. Bot uses Alltalk_tts to generate speech based on the generated article text, outputs a .wav file
6. Bot starts generating 4 images based on the headline and the tone stored from step 2, it will start generating using what ever model is loaded into stable-diffusion-webui, outputs .png files
- If enabled, bot will overlay a watermark to the images, logo_overlay.png in the root is the watermark file, it can be changed to whatever you want.
7. Bot combines the images and speech into a .mp4 file
- If enabled, bot will combine videos together to form a longer video with multiple articles in it, transClip.mp4 in the root is the video that goes in between your generated videos, you can change this to whatever you want
- If enabled, bot will post the .mp4 to a Facebook page.
- If enabled, bot will add the generated hashtags to the end of the video description on Facebook.
8. Bot will continue to generate until the python script is stopped.Models I'm using:
LLM - Awanllm-Llama-3-8B-Dolfin-v0.6-Abliterated-DPO.Q5_K_M.gguf
Stable Diffusion - 512-base-ema.ckpt
TTS - Piper with various RVC models.