Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/kumarde/llm-content-mod
Datasets and code for studying LLMs and Content Moderation
https://github.com/kumarde/llm-content-mod
Last synced: 3 days ago
JSON representation
Datasets and code for studying LLMs and Content Moderation
- Host: GitHub
- URL: https://github.com/kumarde/llm-content-mod
- Owner: kumarde
- Created: 2024-01-10T15:48:52.000Z (12 months ago)
- Default Branch: main
- Last Pushed: 2024-01-10T17:04:04.000Z (12 months ago)
- Last Synced: 2024-04-09T23:06:30.544Z (9 months ago)
- Language: Python
- Size: 9.92 MB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
- awesome_ai_agents - Llm-Content-Mod - Datasets and code for studying LLMs and Content Moderation (Building / Datasets)
README
# llm-content-mod
Datasets and code for studying LLMs and Content Moderation# Datasets
- The data/ folder contains all the subreddit specific data and toxic content data we curated to conduct our experiments
- rule_moderation/ contains balanced datasets for each subreddit we studied. In particular:
- subreddit_rules_w_description.jsonl contains a description and the rule-string we used for prompting each LLM
- subreddit_balanced_datasets contains the actual balanced data files for each subreddit
- toxic_content/ contains 10k_balanced_sample.jsonl, which is the data file with balanced toxic/nontoxic comments curated for our toxic content evaluation# Scripts
- The scripts/ folder has all the base scripts we used to run our data through various commercial models.
- scripts/is_toxic_*.py has all the scripts for toxic content detection
- scripts/rule_based_moderation.py has the code for evaluating rule-based decisions per subreddit