{"id":23427262,"url":"https://github.com/dahsie/spam_classification","last_synced_at":"2025-07-24T07:39:33.696Z","repository":{"id":193072944,"uuid":"688050549","full_name":"dahsie/spam_classification","owner":"dahsie","description":"Ce fut mon prémier projet NLP où j'ai réalisé la détection de spam en utilisant les algorithmes d'embedding pour encorder mes textes. J'ai utilisé Random Forest  et Milti-Layres Perceptrons pour la phase de classification. Ce qui a pemit l'obtension des précisions respective de 97% et 98%. J'ai aussi appris à documenter mes codes via sphinx","archived":false,"fork":false,"pushed_at":"2023-10-03T15:50:52.000Z","size":2000,"stargazers_count":0,"open_issues_count":0,"forks_count":0,"subscribers_count":1,"default_branch":"main","last_synced_at":"2025-04-09T12:29:10.062Z","etag":null,"topics":["doc2vec","fasttext-embeddings","gensim","glove-embeddings","python","scikit-learn","sphinx-doc","word2vec-algorithm"],"latest_commit_sha":null,"homepage":"","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":null,"status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/dahsie.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":null,"code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null}},"created_at":"2023-09-06T14:46:00.000Z","updated_at":"2023-10-03T15:52:01.000Z","dependencies_parsed_at":null,"dependency_job_id":"db501b55-3f9e-470a-a916-cafa61fc0954","html_url":"https://github.com/dahsie/spam_classification","commit_stats":{"total_commits":3,"total_committers":2,"mean_commits":1.5,"dds":"0.33333333333333337","last_synced_commit":"8d92b0b1be27d1cd42f0b35fad68ec57a86277b0"},"previous_names":["dahsie/spam_classification"],"tags_count":0,"template":false,"template_full_name":null,"purl":"pkg:github/dahsie/spam_classification","repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/dahsie%2Fspam_classification","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/dahsie%2Fspam_classification/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/dahsie%2Fspam_classification/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/dahsie%2Fspam_classification/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/dahsie","download_url":"https://codeload.github.com/dahsie/spam_classification/tar.gz/refs/heads/main","sbom_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/dahsie%2Fspam_classification/sbom","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":266808556,"owners_count":23987450,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","status":"online","status_checked_at":"2025-07-24T02:00:09.469Z","response_time":99,"last_error":null,"robots_txt_status":"success","robots_txt_updated_at":"2025-07-24T06:49:26.215Z","robots_txt_url":"https://github.com/robots.txt","online":true,"can_crawl_api":true,"host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["doc2vec","fasttext-embeddings","gensim","glove-embeddings","python","scikit-learn","sphinx-doc","word2vec-algorithm"],"created_at":"2024-12-23T06:18:16.828Z","updated_at":"2025-07-24T07:39:33.677Z","avatar_url":"https://github.com/dahsie.png","language":"Python","readme":"# spam_classification\nCe fut mon prémier projet NLP ou j'ai réalisé la détection de spam en utilisant les algorithmes d'embedding pour encorder mes textes. J'ai utilisé Random Forest  et Milti-Layres Perceptrons pour la phase de classification. Ce qui a pemit l'obtension des précisions respectives de 97% et 98%. J'ai aussi appris à documenter mes codes via sphinx\n","funding_links":[],"categories":[],"sub_categories":[],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fdahsie%2Fspam_classification","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fdahsie%2Fspam_classification","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fdahsie%2Fspam_classification/lists"}