{"id":28473791,"url":"https://github.com/halacoded/hyperparameter-tuning","last_synced_at":"2025-07-02T07:31:54.062Z","repository":{"id":297421556,"uuid":"996703693","full_name":"halacoded/Hyperparameter-Tuning","owner":"halacoded","description":"Instead of checking every combination like Grid Search or picking randomly like Random Search, Optuna uses a smart method called Bayesian optimization. This means it learns from previous tries to find better settings faster.Part of CODED Data Science Bootcamp","archived":false,"fork":false,"pushed_at":"2025-06-05T10:40:04.000Z","size":11,"stargazers_count":0,"open_issues_count":0,"forks_count":0,"subscribers_count":0,"default_branch":"main","last_synced_at":"2025-06-05T11:37:59.045Z","etag":null,"topics":["classification","coded","descision-tree","hyperparameter-tuning","kuwait","kuwait-codes","machine-learning","optun"],"latest_commit_sha":null,"homepage":"https://colab.research.google.com/drive/1GiVP77A5iIJKOpJ01u2cultWzVezcpHY?usp=sharing","language":"Jupyter Notebook","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":null,"status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/halacoded.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":null,"code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null,"zenodo":null}},"created_at":"2025-06-05T10:37:13.000Z","updated_at":"2025-06-05T10:44:03.000Z","dependencies_parsed_at":"2025-06-06T18:47:00.772Z","dependency_job_id":null,"html_url":"https://github.com/halacoded/Hyperparameter-Tuning","commit_stats":null,"previous_names":["halacoded/hyperparameter-tuning"],"tags_count":0,"template":false,"template_full_name":null,"purl":"pkg:github/halacoded/Hyperparameter-Tuning","repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/halacoded%2FHyperparameter-Tuning","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/halacoded%2FHyperparameter-Tuning/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/halacoded%2FHyperparameter-Tuning/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/halacoded%2FHyperparameter-Tuning/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/halacoded","download_url":"https://codeload.github.com/halacoded/Hyperparameter-Tuning/tar.gz/refs/heads/main","sbom_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/halacoded%2FHyperparameter-Tuning/sbom","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":263093823,"owners_count":23412882,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["classification","coded","descision-tree","hyperparameter-tuning","kuwait","kuwait-codes","machine-learning","optun"],"created_at":"2025-06-07T12:30:35.949Z","updated_at":"2025-07-02T07:31:54.047Z","avatar_url":"https://github.com/halacoded.png","language":"Jupyter Notebook","readme":"# Hyperparameter-Tuning\nInstead of checking every combination like Grid Search or picking randomly like Random Search, Optuna uses a smart method called Bayesian optimization. This means it learns from previous tries to find better settings faster.\n","funding_links":[],"categories":[],"sub_categories":[],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fhalacoded%2Fhyperparameter-tuning","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fhalacoded%2Fhyperparameter-tuning","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fhalacoded%2Fhyperparameter-tuning/lists"}