{"id":14979678,"url":"https://github.com/smahesh29/web-scraping-python","last_synced_at":"2025-10-28T19:31:55.921Z","repository":{"id":106546214,"uuid":"226179058","full_name":"smahesh29/Web-Scraping-Python","owner":"smahesh29","description":"It contains some web scraping examples implemented using Python.","archived":false,"fork":false,"pushed_at":"2023-02-05T15:25:33.000Z","size":6237,"stargazers_count":29,"open_issues_count":2,"forks_count":11,"subscribers_count":3,"default_branch":"master","last_synced_at":"2024-09-28T03:21:24.293Z","etag":null,"topics":["beautifulsoup","beautifulsoup4","flipkart-scraper-python","flipkart-selenium","google-images-crawler","google-images-downloader","internshala","internships","pandas","pandas-dataframe","python","selenium","selenium-python","web-scapping","web-scraping","webscraping","webscraping-search","webscrapper","youtube-scraper","youtube-video"],"latest_commit_sha":null,"homepage":null,"language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/smahesh29.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2019-12-05T20:00:07.000Z","updated_at":"2024-06-19T12:36:56.000Z","dependencies_parsed_at":null,"dependency_job_id":"7bf71ea9-28f7-4594-9407-69b5b5775c3b","html_url":"https://github.com/smahesh29/Web-Scraping-Python","commit_stats":{"total_commits":12,"total_committers":1,"mean_commits":12.0,"dds":0.0,"last_synced_commit":"a918f84b58fb267f6753e8f602175347945f3d61"},"previous_names":[],"tags_count":0,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/smahesh29%2FWeb-Scraping-Python","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/smahesh29%2FWeb-Scraping-Python/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/smahesh29%2FWeb-Scraping-Python/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/smahesh29%2FWeb-Scraping-Python/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/smahesh29","download_url":"https://codeload.github.com/smahesh29/Web-Scraping-Python/tar.gz/refs/heads/master","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":219875834,"owners_count":16554708,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["beautifulsoup","beautifulsoup4","flipkart-scraper-python","flipkart-selenium","google-images-crawler","google-images-downloader","internshala","internships","pandas","pandas-dataframe","python","selenium","selenium-python","web-scapping","web-scraping","webscraping","webscraping-search","webscrapper","youtube-scraper","youtube-video"],"created_at":"2024-09-24T14:00:28.883Z","updated_at":"2025-10-28T19:31:49.841Z","avatar_url":"https://github.com/smahesh29.png","language":"Python","readme":"# Web-Scraping-Python  \u003cimg alt=\"GitHub\" src=\"https://img.shields.io/github/license/smahesh29/Web-Scraping-Python\"\u003e\nIt contains some web scraping examples implemented using Python.\n\n\u003ch1\u003e1. Internshala :\u003c/h1\u003e\n\u003cb\u003ePython codes for Scraping Internship from \u003ca href=\"https://internshala.com/internships\"\u003eInternshala\u003c/a\u003e\u003c/b\u003e\n\n\u003ch2\u003eAdditional Python Libraries Required :\u003c/h2\u003e\n\n\u003cul\u003e\n  \u003cli\u003eBeautifulSoup\u003c/li\u003e\n  \n      pip install beautifulsoup4\n\u003c/ul\u003e\n\u003cul\u003e\n  \u003cli\u003ePandas\u003c/li\u003e\n  \n      pip install pandas\n\u003c/ul\u003e\n\u003cul\u003e\n  \u003cli\u003eRequests\u003c/li\u003e\n  \n      pip install requests\n\u003c/ul\u003e\n\u003ch2\u003eUsage :\u003c/h2\u003e\n\u003cb\u003eThis directory contains 2 python files :\u003c/b\u003e\n\u003cdl\u003e\n  \u003cdt\u003e1. python_django_internship.py :\u003c/dt\u003e\n  \u003cdd\u003eThis program is used to get details of all the \u003cb\u003ePython/Django\u003c/b\u003e category internships from first page. It will store the details in \u003cb\u003epython_django_internship.xlsx\u003c/b\u003e sheet. To Run the program :\n    \u003cul\u003e\n         \u003cli\u003eOpen your Command Prompt or Terminal and change directory to the same folder where you have saved the program.\u003c/li\u003e\n         \u003cli\u003eUse Command :\u003c/li\u003e\n  \n                        python python_django_internship.py\n                  \n   \u003c/ul\u003e\u003c/dd\u003e\n  \u003cdt\u003e2. keyword_internship.py :\u003c/dt\u003e\n  \u003cdd\u003eThis program is used to get details of all the internship of specified \u003cb\u003eKeyword\u003c/b\u003e. The keyword should be specified during the execution of program. The program will ask for keyword, after running it. It will store the details in \u003cb\u003ekeyword_internship.xlsx\u003c/b\u003e sheet. To Run the program :\n    \u003cul\u003e\n         \u003cli\u003eOpen your Command Prompt or Terminal and change directory to the same folder where you have saved the program.\u003c/li\u003e\n         \u003cli\u003eUse Command :\u003c/li\u003e\n  \n                      python keyword_internship.py\n                  \n   \u003c/ul\u003e\u003c/dd\u003e\n\u003c/dl\u003e\n\n\u003ch1\u003e2. Google image Downloader :\u003c/h1\u003e\n\u003cb\u003eA Python program for downloading images from a Google Search.\u003c/b\u003e\n\n\u003ch2\u003eAdditional Python Libraries Required :\u003c/h2\u003e\n\n\u003cul\u003e\n  \u003cli\u003eBeautifulSoup\u003c/li\u003e\n  \n      pip install beautifulsoup4\n\u003c/ul\u003e\n\u003cul\u003e\n  \u003cli\u003eRequests\u003c/li\u003e\n  \n      pip install requests\n\u003c/ul\u003e\n\u003ch2\u003eUsage :\u003c/h2\u003e\n\nOpen your Command Prompt or Terminal and change directory to the same folder where you have saved the program. Type the following command in your CMD or Terminal:\n\n     python image_downloader.py -n 50 -s \"apples\" -d \"E:\\images\"\n     \nHere, change the following as per your needs:\n\n-n: Number of Images to be downloaded (Default = 10).\n\n-s: Search Query (Default = red roses).\n\n-d: Directory where to save the images(Default = Download - the directory present with code).\n\n\u003ch1\u003e3. Flipkart\u003c/h1\u003e\n\u003cb\u003eA Python program to extract the Price, Name, and Rating of Laptops from Filpkart website.\u003c/b\u003e\n\n\u003ch2\u003ePre-requisites :\u003c/h2\u003e\n\u003ch4\u003e1. Additional Python Libraries Required :\u003c/h4\u003e\n\n\u003cul\u003e\n  \u003cli\u003eSelenium\u003c/li\u003e\n  \n      pip install selenium\n\u003c/ul\u003e\n\u003cul\u003e\n  \u003cli\u003eBeautifulSoup\u003c/li\u003e\n  \n      pip install beautifulsoup4\n\u003c/ul\u003e\n\u003cul\u003e\n  \u003cli\u003ePandas\u003c/li\u003e\n  \n      pip install pandas\n\u003c/ul\u003e\n\u003ch4\u003e2. \u003ca href=\"https://chromedriver.chromium.org/downloads\"\u003echromedriver\u003c/a\u003e (For Windows operating system you can download the chromedriver from this repository)\u003c/h4\u003e\n\u003ch2\u003eUsage :\u003c/h2\u003e\n\u003cul\u003e\n         \u003cli\u003eOpen your Command Prompt or Terminal and change directory to the same folder where you have saved the program.\u003c/li\u003e\n         \u003cli\u003eUse Command :\u003c/li\u003e\n  \n                      python laptop_details.py              \n\u003c/ul\u003e\n\u003cul\u003e\n  \u003cli\u003eThe laptops details will be stored in laptop_details.csv file.\u003c/li\u003e\n\u003c/ul\u003e\n\n\u003ch1\u003e4. Youtube :\u003c/h1\u003e\n\u003cb\u003eA Python program to extract the information such as title, duration, channel, url, views, description, etc. of youtube trending videos.\u003c/b\u003e\n\n\u003ch2\u003eAdditional Python Libraries Required :\u003c/h2\u003e\n\n\u003cul\u003e\n  \u003cli\u003eBeautifulSoup\u003c/li\u003e\n  \n      pip install beautifulsoup4\n\u003c/ul\u003e\n\u003cul\u003e\n  \u003cli\u003ePandas\u003c/li\u003e\n  \n      pip install pandas\n\u003c/ul\u003e\n\u003cul\u003e\n  \u003cli\u003eRequests\u003c/li\u003e\n  \n      pip install requests\n\u003c/ul\u003e\n\u003ch2\u003eUsage :\u003c/h2\u003e\n\u003cul\u003e\n         \u003cli\u003eOpen your Command Prompt or Terminal and change directory to the same folder where you have saved the program.\u003c/li\u003e\n         \u003cli\u003eUse Command :\u003c/li\u003e\n  \n                      python youtube_trending.py             \n\u003c/ul\u003e\n\u003cul\u003e\n  \u003cli\u003eThe videos details will be stored in youtube_trending.xlsx file.\u003c/li\u003e\n\u003c/ul\u003e\n","funding_links":[],"categories":[],"sub_categories":[],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fsmahesh29%2Fweb-scraping-python","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fsmahesh29%2Fweb-scraping-python","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fsmahesh29%2Fweb-scraping-python/lists"}