An open API service indexing awesome lists of open source software.

https://github.com/methyldragon/fb_embedded_comment_scraper

A scraper for gathering data from Facebook's embedded comment widgets for all pages on any number of URLs! It bypasses the Facebook graph API (you don't need an access token) so there's little risk of throttling.
https://github.com/methyldragon/fb_embedded_comment_scraper

dataset-generation python3 scraping script website-scraper

Last synced: 11 months ago
JSON representation

A scraper for gathering data from Facebook's embedded comment widgets for all pages on any number of URLs! It bypasses the Facebook graph API (you don't need an access token) so there's little risk of throttling.

Awesome Lists containing this project

README

          

# fb_embedded_comment_scraper
Now updated to be able to pull every comment in every page of a domain (if you input it correctly!)

A scraper for gathering data from Facebook's embedded comment widgets on any number of URLs! It bypasses the Facebook graph API (you don't need an access token) so there's little risk of throttling.

This script extracts JSON data that Facebook uses to power its embedded comment
platform, recursively picking out comments. (TO A MAXIMUM OF 100 TOP LEVEL COMMENTS PER LINK!)

It can display nested replies, all in order too, so the .csv reads naturally like the Facebook comment feed!

It writes to a specified .csv file the parsed data with columns:
Reply Depth | Link | Description | Total Comments | Username | City | Date-Time | Comment | Likes | Replies

### Usage

Just change the relevant parameters!

```python
# =============================================================================
# SET PARAMETERS
# =============================================================================

if __name__ == "__main__":

"""
NOTE: Make sure there's no http:// or https:// in the source_URL list

If you want to crawl something like
"troublesome_site.com/page=?>/other_stuff"

Write it as a tuple! ("troublesome_site.com/page=","/other_stuff)
"""

# [[INSERT YOUR DOMAINS HERE!]]
source_URLs = ["somesite.com/page/", "methylDragon.com"]

# If you want to pull all pages from a site, ensure the site's page
# structure is numeric (Eg. somesite.com/page/ )

start_page = 0
```

### How it works

I noticed that you can access the source of the iframes that Facebook dynamically generates via embedded JS in a webpage. Said source has JSON data hidden in a mess of text that can be parsed accordingly!

### Applications

I wrote this as part of a bigger project where we wanted to do natural language processing on a dataset of comments. This can help build that dataset, 100k+ comments strong!