Now updated to be able to pull every comment in every page of a domain (if you input it correctly!)
A scraper for gathering data from Facebook's embedded comment widgets on any number of URLs! It bypasses the Facebook graph API (you don't need an access token) so there's little risk of throttling.
This script extracts JSON data that Facebook uses to power its embedded comment platform, recursively picking out comments. (TO A MAXIMUM OF 100 TOP LEVEL COMMENTS PER LINK!)
It can display nested replies, all in order too, so the .csv reads naturally like the Facebook comment feed!
It writes to a specified .csv file the parsed data with columns:
Reply Depth | Link | Description | Total Comments | Username | City | Date-Time | Comment | Likes | Replies
Just change the relevant parameters!
# =============================================================================
# SET PARAMETERS
# =============================================================================
if __name__ == "__main__":
"""
NOTE: Make sure there's no http:// or https:// in the source_URL list
If you want to crawl something like
"troublesome_site.com/page=<??>/other_stuff"
Write it as a tuple! ("troublesome_site.com/page=","/other_stuff)
"""
# [[INSERT YOUR DOMAINS HERE!]]
source_URLs = ["somesite.com/page/", "methylDragon.com"]
# If you want to pull all pages from a site, ensure the site's page
# structure is numeric (Eg. somesite.com/page/<NUMBER> )
start_page = 0
I noticed that you can access the source of the iframes that Facebook dynamically generates via embedded JS in a webpage. Said source has JSON data hidden in a mess of text that can be parsed accordingly!
I wrote this as part of a bigger project where we wanted to do natural language processing on a dataset of comments. This can help build that dataset, 100k+ comments strong!