![]() If you're curious about the current state of this script, you can see it at the niffler repo. ![]() But after adding only a 2 second pause between calls, which should still exceed the published limits, the dashboard doesn't show any calls, so seems Facebook's not too strict before throttling.īefore I invest too much in making an efficient script and getting #allthedata, I want to check out a sample in Pandas and make sure I'm getting things in the right format. I can also view the API calls in the Facebook App Dashboard, shown here before I added pausing between calls.Īfter the last calls shown I got an error when trying to grab more data, because it had already past 200. As you can see, just a week's posts totals 26- to retrieve comments for each post, I'll have to make an API call, and for any replies to each comment- assuming each have just 10 comments and replies, that's more than 200 calls.Ī proper way to manage this is by accessing the X-App-Usage headers, to always check the percentage of calls left, but standard requests functions to get header data are limited with the facebook-sdk and I'm anxious to explore the data, so I just used the old time.sleep() function in between API calls. The task will change to Running status, so wait for the scraper to finish. The next problem is preventing errors from the 200 calls an hour limit facebook places. Once youre ready, click the Start button to begin the data extraction from Facebook. I also tried using logger, and actually catching exceptions, but I don't think I quite have the hang of it. Config file always sounds so intimidating to me, like if you touch the file wrong a space shuttle will blow up, but in this case it's just a couple variables. 1K views, 16 likes, 2 loves, 66 comments, 2 shares, Facebook Watch Videos from Mega Soft: Item code 39 Facebook Data Extractor Capture Thousand Real Email & Phone Number from Facebook in few. One screen dash board will give ease in control and operations. One grown-up programmer thing I did in my code was use a config file to store the group_id and access_token, just by importing and doing config.access_token, and then adding the config file to my gitignore file so that it's ignored. Facebook is a library for scraping Facebook data, including profile detail, posts, story, search, and many more. ![]() Note: it's important to have a good json viewer. facebook csv scraping growth growth-hacking facebook-scraper facebook-data-extract facebook-scraping facebook-data-scraper Updated TypeScript. Then all that's left is exporting the dictionary to json and making sure I sorted the replies/comments/posts properly, with a json viewer. Next, I added a function to grab comments and replies for each post, by searching for comments on the post id. And voila, you get the posts in a dictionary format!Įvery post is stored in 'data', as a message with an id. Facebook has not notified the more-than 530m users whose details were exposed on a hacker forum in 2019 and has no plans to do so, according to company representatives.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |