## [t3_1328pmj](https://www.reddit.com/r/DataHoarder/comments/1328pmj/improved_version_of_redditscrape_for_backing_up/) Thanks to the help of a fellow anonymous Redditor I've released a new version of [RedditScrape](https://github.com/NSFWUTILS/RedditScrape). This new version now uses the push shift API to gather gigantic levels of data for you to download. This means we no longer need to provide any form of reddit credentials. While the previous version was hard capped at 1,000 posts using the Reddit API, this new version has no limits at all, other than what resources and disk space you have. For example, if you're brave enough to try and scrape something like *gonewild*, you'll find it takes DAYS just to get all of the data back from push shift. The JSON text alone is over 9 gigs (3.3 million posts) and climbing. Running this is now a two step process, but results in a substantially larger set of media from your favorite subs. Instructions can be found [here](https://github.com/NSFWUTILS/RedditScrape#setup-the-script). I hope I've fixed a few of the problems that people had with the first iteration along the way. --- submitted to [r/It's A Digital Disease!](https://www.reddit.com/r/It's A Digital Disease!) by [u/nsfwutils](https://www.reddit.com/user/nsfwutils)