The Ultimate Guide to Scrapebox SEO

Social Signals

This guide is going to teach you how to become a Scrapebox master, so brace yourself. For years the SEO community has been needing one true ultimate Scrapebox tutorial, however, no SEO has been brave enough to see it all the way through. At first, I thought it would be impossible to complete. But then five weeks and 9,000 words later it was finally here, enjoy everyone.

Chapter 1: Introduction to Scrapebox

If you are experienced with Scrapebox then please feel free to skip straight to other sections, but for the complete newbies out there we will walk through everything. Ok, so you have downloaded and installed your copy of Scrapebox (this can be done locally or on VPS). It is now crucial that you purchase a set of private proxies if you are going to do some serious scraping.

Update: Scrapebox Discount Code for $40 off
If you don’t already own SB, here is the Blackhat World $40 off discount link that most people don’t know about. Simply go to and the price will automatically have the $40 off coupon code applied.

What are proxies and why do we need them?

A proxy server acts as a middle man for Scrapebox to use in grabbing data. Our primary target Google, does not like it when their engine is hit multiple times from the same IP in a short time frame, which is why we use proxies. Then the requests are divided amongst all the proxies allowing us to grab the data we’re after.

So pick yourself up a set of at least 25 private ScrapeBox proxies. Personally I use 100 but I go hard. Start with 25 and see if that works out for you. Get acquainted with the Scrapebox UI. It can be quite intimidating at first, but trust me, after some time you will become very comfortable with the interface and understand everything about it.

Scrapebox Introduction

See the field where it says “Proxies go here”? That is where you paste in your proxies.

The required formatting is – IP:Port:Username:Password

Depending on your provider you might have to rearrange your proxies so they follow this format. If your proxies don’t have passwords attached and are activated through browser login, then just enter the ip:port portion after logging in.

Then we will click manage.

testing scrapebox proxies

Now the proxy test screen will pop up and we will click “Test all proxies“.

test proxies 2

If everything is good to go, you will see nothing but green success and Y for “yes” on the Google check. This is crucial! If your proxies aren’t working, you are dead in the water. So make sure you use a reliable provider with quick proxies, otherwise this is going to be a useless endeavor. First click the filter button and then “Keep Google proxies” to remove any bad proxies.

Good proxies are everything when it comes to using ScrapeBox effectively, so invest in a set from SquidProxies if you’re serious about scraping.

Manage proxies

Now click “Save to Scrapebox” and it will send all your working proxies back to Scrapebox (if they are all working just close).

save proxies to scrapebox

Ok. So our proxies are good to go, now for our settings.

Everything is good at default for the weekend scrapers out there. If you want to turn the heat up then go to “Adjust Maximum Connections” under the Settings tab. From here you can tweak the amount of connections used when hitting Google under “Google Harvester” settings. The amount in which you can push depends on the amount of proxies you are using. I usually run 100 proxies at 10 connections, do the math. But also keep in mind the number of connections allowed depends on the type of queries you are doing. More on that in a minute.

For a massive list of footprints all using site: operator, you should turn it down. i.e. the Google index check.

And to learn more about proxies, here is a comparison of the top providers I recently ran.

adjust maximum connections

Chapter 2: Building Footprints

What is a footprint?

A footprint is anything that consistently come up on the webpages you are trying to find in the search engine index.

So if you are looking for WordPress blogs to comment on, the text “Powered by WordPress” is something very common on WordPress blogs. Why is it common? Because the text comes on the default theme.

Bingo, we’ve got ourselves a footprint. Now if you combine that with our target keyword then you can start digging up some WordPress blogs/posts in your niche. And yes, we will go way more in depth but for now understanding this simple example will be enough.

Good footprints are now your best friend as a Scrapebox user. Building them is very simple but takes some focus and attention. This is where you’re going to be better then the average Scrapebox user. If you are any type of white hat link builder then you have certainly used some sort of footprint before, you just might not have called it a “footprint”.

Have you tried searching out guest post opportunities or link resource pages before? You are using footprints.

But in this section we are building footprints and for strategic reasons. We will build sets of footprints and use them again and again for specific purposes. As a quick side note let me remind you that replication is one of the keys to success in SEO, so let’s build some badass footprints and start using them over and over again.

Fortunately I have included a massive list of footprints categorized by target platform that I’ve spent years digging up. They are enclosed below.

Once you understand the goal, building footprints is quite simple. Pull up some examples of the target site you are trying to find. Looking for link partner pages? Well bring up a handful that you can find and open them in a bunch of tabs. Compare each one and look for consistent on page elements.

See a phrase that comes up all the time? You might have yourself a footprint.

And if you haven’t yet, and you call yourself an SEO, become an expert with advanced Google search operators.

This knowledge is key to being an effective search engine scraper. So take some time, study, and become a search modifier guru. Then apply that to your footprint building and build some killer prints.

There are two main elements to hunt for when building footprints.

Either in the url structure or in content somewhere.

Here are my goto operators.




How to Test Footprints:

After you think you’ve created a footprint testing them is incredibly simple. Just go Google them!

First note how many results come up. If it’s under 1,000 your footprint sucks.

We are trying to create footprints that will dig up tons of sites based on platform so the number should be decent.

Comb through the results and see how much honey your footprint is finding for you. See a bunch ofthe site types you’re searching for? Good, bank that footprint and continue build more. Save your footprints with titles for their specific purpose, so say “Vbulletin Footprints” for finding Vbulletin forums. Now that you have some footprints ready, let’s move on to massive scrapes.

Chapter 3: Massive Scraping

Now you may or may not know what you’re looking for, so let’s get a ton of it.

If you want to scrape big, you’re going to have leave Scrapebox running for a good amount of time. Sometimes even for several days. For this purpose, some may opt for a Virtual Private Server or VPS. This way you can set and forget Scrapebox, close the VPS, and go about your business without taking up resources on your desktop computer. Also know that Scrapebox is PC only but you can run it with Parallels. If you do run SB on Parallels, be sure to increase your RAM allocation. Hit me up if you need some help getting a VPS set up.

Here’s are the different elements you need to consider with big scrapes:

  • Number of proxies
  • Speed of proxies
  • Number of connections
  • Number of queries
  • Delay between each query

With the default settings everything should be golden, so the determinant of how long your scrapes will take will be mainly on how many “keywords” you put in.

You can change the number of connections – This depends on if you are using private or public proxies, and how many working ones you have.

As I mentioned before I usually run with a set of 100 and set my Google threads to 10.

The keyword field in Scrapebox is where you paste in your keywords and merge in your footprints.

Merging is very simple. All we are doing is taking what ever is listed in scrapebox and merging it with a file that contains the list of our footprints, keywords, or stop words. So say taking keyword “powered by wordpress” and merging it with “dog training” to create.

“powered by wordpress” “dog training”

Ahh yes, this Scrapebox thing is starting to make some sense now.

Now we’re after some urls from some of our favorite search engines, which one is up to us.

check google harvest

See how only Google is checked? This means Scrapebox will only harvest urls from Google. If you want to hit the other engines just select them. Also be sure that you have Use Proxies checked.

Note: You can also add foreign language Google engines by clicking the dropdown and “add more google“.

add foreign language google

Simply add the extensions for the languages you are going for and click save.

google extensions screen

The final thing to note before starting is the Results field.

number of results

Very straight forward, this is the number of results (or urls) Scrapebox will grab from the specified search engine(s).

Depending on your goals, set this accordingly. If I am scraping for some sites to link out to in some of my link building content, I will only go 25 results deep for each keyword. But if I am trying to find every possible site out there for a certain platform I will do 1,000. And this brings us to our next problem.

What if our query yields more than 1,000 results?

This is where merging in stop words comes into play.

Manually try the query “dog training” “powered by wordpress”.

You will see there are over 500,000 results.

G footprint screenshot

Now see what happens when I add the word “there”.

G footprint screenshot2

Besides that stupid ad, the organic results are different now. By using stop words combined with our footprints we can effectively scrape deeper into Google’s index and get around that 1,000 result limit.

Don’t worry, you can download my personal list of stopwords by sharing this guide below. Keep reading!

Once you have some quality footprints and stop words ready, the rest is easy. We’re going to let Scrapebox rip and come back when complete. If you’re running on your desktop then scrape overnight to minimize downtime on your system.

After Scrapebox is complete you will see the prompt saying Scrapebox is complete.

scrapebox is complete

Now if you stop the harvester prematurely a prompt will appear showing you the queries that have been successfully run and the ones that have not.

stopping scrapebox

Non-complete queries can mean one of two things.

1. There were zero results for that query.


2. That query has not been hit yet.

If you want to complete this harvest later then be sure to export “Non-Complete Keywords”and set them aside. If you inputted a list of 10,000 queries, stopped after 2,000, then you just save the remaining 8,000 queries for later.

exporting non complete keywords

One of the keys to massive scrapes is understanding that Scrapebox only holds 1,000,000 urls in the urls field and stacks files in the “Harvester Sessions” folder.

For each scrape, the software will create a time stamped folder containing txt files with each batch of 1,000,000 urls. And this is great but if you don’t know about Duperemove then you are burnt.

Duperemove is an amazing free add-on from Scrapebox that allows you to merge list of millions of ulrs and remove dupes and dupe domains. This way we can run massive scrapes and process the resulting URLs.

We can also use Duperemove to split a massive file into smaller files so we can further process the resulting urls. We can take 100,000 urls and split them into ten files with 10,000 urls for example.

After finishing a massive scrape, open dupe remove.


Start by clicking “Select source files to merge” and navigating to your harvester folder with your batch files of 1,000,000 URLs. Also be sure to save the urls left in the Scrapebox harvester when stopped, and put this file with the rest of batch files.

Select all the files and give the output file a name, I like to call it “Bulking up”. Now click “Merge files”.

Duperemove will merge everything into one enormous txt file so you can then remove dupe urls and dupe domains.

Below the Merge lists field, select the previous file “Bulking up” and chose a file name for the new output, I like to call it “Bulking down” .

Then click Remove Dupe URLs and Remove Dupe Domains. Now you have a clean list of Urls without duplicates. Depending on what you have planned for this giant list I will use the split files tool and split the large file into smaller more manageable files.

Below I have compiled the largest footprint collection of anywhere anywhere on the web. Everything is broken out into platform type, ready for scraping domination.


And now that we have covered everything about footprint building and massive scrapes, let’s move onto keyword research.

Chapter 4: Keyword Research

Having fun yet? Now that we’ve gotten all the introduction shit, things are going to start getting good.

With keyword research Scraebox continues to be one of my “go to” tools. It has two main weapons; suggesting tons of Keyword suggestions and giving us Google exact match result numbers.

Keyword Research Weapon #1 – The Power of Suggestion

With this method we will be using Scrapebox to harvest 100s or 1000s of suggestions related to our keywords. Then we will use the Google keyword tool to get volume and move on to our research weapon #2.

First we will explore the suggestion possibilities and how the keyword scraper works.

Start by clicking the Scrape dropdown, and then Keyword Scraper.

Scrapebox keyword scraper

Now after you get the keyword scraper open, type in the keyword you would like to scrape suggestions for.

Next you can select the sources you for which the scraper will grab for suggestions.

Keyword scraper sources

Protip – Tick the YouTube box if you’re doing keyword research specifically for YouTube videos. Searches can be very different on Youtube compared to typical Google queries.

After you have finished the first run through scraping keywords, remove duplicates, and then you have two options.

You can send the results straight to Scrapebox and move on or you can transfer them to the left and scrape the resulting keywords for more suggestions. You can repeat this process over and over again until you get the desired amount of keywords. Scrape, remove dupes, transfer left, scrape again, crack beer. It’s actually quite enjoyable.

remove dupes from scraper

So now that you have keyword scraping/suggesting down we will move on to one of the simplest and most powerful free addons for Scrapebox. If you haven’t yet, click “addons” in the top nav, then “show available addons”. Now install the Google Competition Finder addon.

Keyword Research Weapon #2 – Google Exact Match Results

After you open the competition finder the first step is to import the keywords from Scrapebox. Click Load Keywords and Load from Scrapebox.

Also be sure that the Exact match box is ticked. This way Scrapebox will wrap your keywords in quotes and get the exact match results for each. You can also change the number of connections for large keyword lists but I would recommend keeping it at the default of 10. Give your proxies a chance to breath.

Google competition finder

When all the results are in, click the Export dropdown, and Export content of grid as csv.

export competition results

Now you will have a nice csv with all your keywords and the corresponding results. The next step is to open the grid with excel and sort the data from low to high. Delete the proxy used and status column, then click the Sort dropdown and “Custom Sort“.

excel custom sort screen 1

Now that the custom sort screen is open, select the column with the results and sort from smallest to largest.

excel sort screen 2

After you click OK you will have a nice sorted list of keywords with exact match results from low to high.

Depending on the yield I get, I will break the keywords down into ranges of exact match results.






From there I will paste each range into the keyword tool, gather volume, and sort again, this time from high to low on the search volume. Then you can comb through and find some easy slam dunkable keywords.

Now this is by no means a 100% indicator of Google competition but it’s a good rough estimate. And when the number is REALLY low, it becomes a more accurate indicator of an easy to dominate keyword. This method can be extremely helpful when you have a massive list of keywords and you are trying to figure out which ones to target with some supporting content, boom, go for the ones with volume that you can easily rank for. This method will unlock those.

Chapter 5: Expired Domaining

This is by far one of the most powerful grey hat SEO areas in the game. Expired domains can hold a ton of juice, you just need to know how to find them and how to properly relaunch them. Before diving into the Scrapebox methods we will go over the basics of expired domaining.

There are three areas you can focus your domaining efforts or some combination of the three; Building a blog network, creating money sites, and link laundering.

1. Building a blog network

Building a network is one of the most powerful SEO techniques in the business. Owning a private network of over 100 sites PR 1-6 is quite nice, think about it.

Private Blog Network 101

There is nothing wrong with building a private blog network. This SEO strategy is not flawed in anyway. The only flaw is from the creator.

If you leave a footprint, that allows Google to identify the network and your network becomes useless. And like many other things, after the Google propaganda disseminated throughout the community, people deemed PBNs worthless and ineffective. But when done right, links from your private network will be just as effective as naturally occurring links on authority sites.

Main Points:

*Use many diverse IPs and hosting accounts
*Use different themes, category structures, permalinks, and www. vs root
*Vary the extensions! .com, .net, .info, .org, .etc
*Use different domain registrars with some private registrations and some with old owner’s information. Godaddy, namecheap, etc. some private and some with joe schmoe.
*Build some good links to each site.

2. Creating money sites

Occasionally you will find a nice domain that is fitting for a money site. In this case, congrats, you just found yourself an SEO time machine.

I’ve gone back as much as 10 years before and gained myself 40,000 natural links!

How about building a brand new site and working with a domain like that?!

These are rare but they’re out there. Most likely you’re going to have to pay for it in a small bidding war unless you get lucky. But if you know it’s a winner, then go for it.

Always be cautious with drastically changing the old content theme of the site. If you have a money domain about dog snuggies, figure out a way to rank and monetize it while keeping the content semantically relevant to that topic. Used effectively you will easily exceed the results from the same exact efforts on a fresh domain. Also if you get an aged domain with a diverse natural link profile you will be much safer blasting some links at the site. An existing diverse link profile can effectively camouflage grey hat link building tactics.

3. Link laundering

This is by far the dirtiest method of all when it comes to expired domaining shenanigans. With this technique we will be using our friend the 301 redirect to redirect pages, subdomains, or entire sites at the site or page we are trying to rank. Effectively sending tons of link juice while also cloaking our link profile a bit.

See Bluehatseo for more info on link laundering in the traditional way, with this technique we will be link laundering through server level redirects, specifically the 301.

Step 1. Acquire expired domain

Step 2. Relaunch domain and restore everything.

Step 3. Redirect domain via 301 redirect.

Step 3. Aggressively link build to the now redirected domain.

Here is the redirect code to use in you .htaccess file to execute the redirect:

RewriteEngine on
redirectMatch 301 ^(.*)$$1
redirectMatch permanent ^(.*)$$1

After you set the redirect, start blasting some links and enjoy.

Expired Domaining 101

In this section we will step away from Scrapebox a bit and discuss SEO domaining domination. But don’t worry, we will be back to Scrapebox shortly.

Buying expired domains takes some skill but it’s not rocket science. The thing is, for every good domain there is ten shitty ones out there that we must avoid.

Here is an overview of the process:

Part 1. Finding domains

Part 2. Analyzing your finds

Part 3. Smart Bidding

Finding Killer Domains with Shit Tons of SEO Juice

Ok, so Scrapebox has the TDNAM scraper addon that we are going to discuss in a moment but it is limited to only Godaddy auctions. So while this is a free addon, you are not accessing the entire expired domain market.

In order to do that you are going to have to use some sort of domaining service. These services pull expired feeds from all different sites on the web and also offer some metrics that Scrapebox does not.

Here are my recommended domaining services that I have personally used to snag domains for over 100x the initial purchase price.

Freshdrop – This is the top dog, and the price comes with. $99 per month but this is definitely the king of expired domain buying tools. If you are trying to build a network then the subscription will only be short term until you have completed all your domain buys. Recently they have added the MajesticSEO API so you can filter results by backlinks right in Freshdrop, pretty awesome.

If you can’t afford this tool then you can still land a whale on Godaddy auctions. Open the TDNAM addon and enter a keyword for domains to lookup.

TDNAM Scraper

At default ALL extensions are selected but you can specify between, .com, .net, .org, or .info. Click start and if you don’t already begin feeling like a boss.

After the scraper is finished, click the Export dropdown and Send to Scrapebox.

Analyzing your Domains and Confirming Their Greatness

After we pull up a list of potential prospects it’s time to take things a step further and be certain we have a winner. We will be using the following tools to validate which domains are worth purchasing.

-Scrapebox (of course)
-SEOMoz Api (sorry but for this it’s worth it)

First step is to check the pagerank of each domain prospect (if you haven’t sorted from a tool above already.

Click the Check Pagrank dropdown and click Get Domain Pagerank.

Pr check screeny

Now chuck everything with no PR.

Next open the Fake Page Rank Checker addon. This will confirm that each domains has legitimate Pagerank and not a false redirect.

Open the addon and load your list from Scrapebox. click Start, filter out the trash, and grab a beer.

Open a beer and take a nice chugg, you’re about to get an edge on your competition.

You can now scan through your domains with PR and use your judgement to identify domains with potential and that you are interested in.

But let’s put this process on steroids shall we?

Now we can use one of the newest free addons, the Page Authority addon. Using the moz api to scan DA (domain authority) and PA (page authority) we can quickly identify high quality prospects.

Since we will be using this tool several times later let’s set it up.

After you open the addon, click Account Setup and paste in your access id and api key in the following format.

Access ID|Secret Key

Moz page authority addon

Now click Start and get some great insight from SEOMoz’s internal scoring system. Sure it’s not perfect but gives us a quick and dirty evaluation of the domain prospects. Just enough screening to allow us to move on to the next phase of analysis.

Now we need to research the history of the domains and their backlink profiles.

Domain History, What we want:

-The shorter the time frame the site has been down the better

-Make sure the domain has not changed hands multiple times. Look at the whois history via domaintools to verify this.

-Check to see what the site used to be. Something you can roll with?

Backlink Profile, What we want:

-Take the domains you’re interested in and start putting them one by one into backlink checking tool

We want domains juiced with good links, not some piece of shit that someone blasted 10,000 viagra links at and threw out after they were done with it. You will also be able to spot an “SEO’d” link profile, just look for an abundance of keyword rich anchors or anchors with lack of natural anchor text distribution and diversity. I avoid these at all costs. Typically SEOs have no idea what they’re doing, so 99% of expired domains that previously had a “link builder” behind them will be complete shit.

Also keep an eye out for some familiar super authority links, like .govs, .edus, and big news sites. Cnet, WSJ, NYtimes, etc. A few of these areusually an indicator of a once legit domain.

Step 3: Smart Bidding

Smart bidding is a very simple process that beginners will neglect.

The process is simple, wait until the last minute and start bidding like a beast.

When you find that money domain with links from and huff po, contain your excitement and don’t go nuts quite yet.

Depending on the domain auction you’re using, watch the auction, and also set a reminder on your calendar and cell phone.

Whatever works for you, I usually set two timers, the first one hour before the auction closes, and the second 15 minutes before the auction closes.

Use the TimeandDate calculator to find the time in which the domain is going to close. Be ready and pounce.

Also keep in mind that early bidding will alert guys like me who occasionally just sort out domains by # of bids and analyze from there.

So your preemptive $50 bid just alerted me of a quality domain you found that I should throw on my calendar. Then when the time is right I strike like a hungry pit viper out for Pagerank and domain authority.

Conclusion – chill out and bid smart.

This Guide is Originally from

Chapter 6: Link Prospecting

In this chapter we will be analyzing related SERPs to our keyword and looking for places to drop links. Say there is a forum powered by Vbulletin ranking on the 5th page for a relevant keyword. It would be easy to go and drop a link on that page right? First register for the forum, make a legit profile, go post a few times in other threads, then go drop a nice juicy link on an already indexed page.

Or if you’re feeling real ambitious, train a VA to run this entire process for you.

Because you see, this same methodology can be applied on a massive level by scanning for multiple platform types.

Using a list of the most popular community and publishing platforms, you should be able to create simple html footprints and scan all the urls to identify the potential link drop opportunities.

There are two main approaches that we can use this technique for.

1. Simply analyzing urls related to the target keyword for link dropportunities (see what I did there).

2. Performing deeper analysis on targeted scrapes

For both methods we will be using the page analyzer plugin to analyze the html code of all the pages we dig up.

Method #1 – Find Ranking Related Link Dropportunities

Start by scraping a bunch of keyword suggestions closely related to your target keyword.

Set the results to 1,000 and harvest.

Remove dupe urls and open the page scanner addon.

Once the page scanner is open you will need to create the footprints for it to scan with.

Page scanner addon

Here are some example footprints:

Platform – WordPress

Platform – Drupal

Platform – Vbulletin

Platform – General Forum
All times are GMT

Note that these footprints are different than the traditional footprints we are building when scanning for onpage text. We are taking it one step further and scanning the actual source code of the returned pages for a common html element. If you invest the time, you can build extremely accurate footprints and basically find any platform out there.

After you have inputted the footprints and run the analyzer, export your results. All of the results will be exported and named by the footprint name. So your Vbullletin link dropportunities will all be one file name Vbulletin.

Now continue your hunt and perform further link prospecting analysis on the page level.

Check PR, OBLs PA/DA, etc. When completed you will have a finely tuned list of relevant potential backlink targets to either hand over to a VA or run a posting script on.

Method #2 – General Page Scanning for Targeted Link Dropportunities

With this method I’m going to show you an actual exploit that I discovered the other day to clearly explain this technique.

We are going to be finding blogs with the Comment Luv platform and do-follow links enabled.

All you will need is a few bogey Twitter accounts to tweet the post and get a choice of the post you want to link to.

*Note – This technique requires your site having a blog feed.

comment luv logoTo start we are going to be using an onpage footprint to dig up these potential comment luv dofollow drops.

Here is the footprint I created, a common piece of text found right by the comment box, comes default on all Comment Luv installs.

“Confirm you are NOT a spammer” “(dofollow)”

And a bit of SEO irony there!

Now save that badboy to a txt file as “Comment Luv Footprint” or something dear to your heart.

Bust out the keyword scraper and start scraping a shit ton of related suggestions.

Now click the M button and merge that beast in with all your freshly scraped keywords. Click start and get ready to unleash the hogs of war.

When the results are in, remove dupes, and open up the page analyzer addon.

Now create a new footprint called “Comment Luv”

And here is the Gem of an html footprint that my buddy Robert Neu came up with.

Sorry code wrap not working, check back later.

Thanks Robert!

Now run the analyzer and you’ll have some crisp comment luv enabled dofollow blogs to go link drop your face off.

Hopefully you are starting to see the potential of the page scanner and the wheels are turning. Maybe an evil laugh also?

Chapter 7: Guest Posting

Contributed by Chris Dyson from

If you want to find link building opportunities beyond blog comments, then you can use Scrapebox for its primary function which is scraping search results on an industrial scale.

A lot of white hat SEO blogs tell you to run individual searches in Google for inurl:”write for us” + Keyword and use free tools to scrape up to 100 links at a time.


This is a sure fire way to:

a)       Get your IP blocked by Google

b)       Bore you to death

c)       Waste your time and money

d)       Did I say bore you to death?

Thankfully Scrapebox will come to the rescue here to save your sanity.

#1 – Load up your list of footprints into a custom list in Scrapebox

If you are not sure what to do here please refer back to the “massive scraping section” 

#2 – Go grab another cold beer from the refrigerator


Jacob’s office on a Monday Morning

# 3- Now we want to remove any duplicate URLs, in the Remove/Filter drop down you want to select “Remove Duplicate URL’s” and then “Remove Duplicate Domains”


# 4 – Look up the PageRank

# 5 – Export the results and hand our list over to the VA to check the website is of suitable quality.  You also want them to locate the blogs contact information such as name, email address/contact form and whether the site meets the criteria we have for the project.

If you haven’t got a web researcher then create a job listing on an outsourcing site such as oDesk to have the links checked against your requirements.

Here is a useful outsourcing guide from Matt Beswick

# 6 – Once your list is cleansed you want to upload the information in to your CRM of choice and start outreach

Common Guest Blogging Footprints
Here is a list of common guest blogging footprints to get you started for free…

  • guest blogger wanted
  • guest writer
  • guest blog post writer
  • “write for us” OR “write for me”
  • “Submit a blog post”
  • “Become a contributor”
  • “guest blogger”
  • “Add blog post”
  • “guest post”
  •  “submit * blog post”
  • “guest column”
  • “contributing author”
  • “Submit post”
  • “submit one guest post”
  • “Suggest a guest post”
  •  “Send a guest post”
  • “contributing writer”
  • “Submit blog post”
  • inurl:contributors
  • inurl:”write for us”
  • guest article OR post”
  • add blog post
  •  “submit a guest post”
  •  “Become an author”
  • submit post
  • submit your own guest post”
  •  “Contribute to our site”
  •  “Submit an article”
  • “Add a blog post”
  •  “Submit a guest post”
  •  “Guest bloggers wanted”
  • “guest column”
  • “submit your guest post”
  • “guest article”
  • inurl:”guest posts”
  • “Become * guest writer”
  • inurl:guest*blogger
  • “become a contributor”

Beyond Guest Posting
As you can imagine any search query can be added to Scrapebox to harvest URL’s for Link Prospecting for example:

  1. Sponsorships
  2. Scholarships
  3. Product Reviews
  4. Discount Programmes
  5. Resource Lists/Link Pages

It’s quite easy to load your footprints for these types of link building opportunities into Scrapebox and build some high authority links on these types of pages.

  • keyword + inurl:sponsors
  • keyword + inurl:sponsor
  • keyword + intitle:sponsors
  • keyword + intitle:donors
  • keyword + intitle:scholarships site:*.edu
  • keyword + intitle:discounts site:*.edu
  • “Submit * for review”
  • keyword + inurl:links
  • keyword + inurl:resources

If you are an experienced link builder then you can use other add-ons in the Scrapebox tool-belt to find broken links or help webmasters fix malware issues on their site.

Chapter 8: Comment Blasting

No Scrapebox guide would be complete without a legit walkthrough on comment blasting.

I know what you’re thinking, comment blasting is so 2006.

Well it is, but only on the first tier. I recommend using blog comment blasts as a third tier link more for force indexing.

Since you are dropping comments on indexed and sometimes regularly crawled pages by Google, they will crawl your comment link back to whatever tiers you have are linking to thus indexing it.

As in most cases with link blasting, it’s all in the list. So you need to be sure you have a decent auto approve list and aren’t swimming in the gutter too much.

The big determinant is # of outbound links (OBLs) and pagerank. The less OBLs and higher the PR the better. The thing to be cautious of is if you don’t deeply spin your comments they will leave an awful footprint which can easily be found with a quick Google search using a chunk of your comment output in quotes.

And you can bet your ass if I can dig it up with a few queries than those PHD having algorithm writing sons of bitches can too. So keep your game tight.

Here is what you need to run a comment blast:

*Spun Anchors
*Fake Auto Generated Emails
*List of Websites for Backlinking
*Spun Comments
*Auto Approve Site List

Spun anchors – To prepare your anchors use the scrapebox keyword suggestions. Select all sources and scrape a shit ton of keywords. The more comments you plan to blast, the more anchors you should scrape. Get at least a few hundred.

Save this file as names.txt

Optional – Mix in some generic anchors in your list. Simply paste your keyword rich anchors into excel and count them, then paste in the desired quantity of generic anchors.

Fake Emails – Under the tools tab you will see “Open Name and Email Generator“, open that little gem.

Name and Email Generator

After you get this little beauty opened up, type 100,000 in the quantity field, check “Include numbers in emails” and select Gmail under the dropdown for “Domains for emails @”

name generator

After you generate the 100,000 names, just click generate emails, save them as emails.txt and you’re good to go.

List of Websites for Backlinking – If you’ve already built links, check them with the link checker, and save those as websites.txt.

Spun Comments – Generating spun comments is actually quite simple. We will simply grab comments from relevant pages and spin them together.

In the scrapebox harvester, check the WordPress button.

Take you relevant keywords from before and surround them with intitle:”your keyword”
*Click start harvesting
*Remove duplicates urls when completed
*Click on Grab, Grab comments from harvested URL list
*Tick Skip comments with URLS
*Select to Ignore comments with less than 10 words and URLs in them
*Click Start

Now open your favorite text editor and find and replace the page breaks with a space.

For spinning we will be using theBestSpinner.

Copy and paste the exported comments into TheBestSpinner and Click Everyone’s Favorites

*Select Better from the dropdown
*Uncheck Replace Everyone’s Favorites inside spun text o Tick Keep the original word found in the article
*Uncheck Only select the #1 best synonym
*Spin levels All to All with max synonyms set to 4
*Click Replace
*Once complete, highlight all, and select the Spin Together button
*Click do not include a blank paragraph

Congratulations, you have some spamtacular comments ready, save them as comments.txt

Auto Approve Site List – Trying Googling some shit like “scrapebox auto approve list”. Have yourself a field day, gather up a ton of lists, and open Duperemove.

Place all the AA list in one folder, select them all and merge together into one monster list. Remove dupe urls and it’s time to blast away.

Blast Settings:

First you need to get your setting right. Under the Settings menu, go to “Adjust Timeout Settings”.

Move the Fast Poster time out to max, 90 seconds. This way the poster will be able to load massive pages with tons of comments and slow load times without timing out.

Time out settings

Check the “Fast Poster” box. And begin opening each of the files you created from above. Names, Emails, Target Websites, Comments, and AA list all in txt format.

Comment Blatser screenshot

Click Start Posting and open beer. Drink beer and continue reading this guide.

Chapter 9: Niche Relevant Comments

Contributed by Charles Floate

There’s a cool thing you can do with ScrapeBox to make highly approved and more specifically niche relevant comments.

Preparing Comments

Firstly, you’re going to need to make 3-5 different comments per 500 harvested URLs around the same topic.

For example if you’re link building for white hat SEO I could make a comment like:

“Content has always been king, seems the black hats are getting destroyed by the white hat profit making machines”

Then, you need to “spin” the comment, by spin I mean manually spin the comment.

An example of the above comment, manually spun would be:

“{Content|Information} {has always been|has long been|has become} {king|master}, {seems|appears} {the|all the|all of the} {black hats|black hat’s} {are getting destroyed|are getting owned|are getting own} {by the|from the} {white hat profit making machines|white hat profit makers|white hat profiteers}.”

As you can see, it’s perfectly readable in all ways and these kind of comments tend to have a pretty high approval rate.

There’s a few different styles I like to incorporate into my strategies that can boost up both the diversity and the approval rate.

Ego Approval Bait:
This is based on the ego of the writer, I’ve been trying to come up with a solution to add a name to the comment but only looks like I can do this with Xrumer, and this tutorial isn’t based on Xrumer is it ;)

Example Ego Bait Comment:
Always a pleasure to read your content, seems you really do have a talent for creating great content!
(As a split test, adding the exclamation mark increases approval rate by 6%!)

Social Approval Bait:
These tend to be based around asking about social mentions, ask the author how you can connect with them on Twitter for example.

Website Approval Bait:
This is based on the fact that you’re complementing the design (and if you’re posting only to WordPress sites, you already know the answer).

Example Website Bait Comment:
Site’s design is really nice, is it a custom theme or can I buy or download the WP theme from somewhere?


Now once you have all the comments ready, you’re going to want to search for sites related to the niche you’re building for:

niche comments - keywords

Selecting WordPress will find all the WordPress blogs out there, this is great if you just want to build niche relevant nofollow comments, selecting BlogEngine will find tons of different blog CMSs, some being dofollow.

niche comments - harvesting

Posting Comments

niche comments - start posting

Once all your comments are harvested, you are ready to post.

In the Names Area, you need to open a text document with your anchor texts, I always create a mixture of branded, generics and some LSI/Longtail keywords.
In the emails section, either put your actual email (This a lot of the time will receive an email about replies, comment approvals or declines) or just input a list of randomly generated emails so your email doesn’t get flagged for spam.
In the websites list, just input your websites you wish to build links to.
In the comments section, open the text document with all your manually spun comments.
Blogs List:
In the blogs list, add in the harvested blogs, this is pretty easy as you can just click: Lists > Transfer URL’s to Blogs Lists for Commenter.

Make sure you select the Fast Poster. Now click start, it’s as easy as that!

dr evil laser

Chapter 10: PageRank Sculpting

PageRank sculpting, say it, Matt Cutts won’t hear you. Now if you sculpt like a pro, then that dumbass Algo won’t have a clue either. There are many ways to approach PR sculpting, some methods are more aggressive than others such as pointing the majority of your posts, homepage, and category pages at the target you want to rank. My method isn’t quite as risky, actually if done right it’s not risky at all, it’s SEO 101.

We will be analyzing all of your indexed urls and making sure we have taken advantage of all relevant internal link opportunities. This can also be handy for client audits, it’s a quick and easy win.

There are two methods you can use to gather your site’s urls.

1. Use the harvester and the site: command.

2. The sitemap scraper addon, this is necessary for large sites with over 1,000 indexed urls. With this addon you can scrape XML sitemaps.

sitemap scraper

After you gather the urls, simply run a PR check and save all the URLs with PR. Then open the Page Authority Addon if you have the Moz API setup, and analyze each URL. Export to CSV then sort by Page Authority, Moz Rank, or External links to identify your highest juiced pages.

No don’t go dropping heavy anchor text links all over the place like a link happy freak or anything. Be smart about it. Use varied anchors and only where it makes sense. Weave it in naturally not like a drunk Scrapebox toting lunatic. If you find relevant places to drop, do it up.

And don’t go linking to your homepage a bunch of times rook.

Chapter 11: The Automator

Ok, so not only is Scrapebox the most badass SEO tool ever created in almost every aspect, but you can also automate most tasks.

And for a whopping $20 this premium plugin can be yours. Under the tab, click Available Premium Plugins, purchase the plugin through paypal and it will be available for download.

This is where you are going to need to use you imagination. With the automator you can easily string together huge lists of tasks and effectively automate your Scrapebox processes. The beauty of the automator is not only it’s effectiveness but it’s ease of setup. Very low geek IQ required, simply drag and drop the desired actions, save, and dominate.

As an example I will walk through setting up a series of scrapes.

Say you have multiple clients to harvest some link partner opportunities for. You can literally set up 20 and walk away. Come back to freshly harvested and PR checked URLs.

We would start by preparing our keywords, merging with footprints, then saving them all into a folder. Client1, Client2, Client3, etc.

Now open the Automator.

SB automator scrrenshot

Here is the sequence we would use:

Harvest Urls, Remove duplicates, Check Pagerank, clear, wait a few seconds, and repeat. The screenshot below shows three loops.

automator loop screen

After you add the commands, filling out the details should be easy to figure out. You’ll notice I put a wait command in between each loop, just set that to 5 seconds to let Scrapebox take a quick breath between harvests. I also added the email notification command at the end which is the icing on the automator cake.

Chapter 12: Competitor Backlink Analysis

To do this right you are going to need some sort of backlink checking service. Ahrefs, Majestic SEO, or Moz Opensiteexplorer will do.

If you have multiple services, you can use all of them and remove dupes. Yes, this is a bit crazy but will get as many of your competitor’s backlinks as possible.

Now in classic Scrapebox fashion we are not going to just look at one competitors backlinks, we are going to look at them all. Take your top 10 competitors, export ALL of their backlinks and merge together.

Once you get all the links exported and pasted into Scrapebox, you can began analysis.

We can collect the follow information on our competitors links:

URL or Domain PageRank
Moz Page and Domain authority
Moz External links
Social shares
Anchor text
IP Address
Whois info
Platform Type
Dofollow/nofollow links

We can approach this in two ways:

1. Get links from the same places as our competitors.

2. Get a clear picture of what is working for sites currently ranking so to replicate it.

So let’s start with approach one, snagging competitor link opportunities. From here you will be able to break down your competitors links in many ways. This is where we can use our link prospecting techniques via the page scanner addon and spot some easy slam dunk link opportunities. Thanks competitors!

Depending on your niche, you might be able to pick up some nice traffic driving comment links here as well. Bust out the blog analyzer and run all the links through that, it will identify blogs where your competitors have dropped links. Sort by PR and OBLs, viola you’ve got some sweet comment links.

Approach Two, What’s working now…

One of the most powerful SEO tactics around and one that will always live is reverse engineering competitor backlinks to see what is currently working in the SERPs.

There is no one size fits all approach, so understanding what’s ranking the site currently that you’re trying to outrank is key.

Sure finding relevant link opportunities and matching your competitors links is huge, but understanding what Google is favoring is the insight you need.

Using the live link checker you can take the links and check the exact anchor text percentages they are using. Since the “sweet spot” can be niche specific with our pal Google, this is a necessary approach for SERPs you’re very focused on.

This is done on a site by site basis. Start by taking the top ranking site’s backlinks and saving them into a txt file, backlinks.txt

Then create an additional txt file with nothing but the competitors root domain, save that as Backlink-target.txt

In the comment poster section, tick the box “Check Links”.

check links

Now in the Websites field open the Backlinks-target.txt file with your competitors homepage url. Then in the Blog Lists field open the text file with all of the backlinks, backlinks.txt.

link checking 2

Click check links, let roll, then export as csv.

Open the file and sort the anchor text column fro a-z. From here you can easily see the % distribution of their anchor text. Take the number of occurrences and divide it by the total backlinks. Boom, you know exactly what the anchor text percentage is for the currently top ranking site. Use that information how you will.

Now we could continue to go wayyyy more in depth on competitor links and how to leverage this intelligence in hundreds of different ways but I’m running out of gas here. The best way to learn this stuff is by getting your hands dirty. So bust open your backlink checkers, roll up your sleeves, and fire up Scrapebox already.

Start making your competitors wish they would have blocked the backlink crawlers like you did. Well, hopefully ;-)

This Guide is Originally from

Chapter 13: Free Scrapebox Addons

Social Checker – Bulk check various social metrics; Facebook, Google +1, Twitter, LinkedIn, and Pinterest. Results can be exported in multiple formats, .xlsx, .xls, .csv, .txt, .tsv, and others. Also supports proxies.

Unicode Converter
– Convert text in different languages such as Chinese, Russian, and Arabic into an encoded format that cane be used in the Google URL harvester keywords and footprints inputs.

Backlink Checker 2
– Download up to 1,000 backlinks for a URL or domain via Moz API.

Google Cache Extractor – Fetch the exact Google cache date for a list of URLS and export the URL and date.

Alive Checker
– Take a list of URLs and check the status of the website, alive or dead. You can also customize what classifies dead urls by adding response codes like 301 or 302. Will also follow redirects and report the status of the final destination URL.

Alexa Rank Checker – Check Alexa rank of your harvested urls.

Duperemove – Merge multiple files together of up to 180 million lines and remove dupes. Work with enormous files and split results however you’d like.

Page Scanner – Create custom footprints as plan text and html, then bulk scan URL’s source code for those footprints. You can then export the matches into separate files.

Google Image Grabber – Harvest images directly from Google image search in small, medium, and large outputs.

Rapid Indexer
– Submit your backlinks to various statistic, whois, and similar sites to help force indexing.

Audio Player – Bump some tunes while you scrape.

Port Scanner – Display all active connections and corresponding ip addresses and ports. Useful for debugging and monitoring connections.

Article Scraper – Scrape articles from different article directories and save them as txt files.

Dofollow Test
– Load in a list of backlinks and check if they are Dofollow or Nofollow.

Bandwith Meter – Displays your up and downstream speed.

Page Authority
– Gather page authority, domain authority, and external links for bulk URLs in the harvester.

Blog Analyzer
– Analyze URLs from harvester to determine blog platform (WordPress, blogengine, moveable), comments open, spam protection, and image captcha.

Google Competition Finder – Check the number of indexed pages for given list of keywords. Grab either broad or exact match results.

Sitemap Scraper
– Harvest urls directly from sites XML or AXD sitemap. Also has “deep crawl” feature where it will visit all urls on the sitemap and identify and URLs not present in the sitemap.

Malware and Phishing Filter
– Bulk detect websites containing malware, or that have contained malware in the last 90 days.

Link Extractor
– Extract all the internal and external links from a list of webpages.

Blogengine Moderated Filter – Scan large lists of BlogEngine blogs and determine which are moderated and which are not. Then load into the fast poster and blast away.

Domain Resolver – Resolve a list of domain names to the IP addresses(s) they are hosted on and check location.

Outbound Link Checker – Easily determine how many outbound links each URL in a list has and filter out entries over a certain threshold.

Mass URL Shortner – Shorten massive URLs using some common shortening services such as tinyurl.

Whois Scraper – Retrieve whois entries from harvested URLs, get names, emails, and if available, domain creation and expiration date.

TDNAM Scraper – Harvest soon to expire domains straight from Godaddyauctions.

ANSI Converter – Export URLS from harvester as unicode or UTF-8 to use Learning poster in other languages.

Fake PR Checker – Check fake Pagerank of harvested urls.

Chess – Play chess, it’s good for the mind.

Chapter 14: The END

Final Thoughts and General Ass Kicking Advice

Now that your eyes have been opened to the power of Scrapebox you might find yourself in brief SEO shock. My hope is that not only will you see the benefits of Scrapebox but this will also change the way you look at playing the game we call SEO.

If you are guilty of manually combing through Google SERPs for link opportunities then I will forigve you if you promise to change your ways.

The data is at your finger tips, leave no stone unturned and don’t let something silly like Google’s 1000 result limit stop you. One of the prerequisites to being a “good” SEO is being able use search engines better than any other human can. And without some sort of scraping tool you’re going to get your ass handed to you.

There are always ways to improve your processes, even when you think you have it mastered and 100% optimized. SEOs neglecting the power of Scrapebox is just one example. Keep your eyes open and get money!

And after all that who doesn’t like free stuff??

I have compiled the largest footprint collection you’ll find. Everything is broken out into platform type, ready for scraping domination. Simply click the image below to access them:


Wow, you made it to the end, good job. Now please share this damn guide that I dedicated a substantial chunk of my life to creating!!


  1. says

    I have to thank for such a great post and comment. I also sent you an email for your Proxies recommendation but now I have a last question :
    I am very intereste by EXPIRED (not expiring) domains in .fr and .es
    Any idea how to scrape to find these kind of expired domains ?
    Thanks for any comment !

  2. says

    Brilliant tutorial! I also notices my scrapebox updated yesterday and now I have version 1.16.0. Also there is a platform button next to the Custom Footprints and Platforms radio buttons. Click on that and click “Check for more platforms”. My success rate has gone up with the upgrade and I’m stoked.

  3. Christy Kunjumon says

    This is really epic information & very very useful. I have been personally using scrapebox from past few months for finding dead domains. I feel this too very useful for finding such domains. I started with this post is explaining about how to use scrapebox for finding dead domain.

    Thanks for publishing such an informative blog. (Y)

  4. says

    Having never used ScrapeBox, I just wanted to make sure…have you personally seen success using ScrapeBox?

    What would you say to people who claim that it’s a Black Hat tool and using it will help you get Google slapped?

    Thanks for writing all this, I don’t think you would unless you really did believe in the tool.

    • says

      Hey Eli, thanks for your direct and to the point question.

      Yes I have seen lots of success with using Scrapebox.

      But it’s not the only piece of the puzzle, it’s more of a tool I use for everyday SEO activities. If I need to generate some fake names, use SB. If I need to dig up some relevant link prospects or do any kind of big scrape, que SB.

      I’ve heard the software referred to as the swiss army knife of SEO and I fully agree with that analogy. Also keep in mind that my guide doesn’t even cover half of the potential uses for Scrapebox, it’s almost endless. And to the people who say it’s a black hat tool I would say they are ignorant, sure there are some black hat applications, but the majority are perfectly white hat and simply increase efficiency.

      Who want’s to manually search for guest post opportunities when Scrapebox can dig up 1000s and filter down the best in minutes? I know I don’t. If you really dig into to the program you will see that any SEO who shuns Scrapebox is making a critical error. Unless they are a coding geniuses who create custom scraping tools.

      For the average SEO hustler like myself, Scrapebox is the go to weapon of choice.

  5. says

    Awesome guide Jacob! I wish Scrapebox would run on mac. I had to buy a poc windows xp laptop from craigslist just for Scrapebox! Now I’m ready to go. My only concern is that my Ip may get banned from Google. I know that this is the what the built in proxy feature is for. Do I have anything to worry about? Is it possible to get in trouble for using Scrapebox?

    • says

      If you are using proxies it will be fine.

      The only risk is banning the proxy. With any tool if used maliciously then you can run into trouble, keep to yourself with a set of private proxies, you’re golden.

      • Kamil says

        You can use several sites to obtain a list of proxies. I get mail every day is such a list. It also gets a few thousand others, so at the beginning you have to check the proxies that are not blocked. In this way I have a running server and I can move the job.

  6. says

    Great post. I have enjoyed every word.

    I just recently started again using Scrapebox after some time without using it and I still find it quite useful.

    Thanks for sharing this information and the goodies.

  7. says

    Great guide, thanks for taking the time aggregate all this info. The only thing I think is inaccurate would be the guest posting section… I mentioned the same to Neil Patel. If sites are advertising guest posting, you don’t want to be guest posting on those sites.


    • says

      Thanks for the comment Danny.

      While I can see your logic behind the site mentioning guest posts leaving footprint I don’t think G would solely rely on that to devalue links or penalize links.

  8. says

    Great tutorial Jacob, thanks for all the hard work.

    Thinking outside the box with ScrapeBox (excuse the pun) is what makes the difference. But what a powerful tool!

  9. Nacy says

    Hello Jacob,

    I like your posts and jokes, such as “start blasting and drinking beer”. In another post “why I love blog spam”, you made some guy spill coffee on his keyboard :)

    Thanks for your hard working and sharing!

  10. says

    Just listened to you being interviewed at Halo 18. Thought I would check out the tutorial. Blown away, thanks for the info. Scrapebox looks a cool tool.

  11. riya says

    Holy shit! This is awesome. I will have to set a weekend to go through the wealth of information you have provided. Thank you very much :)

    One question – what’s your take on instead of buying proxies?

  12. says

    Jacob: I have been hesitant to jump on Scrapebox, but you have convinced me to make the move. Thanks for the post — I greatly appreciate all the work you put into it.

  13. says

    Sick tutorial man!! Wow, lots of work involved in this post.. really as HQ as it gets.

    In b4 someone rips it, packages it as an ebook and sells it to newbs on WF for $29.99 :P

    • says

      Thanks Chris, many sleepless nights went into this.

      Ha I thought about that and someone ripping it. Risk I decided to take I guess. Hopefully nothing like that ever happens :-)

  14. Gus L. says

    Great article! Some of this I’ve done, a lot I haven’t–but look forward to implementing. One question: for redirecting expired domains, you have four steps (below)…but why is #2, restoring pages, necessary? Why not just identify the old pages with and 301 immediately to similar pages on your money site?

    Step 1. Acquire expired domain
    Step 2. Relaunch domain and restore everything.
    Step 3. Redirect domain via 301 redirect.
    Step 3. Aggressively link build to the now redirected domain.


    • says

      Hey Gus, in the past it seemed to work just fine immediately redirecting and blasting. From my experience though, it seems Google is tougher with reattributing the old link profile back to the newly resotored recently expired domain.

      This is why I like this approach, I am trying to ensure that the domain regains its former glory.

      • Gus L. says

        Ah, that makes sense, thanks!…I’ve been either building out a tier 1 site with expired domains’ old pages, or 301 a domain that I don’t have time to build (or link profile isn’t worth it).

        One more question if I could: if you 301 a domain with a spammy, Penguin slappable backlink profile, does Google also slap your money site?

        • says

          Ooohh, breaking out the big guns. They totally can and I’ve seen it personally, but it depends on how the move occurs. If you literally move a site and 301, the penalty typically will follow. If you just 301 a shit domain and something random, I’m not sure on that.

          I’ve heard many success stories of people redirecting away from a penalty but haven’t had too many personally. I usually just crack some beers and spam more.

  15. Mark says

    I knew Scrapebox could do a lot more than I was using it for, but I have never taken the time to try to figure it out. This is going to take a few reads to fully digest, but thank you! Seriously awesome stuff here.

  16. Iman says

    Although Google is using the social signal as new algorithm to the SERP since last year. Backlink factor is still used for improving ranking at Google. Then use ScrapeBox with these tutorials!

  17. Febrodroid says

    It’s so f*ckin über especially Chapter 6.2. …. I can’t even stop laughing evil…


  18. Matt says

    Wonderful resource! Thanks very much for putting in all that work.

    Just an FYI The Automater is $20, not $15, but still a good deal none the less.

    All the best,


  19. Hans says

    Hi Jacob, thanks for making the tutorial! I bought scrapebox and proxies and followed everything you said, including buying the squidproxies. But they are not working for me. When I test them, they are okay. But when I start harvesting I don’t get results. When I start harvesting without the proxies I get results.

    I don’t know why it’s not working. Do you have an idea?

    Thanks in advance,

    • says

      Hey Hans, at first they did not work for me either.

      They told me to try using port 3128 instead of 8800 and then they worked fine. Let me know if that helps and also submit a support ticket if it doesn’t work.

  20. says

    Thanks so much for this Jacob.

    Im somewhat new to IM & Scrapebox in general, and this really made a complicated program easy to understand!

    Will be reading & following in the future :)


    • says

      Hey Brian, glad you enjoyed it. It was tough striking a balance offering the right mix of advanced tactics and general SB knowledge for beginners. Glad you enjoyed it dude.

  21. Paul says

    Thanks for this detail breakdown of scrape box. There is really a lot to a achieve with this powerfully tool. But please do you know of any tool that can do this scraping of data directly online. also Jacobs do you offer seo mentoring.


    • says

      Hello Paul

      You question : is it because as me you have a Mac and don’t want to use a win software ? I have the solution for Mac. A french application (with english vers.) even better than Scrapebox and exporting file with SB extension.. with better use of proxies, and some options that SB realize very slowly. I discover this soft a month ago… and I am totally amazed !
      I dont want to make ad in this thread so send me a message from my site and I will give you the name.

  22. says

    Great guide Jacob

    I read it all in one go whilst drinking four pints of real ale – so forgive me if you covered this and I missed it…

    Can you use Scrapebox to search for broken links across the Web based on keywords alone? For example, I type in ‘blue widgets’ and it uses a series of preset searches to find me URL’s with broken links on them in that niche?

    I just also have to say you’ve got a great approach to SEO and this is a great blog – care to be a guest on my podcast?



    • says

      Awesome, drinking while reading it is definitely encouraged. Maybe I should have noted that up top.

      Hrmm, I’d have to think about that one.

      I know you can use Scrapebox to dig up link resource pages and then scan those for broken links.

      Glad you like the blog! I’m down for a podcast, hit me up jacob @

  23. Nak says

    Overall a really great article .. unfortunately the noobish me still trying to digest the 1st half of it.

    Just a question on the “Comment Blasting” Section .. have u known of anyone who got into any issue with Google’s Webmaster’s tool feature called “Manual Actions”?

    Any tip on ways not to get into trouble with that?

    PS: This is not auto-posted by SB :)

    • says

      You my friend have just entered into the spam comment hall of fame ^^

      This is amazing, for anyone who has been brave enough to scroll this far, let’s analyze this comment a bit.

      It almost got me, but the link…the link to was a spammy porn web2 blog. Ahah, take a deeper look at the comment.

      It is possible to create an spun comment like this and pull page elements in the comment.

      See below:

      “Overall a really great article %authorname%… unfortunately the noobish me still trying to digest the 1st half of it.

      Just a question on the “%subheader%” Section .. have u known of anyone who got into any issue with Google’s Webmaster’s tool feature called “Manual Actions”?

      Any tip on ways not to get into trouble with that?

      PS: This is not auto-posted by SB :)”

      Spin the hell out of that, get your poster to pull in those elements, and bang, manual approval for tier 2.

  24. says

    Hi Jacob,

    Really Nice post, you make a great Ultimate Guide for Scapebox !
    I have a question related to scrapbox, i’m not a pro in SEO, and i’m trying to understand all this tools.
    But you are an expert so i would like to have your opinion on this :

    Should i use Scrapbox or RDDZ Scrapper ? , im about to buy a license for one tool, and im a mac user (RDDZ is compatible mac).
    Do you know this tools ? SB is more powerful or its the same ?

    Thanks for your answer, and sorry for my bad english (im French)

      • Elie says

        Thank’s for your reply Jacob,

        RDDZ is a scraper just like SB, but the difference for me, is that RDDZ is working on Linux and Mac and im using mac for work..
        So if you have some time to check it out what is RDDZ maybe you will like it and give some advice.

  25. says

    Hey Jacob,

    Great blog and post! I just went through the entire post and enjoyed it. Scrapebox seems like a a necessary item to add to the SEO ninja tools. I’m learning so much right now also has a question. Have you ever looked into These guys are about the same price as your Squid Proxies, but they multiple locations and dedicated ip’s for even the 10 package. Let me know what you think cause I just purchased Scrapebox I want to get to scrapin!!

  26. Anshul says

    Wow, the most informative post, i think it was worth i stopped her and learnt about the new tool which i will surely use in my website. Thank for such a great post. Keep posting let the newbie like me aware for such great tools

  27. Len says

    Great instructions you put together for us all here as its a no brainer to follow for newbies wanting to know how to use scrapebox properly and all its functions.

  28. Lalit says

    Hi Jacob,
    It was nice to read entire article on scrapebox. I have a SENUKEXCR tool to use. But as per discussion over the internet, I found SENUKE will ruin my IP and domain image in the long run infront of major search engine so i was looking for a white hat method like scrapebook.I would like to understand what will be the exact budget that I need to show my organization for this software including major plugins required?

    Please let me know,can we use it interact with social media and can to analysis for web traffic?

    Thanks for sharing Valuble Knowledge


  29. Michael says

    Hey Jacob, how do you randomly merge your custom list of stopwords with your list of keywords that you’re using for scrapebox? For example, I have a list of footprints that I exported from the articles engine in GSA, pasted that into textmechanic and added %KW%, and then imported this into Scrapebox and merged my footprints list? How would I add the additional step of randomly merging your list of stopwords?

  30. Skye says

    Amazing that scrapebox is still up until now. This is a promising software along with MS and SEMrush. Its just that I can’t afford it anymore! LOL

  31. says

    Amazing tutorial you have created! Thanks SO MUCH. It is priceless for a newbie like me. I have a quick question though. Regarding the competition analysis you say and I quote:
    After you click OK you will have a nice sorted list of keywords with exact match results from low to high.

    Depending on the yield I get, I will break the keywords down into ranges of exact match results.






    From there I will paste each range into the keyword tool, gather volume, and sort again, this time from high to low on the search volume. Then you can comb through and find some easy slam dunkable keywords.


    I do not quite understand this last part. If you already have words with low competition (0-50) why would you want to sort again and why from high to low? You mean from high to low in that group (0-50) or that group compared to the others with more competition? Which would be the words with the “easy slam dunkable keywords”?

    Mind you, your tutorial is excellent, just that this part throws me off a little bit, but it is my inexperience talking.

    • says

      At the point you are grabbing the search volume then sorting. Say you have 1,000 keywords with under 50 exact match, the next way to sort is by search volume.

  32. says

    Hey man thanks for putting all rad, Internet-dominating knowledge together. Unless I missed it, you didn’t mention much about scraping for email contacts from a list of domains. Do you ever use SB for that? If you do, care to share any advice?



    • says

      Nope, nothing on scraping emails. Haven’t done too much emails spamming, although I’d like to. I want to inflate penny stocks with massive email spam, sounds like a good time.

  33. says

    Well, this one is amazing Jacob. It really helps understanding what scrapebox is really about and how it ultimately facilitates seo practices at once.

  34. Prachi says

    Does Scrapebox harvests only wordpress links? Or even links for social bookmarking, wiki, web 2.0, rss submissions, etc., are provided?

  35. Chris says

    Hey Jacob, I am trying to leave “footprints” for people looking for guest posting oppurtunities on my network. I want to charge a small price to post their content. Also to have SEO agencies contact me through the contact page for advertising inquiries and such.

    Whats the best way to leave them footprints to find??

    * I thought maybe having the words “guest post” and “advertise” in the title tag on a separate page would help.. Should I have a live “email link” that they can scrape on that page too? I’m still trying to figure out how to use Scrapebox so I’m just trying to figure out the best way for people to find these pages for transactional meaning.

    • says

      This is a first, I get you, you want people to find your stuff for guest post related terms. In essence leaving footprints intentionally on your site.

      You are on the right track, see you should build a list of some really common guest post footprints. Then do onpage SEO for those, maybe if link build it a bit, but if you’re in the top 1000 results for a ton of those “guest post accepted” type terms, you’ll get flooded. Not sure why exactly you’d want that, but that would work. Setting up a honeypot?

  36. Bell says

    I see in the picture you are getting like 10 urls a second I am only getting like 2 urls a second what am I doing wrong? I got it set at 21 connections 40 sec timeout using public proxies. Thank you.

    • says

      You would ideally build some custom footprints for WordPress and either merge them in or put them in your footprints.ini file so they come up under the dropdown. The default wordpress footprint is very limited.

  37. says

    OMG this guide is awesome! It got me days to read through your guide properly (and I still need to keep coming back here for the details). Thank you for putting everything together in this guide and share it for public. I’ve been wondering how to properly use Scrapebox. But I’m only using MacBook Air. Is it possible if I use Scrapebox using Parallels with this MacBook Air? I won’t do heavy scraping, though. Or do you suggest to just use a desktop computer?

    Thank you, Jacob!

    • says

      Hey Elena, thanks, glad you enjoyed. This thing is like a freaking book. It is definitely possible I just wouldn’t go crazy with it. If you’re doing some white hat work and just need to run some small scrapes you should be fine. I wouldn’t leave it running all day though, not very nice to your Macbook air. You can also consider a cheap VPS if you decide the burden on your Macbook is too much.

  38. says

    Hi, I am using scrapebox since last 3 years and found it very useful. The best part of your tutorial is chapter 15 i.e. you have included links to worlds best tutorial.

    I must say this is a complete collection of ScrapeBox tips under one roof.

    Thanks Jacob.

  39. Mark says

    Or you can choose the easy way and buy cheap seo backlinks for just $1 from here.

    Anyway, great scrapebox guide, thank you !

  40. says

    Jacob thank you so much for the SEO tips! I have scrapebox already, will be trying the Article Spinner that you recommended. I look forward to following you!!!
    Justin Merced

  41. William says

    Hi Jacob,

    Thank you for this great tutorial, just got started with Scrapebox :)
    I am trying to harvest a custom footprint – 10 Results only.

    “powered by wordpress” + “post comment” OR “post a comment” Cars

    Without proxies, my results are the same as in Google.

    As soon as I use my proxies I get no results – zero results :(

    Am I doing something wrong? Or is something wrong with my proxies and I should I ask for my money back?

  42. Conor says


    That’s one boss post! Kudos for putting it all together.

    You should consider making a downloadable pdf version too, that way we can share it about and print it out :)

  43. Ray says

    Great article. I just downloaded your footprints file. In Word, it’s 33 pages long! If I were to scrape sites to post on–say, for the word “bicycle”–do I merge my scraped keywords with that entire 33-page footprints file?! Or do I pick and choose which footprints I should use (i.e., cut and paste the various footprints into a more manageable file)? Thanks.

    • says

      There are many footprints for many different platforms. So if you where scraping for ALL of those platforms you would use them all I guess, but they are meant to be used platform specific.

  44. harish shah says

    Great article. I downloaded scrapebox and got harvester but now I am confused how I can build links. Do I have to send mail to webmaster and asking to allow build the link or there an automatic method.

    When I see competitor’s backinks it goes to thousands and think how they build ?

    How people get link juice ?

  45. Dave says

    Hello Jacob.

    I am the Executive Director of the world’s most published PR firm for industrial and technology companies.

    We are trying to scrape email addresses from trade show websites — what has to be done is to scrape the URL’s of the exhibitors from the trade show websites and then scrape the email addresses from those the URL’s.

    We will also need to scrape email addresses directly from the web using key words to locate the companies we want to bulk email.

    I have played around with Atomic Email Hunter and some other software, but am a bit confused.

    And I am in touch with a company called Datafiniti which may be able to provide what we need.

    Any suggestions?


    • says

      Hey Dave, I don’t have anything for you there. I know Scrapebox has the grab emails function but that is only going to get emails visible in text onsite.

      Not sure about the tools you mentioned, but I’m sure there are some kick ass ones out there. Gonna have to keep hustling on that one…

  46. Bert says

    This is so far over my head, I had 3 Capt. Morgans just to get through it. I am such an amateur, I don’t know where to begin. Everyone says they have the answers to me getting my new website seen. Just pay them $200, 300, 400+ per mo and they promise nothing. Very confusing and even more depressing.

    But, your tutorial here is very thorough and written in a way that folks with some knowledge will probably understand. It seems to be quite advanced and dangerous if not used properly. I am reading your other articles as well to become more informed. Thank you.

  47. says

    Dude… I was about to go to bed and seen this.. WTF?! Serious post about Scrapebox!! Im having doubts about my decision to buy a Mac..

    SEO tools – Senuke, The Best Spinner, Scrapebox… All Windows..

    Need a new laptop just for SEO!

    • says

      Dude… I was about to go to bed and seen this.. WTF?! Serious post about Scrapebox!! Im having doubts about my decision to buy a Mac..

      SEO tools – Senuke, The Best Spinner, Scrapebox… All Windows..

      Need a new laptop just for SEO!

  48. julius says

    I am really satisfied with your blog content, your posts are really good and you are keeping it well. I would like to publish my post on your blog (as guest post) with my website link. Mostly I create about educational and student related subjects. Please let me know if you are recognizing guest posts and I’m ready to discuss my content, I guarantee it with unique and quality content.

    Looking forward to get your reply.

    • says

      Oh thanks Julius, and yes I recognize guest posts, they’re pretty spammy usually. Like 99.9% of the time. Are you in the .01%? Are you gonna make me a believer?

  49. George says

    I searched for Scrapebox in google and your site came up second in the list.Read the article I loved this automated portion in this.

    I have one doubt what does page authority means my blog now has page authority of 31 ?

  50. Jeff says

    Hey Jacob,

    Thanks for this post… it enticed me to purchase SB. I have a quick question… for you or a random reader.

    I’m using 10 private squid proxies strictly for harvesting URLs from Google. I have been using it a lot with advanced search parameters (inurl & intitle). I noticed that I was getting a lot of “banned (error 302 IP blocked)” errors. However, when I do queries without these advanced search parameters, I don’t get the errors. I used firefox to check the proxy and the captcha was showing up for only the advanced search parameter queries. I’m able to enter the captcha through firefox and SB will work on that proxy for a while but that damn catcha comes back. Have you experienced this or know the best fix? Oh, and I’m using the settings it came with.

    • says

      Hey Jeff, yeah the advanced operators make for some more trouble. It does that shit to me when I’m googling stuff manually so I imagine the proxies run into it as well. That’s why you might want to test with 25 proxies instead of 10.

      Adjust the RDN delay setting, move up to 1 second or more. Then reduce your threads to only 1, this is under maximum connection setting. inurl is a bastard.

  51. says

    I came here with the suggestion of Matthew. You nailed it completely, Jacob.

    I have bookmarked all the links mentioned in this article and going to read them in next 2-3 days..

    Thanks a ton. (Y)

  52. Ron says

    Wow, I just bought SB a day ago and have been reading A LOT, and there are so many videos too… Wow, really good stuff here. Thanks for all the info, it’s gonna take me a while to go through all of this information and I’m looking forward to it!!!

  53. Wazza says

    Just bought SB from your link and got it for $57 that is awesome! Thanks for that.
    Now I have SB, GSA SER and Capture Breaker to help rank my sites.

    I have a LOT of learning ahead of me and I am loving your tutorials as they are very easy to understand. Plus your articles aren’t boring like most I’ve read.

    Great stuff Jacob, keep up the great work on your blog. Can’t wait to read more

  54. Sunaina says

    Hey Jacob! Very informative guide and your step by step tutorial has made me so inspiring to check out the Scrapebox today! I’m so excited and as always I love to read your blogs.

  55. Sean says

    Hi, this a great tutorial and I’m learning a lot from your website. I just bought SB last night (through your link of course!), and I’m now trying to familiarise myself with some of the functions. Just now I was doing some keyword research but were a few things I was unsure of. I was getting a lot of errors which I assume is because I’m using free proxies? Is that the most likely cause of the problem? Also, what data is the Google competition finder returning? Is that the number of competing pages in the SERPs? You said you break down the results into ranges and the highest is 1,000 – 5,000 but I don’t think I’ve ever done a Google search with so few results (even exact match results).

    • says

      Ey Sean, yes the errors are 100% from the free proxies, those are shit and never gonna get you far. The competition finder is pulling the number of Google results, always do exact match, otherwise like you said the number will be crazy high.

      You could even do intitle:”keyword” and get an even more accurate measure.

  56. mehfoz says

    i have SB working but in addon i want to install rapid indexer from list i select it but install butto not highlighted tell me is software problem or addon.mnu file problem all addons not working can you solve this?

  57. ClickMinded SEO Training says

    ScrapeBox is really helpful! I’m really grateful I reached you’re blog. I’ll remember and understand the steps one at a time. Thank you!

  58. David says

    Wow thanks for this EPIC blog post. Love people who share things this clearly.

    One question, I’ve only been using it for a month and I’m having trouble with this..can Scrapebox eliminate blog posts over two years old, and with closed commentS?

    • says

      Thanks David, I have an answer for you. Scrapebox can’t do this on its own but YOU can filter them out over time as well as use modified footprints to exclude them. Such as -“comments are closed”

      You can even do a “dry run” where you just blast link at something dumb like purely to gauge the amount of successful submissions you list gets and filter out the junk.

  59. Krish Murali Eswar says

    I shared your post in in twitter by clicking the social-locker button. Yet, the lock did not open to your 1400 foot prints. Please fix it. Thanks.

  60. BackY says

    Hello Jacob, first of all, million of thanks to you, for this amazing tutorial. It helped me a lot.
    My question is, is there a way to make a .csv file with same monthly searches&results for one keyword?
    I can only manage to make two seperate .csv files with these details (one for monthly searches for that keyword, from Google Keyword Tool) and one with the Competition Tool (for results for that keyword)

    Is there a way I can combine these two infos for each keyword on the same file? That would be amazing.
    Thank you, and please keep posting great stuff!

    • says

      Well, the obvious way I would think would be to paste one column into the other csv and just save it. Probably and automated way to do that also if that’s what you’re asking.

    • says

      Yeah I’m always nervous about approving a link that’s not totally related because retards out there will spam blast the comment url. Seen it many times, just not nice. So I assume they link check the blogs they comment dropped, take those urls and start spamming em, pretty dumb on their part but sucks for the site owner.

  61. Rais Zada says

    Thanks for the guide and you are using some plugin for Contents style ? If yes, than please let me know name.

  62. duane bonas says

    Very good guide, i hope to create something like this in the future lol, just getting into seo and scrapebox was the first place i was gonna start, can anyone recommend xrumer ? just i know it is supposed to be a big learning curve ?

  63. Jakub says

    I am scraping google with your footprint file(about 500k operators) I use 40 private proxies and 1 thread and every time I only manage to scrape about 30k urls before all proxies get blocked. I even set delay for 2-3 seconds. Still does not help and the speed of harvesting gets pretty low there. I use single threaded harvester. Do you have any ideas what can i do to scrape constantly with no or just a few proxy bans? Also when I try using custom harvester and find public proxies the program does not scrape at all. I think it can’ t find any proxies. And I have “Auto harvest proxies” ticked and the whole process is finished in less then 1 minute with no urls harvested. I use GSA and the program has no problem with finding public proxies so why does scrapebox?

  64. I'm a spammer says

    I hardly create comments, however i did some searching and wound up here Ultimate Guide to Scrapebox SEO and Link Building.
    And I do have a couple of questions for you if you don’t mind.

    Is it simply me or does it seem like a few of the comments come across like they are
    left by brain dead visitors? :-P And, if you are writing on other sites, I would like to follow
    anything new you have to post. Could you list of all of all your
    public sites like your linkedin profile, Facebook page or twitter feed?

    • says

      Damn this comment must get sick approval. Me like it a lot, published for those who happen to stumble this far down and notice this advanced comment dropping technique.

      • RitzP says

        Wow, Jacob – this is a nice one! Im glad I started reading your awesome write up on scrapebox and looked at all the new comments with in the last 30 days. But Im new at this and I have a lot more reading up to do before I start, but I see why this comment is good! Im learning something lol

  65. ganhe dinheiro says

    I am truly grateful to the holder of this site who has shared this fantastic paragraph at at this time.

  66. says

    Hey Jacob,

    I was pretty clueless when I first started using scrape box but thanks to this guide, I’m pushing to get my first keyword ranked.

    I’m a little bit wary of auto approve sites. But niche comments are of interest to me. I have a good list of URLs ready now for commenting within my niche. But can’t think of any relevants comments that could be generic enough to post on all websites. Why do you recommend 3-5 different comments per 500 URLs? can’t see how that would make any difference within a niche topic.

    • says

      Hrmm I didn’t write that section actually.

      You’re going for something generic enough to pass and creating several versions for uniqueness along with spinning the words.

      SEO for example, could be some dumb white hat shit like, “It seems that content quality just keeps gaining more importance with Google’s algo, just my two cents.”

      Write 5 more of those, spin the hell out of em, I think that is what he was getting at.

  67. Reven says

    Wow such a long guide…..Thanks for it
    I have read 4 chapter of this guide till i have two question in two chapter please clear it for me.

    What does these lines means.
    Chapter 2: For a massive list of footprints all using site: operator, you should turn it down. i.e. the Google index check.
    Chapter 3: Scrapebox is PC only but you can run it with Parallels.

    • says

      1. If you’re doing the Google index check on high threads it will get your proxies banned quick, this is a more quick to flag type query.

      2. Yes Scapebox is PC only, if you’re running Mac then you need Parallels, in that case it still runs fine tho.

  68. says

    why not you use pages for comments ?
    currently comment list is a bit long.

    +not truse , search in google for “ DHsellerUk” you will know what i mean.


  69. Kristina Lacida says

    Hey Jacob! Very informative guide and your step by step tutorial has made me so inspiring to check out the Scrape box today! I’m so excited and as always I love to read your blogs.

  70. william says

    Wow thanks for the great scrapebox info! I have had it but didn’t realize how powerful it is till i read your blog. Is it still a good tool in 2014. I am so glad I came across your blog and didnt erase it. All the expensive software like senukexcr seems great but didnt give me any links at all worth keeping plus I think google is smarter then that I am trying to do things honestly but use software to keep up with the bigger guys. What software do you suggest. I have domain samurai, traffic travis, scrapebox, pretty much everything you can name I just get lost trying to figure out witch ones to use.. Anyone willing to help me with aff marketing a cpa offers which I am approved for I would be more thne happy to pay a percent in exchange for some good seo info thanks will

  71. Giovanni says

    I’m speechless! Nice little Pandoras Box. Reading this even game enough confidence to ask the hot chick at the club out. Kidding! I can’t see in the dark.

  72. Mary says

    Hi Jacob! I really appreciate the effort that you put into creating this post! I’m now officially addicted to your website!

    Quick question.

    Apart from checking the option “use proxies”, is there anything else in the settings I need to change in order to tell Scrapebox NOT to use my own IP if my semi-dedicated proxies stop working?

    (In other words, I’d rather have scrapebox stop working on the projects when my semi-dedicated proxies go down.)

  73. Pete says

    I’ve been reading on your blog for the past 2 hours and save $497 from buying those useless hypo courses (you can get free from BH forums). Thanks dude, …..Just wondering… you make your money by exposing so much wheat to the pigeons. :)

  74. Jaison says

    Hay Jacob

    Thanks for the amazing article.. i had lot of problems with the proxy now i have purchased few proxy let me try once again .. Your blog is a really amazing stem by step details on using scrape box.. thanks again

    • says

      Thanks, toss up between your comment being legit and spam, I’m on the fence. Chopped your random ass link and now I’m starting to get the feeling like I’m talking to myself.

  75. Red Chen says

    You made big mistake!
    “Now open your favorite text editor and find and replace the page breaks with a space.”
    Please don’t replace pagebreaks or spintogather won’t work.

  76. says

    alot of typos but fuck it, read through this post like twice. a lot of useful information and resources.. just a few questions

    how does one go about getting auto approved links and how does one know they are auto approved?

    not sure if you are familar with GSA but how should we import the list to GSA, should we only import ‘successful’ comment links to GSA? since the engines are somewhat different some comments wont be able to be posted through scrapebox since it has a limited capabilites of commenting, or am i wrong?

    i just bought SB and would love some feedback..Thanks buddy

    • says

      Well shit, apologies for not being more meticulous with my editing. Yes I’m familiar with GSA, I practically brush my teeth with it and I was going to answer your question but the first 6 words fucked it up for you.

      • Bryan says

        wasn’t trying to be a a-hole. it was just funny for me, i didn’t care if there were typos, actually made me re-read it.. don’t take it the wrong way..

        • says

          My bad, maybe I’m a bit sensitive about this guide cuz I worked a ton on it and I know there’s typos, passive aggressive, mad at myself.

          Let me try to redeem myself with a helpful response.

          There are a couple methods for finding auto approve sites, the first one that I like is extracting people’s poorly spun comments putting them in quotes and using those to search google for other sites they dropped a comment on. Usually pretty effective.

          Here is one I just pulled from my spam que here, take a chunk of the comment – “Getting the information from Ceiling fans reviews helps you to determine” and Google that shit.

          That sentence chunk alone has 1,810 results that are most likely auto approve.

          Now that is a really good way to find them, and also sort of funny. The next way would be by scraping for some sort of inurl or onpage element to find a CMS with auto approve enabled. Which leads me to the next part of this process, test blasting…

          How do you know which blogs are auto approve in your scraped list? Blast a test run at or something with verification on, get some verified links? Congratulations you’re building an AA list.

          • Bryan says

            Appreciate it. sorry if i insulted you.

            The methods are pretty slick and aren’t very complex. I thought they would have been a little more in depth. As of now i’ve been trying to get my hands on some good AA links but wasn’t sure on how or what way to test them out for GSA.

            next step for me is to gather all my lists and run it on a popular site and see what my success rate is. Thanks for the tips. great info too

          • says

            Can you direct me to any resources that tell me what GSA is and how to use it?

            PS: This resource you created is really great! Thank you!

  77. Jared Jammer says

    Hello Dr. King,

    I have a question about scraping, so, naturally, I decided to ask an SEO expert. None of them replied, so I thought I’d try you.

    Basically, the overwhelming majority of web 2.0 sites I’m scraping are unusable (at least in UD) thanks to those annoying spam prevention questions.

    So, my question is:

    Are there any footprints we can use to help filter out those annoying sites? The best my pea-size brain can come up with is using the hyphen (-) operator to filter out sites with phrases like “what’s * + *,” “who’s the *,” etc., but I’m skeptical that it’s very effective.


    • says

      I dont think there will be any specific queries for this, just gotta make sure you’re running everything right on submission and you should be able to determine which are stopping you during the registration phase and which get successful submissions.

      Are you using Death by captcha? This will help you get past recaptcha. Personally I run GSA SER with captcha break 1st and DBC 2nd. If I wanted I could run a test blast at something and isolate the working sites. As for a specific scraping approach, I don’t think there is one. Just keep scraping and feeding those sites in. Every thought of running a test blast with no filters to see how many successful links you can get?

  78. says

    Mate, what a brilliant tute!

    If you’re ever motivated to do the same for GSA SER, their Captcha Breaker and Indexer . . . let me know.

    I’ll buy it

    Cheers, John

    • says

      Well I’ve definitely considered putting together a paid GSA course, but still not finding the time and worried about how rapidly GSA changes making this type of guide.

      Thanks though, appreciate the vote of confidence.

  79. says

    Nice work Jacob! just bought Scrpaebox and had no clue…until now :)

    I can’t see anything on scraping email addresses by keyword. is there a way to do that?

    Kind regards,


      • says

        Cheers for the quick reply :) By scraping mails, I meant more along the lines of forum members email addresses, mails from directories etc.


        • says

          Well if it’s not publicly viewable then there is no way to do that, which is usually the case with forum emails. When it comes to scraping emails, as far as my link spamming black hat ass knows, the only way is scraping contact pages. Sure there are some other tactics tho.

  80. says

    Whats up Jacob. Thanks for the Guide. I’d love to crack a beer with you sometime :). I figure we can keep a tab for every question I ask thats 1 beer I owe you. It makes sense to buy at least a case don’t you think?

    Heres the first one:

    What did you mean when you talked about using stop words in your query. In the first screen shot you see results of over 500k but then when you used the stop word it was around 400K. U said that you get closer to the 1000 limit, how is 400k = 1000? What if you want to scrape for more then 1000 results, because I want to go through a forum and get every single post and get a page rank on it.

    Also I didnt see the list of Stop Words

    Beer Count = 1

    • says

      Just woke up after one too many beers at Belmont.

      By stop words I mean words that have no meaning and we merge in with our main footprint.

      For example, say we wanted to scrape WordPress blogs talking about Matt Cutts. A simple query would be:

      “powered by wordpress” “matt cutts”

      This returns 68,000 results, but we can only scrape 1,000 of those. So then we could append this with stop words, example:

      “powered by wordpress” “matt cutts” “the”
      “powered by wordpress” “matt cutts” “who”
      “powered by wordpress” “matt cutts” “these”

      The stop words should be in that pack of shit after the social share. Enjoy, and yes I do take payment in beer.

  81. Harry says

    Great post. Many thanks!
    Tried to download your footprint collection, but nothing happened when I tweeted.
    Thought you’d want to know.

  82. says


    I have been coming back to read this blog post about Scrapebox. I love it and there are tons of information I learned from it.

    SB’s interface is pretty intimidating and if you don’t know the basics, you are going to find it hard to master. Basics like, where the proxies go, where to enter keywords, and the resutls.

    Of course, I don’t use SB to spam comments, it’s not working anymore. I just keep on coming to understand the use of SB. There are insights that I got from these, including analyzing your own backlinks as opposed to competitor backlinks reverse-engineering. And how to every separate data gathered from different addons. Like if you are trying to analyze dozens of links, you need to get the PR, DA, PA, IP and OBLs depending on which matters to you. I find it hard at first since, going through each of them and combine the data into one excel. However, the more I read this guide over and over again, I still find some useful idea that comes into my mind.

    I thank you for writing this guide. You rock. I use to have the SB and don’t understand a thing about it but with this guide, and of course, reading it over and over again helps a lot. :D

  83. Sam says

    Hello Jacob,

    I am trying to understand and learn Link Building. I am bit of confusion about

    Chapter 2: Building Footprints

    I have downloaded the Ultimate Scrapebox Footprints Master List – do i copy and replace the default Scrapebox footprint inside configuration? Than start scraping the list.

    Pardon me if this make you laugh but my head is just spinning and really want to understand.

    Many thanks

  84. George says

    After searching a lot about Scrapebox and all of it’s features I always find myself reading some parts of this tutorial and I have to say, it’s actually very helpful. Thank you for taking the time to write this detailed guide, it has really helped me!

  85. Pete says

    So we should use keywords in the name field when submitting comments?

    Excellent guide! And of course I’m a robot…or am I? (Just a little non-robot joke).

    Thankyou so much for this comprehensive guide!

    Pete E

  86. Fahad says

    HI jacob a quick question if you can help. I have been trying to scrape with scrapebox but recently i got into a trouble. I wanted to scrape business listing directories with scrapebox , canadian directories only and i tried different footprints like “add your business” but i am only getting 3 or 4 results with every footprint i try and the results it returns are same.
    I will appreciate if you can tell me a better way to scrape country specific business listing sites.

  87. says

    How do stop words help dig deeper into the googledex?

    When I try out the dog training search you give as an example, adding and removing “there” simply yields a whole different result set for me.

    The only way this digs deeper, is if you do a search with *and* without the stop word and then merge the results.

    Iow… a lot of manual labor.

    • says

      Yes see, it returns different results, allowing you to harvest more results from the root query. Scrape scrape scrape then remove duplicates, now you have thousands of results instead of just 1000. No manual labor, forget that shiz

  88. dmosCO says

    WOW good one dude! My head hurts now though….can’t wait to start trying to master all of this! Any one in particular that you suggest a beginner starting with?

  89. says

    Hey Jacob,

    I’m a total noob trying to get set up the right way. Do you recommend using software like GSA to push links after creating the target list with Scrapebox?

    Thanks for the advice and the tute.

  90. pauled says

    I don’t know if am being stupid but I cant see this post? All I see are the comments and no actual scrapebox tutorial.

  91. John says

    Hi Jacob,

    Many thanks for your article about Scrapebox, but I am still a bit confused about it. First of all its not easy to use, second I want to use it for raise traffic to my blogs and appearance in search results. In this regards do you think it will work for me ?

    Btw am already subscribed to your newsletter, will wait for desperately to see receive more tips and suggestion about SEO.

    Many thanks for your well written article.

  92. Mike Nate says

    Great post ….. have been using scrapebox for few months but your post definitely give me some new techniques to do my research on google boy !!!

  93. Keith says

    Hello Jacob,

    Glad I found your webiste ! These are some statements you have made on above tutorial listed below:

    So pick yourself up a set of at least 25 private ScrapeBox proxies. Personally I use 100 but I go hard. Start with 25 and see if that works out for you.

    Do these 25 proxies have to be dedicated proxies or can they be semi dedicated?

    What are your thoughts if I just used 20 dedicated proxies?



    • says

      Shared should be fine for Scraping, also fine for GSA. I was a bit over the top with my setup in these days, 25-50 shared and you should be golden, just keep your threads to proxies at the ratio I suggested.

  94. Anchit Shethia says

    How do I run Scrapebox on my Mac? Will Virtual Machine work fine? I need to make this sure before I go ahead and purchase it. Thank you for the detailed guide though :)

  95. mec says

    what is your opinion on their premium plugins
    1.automator-i own it its great
    2.ranktracker-can’t command as i don’t own it.
    3.article scraper-works fine
    do they all work fine especially answers are appreciable.

    • says

      1. automator, yeah quality.
      2. Meh why not just use a service for consistent hands off tracking.
      3. Yeah has some potential there, I’ve used it a bit. Just scrapes articles from directories.

Like this Post or Hate It? Let me Know!

Questions, feedback, and everything else.