This is the fifth video to be released in my Ultimate Guide To Tiered Link Building video tutorial series. If you missed the previous parts then check out the episode guide on the right.
This is the longest video in the series so far weighing in at nearly 28 minutes! Thanks to all of those that have watched the video, left feedback and been in touch to say thanks so far.
Important Update: Please read this.
What You Will Learn
- How to export & organize all of your links
- How to get every link indexed including the character spun text
- How to send search spiders through all of your tiers
Hi guys welcome to part 5 of my ultimate guide to tiered link building. In the last video we setup tiers 2 and 3 to post forever, in this video I’m going to show you how to export and organise all of your links.
How to get every single you have created indexed on auto pilot and most importantly how to send the search spiders ripping through the millions of unique paths that lead to your site.
Now if you remember from the last video we created this link tracker spreadsheet and I created some additional tabs here. What we need to do is populate these tabs with tier 2 and 3 links and in the case of link wheel bandit tier 4 links that we have created. And once we’ve got that list together we can move on to getting them all indexed and things like that.
So first of all lets jump over into link wheel bandit so exporting the report from linkwheelbandit is really easy, just select tools, generate report and that gives us this window here and then we can select our link wheel that we want to export the report from. You can select date ranges; I always go for the beginning of time to make sure we get all of the links here. In options we can select which tiers we are going to export, remember this is actually tier 2, this is tier 3, this is tier 4 because we built the link wheels out to our tier 1 links.
So you will need to in order do tier 1 posted article urls and then save that as Tier 1. And this is going to say there’s no data for my options because I didn’t actually build this link wheel out but yours will actually export a list of URLS. So you need to go through and do a report for each of the tiers as so, generate report and again yours will actually export some data here and then tier 3 and generate report here, ok.
So if you come to C link wheel bandit reports here you’ll get the text files of the reports you should have 3 here, tier 1, tier 2 and tier 3. So what we need to do is open this up and select it all and come to our excel spreadsheet and remember tier 1 in linkwheelbandit is actually tier 2, paste them in here and then go through and do the same for the tier 2 text file which goes on the tier 3 tab and the tier 3 text file goes on the tier 4 tab.
So I’m just going to paste in some dummy content here to show you how it all pieces together. Just before I show you how to export the URL’s from GSA I just wanted to show you the value of Captcha Sniper since my last tutorial. Last time the total solved was around 100,000 and since then it’s done another 86,000 captcha solves for me. That an immense number and would cost about the same cost as Captcha Sniper to solve and that is in just a week’s time.
So again I highly recommend captcha sniper it is essential for posting out on mass with GSA so just make sure you pick it up it will save you a lot of money in the long run. So let’s just minimise that.
Exporting our tier 2 links first here is the project we created last time, I didn’t let it run for very long just enough to build a few links so I can export them and show you how this works.
So if we just right click on the project, show urls, verified and in this window here click and press ctrl + a to select all, right click copy URL, come to our spreadsheet and paste it down into the tier 2 tab here.
And then we just need to come back into GSA and find our tier 3 project and the same process again, right click, show URL’s, verified, select all, right click, copy urls and then go to tier 3 of our spreadsheet and drop them in there. Then at this point you want to be saving a copy of all of your exported urls.
Now we just need to export all of these URL’s into their own individual text file. So to do that select all of the tier 1 links, come into your text editor, paste them, save tier 1 dot txt and then tier 2 tab, come up to the top here, select them all and tier 2 dot txt and then come here to tier 3 links, save as tier 3 dot txt and then here the tier 4 links. And again these are just sample ones I’ve tweaked all the URL’s so none of these are actually live URLs and your list will be an awful lot longer here so copy these, come across tier 4 dot txt.
And now we can cross reference all of these text files against each other in Scrapebox to make sure all of the links are still alive and none of them have been deleted before we move onto the next step of getting them all indexed and boosted even further.
So in Scrapebox the first thing that we are going to do is make sure that all of our tier 2 links are still all live and linking to our tier 1 links. So to do that we are going to use the check links function in the websites field here we are going to load our list of tier 1 urls. In the blog lists here we are going to load our list of tier 2 urls and then click on check links.
And what that will do is this is a list of our tier 2 links here and it will check each of these and see if they are linking to any of our tier 1 links. Now this tier 2 list is only 62 URL’s here but your tier 2 lists is going to be much bigger. So once you’ve got that loaded in like this just click start and I’m going to pause the video and resume it when it’s finished.
Ok that finished so we just need to go to export links, export links found and we are going to call this tier 2 found and press save. Then we can close this next in this field here we need to open our tier 2 links and this time down here we are going to open tier 3. Click check links again and this time this is a list of our tier 3 links and its going to check each of these to see if they link to any of our 62 tier 2 links down here so again click start I’m going to pause the video I’ll be back when it’s done.
Ok that’s done so again export links, export links found this one’s tier 3 found and click save and close. And we just need to do it once more this time in here a list of our tier 3 links and in here a list of our tier 4 links click check links click start and I’ll resume the video when that’s done.
Ok that’s done you can see that it only found 7 entries here this is because I didn’t actually build out the full campaign and build all of the tiers in link wheel bandit and GSA I just let it build a couple of URLS before I stopped it this is purely for demonstration purposes. So again that’s why you’re not seeing all the URLs, you’ll have a much bigger list here when you’re doing this final check but anyway export links, export links found and that is tier 4 found.
If we come back over to our link tracker spreadsheet we can actually go through and delete everything out of tier 2, 3 and 4. And then what we need to do is open our tier 2 found, tier 3 found and tier 4 found files that we have just created and then copy those into the relevant tabs.
And once you’ve done that what you will now have is a nice organised spreadsheet of all your live links split into tiers. This will be really handy moving forward so make sure you get everything organised like this and now we can start indexing everything up and sending all the search spiders through all of these links to eventually find our site.
Just before we move on to the next step I wanted to show you a bit of software called Inspyder Backlink Monitor. I’ve only just started playing with it properly I don’t actually use it as part of my main process yet but it’s a really nice way of managing a tiered link building campaign. Now the process I’ve just shown you in excel while it works it is a bit of a ball ache and it does take a bit of time to put together but with Inspyder Backlink Monitor you can just setup a project and paste all of your backlinks into this one box and thats all of your tier 1, 2, 3 and 4 and there’s no need to separate them out in anyway just one big long list of all the links you’ve built in your campaign.
Once you’ve done you can just hit go here and it goes out and checks various things for each page. If the link is still alive, how many outbound links there are, if the links no follow, the domain page rank, the page rank, whether or not it’s indexed in Google, the anchor text and the IP address. And when the run is completed if you come over to this link hierarchy tab you can see here earlier on I think in video 3 I created bookmark links to Google.co.uk and the rest of the links to Yourdomain.com as examples. But you can actually see here we can double click here and this is our money site URL, these are all tier 1 links this is a tier 2 link pointing to the tier 1 link and we can see another example there a tier 2 link pointing to a tier 1 link and if we just open this one up here, here again we’ve got a tier 2 linking to our tier 1 link. If you’d have done the campaigns properly, oh here we go so here we’ve got a tier 2 link, tier 3 link and all the additional tiers under that.
If I’d have built this campaign out fully there would be a lot more here for you to see but this is a really nice way of organising your tiered link building campaign as it automatically puts the link hierarchy and the structure together so you can easily see exactly what’s going on, the PR, the links whether things are indexed or not and then as you build new links for your campaign all you need to do is paste them into this box click Go again and it will go out and update all of these and update the data for your backlinks and recreate the link hierarchy and things like that.
There’s also the selection tool here so you can select things based on criteria. You can say I just want tier 1 links that are not indexed or you might want just tier 3 links that are not indexed and you can see this supports up to 10 tiers of links which is far beyond anything I’m ever going to build. But I just wanted to show you this tool it’s not something I’m using heavily right now I’ve only just started testing it properly this last week or two it might be a better solution for some people out there so I just wanted to go over that and show you how it all works.
So let’s take a look at how we are going to index everything up. It is really important that we get all of the tiered links we have created indexed in Google. We also need to make sure we send the search spiders on a path of discovery throughout the tiers and as the spiders move through the tiers they are going to discover millions of unique and relevant paths to find your money site.
The way we are going to do that is by building more backlinks obviously, this is tiered link building and the best way to index anything is to build more links to it. But there’s a clever way to go about it, we are only going to build links to pages that are already indexed in Google and get regular spider visits already. This means that we don’t have to worry about building links to index links and then we need to build more links to index the links we have created above that. If we just build links that are already indexed by Google and get regular spider visits then it will index everything underneath it automatically, we don’t need to worry about indexing any more.
So there’s two ways that we can do this one is to use auto approve blog comments and we can use software like scrapebox to achieve that quite easily and the second one is guestbook submissions. Personally I use xrumer to do these just because it’s much faster but I appreciate that not everyone’s going to be able to afford the costs of xrumer nor the server that you’ll need to run it on so I’m going to show you how to do it with the Scrapebox learning mode poster and also GSA search engine ranker so you’ve got 2 ways to go about doing the Guestbook submissions.
So just before we jump to scrapebox we need to make a master list of our tier 2, 3 and 4 links. So to do that if you just select all of your tier 2 links and bring them to a new tab here, all of your tier 3 links and the tier 4 links which is actually tier 3 from linkwheelbandit if you remember correctly and drag that here. And if we just rename that Master List and give that a save. Next we need to go and get a big long list of related keywords so if we fire up Scrapebox.
And once Scrapebox is open go to scrape, keyword scraper and just enter a few of your root keywords into here and hit scrape and I’m just going to pause the video while that completes. Ok that’s completed so hit ok, remove duplicates and we actually need a really really long list of keywords so take those scraped keywords and we are going to add them to the search box here and hit scrape again and that will find even more keywords for us so I’m just going to pause the video.
Ok that’s done and then if we just remove duplicate keywords, transfer results to main keyword list, close and then we’ll just save that as keywords. Ok and what I should have done before with the master list from excel is if you just take a copy of the master list we created and save that as a text file as well and that can be called websites, save.
So then we need to generate a list of names and e-mail addresses and we can do that up here and I’m just going to generate a couple thousand names and we are not actually going to use the names for posting we are going to use the keywords we just saved for posting so we can skip the names files and just come across here and generate some e-mails and save those. Ok.
So close that and then we just need to setup the actual comment posting project. Make sure your using proxies this is important in this example I’m not going to both but you should be using proxies here and setting the project up for names we are actually going to select our keywords file. E-mail our E-mails file. Websites our websites file which is a master list of our tier 2, 3 and 4 links, Ok. Your comments file, I have a pre-prepared spun comments file, I’ll show you what that looks like quickly. This is something you should prepare in advance but it’s just a list of generic type comments that are suitable for any site. And finally your blog lists or target lists of sites that we are going to post to and this is a huge auto approve list that I’ve created over time. If you need to create your own auto approve list then have a look around some forums and things like that you can usually find some lists going about. Spend some time collecting lots of different ones and you can merge them into one list.
As you can see here mines over 180MB in size I think there’s over 3 million entry’s in it at the moment but just go around the web, collect your own list and make sure you remove any that constantly fail. So open. And that might take a while to load because of the size of my list so I’ll pause it while it’s having a think. Ok and once that’s loaded all you need to do is press start posting and that will go out and start submitting a load of blog comments to your targets URL’s and if I just press Ok there and we can see that’s working through the list already. So we can just leave this running now in the background and this will build lots of auto approve blog comment links to our tier 2, 3 and 4 properties and next we can move over to guestbook submission.
So with guestbook submissions I normally use xrumer but it’s pretty expensive and you need a dedicated server or at least a VPS to run it really so it’s probably going to be out of most people’s price range. So we are going to look at two different ways we can post to guestbook’s without a huge cost. The first one is going to be with the Scrapebox learning mode, now you can install the learning mode in the show available add-ons toolbox and once you’ve got that installed just open it up. And once it’s open you can come across to the learning tab here.
Now you can pretty much teach Scrapebox any platform you want with the learning mode and it’s really easy to do. So I’m going to teach it the Lazarus guestbook platform here and I’ve already got a list of Lazarus guestbook’s as you can see here so if we go to load urls from file and select our list of Lazarus URLS and just open it up and here it will say detected forms, name = book that looks like a guestbook form to us and click select and then it’s just a case of clicking in these fields and it will come up with a list of variables to assign, so in the name field we are going to want username, Ok. For e-mail field user email, Ok. Location it isn’t a required field so we can skip that. Homepage userurl, ok. Your message, comment, ok.
And Captchas this is a text based question captcha so we can just click here and it says select the part before the question that will identify the captcha which it has already done for us there. Select the part after the question which will identify the captcha and it’s already done that for us there as well so just click Ok. And in here we have got a text captcha result and press OK, and once you’re happy with those we can move onto the next URL up here. Detected the form, select and as you can see it has already filled out these fields for us based on what we filled out for the last one.
So let’s take a look at one more and check they are all the same. Oh that one’s got an image based captcha so we’ll skip that you can’t actually mix image and text based captchas in the learning mode you have to separate them out so let’s try this one instead. That’s an image based one, here we go another text based captcha so select username, useremail, user URL and comment. It’s already looking pretty trained to me lets double check all of this should be filled out and that’s it, so that is scrapebox we’ve taught it the Lazarus platform here you can go through and teach it lots of other different platforms but once you’ve done that you just need to save the learned form data, Lazarus posting config and hit save and let’s quickly have a look at how we post out to all of these guestbook’s.
So submitting to the guestbook’s is really easy if we just come across to poster here and the definition file is the file we have just saved the posting config open that up and then just fill out the rest of these quickly. Name and E-mail, comments that’s it there and load URL list this is our list of Guestbook’s we want to target which is there and that is pretty much it. So we taught it the Lazarus guestbook platform, exported that config, loaded it all into the poster here and then you’ve just got to click on start posting and that will go out and submit your link to all of the guestbook’s.
This is kind of a slow and sluggish way to achieve guestbook posting as you’ll have to go through each guestbook platform and teach it to scrapebox which is going to take you a while and it’s a bit of a ball ache managing different lists for different platforms and things like that. There’s a much easier way to do it with GSA Search Engine Ranker if you’ve got it but if not Scrapebox, you can do it with. So let me show you how to do it with GSA because it’s much easier.
So if you open GSA and click new, untick everything here except for guestbook we want that ticked and you can see how many guestbook platforms are supported there, there’s quite a few. So let’s fill out this form quickly our target URL’s which are our tier 2, 3 and 4 links and we are going to import our master text list of those urls which is here.
Our keywords is a comma separated list of keywords and anchor text is a spun list of our anchor text. Tick these 3 boxes here it comes with some default text inserted there but I like to use my own custom written and hand spun comments so if you grab that and chuck it in there and also replace the German one and the guestbook titles which you can leave at default.
Come across to options the only thing that you want to untick is that so we only post guestbook links otherwise we end up getting some of the other types come through and right click check all with English language and click on Ok and this is going to ask us for a name so I’m going to call it video tutorial guestbook spam and hit ok.
And as you can see our project has been created there now what GSA will actually do is it will go out and scrape its own target list of guestbook’s before it submits to them. What we can actually do is use Scrapebox to scrape a huge list of all the different types of Guestbook’s and then import that URL list into GSA so that this project only has to focus on the submission of our URL’s and it’s just a really nice way to speed things up. So underneath this video on my blog you will find a list of guestbook’s that I have previously posted to and used you can download that and import as target list here free. But what you should do is use Scrapebox to scrape your own targets, now I’ve included a list of footprints as well under this video that you can use for scraping so let’s just head over to Scrapebox and I’ll show you how to use those very quickly one moment.
So in Scrapebox to complete the guestbook scrape you need 2 files, one is the list of footprints for all of the guestbook’s that are supported in GSA this file you can get underneath this video on my blog and then we also need the merge list of common words which I made available in video 2 I believe it was. So with those 2 files in place here we can import from file and that’s going to be our actual footprints list here and then if we click the merge button we can merge that with our merge list and that creates a list of things to go out and scrape for. You will need to use proxies here to complete the scraping process I’m not going to just for this example I’m just going to do a very brief harvest here and just see what we can get so I’m just going to pause the video while that completes and we’ve got a couple of thousands results in now so I’m just going to stop harvesting, remove duplicates and then just export that URL list as GSA Guestbook Targets.
Now obviously that will take a lot longer to run as there’s nearly 10,000 keywords here and you’re going to end up with a huge list but once you’ve got a list of guestbook’s exported you can head back over to GSA.
Once you’re in GSA to import that list of Guestbook’s you can just right click on the project, import target URLs from file and select that target list and GSA will automatically parse that list and submit to them all as soon as you press start.
So let’s quickly sum up what we have done in this video.
First of all we learnt how to extract and organise all of our tiered backlinks with scrapebox and a spreadsheet. Don’t forget to take a look at inspyder backlink monitor as well.
Then I showed you how to setup an auto approve comments blast with Scrapebox
Next we looked at 2 different ways to post to guestbook’s, one way was with the Scrapebox learning mode and the other was with GSA search engine ranker
The end result is you now know how to index all of your tiered links and most important get the attention of the search spiders.
All of the comment and guestbook’s links have been placed on pages already indexed in Google so now all we have to do is wait for Google to find our links and send it’s spiders through all of our tiers.
In the next video I’m going to reveal my personal advanced tier 1 tactics to help throw you up the rankings in no time at all.
You’re also going to learn how to automate natural social signals in just a few clicks
And I’m going to show you how to buy and use new or aged domains in your campaign to create the perfect tier 1 link.
Well that wraps up part 5 of the ultimate guide to tiered link building tutorial.
If you have enjoyed it and want to see more like this then please subscribe to my blog at www.matthewwoodward.co.uk
Resources In The Video
GSA Search Engine Ranker FREE TRIAL – An absolutely essential piece of software for any link builder.
Captcha Sniper – Solves captchas automatically, saves huge amounts of money on Captcha fees.
Inspyder Backlink Monitor FREE TRIAL & Exclusive Discount – The only piece of software capable of tracking a multi tiered link campaign.
Update: The Inspyder team got in touch to offer a $10 discount for my readers, this gets added on automatically at checkout.
Xrumer FREE TRIAL – My personal choice for posting to Guestbooks (I show you how to do it with Scrapebox/GSA in the tutorial).