Guest Blogging for SEO - How to Find Websites that Accept Guest Posts
Guest Blogging for SEO - How to Find Websites that Accept Guest Posts
The Search Engine Scraper and Email Extractor by Creative Bear Tech can be used to find niche websites that accept guest blog posts. You can then contact these websites with your custom guest post proposal.
What Is Guest Blogging in SEO?
In very simple terms, guest blogging is the process of publishing your articles on other websites. Typically, guest posts contain backlinks pointing back to the target website.
Why is Guest Blogging Important for your Website Search Engine Rankings, Traffic and Sales?
In SEO terms, if you have quality backlinks coming from niche-related websites, popular search engines such as Google are going to rank you higher on the Search Engine Results Pages (SERPs) which will lead to more traffic and sales. Posting unique, SEO-optimised and quality articles on niche-related authority websites is considered to be the best, safest and most effective backlink building strategy.
How to Find Websites that Accept Guest Posts
Finding niche-related websites that accept guest posts can be a tricky and time-consuming process. Our Website Scraper will allow you to find websites from your niche that accept guest posts and collect all the business contact details that you can use to contact your list of sites with your guest post proposal.
A Step-by-Step Guide on How To Find Guest Blogging Websites with Our Scraper
Part 1: Website Scraper Settings Configuration
Step 1: Configure your proxies
Open the "Proxy Settings" tab. Here, you can add your own list of proxies. The search engine scraper supports many types of proxies, including:
Backconnect Rotating Proxies
Private Dedicated Proxies
You will need to enter your proxies in this format:
IP:PORT:USERNAME:PASSWORD (for proxies requiring authentication)
IP:PORT (for IP authenticated proxies)
You can paste your proxies into the text box or you can upload them from a notepad file (one proxy per line)
You will then need to test your proxies to make sure that they are all working.
If you would like to use proxies, do not forget to check the "use proxies" option on the main GUI screen.
We recommend that you use quality private proxies if you intend to use many threads. This typically applies to users running the software at more than 10 threads. Likewise, proxies are recommended if you would like to scrape data from a particular region such as United States. In this case, you should aim to use local USA proxies.
For quality proxies, we recommend shared proxies by Act Proxy: https://actproxy.com/aff.php?aff=252
Step 2: Select your Search Engines and Websites to Scrape
You will need to open "Search Engines/Disctionaries" tab and select the search engines and the websites that you would like to scrape. In order to scrape a list of websites that accept guest blog posts, we recommend that you only select search engines such as Google, Bing, Yahoo, Yandex, Ask, Ecosia, DuckDuckGo. You should not select business directories or social media sites because they will not contain any website pages dedicated to finding guest posting sites.
It is usually enough to select a couple of search engines.
Step 3: Captcha Settings
Here, you can enter your 2captcha remote captcha solving API key. If too many requests are made from the same IP address, search engines such as Google can throw up a Google image recaptcha to verify that you are not a robot. Unless the captcha is solved, you will not be able to do any scraping from that IP address. One way to solve Google image recaptchas is to use a remote captcha solving service 2captcha or XEvil Captcha Solving software. Please check our separate guide for configuring XEvil with the Search Engine Scraper.
You can buy XEvil captcha solving software here: http://www.botmasterlabs.net/product70741/
If you are not using many threads (2 to 4 threads) and are applying at elast 1000 milisecond delay between requests (see main GUI screen), you should be ok and not need to solve captcha. However, for smooth and uninterrupted scraping, we strongle recommend that you use either XEvil or 2captcha solving service.
Step 4: Speed Settings
Open the "Speed Settings" tab. These settings will determine the speed of your scraping and the scope of your results. Configuration include:
"The total number of search results (websites) to parse per keyword" - this is how many websites should be scraped for each keyword from the search engine results page (SERPS). Generally, 100 to 200 websites is the gold standard as you are not likely to find relevant results beyond page 20 of Google which is more or less the search engine graveyard.
"Maximum Number of Emails to Extract from the same Website" - you should never select more than 5 emails as you are likely to get a lot of spam or duplicate emails.
"Do not show pictures in integrated web browser" - by enabling this option, you will use less CPU resources as the scraper will not need to load images, which takes time and more computer resources.
"Enable Application Activity Log" - by enabling this option, we will be able to see what went wrong in case of a crash.
"Enable individual threads activity log" - by enabling this option, we will be able to track the work of every single thread and get to the root of the problem in case of a software crash. By enabling this option, the scraper will consume more CPU resources and run a bit slower.
"Scrape Facebook in case emails not found on the target website" - the scraper will extract data from a Facebook business page in case data is not found on the website (email, tel, address).
"Always Scrape Facebook for more emails" - by having this option enabled, the website scraper will always check the Facebook Business Page and try to retrieve more emails.
"Scrape Twitter" - you can also choose to scrape Twitter for extra data.
Step 5: Domain Filters
This setting will control the topical relevancy of your target websites.
Here, you can enter
1) Keywords that website URLs must contain
You should be very careful when using this filter and have a think about your niche and whether the target urls are likely to contain your set of keywords. Invariably, some niches such as cryptocurrency have a tendency to contain regular patterns of the same keywords such as crypto, cryptocurrency, bitcoin, blockchain. The same applies to CBD niche where each website URL will contain CBD or Hemp. However, niches such as beauty and cosmetics have a very broad list of category keywords and it would be recommended not to use the url filters. Ultimately, it is your judgment call and there are not hard and fast rules.
For example, https://justcbdstore.com/product-category/cbd-gummies/ contains CBD in the url.
2) Keywords that website URL must not contain
Here, you can enter the keywords that must not be present in the website url. You can add spammy keywords such as porn, gambling, viagra and so on.
3) Website Blacklist
Here, you can enter a list of sites that the website scraper should skip. You can include sites such as Amazon, Ebay, Newspapers, Magazines and so on.
TIP: DO NOT ENTER TOO MANY KEYWORDS OR URLS IN THE FILTERS AS THIS WILL SLOW DOWN THE SOFTWARE. TRY TO PICK KEYWORDS AND WEBSITES THAT ARE MOST LIKELY TO COME UP FOR YOUR SEARCH. FOR EXAMPLE, IF I AM SCRAPING THE BEAUTY NICHE, I KNOW THAT FAMOUS MAGAZINES AND NEWSPAPERS ARE LIKELY TO COME UP SO I CAN SKIP THEM.
Step 6: Content Filters
This section is very important as it plays an instrumental role in determining the relevancy of your websites. Inside the text box, you can enter your set of keywords that must be present inside a website's meta title, meta description and even HTML and visible body text. You should spend some time in thinking and researching your keywords as this will determine the scope of your results.
You can tell the software how many keywords from your list a website must contain and "match the exact keyword" (the keyword must be the same on the website meta title and description AND/OR body as it is on your list. We recommend having this option enabled). If you enable the option to check for keywords in a website's HTML or body content, you are likely to get more but less relevant results.
Think about all the possible variations of your keywords. Here are some examples:
Step 7: Save & Login Settings
You should specify the path to the folder to which the results should be saved. Please note, it is recommended to keep this path as short as possible. Saving to Disk C or Disk D would be ideal.
In the login account section, you should enter your Facebook login details. We recommend that you create a Facebook account using your local IP address. The scraper will access every Facebook business page and try to extract data from it. Sometimes, Facebook requires for a user to be logged in in order to view a Facebook Business Page. Without a Facebook login, your results could be reduced. It is worth pointing out that the scraper will always run Facebook on a single thread using extended delays in order to emulate real human behaviour and to avoid Facebook bans. Facebook will be accessed via your local IP address (no proxies) so make sure that you are NOT running a VPN service such as HMA VPN when using your Facebook login.
Part 2: Setting Up your Keywords and Footprints
In this section, we are going to configure our keywords and footprints.
Recommended Search Footprints to Find Blogs Accepting Guest Posts
"Guest Blogging Spot"
"Submit a Guest Post"
"Become a Guest Blogger"
"Guest Post Guidelines"
"Submit an Article"
"Want to Write for"
"Blogs that Accept Guest Blogging"
"Blogs Accepting Guest Posts"
"Suggest a Post"
"Become an Author"
"Become a Contributor"
"Places I Guest Posted"
"Publish Your News"
"Guest post by"
"This is a guest article"
"Add Guest Post"
"Guest Bloggers Wanted"
"Guest Posts Roundup"
"Write for Us"
"Submit Guest Post"
"Submit a Guest Article"
"Guest Bloggers Wanted"
"Group Writing Project"
"Blogs that Accept Guest Posts"
"Blogs that Accept Guest Bloggers"
"Become a Contributor"
"Submit Design News"
"Submit Blog Post"
"Suggest a Guest Post"
"Contribute to our Site"
"Become a Guest Writer"
"My Guest Posts"
"This guest post was written"
"This guest post is from"
"Now Accepting Guest Posts"
"The following guest post"
We will now need to combine this footprints list with your main keywords.
Inside the main GUI screen, click on "Add Footprint" button just under the keywords box.
Inside column 1 (Keywords), enter keywords relevant to your niche. For example, you could enter
Inside column 2, you should enter your footprint list for finding guest blogging sites (you can use the above list. Add the keywords with quote marks).
Now click on "merge" and "OK". The software will merge your keywords with every single footprint list and you will have keywords that will look like this:
CBD "Submission Guidelines"
Hemp "Submission Guidelines"
CBD "Write for Us"
Hemp "Write for Us"
The merged keyword list should now appear inside the keywords box.
You are good to go. Before we start the scraper, let us quicky run through the options on the main GUI.
"Crawl and Scrape E-Mails from Search Engines" - Check this box as we will be scraping search engines.
"Scrape Emails from your website list" - you should only use this option if you have your own list of urls and you would like to extract data from these urls. For example, you may have scraped your own url list using scrapebox and would like to now extract data from your list.
"Use Proxies" - check this box if you want to enable your proxies
"Invisible Mode" - the browser windows will be hidden.
"Fast Mode" run the scraper on multiple threads.
"use an integrated web browser instead of http request" - you should only use this option if you are using a VPN service such as Nord VPN or HMA VPN.
"Real Time View" - enable to view the results in real time (will consume more computer resources).
"Delay between requests" - keep this at the default value. This is to avoid IP bans and captchas.
"Delete results" without "emails" or "tel" - you can delete results that do not have an email or a telephone number.
"Complete Previous Search", you should launch the scraper and check this box and click on run to resume your previous search (useful in case you want to power off your machine or your software or operating system crashes or restarts). The software will use your previous settings and pick up from where it left off.
Once you have everything configured, hit on start and watch your results become populated with guest posting opportunities. You can then use your list to contact websites with your guest blogging proposal via email, telephone, social media, website contact form or otherwise.
Part 3: Results
Results will be saved in real time inside the save folder that you have specified in the settings. Once the scraper finishes running, it will create a results file with the word "DataResults" inside the filename. The software is going to create a separate line for every unique email as sometimes businesses/websites can have more than one email and we want to have one email inside one cell at a time in case we want to copy all emails inside an export file. So if you see duplicate values, these are not duplicates but rather unique emails per website/business.
Email List Cleaner
You can click on the pink button on the main GUI "email list cleaner" to filter the results. Inside the email list cleaner, you can filter your emails according to keywords. You can choose for "email to match domain" to keep only company emails. Likewise, you can remove emails containing or not containing your set of keywords.
Click on "Export Data" to export all results in an Excel Format.
Click on "Export Emails" to only export email address / one email per row. This is useful if you need to use only emails for mass email campaigns or newsletters.
TIP: You can always import the csv file into the scraper and scrape new results on top of your previous work. Make sure that this is a CSV file and not the results file (the results file has different formatting). The results file will have the word "DataResults". You should never import this file because it is formatted differently. Instead, import the latest CSV file from the save folder without "DataResults".