Search Engine Scraper by Creative Bear Tech Change Log

Our tech wizards are constantly making new updates to our Search Engine Scraper. Below you will see the latest changes to the software. If you have purchased a copy of the software, you will be able to update your software automatically. You do not need to download any additional files.


Version 1.1.1

  • Added licensing
  • Created "Creative Bear Tech Manager" that will keep the application running in case it's interrupted.
  • Added auto-update
  • Updated the application GUI
  • Updated LinkedIn, Trust Pilot, and Duckduckgo scraper.

Version 1.1.2

  • Adding XEvil service to resolving the captcha issue.
  • Enhance the speed of scraping by update multithreading mechanism.
  • User controls the number of threads running by the scraper.
  • Enhanced proxy routing mechanism, each sub scraper form has its own proxy.

Version 1.1.3

  • Updated Google Map scraper
  • Add setting for enabling/disabling the application logs [this can enhance the scrapers speed]

Ideas for Future Updates


Privacy and Proxies


  • Scraping via a TOR browser. This is where inside the proxy settings, the app would simply connect and scrape using the TOR browser.
  • Scraping using Google Chrome Incognito window.
  • Both options would be available by way of check boxes under the proxies tab.
  • Under the public proxies tab, the app should have an option to auto check and verify the public proxies every X number of minutes (make sure that they are working), automatically remove non-working proxies and scrape new proxies every X number of minutes OR when the total working proxy number falls below X number of proxies. As well as allowing the user to upload and enter their own public proxy sources, we can have a list of ALL proxy source urls and the user can simply check using checkboxes which proxy sources they would like to use. The idea here is that the app will constantly monitor the proxies by removing non-working ones and scraping and adding new ones every so often to ensure that the app has enough proxies to run on at all times.

Keyword tool


  • We can have an integrated keyword tool. In the field where a user enters their keywords, we could have a button "Search Keywords". The app would then go to Google and search for that keyword and pick out all the suggested keywords and automatically add them to the keywords list to scrape. The idea is to save a novice user some time from having to think of many keywords for running their searches.

Footprints option


  • The Footprints function should be simplified. Once a user opens up the footprints button, the app should simply give them a field to enter their own footprints inside a pane or upload a file. Next to the pane, we should have a check box to "Combine footprints with keywords". The app would then merge their keywords with every single footprint. For example, if we have 1 keyword and 20 footprints, this would give us 20 unique keywords: root keyword + footprints. The idea here is to save a user time and effort. For example, the footprints section could be used to search for guest posting opportunities or whatever a user likes.

Search results pane


  • We could add an option to disable the real time view of results / disable GUI to reduce the consumption of processing power. We can simply add a check box with something along the lines of "Disable GUI for faster speeds".
  • Inside each column name i.e. url, email, website, address we should add a check box so that a user can select exactly what data to scrape. And in the first column we could have one checkbox to select all or select none.
  • We should add the name of the business name to our results. We can get this via Facebook business page.

SPeeds and threads


  • We could add an option to automatically change the thread numbers if the CPU usage exceed X% and ram exceeds X%. We could allow the user to enter / select any values they want. This is a good way to ensure the stability of the app.

DOMAIN FILTERS


  • We could add an option to "Skip subdomain sites" as those tend to be web 2.0 and contain a lot of spam. This could allow us to save time and processing power.

Extra scrapers / dictionaries


  • We could add Instagram to the list. However, the specifics of how Instagram works are different to other sources. We should add some simple options under Instagram drop down whether to search for users or hashtags on Instagram or both. We should also add an ability to login / add login details to an Instagram account under the last tab inside the settings.

adding accounts under the last settings tab


  • We should also add a field for adding a private proxy that will tie that account. This will ensure that the user will always access social media accounts through one private proxy. Perhaps it would be a good idea to all a user to enter their accounts inside a plain text pane in a format like username:password:proxy:platform (platform would be our shortcode that would identify the social network. The could be LinkedIn, Facebook, Twitter. This would make it quicker to enter accounts. If a user enter more than 1 account. The app could switch between accounts every X number of minutes.