SEO

Netpeak Spider 3.0: Brief Overview

2
0

Just yesterday Netpeak Software team released new version of our flagship product — Netpeak Spider 3.0. It is the greatest update that took place in the lifetime of our program with more than 300 improvements: from super features to small usability upgrades. This update is currenlty available only for Windows, but we are already working on version for Mac OS.

Let’s go over the major updates you can try for free during the week (until April 4th inclusive).

By the way, a code name for new version of our crawler is ‘New SEO Hope’, which is explained in the video below.

And now let’s focus on the key new features.

1. Up to 30 Times Faster Crawling, Compared to 2.1

We have significantly optimized crawling speed in Netpeak Spider 3.0. We conducted tests to compare crawling 100,000 URLs with Netpeak Spider 3.0 and the previous version, Netpeak Spider 2.1, with the same conditions and got great results. The new version shows up to 30 times increase in speed.

Comparison Netpeak Spider 2.1 with Netpeak Spider 3.0 using large website

2. Optimized RAM Usage

When analyzing medium sites (10K pages), we reduced RAM consumption by 3 times and accelerated crawling by 8 times.

Large websites crawling (100K pages) highlights strengths of the new version – RAM consumption has decreased even more (4 times), and data analysis took 30 times less time.

3. Option to Resume Crawling After Opening the Project

Now you can stop scanning, save the project and then, if you want, resume website crawling:

  • For example, the next day → you just need to open the project and click on the ‘Start’ button.
  • On another computer → to do so, just transfer the entire folder with the project to the necessary computer (for example, if your device reaches RAM limit and you want to continue crawling on more powerful one), open the project there and click again on the treasured ‘Start’ button.

4. Rescanning a URL or a List of URLs

Previously, to check whether  previously found issues were properly fixed, you had to scan the entire website or the URLs list onceagain. Now you can select an issue in sidebar and start rescanning pages with this particular issue → and you will get a completely updated report of the entire website.

Rescanning the list of URLs in Netpeak Spider 3.0

You can also rescan both single URL and URL list on ‘All Results’ or ‘URL Explorer’ tabs.

5. Deleting URLs from Report

If you’ve crawled some extra data and it made your report too heavy, now you can easily delete this information, and all your basic reports will be updated.

Deleting URLs from results in Netpeak Spider 3.0

Moreover, at this stage you can just clear the results without deleting URLs. It is very useful when you need to rescan same pages with new parameters.

6. Changing Parameters During Crawling

Option to change parameters during crawling is one of the most awaited features of Netpeak Spider. This is what it allows to do:

  • If you’ve started crawling and then realized that you forgot to include some significant parameters or parsing of necessary HTML fragment, now you can stop crawling, enable necessary parameters and then continue scanning. In this case, scanned pages will not be recrawled and included parameters will be applied only for new pages.

  • If you’ve started crawling and understood that there is too much data (that is blowing your RAM), you can stop scanning, turn off unnecessary parameters, and resume it. And, voila, you freed up RAM and now can parse more data. By the way, the good news is that received data is not deleted, but just hidden. It means that if you turn these parameters back on, the data will appear in the table and be in all reports.

7. Data Segmentation

Segmentation is a unique feature on the market allowing you to change data view using a certain filter. Segments affect all reports in the program, including ‘Issues’, ‘Overview’, ‘Site Structure’, ‘Dashboard’, etc.

8. Site Structure Report

We’ve implemented a new report with the full site structure in the tree view, where you can filter pages at any level.

New report with full site structure in Netpeak Spider 3.0

Try to select a category and press ‘Advanced copy’ button. Results will be copied to clipboard. Now you can paste them into external table (for example, Google Sheets or Microsoft Excel).

9. Dashboard (Information about Crawling Process and Diagrams After It’s Complete)

In a new version of Netpeak Spider you’ll find a new ‘Dashboard’ tab containing information summary about crawling settings and process, diagrams with insights about site structure, etc., SEO-insights and default filters. During crawling it will show you an overview of the scanning process and important settings. If necessary, you can download full screenshot of this panel using ‘Export…’ button.

Dashboard during crawling in Netpeak Spider 3.0

When crawling is complete or paused, settings take a back seat and diagrams with useful insights about crawled pages take first place. All diagrams are interactive: you can click on any part of the chart or any value near it and go to results filtered for that value. Also, you can download full screenshot of this panel using the ‘Export…’ button.

Dashboard after crawling is stopped in Netpeak Spider 3.0

10. Exporting 10+ New Reports and 60+ Issue Reports in Two Clicks

In Netpeak Spider 3.0 we’ve developed default export templates. You can export the following data:

  • All issues → exporting all issue reports into one folder;
  • Custom issue reports → here are reports with convenient current data view on certain issues:
    • Broken links
    • Missing images ALT attributes
    • Redirect chain
    • Redirect blocked by Robots.txt
    • Bad URL format redirects
    • Canonical chain
    • Canonical blocked by Robots.txt
    • Links with bad URL format
  • All links → export external and internal links on scanned pages separately;
  • All unique URLs and anchors → in the same way, export separately all external and internal links. Data in the first report is grouped by URL, in the second one – by URL and anchor.

Export templates in Netpeak Spider 3.0

11. Determining Compliant URLs

Now Netpeak Spider determines 3 types of page compliancy:

  • Compliant — HTML files with 2xx status code, not disallowed in indexation instructions (robots.txt, Canonical, Meta Robots, etc.). They are the most important pages on your website, because they can potentially lead organic traffic.
  • Non-compliant — HTML files with non-2xx status code or disallowed by indexation instructions. These pages usually do not lead traffic from search engines and also waste crawling budget.
  • Non-HTML — other documents with lower chances for high ranking in SERP.

Compliance becomes a unified concept for tools and analytics inside the program, most issues can be detected only for compliant URLs.

12. Special Reports for Every Issue

To filter pages with some issues in more in the most convenient way we created a a quick access to special reports for the following issues:

  • Broken images
  • Bad URL format
  • Redirect blocked by robots.txt
  • Canonical blocked by robots.txt
  • Missing images ALT attributes
  • Internal Nofollow links
  • External Nofollow links

13. Rebuilt Tools ‘Internal PageRank Calculation’, ‘XML Sitemap Validator’, ‘Source Code and HTTP Header Analysis’, and ‘Sitemap Generator’

Internal PageRank Calculation

  • Added ‘Internal PR’ parameter for each link.
  • Internal PageRank calculation is processed only for compliant URLs. Thus, images and other content types different from HTML will not be accounted for.
  • We added ‘Link Weight’ parameter to show (you won’t believe it) how much link equity is distributed across the website. It helps you optimize internal link weight distribution if you use this structure on your website after simulation in our tool.
  • We created new ‘PageRank changes’ table. During each iteration program calculates the ratio of internal PageRank on current iteration to PR on zero iteration.

PageRank changes table in Netpeak Spider 3.0

XML Sitemap Validator

In previous version of Netpeak Spider, XML Sitemap validator was one of the crawling modes, but now we transformed it into separate tool. You can solve the following tasks using it:

  • Parsing links from Sitemap and importing them to the main table for further analysis.
  • XML Sitemap validation → we completely updated all issues (now there are 34 of them) according to main Google recommendations.
  • Ping XML Sitemap to search engines to draw their attention to changes in it.

Note that crawling rules also work for link parsing from XML Sitemaps. For comfortable use, we save your previous requests here: start entering URL, and tool will show you previously crawled XML Sitemaps.

XML Sitemap validator in Netpeak Spider 3.0

Source Code and HTTP Headers Analysis

We also improved interface of this tool and added ‘Extracted text’ tab showing content of the page excluding HTML tags. This feature doesn’t replace Google cached ‘text-only’ view (as more reliable source), but can show you how exactly Netpeak Spider analyzes text on webpage to count characters and words.

Source code and HTTP headers analysis in Netpeak Spider 3.0

Sitemap Generator

In addition to the interface changes of this tool, it now creates Sitemaps only for compliant URLs. You no longer have to worry about disallowed pages being added to Sitemap by mistake.

Sitemap generator in Netpeak Spider 3.0

14. Custom Templates for Settings, Filters / Segments, and Parameters

In Netpeak Spider 3.0 we released default templates and opportunity to create your own templates for:

  • Settings → These templates save information about all crawling settings tabs (except for ‘Export’, ‘Authentication’, and ‘Proxy’) and allow you to switch between them in 2 clicks. Note that we have default templates you can use for various tasks.
  • Filters and Segments → Create your own templates if you often use the same conditions for data segmentation and filtering.
  • Parameters → You can find these templates on ‘Parameters’ tab of sidebar.

15. Virtual robots.txt

If you need to test new robots.txt file and check how search engine robots will understand it, you can try ‘Virtual robots.txt’ setting. Netpeak Spider will use this virtual file instead of the file in root directory of your website.

16. Combining All of the Crawling Modes into One

In a new program version you can add URLs for crawling in the following ways:

  • By crawling the website → simply add a URL into the ‘Initial URL’ field and press ‘Start’ button: pages will be added according to the crawling progress.
  • Manually → this way a separate window with text input will be opened for you to add URLs list with each URL on a new line.
  • From file → we significantly modified this function and added URL import from files with such extensions: .xlsx (Microsoft Excel), .csv (comma-separated values), .xml (Extensible Markup Language), .nspj (Netpeak Spider project) and .ncpj (Netpeak Checker project).
  • From XML Sitemap → XML Sitemap validator will be opened, so that you can import URLs for crawling in the main table.
  • From clipboard → simply use Ctrl+V combination with program’s main table opened, URL list from clipboard will be added to the table and notification with brief summary information will appear (what is successfully added, what is already in the table, what is not added because of bad URL format).
  • Drag and Drop → you can simply transfer project or any other file with extensions mentioned above from file straight to the main table: Netpeak Spider will analyze files and upload necessary data.

17. List of URLs, Skipped by Crawler

For the cases when your site can’t be properly crawled, we’ve released the following features:

  • URL you’ve entered to ‘Initial URL’ field will be always added to the table. This way, you will know if this URL redirects to another website or simply unavailable at the moment.

  • ‘Skipped URLs’ tab will appear only if some pages were ignored while crawling. Here you will see URLs and the reason why they were skipped: for example, you’ve enabled considering indexation instructions from robots.txt file, applied crawling rules or disabled checking some of content types.

Table of URLs skipped while crawling in Netpeak Spider 3.0

18. Quick Search in Tables

Quick search in table is implemented in all interfaces of the program – just press Ctrl+E and enter search query. If you need to filter data by certain parameter or segment all data, use the following functions:

  • Ctrl+F → filter settings
  • Ctrl+Shift+F → segment settings

19. Delaying Comprehensive Data Analysis

All comprehensive data arrays were moved to a single module – one more unique feature of new Netpeak Spider. Using this module, you can view massive lists of:

  • incoming, outgoing, internal, and external links
  • images
  • issues
  • redirects
  • Canonical,
  • H1-H6 headers

20. New ‘Parameters’ Tab with Search and Scrolling to Parameter in Table

Parameters setting has been moved from crawling settings to ‘Parameters’ tab in sidebar. Now, just like in Netpeak Checker, you can search by all parameters and view a detailed tip about them on ‘Info’ panel.

If you have already got results in the main table, you can click on a parameter (or use ‘Scroll to’ button on this panel) to quickly reach the corresponding parameter in the current table.

21. Monitoring memory limit for data integrity

New version of Netpeak Spider checks the amount of free RAM and disk space: there should be at least 128 MB of both available for program to run. If limit is reached, the crawling stops and the data remains intact.

Summary

Netpeak Spider 3.0 has transformed into a super fast tool, allowing you to work with huge websites and conduct comprehensive SEO analytics without any setbacks. Compared to the previous 2.1 version, we've improved memory consumption by 4 times and 30 times reduced crawling time (by the example of a large sites scanning). And of course, you can find information about all Netpeak Spider improvements and crawling comparison by technical specifications between Netpeak Spider 3.0 and its main competitors in a detailed post on Netpeak Software blog.

Found a mistake? Select it and press Ctrl + Enter

Comments (2 )

  1. 0
    11 hours ago

    This was so useful in building our site. The article helped me to learn something new. Great content thanks for sharing this informative blog which provided me technical information keep posting.

    Regards

    Hari

  2. 0
    13 days ago

    Great updates! Both uper features and small usability upgrades.

    gmail sign up


To leave a comment, you have to log in.

Subscribe

to the most useful newsletter on internet marketing

Most

discussed popular viewed

This site uses cookies and other tracking technologies to help navigate and your ability to provide feedback, analyze the use of our products and services, helps in our advertising and marketing efforts for better user convenience.