Facebook Twitter Telegram Pinterest Linkedin Product Hunt

Version History

The list of SiteAnalyzer updates with release dates

Version 2.5 (build 280), 26.07.2021:

  • added data scraping based on XPath, CSS, XQuery, RegEx
  • added a website content uniqueness checker
  • added Google PageSpeed score checker
  • added the group sites by folder feature
  • added the list of projects filter by name
  • fixed incorrect processing of URL exceptions
  • fixed incorrect processing of website’s crawl depth
  • the display of redirects for URLs imported from a file has been restored
  • the ability to rearrange and save the order of columns on tabs has been restored
  • the processing of non-canonical pages has been restored, fixed an issue with empty meta tags
  • the display of anchor links on the Info tab has been restored
  • increased the import speed for a large number of URLs from the clipboard
  • fixed an issue with incorrect parsing of title and description
  • the display of alt and title for images has been restored
  • fixed an issue when a freeze occurred when switching to the "External Links" tab while scanning a project
  • fixed a program freeze that occurred when switching between projects and updating the nodes of the "Crawl Stats" tab
  • fixed incorrect definition of nesting level for URLs with parameters
  • fixed data sorting by HTML-hash field in the main table
  • the processing of Cyrillic domains has been optimized
  • the interface of the program settings has been updated
  • the logo design has been updated

Read more about v2.5 >>

Version 2.4.1 (build 240), 10.12.2020:

  • fixed wrong counting of duplicate pages
  • it now accounts Noindex tag in Meta Robots and X-Robots-Tag
  • optimized parsing of several descriptions on the page – now it starts from the first on the list

Download >>

Version 2.4 (build 239), 30.11.2020:

  • added data export on Custom Search and Custom Filters tabs
  • added data filtering on the Custom Search and Custom Filters tabs
  • in Custom Filters section (3xx redirects) added export of target canonical URLs
  • restored ability to scan sites hosted on local servers (localhost)
  • fixed incorrect display of actual information about duplicate site pages
  • fixed incorrect display of Alt and Title for images
  • fixed incorrect accounting of external links, when the URL of the investigated domain is included in the URL of the external link
  • fixed a bug with displaying incorrect information when switching between projects
  • fixed a bug with incorrect display of infinite redirects
  • fixed a bug where the progress bar freezes during data export
  • fixed a bug where the project opening process freezes at 90% completion
  • fixed incorrect data accounting of Robots.txt rules
  • fixed incorrect data accounting in the "Internal Links Chart" module
  • fixed typo "Content T1ype" in section "Crawl Stats"
  • accelerated pause of scanning process when importing URL from file
  • added displaying links to images from external domains in the "Images" tab

Version 2.3 (build 231), 19.10.2020:

  • added the Page Load Performance graph, it can be used to analyze the load speed of a website
  • added the Content tab, which displays the content distribution stats per symbols and words on each page
  • additionally, the links chart is interactive now, it is connected with the Low Number of Backlinks report in the Custom Filters section
  • custom Filters tab features a new Links node
  • added an Image Sitemap generator
  • fixed the bug when scanning of large websites with many redirects caused a program crash
  • fixed the bugged Start button in the empty main window, now it correctly reacts to changes in the URL field
  • fixed the bug that caused the incorrect H1-H6 headers display

Read more about v2.3 >>

Version 2.2 (build 227), 22.09.2020:

  • duplicate pages are displayed properly now
  • optimized parsing for H1-H6 headers
  • 307 redirects are processed properly now
  • fixed a text search bug in the Custom Search module
  • the "Exclude / Include URL" tabs in the program settings are now automatically saved for each specific project
  • "Crawl Subdomains" option was added to the program settings. It can be used to scan subdomains according to the scanning rules for the main domain
  • the "Visualization Graph" module no longer has "end nodes" – code 200 pages are colored green now
  • new version notifications now offer the automatic updates feature

Version 2.2 (build 225), 15.09.2020:

  • we added a custom filters module to search for content on websites during crawling
  • we added internal links chart of a website
  • when dragging a node of a graph, its child elements are dragged together with it
  • when clicking a node of a graph, inbound and outbound links have different colors
  • the visualization graph legend is interactive now (when clicking the elements in the legend, the corresponding nodes are highlighted on the graph)
  • X-Robots Tag option added to the website crawling settings
  • optimized parsing for H1-H6 headers that utilize classes
  • software hang in the final stage of large projects crawling is eliminated
  • fixed incorrect statistics representation of duplicate meta descriptions
  • fixed incorrect statistics representation of 404 pages
  • blocked URLs in Robots.txt now return status code 600
  • Response Time parameter is calculated more accurately now
  • fixed incorrect Sitemap.xml generation
  • redirects are displayed more accurate now
  • sorting by URLs is more accurate now

Read more about v2.2 >>

Version 2.1 (build 216), 13.07.2020:

  • added the ability to export every external link, 404 error, and image in a single report together with all the pages that contain them
  • added the ability to manually rescan the URLs that return a response code of 0 (Read Timeout error) in one click from the context menu
  • added the ability to hide and display certain columns and tabs in the main data block
  • added the "Read Timeout" and "Blocked by Robots.txt" status updates for the URLs that return a status code of 0
  • added a display mode for the graph visualization that sets the size of the nodes based on their PageRank
  • added an extra button to calculate PageRank on the corresponding tab
  • added the ability to rearrange and save the order of columns on tabs
  • the active tab is now saved when switching between projects
  • added the ability to parse the URLs protected with a .htpasswd file
  • added the ability to export the data from the Dashboard tab to Excel
  • fixed the AV error caused by switching to the "SERP Snippet" tab with no active project selected
  • fixed the incorrect accounting of the inbound and outbound links and their mapping in different blocks of the program
  • fixed the incorrect mapping of the percent indicators in the "Level" node
  • fixed the incorrect mapping of the "Content-Type" node data in the project scan statistics
  • fixed the unstable data filtering for an exact match
  • fixed the problem of continuing the paused site scans
  • fixed the incorrect HTTP/S sign when exporting reports to Excel
  • fixed the incorrect accounting of canonical URLs during the website scanning
  • fixed the incorrect accounting of Allow and Disallow statistics for robots.txt
  • fixed the incorrect accounting of images ALT and TITLE attributes
  • fixed the unstable PageRank calculation
  • corrected the typo "Alernate" to "Alternate" on the Info tab

Read more about v2.1 >>

Version 2.0.2 (build 208), 03.05.2020:

  • restored the correct work of the web crawler with Cyrillic domains
  • fixed duplication of identical pages on the Duplicates tab
  • fixed a bug when the data on the "SEO Statistics" panel was not updated during the first site scan
  • added a new column on the "Media / HREFLANG" tab that displays the number of alternative URLs found

Version 2.0.2 (build 207), 26.04.2020:

  • fixed bug on the Dashboard tab (found errors were not displayed)
  • fixed bug checking randomly selected URLs on the page indexing check tab in Yandex SE (only the first URL was scanned)
  • fixed a bug that occurred when deleting crawled pages added through the function of importing arbitrary URLs

Version 2.0.2 (build 206), 16.04.2020:

  • fixed incorrect display of incoming links for URLs in the Info panel
  • fixed incorrect display of statistics for doubles title, description and h1-h6 on the "SEO Statistics" tab (relevant for the first site scan)
  • optimized rules to follows Robots.txt directives

Version 2.0.2 (build 205), 13.04.2020:

  • increased parsing speed (growth of about 10%), increased stability of the scanner
  • the speed of drawing the nodes of the graph is increased by an order of 10 times (now the graph can easily display up to 100,000 nodes)
  • optimized time for the final saving of the project to the database (relevant for large sites)
  • optimized display of duplicate headers H1-H6 (duplicate pages removed)
  • added the ability to turn off alerts about the release of a new version when the program starts
  • added a column indicating the number of incoming links to a page on the PageRank tab
  • added ability to rotate the graph and change the scale by an arbitrary value
  • fixed not always correct definition of non-canonical pages (Wrong Canonical)
  • «Access violation» error occurred when deleting a large number of URLs
  • URL disappearance bug fixed when pausing the project / URL list crawl process
  • fixed a bug where the Delete button did not work in the Quick filtering field
  • minor usability errors fixed, minor bugs fixed

Read more about v2.0.2 >>

Version 2.0.1 (build 201), 07.01.2020:

  • fixed a program freeze that occurs at the end of website parsing
  • restored the ability to scan large arrays of links when importing URLs from a file, clipboard, or by link
  • fixed incorrect export of site structure to CSV

Version 2.0 (build 198), 23.12.2019:

  • added website structure visualization on a graph (similar to Screaming Frog)
  • added a feature to crawl local websites by URLs, not just by the main page, as it was before
  • added a feature to crawl local websites running on LiveServer (and other servers) at URLs like
  • added automatic re-crawling of pages if their loading is timed out (for example, when the website blocks a large number of simultaneous requests and returns undefined response codes)
  • added automatic update of program versions (at program startup or through the main menu)
  • added the option to copy text from the URL entry field by Ctrl-C
  • added counter for the remaining time of the current project crawling
  • optimized control of load speed and data recording in the database (now saving data to the database will not be so long with a large number of crawl threads)
  • restored the option to scan large arrays of links when importing URLs from a file, clipboard, or by link
  • fixed a bug with an incorrect server response when parsing pages containing a plus sign (+) in the URL
  • fixed incorrect export of the website structure to CSV (the number of webpages was not completely exported)
  • fixed incorrect accounting of the BASE tag content when parsing pages (when BASE HREF was a domain)
  • fixed a bug that occurred when deleting arbitrary lines immediately after project crawling (also added URL deletion by the Delete button)
  • fixed a bug where the scanner took into account the contents of "noindex, nofollow" meta tags when disabling this option in the program settings
  • fixed a bug where the area of the selected cell range was not displayed upon the cursor removal in the table with the basic data

Read more about v2.0 >>

Version 1.9.2 (build 186), 03.10.2019:

  • switch to the 64-bit version (crawling websites of up to 10 million or more pages)
  • increased website crawling speed and reduced memory consumption
  • added an option to run multiple program copies at the same time
  • added an option of saving advanced filtering parameters when switching between tabs of the active project
  • added memorization of column widths on tabs
  • added an option of crawling specific URL groups, without having to crawl the entire website
  • added an option to pause crawling a list of imported URLs
  • the "lastmode" parameter is taken into account when generating a Sitemap
  • added Server column indicating the type of server hosting the website (example: nginx/1.16.0)
  • added display of states <null> and <empty> for H1-H6 headers
  • added an option to crawl pages that do not respond with the Content-type header value
  • optimized calculation of H1-H6 duplicates (null or empty H1-H6 are now not taken into account)
  • contents of the "Exclude URL" tab of the general program settings have become common to all projects
  • improved page encoding detection
  • fixed incorrect display of pages with 3xx redirects
  • fixed a bug related to incorrect accounting for "robots.txt" rules
  • fixed incorrect parsing of pages containing a plus sign (+) in the URL
  • fixed accounting of images when crawling a website when only HTML parsing is allowed in the settings (if, for example, disabled objects issue a 404 response code)
  • fixed a crash that occurred when loading large projects
  • fixed a bug that occurred when loading large projects when the lower progress bar showed more than 100%
  • assigned a blue icon to 301 and 302 redirects instead of an orange when calculating SEO statistics, and set 1 weight point instead of 3

Read more about v1.9.2 >>

Version 1.9.1 (build 172), 19.07.2019:

  • accelerated the calculation of PageRank (now even for hundreds of thousands of pages the calculation of PageRank began to take a few seconds)
  • when calculating PageRank, weight transfer accounting for 301 redirects has been added
  • fixed bug with incorrect consideration of "robots.txt" rules
  • fixed a bug in which files like CSS and JS got into the database if scanning was disabled in the program settings
  • fixed a bug where outgoing links from a previously processed page remained in the outgoing links of the page
  • fixed a bug where the default number of PageRank calculation iterations was not saved in the program settings
  • optimized statistics of displaying duplicate headers H1-H6 in the section "SEO Statistics"
  • abolished stopping scanning when checking the BASE HREF tag, if they were not specified correctly
  • now, when the scanner is stopped due to a lack of RAM, the scan is paused and does not stop
  • fixed Access violation errors that periodically occur when closing the program
  • accelerated writing data to the database when scanning is stopped
  • accelerated removal of projects from the database
  • improved display of CANONICAL chains

Read more about v1.9.1 >>

Version 1.9 (build 154), 20.06.2019:

  • added the ability to scan a list of arbitrary URLs (using the clipboard or downloading a URL from a file on disk)
  • added the ability to scan Sitemap.xml files (classic Sitemap or index with a list of XML files)
  • added ability to select and copy cell values to the clipboard by Ctrl+A
  • speed up the operation of deleting projects (for complete removal of projects must be done by compressing the database through the program menu)
  • fixed problem with not always correct counting of empty H1 tags
  • fixed problem with not always correct parsing of <title> attribute in images
  • fixed program hang when moving through records during scanning

Read more about v1.9 >>

Version 1.8.3 (build 148), 16.05.2019:

  • added Dashboard tab, which displays a detailed report on the current site optimization quality (the report is generated based on the SEO Statistics tab)
  • added the ability to export Dashboard tab data to a handy PDF report
  • in the main settings of the program, the ability to disable the <base> tag accounting has been added
  • for more convenient cell copying, Shift and Ctrl events are added
  • when setting the parsing of HTML pages only, the mixing of other types of pages has been fixed
  • fixed not always correct handling of robots.txt
  • restored display of external links (outgoing links)

Read more about v1.8.3 >>

Version 1.8.2 (build 145), 23.04.2019:

  • added the ability to copy cell values to the clipboard when multiple elements are selected (main table + Info tab)
  • added visualization of loading projects at program start
  • added visualization of export projects in Excel/CSV
  • added automatic database cleaning when deleting projects
  • in the section "Site structure" added the display of the number of pages in folders
  • fixed not always correct data export to Excel/CSV
  • fixed freezes and emergency shutdowns of the program when navigating the Info tab

Read more about v1.8.2 >>

Version 1.8.1 (build 140), 09.04.2019:

  • in the list of projects added automatic grouping of subdomains relative to the main domain
  • added display of the source code of the page types: html, css, text, js (Info tab -> View Source)
  • the legend of the graph with data has become interactive and now also filters data based on the values of the legend
  • added the ability to export reports on the installed filters in the site master data table
  • added ability to export reports on the Info tab for the selected URL (incoming, outgoing, redirects, images, etc.)
  • added X-Robots-Tag column on the <All Links> tab
  • fixed incorrect handling of robots.txt
  • fixed HTTP/S protocol mapping for pages that actually do not have them
  • fixed checking of indexing pages in Yandex (any page was shown in the index, even if it is not)
  • fixed not always correct display of HREFLANG links
  • fixed incorrect display of response codes for Youtube-video
  • fixed incorrect display of external Nofollow links

Version 1.8 (build 136), 13.03.2019:

  • corrected not always correct scanning of sites using the HTTPS protocol (hangs and incorrect processing of a slash)
  • added correct operation of the scanner with sites that support the HTTP/2 protocol
  • corrected not always correct accounting tag CANONICAL when scanning pages of the site
  • corrected not always correct accounting ROBOTS.txt
  • added accounting for CANONICAL tag when exporting Sitemap.xml
  • fixed incorrect accounting of the site depth scan settings
  • corrected not always correct PREV and NEXT parsing
  • incorrect parsing of links containing uppercase characters has been fixed
  • on the tabs Images, Video and Documents added display of the corresponding files from external resources
  • in the Type of Technical Audit Content section, the bug of displaying statistics for elements displayed without specifying the content type was fixed
  • when a new version of the program is detected, a window is added with a display of the full change history for all time
  • fixed other minor bugs

Version 1.8 (build 135), 11.03.2019:

  • the section "SEO statistics" has been completely redesigned, intended for conducting a site audit
  • now SEO audit checks 50+ main SEO parameters and identifies 60+ key internal optimization errors
  • added function to calculate the internal PageRank for any page of the site
  • added accounting for X-Robots-Tag instructions in page headers
  • the work of the program with Cyrillic domains was restored (the problem arose due to incorrect processing of CANONICAL)
  • restored display of incoming links to pages with the response code 404 (Info tab)
  • restored display of all site redirects (in the previous version, some of them were not displayed)
  • fixed bug with exporting H1-H6 headers
  • fixed minor gaps in program localization

Read more about v1.8 >>

Version 1.7 (build 128), 11.12.2018:

  • added the ability to free check the indexing of pages in the Yandex by XML-limits using the Yandex XML and Majento.ru services
  • added the ability to pause projects during scanning and continue scanning after restarting the program
  • added parsing of the meta-tag rel="alternate" and displaying the contents of the tags "hreflang" and "media"
  • added the ability to parse websites that give data in Gzip format
  • corrected not always correct counting of pages in the site statistics scan filters

Read more about v1.7 >>

Version 1.6.2 (build 120), 13.11.2018:

  • added ability to work with proxy lists (group add, delete, work check)
  • fixed EOutofresourse exception when group scanning sites, when more than 1000 sites in project list

Version 1.6.1 (build 119), 31.10.2018:

  • added SEO-audit of the site by length title, description, h1-h6, the number of external and internal links, the length of content (Settings -> SEO)
  • added consideration of restrictions on the maximum number of redirects when parsing a site (Settings -> Basic)
  • added display chains of redirects to the final page (Info tab -> Redirects)
  • fixed duplication of pages when displaying duplicates in the "All" mode (via the context menu)
  • fixed incorrect encoding when parsing meta tags of some types of sites
  • fixed memory leaks when loading site icons

Version 1.6 (build 112), 09.10.2018:

  • added the ability to filter data on any field using the "quick" filter
  • for more detailed filtering, you can use a custom filter with advanced data selection settings
  • in the panel with additional data was added a tab of the technical statistics of the site (links, meta tags, page response codes, "robots" directives, etc.), as well as the tab of SEO statistics – ß-version (duplicates of meta tags, blank page headings , the presence of descriptions in the tags "alt") with the possibility of automatic filtering for each parameter
  • for the HEAD block added crawler attributes "Next / Prev" and "Refresh" with the ability to display the contents of these fields in the statistics of scanning
  • the date of the "Last change" column document is now displayed in the format of the current country
  • added "Snippets are only displayed for HTML format documents" warning for non-text/html format URI in the "Google snippet" tab
  • added icon mapping for Cyrillic domains
  • the maximum number of scan threads is increased to 100
  • fixed export to Sitemap.xml: it is now validated (ampersand & replaced with its HTML-entity "&amp;")
  • fixed problem with hangup when sorting data by clicking on column header

Read more about v1.6 >>

Version 1.5.1 (build 104), 08.08.2018:

  • increased speed of data export to Excel + eliminated bugs that occur when exporting large projects
  • increased speed of generating a sitemap "Sitemap.xml"
  • setting up the program for the "Scan" and "Exceptions" tabs became individual for each project
  • fixed a bug where objects with the type "text/javascript" were recognized as documents, rather than files like "Javascript"
  • fixed incorrect parsing of the meta-tag TITLE with the SVG-formatted image on the page (in which this tag is also present)
  • fixed incorrect sorting of the main site data in the table, boolean type output (for example, sorting in the HTTPS column)
  • added a pie chart showing the percentage of objects on the site (pages, images, styles, scripts, etc.)

Version 1.5 (build 101), 12.06.2018:

  • increased the speed of parsing websites, increased stability of the program
  • added the ability to download reports in Excel format (all results in one file, with tabs)
  • added parsing video (avi, mpeg, mov, links to Youtube, etc. formats). New column "Video" in the main report + tab
  • added parsing of documents (doc, xlsx, pdf, zip, mp3, etc. formats). New column "Documents" in the main report + tab
  • added highlighting of lines around content types (URL, styles, scripts, images, videos, documents, etc.)
  • added a visual display of the progress of data loading for "large" projects
  • added the ability to copy selected URLs in the list of domains
  • restored parsing and displaying files in *.css format
  • corrected incorrect accounting of the rules of the file "robots.txt" when parsing websites
  • optimized database structure, the removal of projects began to occur many times faster

Version 1.4.5 (build 96), 24.04.2018:

  • increased the speed of parsing sites (compared to the previous version, the speed increased up to 10 times)
  • fixed errors that occur periodically when writing data to the database
  • added the ability to select the types of analyzed objects (pages, images, styles, scripts)
  • added the ability to select multiple URLs and launch them to a re-scan, or group delete them
  • when exporting data, the automatic specification of the section name is added, which is unloaded (export_site.com_h1.csv)
  • when data exporting, was added request to overwrite existing files

Version 1.4.4 (build 91), 18.03.2018:

  • added the function of generating a site structure from parsed pages, with the ability to upload data to Excel
  • added the ability to compare the speed for each site on the chart

Version 1.4.3 (build 89), 26.02.2018:

  • it became possible to add sites to the list of projects
  • added display of data download speed when parsing sites (URI/sec)
  • added showing of favicon sites in the list of projects
  • restored the display of data in the "Info" tab

Version 1.4.2 (build 85), 11.02.2018:

  • added limit the number of pages scanned
  • added the ability to scan selected projects (scan several projects in turn)
  • added the ability to delete selected projects

Version 1.4.1 (build 83), 20.11.2017:

  • added parsing and displaying headers h1-h6, as well as their duplicates
  • added multilanguage support (the program became available in English, German, Italian and 14 other languages)
  • in the program interface the storage of the sizes of the left and right panel is added, and also the possibility of their full concealment
  • in the program settings are combined into one two items on Meta Robots (accounting NoIndex and NoFollow)
  • fixed incorrect definition of text encoding for some sites

Version 1.4 (build 79), 18.10.2017:

  • added an information panel with detailed information about the interesting URL (internal links, external, images, snippets)
  • added tab with displaying the contents of the meta-tag "robots" (index, nofollow)
  • added the ability to limit the time of scanning the project (indicated in hours)
  • corrected an incorrect indication of the level of nesting of pages (now it is counted from 0, not from 1, as it was before)
  • fixed minor errors in the scanner's logic, optimized program interface

Version 1.3.2 (build 76), 03.10.2017:

  • added display of hints for cells whose contents were larger than their width
  • the project re-scan now starts with the original URL entered, not from the root host, as before
  • fixed incorrect adherence to the directives "robots.txt", as well as the following Robots User-Agent

Version 1.3.1 (build 74), 28.09.2017:

  • added the ability to scan Cyrillic domains
  • added default sorting of the list of projects by name
  • incorrect registration of the contents of "robots.txt" when scanning a project
  • added follow-up to the rel="canonical" and "base href" directives on the site
  • added the ability to manually specify the number of pages to stake out "sitemap.xml" not a few files
  • added the storage of the active project when the program is closed and its load when it is subsequently launched
  • the correctness of logging user actions in the program was restored (section Log)
  • restored scan progress display on Taskbar

Version 1.3 (build 71), 10.09.2017:

  • completely redesigned source code, increased program speed
  • added multithreading (it became possible to manually specify the number of threads to scan the site)
  • changed the principle of storing data from PostgreSQL to SQLite (greatly reduced the size of the distribution)
  • added work through Proxy, the ability to specify the User-Agent, as well as the exclusion of certain sections of the site when scanning
  • added parsing of Java scripts and CSS-styles (previously they were not parse)

Version 1.2 (build 52), 22.02.2017:

  • fixed incorrect definition encoding TITLE
  • added the ability to pause the scan for later resumption ("Pause / Resume" button)
  • added tab HRAFLANG, showing the linguistic affiliation of pages (when available)
  • added displaying the contents of "the alt" and "title" images (the "Images")
  • added display of external links with the attribute NOFOLLOW (tab "External Links")
  • added display time scanning the site (located next to the progress bar)
  • fixed message "ZLibError" in the column "Status" for some sites
  • fixed zero time values in the column "Load Time"
  • added display of progress scan site to the taskbar

Version 1.1 (build 48), 05.02.2017:

  • fixed bugs that occur when parsing "TITLE" meta tag (for details of the sites displayed in wrong encoding)
  • added account "ROBOTS.TXT" file directives, as well as meta-tag "robots" when crawling pages on your site
  • accounting added rel="nofollow" when crawling pages on your site
  • added generation of cards "SITEMAP.XML" of the site (including "smart" cards splitting into multiple files on 50,000 pages)
  • added the ability to display the "duplicate" pages, meta tags and headers
  • given in the order interface, increased stability of the program

Version 1.0 (build 26), 20.11.2016:

  • the ability to scan all pages of the site (links pereobhod on site)
  • definition of title, description, keywords, h1-h2
  • statistics on the internal and external links
  • definition code server responses for downloaded files and pages
  • page load time definition
  • export reports to Excel (*.xls)
  • autosave data

Our Clients