wespiva — Web Spider Validator

Deutsche Textversion anzeigen

Web Spider Validator, short named wespiva, is a mix of a

  1. Web-Spider (Robot, Crawler) , which traverses between web pages linked together,
  2. and an XHTML-Validator, which proofs whether a page contains valid tags, attributes and allowed attribute-values.



The purpose of this tool is to ensure high-quality standard-complying websites.
With xenu's link sleuth there is a great tool for spidering and finding dead links, but it does not validate a page.
With the w3.org-Validator there is a great validation-tool, but it only checks a single page, and is often overloaded and slow.
The solution to overcome these restrictions is wespiva, which spiders and validates in one rush. This tool assists in the transition of bigger sites to XHTML.


Click here to download wespiva Version 4.2012.6.21689 (166 kb ZIP-File, 2012-06-29)

In spite of being programmed not to harm any computer, there is a chance of a crash by accident or programming-error in the application or one of the .NET-functions used by it. In order not to be held liable for any negative circumstances resulting of the usage of this program (like time losses, data loss, wrong reports, etc.), you may only use the program when accepting the following rules:



wespiva runs on Windows with .NET Framework 4.0 installed.

How to run

Just unzip the single file in the zip-archive and start using it.

Frequently asked questions for wespiva

Does it run on MONO for Windows?
A special version runs on Mono 2.2, but hangs when resizing the form while wespiva is spidering. The reason is unknown, possibly Mono has some bugs with Windows.Forms and Multithreading. If you don't touch the App until the scan is over, all is well.
Will there be a MONO-Version for Linux/OS-X?
Probably yes—if someone pays for it. If no one would pay for it, there is no big demand for it.
How much pages could be checked in one run?
I've used it to check sites with more than 50.000 elements in less than 15 minutes. The duration depends mainly on the line-speed and responsiveness of the page-delivering web server.
Why Validation?
I'll let others speak here:


Main page with progress log wespiva main page
Page list wespiva main page
Option dialog wespiva main page
Report sample
Sitemap sample


runnable from command-line

			c:\wespiva.exe "www.wissing.com" "example@example.not"

Known Bugs

Future Features

Already done:

History / Changes

Other nice Validators

They are really good, but don't let you check whole sites: