From 756855a0d511c6e537a534e4dc639672f3ec7248 Mon Sep 17 00:00:00 2001 From: Kr1ss Date: Mon, 20 Apr 2020 17:40:26 +0200 Subject: update changelog --- .SRCINFO | 2 +- ChangeLog | 194 +++----------------------------------------------------------- PKGBUILD | 2 +- 3 files changed, 9 insertions(+), 189 deletions(-) diff --git a/.SRCINFO b/.SRCINFO index 91f3e9f..b36817d 100644 --- a/.SRCINFO +++ b/.SRCINFO @@ -1,7 +1,7 @@ pkgbase = wapiti pkgdesc = A comprehensive web app vulnerability scanner written in Python pkgver = 3.0.3 - pkgrel = 1 + pkgrel = 2 url = http://wapiti.sourceforge.net/ changelog = ChangeLog arch = any diff --git a/ChangeLog b/ChangeLog index 25d0b58..75cba8b 100644 --- a/ChangeLog +++ b/ChangeLog @@ -1,3 +1,9 @@ +20/02/2020 + Wapiti 3.0.3 + An important work was made to reduce false positives in XSS detections. + That research involved scanning more than 1 million websites to discover those issues. + More details here: http://devloop.users.sourceforge.net/index.php?article217/one-crazy-month-of-web-vulnerability-scanning + 02/09/2019 Wapiti 3.0.2 New XXE module cans end payloads in parameters, query string, file uploads and raw body. @@ -26,7 +32,7 @@ Fixed issue #54 in lamejs JS parser. Removed lxml and libxml2 as a dependency. That parser have difficulties to parse exotic encodings. -03/01/2018 +03/01/2017 Release of Wapiti 3.0.0 02/01/2018 @@ -298,189 +304,3 @@ 25/04/2006: Version 1.0.0 -03/01/2018 - Release of Wapiti 3.0.0 - -23/12/2017 - lswww is now renamed to Crawler. - All HTML parsing is now made with BeautifulSoup. lxml should be the parsing engine but it's possible to opt-out at - setup with --html5lib. - Analysis on JS in event handlers (onblur, onclick, etc) - Changed behavior ot 'page' scope, added 'url' scope. - Default mime type used for upload fields is image/gif. - Added yaswf as a dependency for SWF parsing. - Custom HTTP error codes check. - Fixed a bug with 'button' input types. - Updated pynarcissus with a python3 version for js parsing. - Rewrote "in scope" check. - -29/12/2009 - Version 2.3.1 - Fixed a bug in lswww if root url is not given complete. - Fixed a bug in lswww with a call to BeautifulSoup made on non text files. - Fixed a bug that occured when verbosity = 2. Unicode error on stderr. - -27/12/2009 - Version 2.3.0 - Internationalization and translation to english and spanish when called from - Wapiti. - Ability to save a scan session and restore it later (-i) - Added option -b to set the scope of the scan based on the root url given as - argument. - Fixed bug ID 2779441 "Python Version 2.5 required?" - Use an home made cookie library instead or urllib2's one. - Keep aditionnal informations on the webpages (headers + encoding) - Use BeautifulSoup to detect webpage encoding and handle parsing errors. - Fixed a bug when "a href" or "form action" have an empty string as value. - Better support of Unicode. - -26/03/2009 - Version 2.2.0 - Fixed bug ID 2433127 with HTTP 404 error codes. - Don't let httplib2 manage HTTP redirections : return the status code - and let lswww handle the new url. - -25/03/2009 - Version 2.1.9 - Added option -e (or --export) - Saves urls and forms data to a XML file. - We hope other fuzzers will allow importation of this file. - -24/03/2009 - More verifications on timeout errors. - -22/03/2009 - Version 2.1.8 - Fixed bug ID: 2415094 - Check on protocol found in hyperlinks was case-sentitive. - Moved it to non-case-sensitive. - Integration of a second linkParser class called linkParser2 from - lswwwv2.py. This parser use only regexp to extract links and forms. - -25/11/2008 - httplib2 use lowercase names for the HTTP headers in opposition to - urllib2 (first letter was uppercase). - Changed the verifications on headers. - -15/11/2008 - Fixed a bug with links going to parrent directory. - -02/11/2008 - Better integration of proxy support provided by httplib2. - It's now possible to use SOCKS proxies. - -19/10/2008 - Version 2.1.7 - Now use httplib2 (http://code.google.com/p/httplib2/)n MIT licence - instead of urllib2. - The ability to use persistents connections makes the scan faster. - -09/10/2008 - Version 2.1.6 - HTTP authentification now works - Added the option -n (or --nice) to prevent endless loops during scanning - -28/01/2007 - Version 2.1.5 - First take a look at the Content-Type instead of the document extension - Added BeautifulSoup as an optionnal module to correct bad html documents - (better use tidy if you can) - -24/10/2006 - Version 2.1.4 - Wildcard exclusion with -x (--exclude) option - -22/10/2006 - Fixed an error with url parameters handling that appeared in precedent - version. - Fixed a typo in lswww.py (setAuthCreddentials : one 'd' is enough) - -07/10/2006 - Version 2.1.3 - Three verbose mode with -v (--verbose) option - 0: print only results - 1: print dots for each page accessed (default mode) - 2: print each found url durring scan - Timeout in seconds can be set with -t (--timeout) option - Fixed bug "crash when no content-type is returned" - Fixed an error with 404 webpages - Fixed a bug when the only parameter of an url is a forbidden one - -09/08/2006 - Version 2.1.2 - Fixed a bug with regular expressions - -05/08/2006 - Version 2.1.1 - Remove redundant slashes from urls - (e.g. http://server/dir//page.php converted to - http://server/dir/page.php) - -20/07/2006 - Version 2.1.0 with urllib2 - -11/07/2006 - -r (--remove) option to remove parameters from URLs - Generate URL with GET forms instead of using POST by default - Support for Basic HTTP Auth added but don't work with Python 2.4. - Now use cookie files (option "-c file" or "--cookie file") - Extracts links from Location header fields - - -06/07/2006 - Extract links from "Location:" headers (HTTP 301 and 302) - Default type for "input" elements is set to "text" - (as written in the HTML 4.0 specifications) - Added "search" in input types (created for Safari browsers) - -04/07/2006 - Fixed a bug with empty parameters tuples - (convert http://server/page?&a=2 to http://server/page?a=2) - -23/06/2006 - Version 2.0.1 - Take care of the "submit" type - No extra data sent when a page contains several forms - Corrected a bug with urls finishing by '?' - Support Cookies !! - -25/04/2006 - Version 2.0 - Extraction des formulaires sous la forme d'une liste de tuples - contenant chacun un string (url du script cible) et un dict - contenant les noms des champs et leur valeur par d�faut (ou 'true' - si vide) - Recense les scripts gerant l'upload - Peut maintenant fonctionner comme module - -19/04/2006 - Version 1.1 - Lecture des tags insensible a la casse - Gestion du Ctrl+C pour interrompre proprement le programme - Extraction des urls dans les balises form (action) - -12/10/2005 - Version 1.0 - Gestion des liens syntaxiquement valides mais pointant - vers des ressources inexistantes (404) - -11/09/2005 - Beta4 - Utilisation du module getopt qui permet de specifier - facilement les urls a visiter en premier, les urls a - exclure (nouveau !) ou encore le proxy a utiliser - -24/08/2005 - Beta3 - Ajout d'un timeout pour la lecture des pages pour ne pas - bloquer sur un script bugge - -23/08/2005 - Version beta2 - Prise en charge des indexs generes par Apache - Filtre sur les protocoles - Gestion des liens qui remontent l'arborescence - Gestion des liens vides - -02/08/2005 - Sortie de la beta1 diff --git a/PKGBUILD b/PKGBUILD index 003ed40..ed59f67 100644 --- a/PKGBUILD +++ b/PKGBUILD @@ -5,7 +5,7 @@ pkgname=wapiti pkgver=3.0.3 -pkgrel=1 +pkgrel=2 pkgdesc='A comprehensive web app vulnerability scanner written in Python' arch=('any') -- cgit v1.2.3-70-g09d2