Compare commits
10 Commits
f6e65c0305
...
49662557ca
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
49662557ca | ||
|
|
a1fb45f783 | ||
|
|
efa8f13e79 | ||
|
|
ac457a4e8d | ||
|
|
fb7ac11ae7 | ||
|
|
9f910d838d | ||
|
|
2d142844f4 | ||
|
|
e1d8ff73fa | ||
|
|
3dbadfc78a | ||
|
|
b8def30f55 |
@@ -43,7 +43,7 @@ XSS vulnerabilities are reported in xsscrapy-vulns.txt
|
||||
Dependencies
|
||||
-------
|
||||
``` shell
|
||||
wget -O https://bootstrap.pypa.io/get-pip.py
|
||||
wget -O -u https://bootstrap.pypa.io/get-pip.py
|
||||
python get-pip.py
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
Scrapy==1.1.0rc3
|
||||
pybloom==1.1
|
||||
scrapy
|
||||
pybloomfilter
|
||||
requests
|
||||
beautifulsoup
|
||||
bs4
|
||||
twisted==16.6.0
|
||||
w3lib
|
||||
lxml
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
from pybloom import BloomFilter
|
||||
from pybloomfilter import BloomFilter
|
||||
from scrapy.utils.job import job_dir
|
||||
from scrapy.dupefilters import BaseDupeFilter
|
||||
from settings import bloomfilterSize
|
||||
from xsscrapy.settings import bloomfilterSize
|
||||
|
||||
class BloomURLDupeFilter(BaseDupeFilter):
|
||||
"""Request Fingerprint duplicates filter"""
|
||||
|
||||
@@ -1,9 +1,9 @@
|
||||
from scrapy.exceptions import IgnoreRequest
|
||||
from urlparse import unquote
|
||||
from pybloom import BloomFilter
|
||||
from urllib.parse import unquote
|
||||
from pybloomfilter import BloomFilter
|
||||
import random
|
||||
import re
|
||||
from settings import bloomfilterSize
|
||||
from xsscrapy.settings import bloomfilterSize
|
||||
|
||||
# Filter out duplicate requests with Bloom filters since they're much easier on memory
|
||||
#URLS_FORMS_HEADERS = BloomFilter(3000000, 0.00001)
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
# Don't forget to add your pipeline to the ITEM_PIPELINES setting
|
||||
# See: http://doc.scrapy.org/en/latest/topics/item-pipeline.html
|
||||
from scrapy.exceptions import DropItem
|
||||
import HTMLParser
|
||||
from html.parser import HTMLParser
|
||||
from xsscrapy.items import vuln#, inj_resp
|
||||
import re
|
||||
import lxml.etree
|
||||
@@ -12,7 +12,7 @@ from lxml.html import soupparser, fromstring
|
||||
import itertools
|
||||
#from IPython import embed
|
||||
from socket import gaierror, gethostbyname
|
||||
from urlparse import urlparse
|
||||
from urllib.parse import urlparse
|
||||
from logging import CRITICAL, ERROR, WARNING, INFO, DEBUG
|
||||
|
||||
class XSSCharFinder(object):
|
||||
@@ -839,7 +839,7 @@ class XSSCharFinder(object):
|
||||
# Just make them useless by entering empty tag and putting them at the end of the lxml matches
|
||||
# so a split at tag won't find anything
|
||||
if not tag_index:
|
||||
print ' '*36+'ERROR: Error: could not find tag index location. Element does not exist in root doc.'
|
||||
print (' '*36+'ERROR: Error: could not find tag index location. Element does not exist in root doc.')
|
||||
tag_index = 999999999
|
||||
tag = ''
|
||||
loc_tag = (tag_index, tag)
|
||||
|
||||
Reference in New Issue
Block a user