mirror of
https://github.com/adulau/DomainClassifier.git
synced 2024-11-22 18:07:07 +00:00
Merge branch 'master' of github.com:adulau/DomainClassifier
This commit is contained in:
commit
bcf5f88a48
4 changed files with 178 additions and 65 deletions
|
@ -1 +1 @@
|
|||
__version__ = "1.0"
|
||||
__version__ = "1.1"
|
||||
|
|
|
@ -7,9 +7,15 @@ attributes.
|
|||
import re
|
||||
import dns.resolver
|
||||
import IPy
|
||||
import redis
|
||||
import socket
|
||||
import time
|
||||
from datetime import date, timedelta
|
||||
import os
|
||||
import sys
|
||||
from uuid import uuid4
|
||||
|
||||
from multiprocessing import Process as Proc
|
||||
|
||||
try:
|
||||
# python 3
|
||||
|
@ -23,9 +29,9 @@ try:
|
|||
except:
|
||||
print("pybgpranking is not installed - ranking of ASN values won't be possible")
|
||||
__author__ = "Alexandre Dulaunoy"
|
||||
__copyright__ = "Copyright 2012-2021, Alexandre Dulaunoy"
|
||||
__copyright__ = "Copyright 2012-2024, Alexandre Dulaunoy"
|
||||
__license__ = "AGPL version 3"
|
||||
__version__ = "0.9"
|
||||
__version__ = "1.1"
|
||||
|
||||
|
||||
class Extract:
|
||||
|
@ -34,15 +40,26 @@ class Extract:
|
|||
from a rawtext stream. When call, the rawtext parameter is a string
|
||||
containing the raw data to be process."""
|
||||
|
||||
def __init__(self, rawtext=None, nameservers=['8.8.8.8'], port= 53):
|
||||
def __init__(self, rawtext=None, nameservers=['8.8.8.8'], port=53, redis_host='', redis_port=6379, redis_db=0, expire_time=3600, re_timeout=-1):
|
||||
self.rawtext = rawtext
|
||||
self.presolver = dns.resolver.Resolver()
|
||||
self.presolver.nameservers = nameservers
|
||||
self.presolver.port = 53
|
||||
self.presolver.port = port
|
||||
self.presolver.lifetime = 1.0
|
||||
self.bgprankingserver = 'pdns.circl.lu'
|
||||
self.vdomain = []
|
||||
self.listtld = []
|
||||
|
||||
self.re_domain = re.compile(r'\b([a-zA-Z\d-]{,63}(\.[a-zA-Z\d-]{,63})+)\b')
|
||||
|
||||
if redis_host and redis_port:
|
||||
self.redis = redis.StrictRedis(host=redis_host, port=redis_port, db=redis_db, decode_responses=True)
|
||||
self.uuid = str(uuid4())
|
||||
self.re_timeout = re_timeout
|
||||
else:
|
||||
self.redis = None
|
||||
self.expire_time = expire_time
|
||||
|
||||
self.domain = self.potentialdomain()
|
||||
|
||||
"""__origin is a private function to the ASN lookup for an IP address via
|
||||
|
@ -52,7 +69,11 @@ class Extract:
|
|||
def __origin(self, ipaddr=None):
|
||||
|
||||
if ipaddr:
|
||||
clook = IPy.IP(str(ipaddr)).reverseName().replace('.in-addr.arpa.', '.origin.asn.cymru.com')
|
||||
clook = (
|
||||
IPy.IP(str(ipaddr))
|
||||
.reverseName()
|
||||
.replace('.in-addr.arpa.', '.origin.asn.cymru.com')
|
||||
)
|
||||
try:
|
||||
a = self.presolver.query(clook, 'TXT')
|
||||
except dns.resolver.NXDOMAIN:
|
||||
|
@ -66,19 +87,40 @@ class Extract:
|
|||
return (x[0], x[2], a[0])
|
||||
else:
|
||||
return None
|
||||
|
||||
"""__bgpanking return the ranking the float value of an ASN.
|
||||
"""
|
||||
|
||||
def __bgpranking(self, asn=None):
|
||||
if asn:
|
||||
bgpranking = BGPRanking()
|
||||
value = bgpranking.query(asn, date=(date.today() - timedelta(1)).isoformat())
|
||||
value = bgpranking.query(
|
||||
asn, date=(date.today() - timedelta(1)).isoformat()
|
||||
)
|
||||
return value['response']['ranking']['rank']
|
||||
|
||||
def __updatelisttld(self):
|
||||
def __updatelisttld(self, force=False):
|
||||
ianatldlist = "https://data.iana.org/TLD/tlds-alpha-by-domain.txt"
|
||||
userdir = os.path.expanduser("~")
|
||||
cachedir = os.path.join(userdir, ".DomainClassifier")
|
||||
if not os.path.exists(cachedir):
|
||||
os.mkdir(cachedir)
|
||||
tldcache = os.path.join(cachedir, "tlds")
|
||||
if not os.path.exists(tldcache):
|
||||
print(tldcache)
|
||||
req = urllib.Request(ianatldlist)
|
||||
req.add_header('User-Agent', 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:54.0) Gecko/20100101 Firefox/54.0')
|
||||
req.add_header(
|
||||
'User-Agent',
|
||||
'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:120.0) Gecko/20100101 Firefox/120.0',
|
||||
)
|
||||
tlds = (urllib.urlopen(req).read()).decode('utf8')
|
||||
f = open(tldcache, "wb")
|
||||
f.write(tlds.encode("utf-8"))
|
||||
f.close()
|
||||
|
||||
f = open(tldcache, "r")
|
||||
tlds = f.read()
|
||||
f.close()
|
||||
tlds = tlds.split("\n")
|
||||
for tld in tlds:
|
||||
self.listtld.append(tld.lower())
|
||||
|
@ -97,21 +139,51 @@ class Extract:
|
|||
|
||||
return self.cleandomain
|
||||
|
||||
def text(self, rawtext=False):
|
||||
def __re_findall(self, rawtext):
|
||||
for x in re.findall(self.re_domain, rawtext):
|
||||
if x[0]:
|
||||
self.redis.sadd('cache:regex:{}'.format(self.uuid), x[0])
|
||||
self.redis.expire('cache:regex:{}'.format(self.uuid), 360)
|
||||
|
||||
def __regex_findall(self, rawtext, timeout):
|
||||
proc = Proc(target=self.__re_findall, args=(rawtext,))
|
||||
try:
|
||||
proc.start()
|
||||
proc.join(timeout)
|
||||
if proc.is_alive():
|
||||
proc.terminate()
|
||||
print('regex: processing timeout')
|
||||
return []
|
||||
else:
|
||||
domains = self.redis.smembers('cache:regex:{}'.format(self.uuid))
|
||||
self.redis.delete('cache:regex:{}'.format(self.uuid))
|
||||
proc.terminate()
|
||||
return domains
|
||||
except KeyboardInterrupt:
|
||||
print("Caught KeyboardInterrupt, terminating workers")
|
||||
proc.terminate()
|
||||
sys.exit(0)
|
||||
|
||||
def text(self, rawtext=''):
|
||||
if rawtext:
|
||||
self.rawtext = rawtext
|
||||
self.domain = self.potentialdomain()
|
||||
self.vdomain = []
|
||||
return True
|
||||
return False
|
||||
|
||||
"""potentialdomain method extracts potential domains matching any
|
||||
string that is a serie of string with maximun 63 character separated by a
|
||||
string that is a serie of string with maximum 63 character separated by a
|
||||
dot. The method used the rawtext defined at the instantiation of the class.
|
||||
This return a list of a potential domain."""
|
||||
|
||||
def potentialdomain(self, validTLD=True):
|
||||
self.domain = []
|
||||
domain = re.compile(r'\b([a-zA-Z\d-]{,63}(\.[a-zA-Z\d-]{,63})+)\b')
|
||||
for x in domain.findall(self.rawtext):
|
||||
if self.re_timeout > 0 and self.redis:
|
||||
self.domain = list(self.__regex_findall(self.rawtext, self.re_timeout))
|
||||
else:
|
||||
domains = self.re_domain.findall(self.rawtext)
|
||||
for x in domains:
|
||||
if x[0]:
|
||||
self.domain.append(x[0])
|
||||
if validTLD:
|
||||
|
@ -124,31 +196,66 @@ class Extract:
|
|||
returns a list of existing domain. If the extended flag is true, a set is
|
||||
return with the associated DNS resources found."""
|
||||
|
||||
def validdomain(self, rtype=['A', 'AAAA', 'SOA', 'MX', 'CNAME'], extended=True, passive_dns=False):
|
||||
def validdomain(
|
||||
self,
|
||||
rtype=['A', 'AAAA', 'SOA', 'MX', 'CNAME'],
|
||||
extended=True,
|
||||
passive_dns=False,
|
||||
):
|
||||
if extended is False:
|
||||
self.vdomain = set()
|
||||
else:
|
||||
self.vdomain = []
|
||||
|
||||
for domain in self.domain:
|
||||
if self.redis:
|
||||
if self.redis.exists('dom_class:cache:{}'.format(domain)):
|
||||
passive_dns_out = self.redis.smembers('dom_class:cache:{}'.format(domain))
|
||||
for out in passive_dns_out:
|
||||
if extended:
|
||||
out = tuple(out.split('[^]', 2))
|
||||
self.vdomain.append(out)
|
||||
else:
|
||||
self.vdomain.add(out)
|
||||
else:
|
||||
|
||||
for dnstype in rtype:
|
||||
try:
|
||||
answers = self.presolver.query(domain, dnstype)
|
||||
except:
|
||||
pass
|
||||
else:
|
||||
# Pasive DNS output
|
||||
# Passive DNS output
|
||||
# timestamp||dns-client ||dns-server||RR class||Query||Query Type||Answer||TTL||Count
|
||||
if passive_dns:
|
||||
rrset = answers.rrset.to_text().splitlines()
|
||||
for dns_resp in rrset:
|
||||
dns_resp = dns_resp.split()
|
||||
passive_dns_out = '{}||127.0.0.1||{}||{}||{}||{}||{}||{}||1\n'.format(time.time(), self.presolver.nameservers[0], dns_resp[2], domain, dnstype, dns_resp[4], answers.ttl)
|
||||
passive_dns_out = (
|
||||
'{}||127.0.0.1||{}||{}||{}||{}||{}||{}||1\n'.format(
|
||||
time.time(),
|
||||
self.presolver.nameservers[0],
|
||||
dns_resp[2],
|
||||
domain,
|
||||
dnstype,
|
||||
dns_resp[4],
|
||||
answers.ttl,
|
||||
)
|
||||
)
|
||||
self.vdomain.add((passive_dns_out))
|
||||
if self.redis:
|
||||
self.redis.sadd('dom_class:cache:{}'.format(domain), passive_dns_out)
|
||||
self.redis.expire('dom_class:cache:{}'.format(domain), self.expire_time)
|
||||
elif extended:
|
||||
self.vdomain.append((domain, dnstype, answers[0]))
|
||||
if self.redis:
|
||||
self.redis.sadd('dom_class:cache:{}'.format(domain), '{}[^]{}[^]{}'.format(domain, dnstype, answers[0]))
|
||||
self.redis.expire('dom_class:cache:{}'.format(domain), self.expire_time)
|
||||
else:
|
||||
self.vdomain.add((domain))
|
||||
if self.redis:
|
||||
self.redis.sadd('dom_class:cache:{}'.format(domain), domain)
|
||||
self.redis.expire('dom_class:cache:{}'.format(domain), self.expire_time)
|
||||
return self.vdomain
|
||||
|
||||
"""ipaddress method extracts from the domain list the valid IPv4 addresses"""
|
||||
|
@ -188,7 +295,7 @@ class Extract:
|
|||
orig = self.__origin(ipaddr=dom[2])[1]
|
||||
except:
|
||||
continue
|
||||
if(orig == cc):
|
||||
if orig == cc:
|
||||
self.localdom.append(dom)
|
||||
elif dom[1] == 'CNAME':
|
||||
cname = str(dom[2])
|
||||
|
@ -197,7 +304,7 @@ class Extract:
|
|||
orig = self.__origin(ipaddr=ip)[1]
|
||||
except:
|
||||
continue
|
||||
if(orig == cc):
|
||||
if orig == cc:
|
||||
self.localdom.append(dom)
|
||||
return self.localdom
|
||||
|
||||
|
@ -280,9 +387,14 @@ class Extract:
|
|||
|
||||
return set(self.cleandomain)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
c = Extract(rawtext="www.foo.lu www.xxx.com this is a text with a domain called test@foo.lu another test abc.lu something a.b.c.d.e end of 1.2.3.4 foo.be www.belnet.be http://www.cert.be/ www.public.lu www.allo.lu quuxtest www.eurodns.com something-broken-www.google.com www.google.lu trailing test www.facebook.com www.nic.ru www.youporn.com 8.8.8.8 201.1.1.1 abc.dontexist")
|
||||
c.text(rawtext="www.abc.lu www.xxx.com random text a test bric broc www.lemonde.fr www.belnet.be www.foo.be")
|
||||
c = Extract(
|
||||
rawtext="www.foo.lu www.xxx.com this is a text with a domain called test@foo.lu another test abc.lu something a.b.c.d.e end of 1.2.3.4 foo.be www.belnet.be http://www.cert.be/ www.public.lu www.allo.lu quuxtest www.eurodns.com something-broken-www.google.com www.google.lu trailing test www.facebook.com www.nic.ru www.youporn.com 8.8.8.8 201.1.1.1 abc.dontexist"
|
||||
)
|
||||
c.text(
|
||||
rawtext="www.abc.lu www.xxx.com random text a test bric broc www.lemonde.fr www.belnet.be www.foo.be"
|
||||
)
|
||||
print(c.potentialdomain())
|
||||
print(c.potentialdomain(validTLD=True))
|
||||
print(c.validdomain(extended=True))
|
||||
|
|
|
@ -1,3 +1,4 @@
|
|||
IPy
|
||||
dnspython
|
||||
git+https://github.com/D4-project/BGP-Ranking.git/#egg=pybgpranking&subdirectory=client
|
||||
redis
|
||||
|
|
2
setup.py
2
setup.py
|
@ -1,7 +1,7 @@
|
|||
from setuptools import setup, find_packages
|
||||
setup(
|
||||
name="DomainClassifier",
|
||||
version="1.0",
|
||||
version="1.1",
|
||||
packages=find_packages(),
|
||||
install_requires=['dnspython', 'IPy', 'pybgpranking'],
|
||||
dependency_links=[
|
||||
|
|
Loading…
Reference in a new issue