We need it to write out the source passwords as a reference as some are removed while processing if they don't fit the model that rulegen is good at (such as all-numerical passwords).
1104 lines
50 KiB
Python
Executable File
1104 lines
50 KiB
Python
Executable File
#!/usr/bin/env python
|
|
# Rulegen.py - Advanced automated password rule and wordlist generator for the
|
|
# Hashcat password cracker using the Levenshtein Reverse Path
|
|
# algorithm and Enchant spell checking library.
|
|
#
|
|
# This tool is part of PACK (Password Analysis and Cracking Kit)
|
|
#
|
|
# VERSION 0.0.3
|
|
#
|
|
# Copyright (C) 2013 Peter Kacherginsky
|
|
# All rights reserved.
|
|
#
|
|
# Please see the attached LICENSE file for additional licensing information.
|
|
|
|
from collections import Counter
|
|
from optparse import OptionParser, OptionGroup
|
|
import enchant
|
|
import multiprocessing
|
|
import operator
|
|
import re
|
|
import subprocess
|
|
import sys
|
|
import time
|
|
import logging
|
|
|
|
logging.VERBOSE = 15
|
|
logging.addLevelName(logging.VERBOSE, "VERBOSE")
|
|
logging.Logger.verbose = lambda self, msg, *args, **kwargs: \
|
|
self.log(logging.VERBOSE, msg, *args, **kwargs)
|
|
|
|
LOGGER = logging.getLogger("rulegen")
|
|
|
|
VERSION = "0.0.4"
|
|
|
|
# Testing rules with hashcat --stdout
|
|
HASHCAT_PATH = "hashcat/"
|
|
|
|
# Rule Generator class responsible for the complete cycle of rule generation
|
|
class RuleGen:
|
|
|
|
# Initialize Rule Generator class
|
|
def __init__(self,language="en",providers="aspell,myspell",basename='analysis',threads=multiprocessing.cpu_count()):
|
|
|
|
self.threads = threads
|
|
|
|
self.enchant_broker = enchant.Broker()
|
|
self.enchant_broker.set_ordering("*",providers)
|
|
|
|
self.enchant = enchant.Dict(language, self.enchant_broker)
|
|
|
|
# Output options
|
|
self.basename = basename
|
|
|
|
# Finetuning word generation
|
|
self.max_word_dist = 10
|
|
self.max_words = 10
|
|
self.more_words = False
|
|
self.simple_words = False
|
|
|
|
# Finetuning rule generation
|
|
self.max_rule_len = 10
|
|
self.max_rules = 10
|
|
self.more_rules = False
|
|
self.simple_rules = False
|
|
self.brute_rules = False
|
|
|
|
# Debugging options
|
|
self.word = None # Custom word to use.
|
|
self.quiet = False
|
|
|
|
########################################################################
|
|
# Word and Rule Statistics
|
|
self.numeric_stats_total = 0
|
|
self.special_stats_total = 0
|
|
self.foreign_stats_total = 0
|
|
|
|
########################################################################
|
|
# Preanalysis Password Patterns
|
|
self.password_pattern = dict()
|
|
self.password_pattern["insertion"] = re.compile(r'^[^a-z]*(?P<password>.+?)[^a-z]*$', re.IGNORECASE)
|
|
self.password_pattern["email"] = re.compile(r'^(?P<password>.+?)@[A-Z0-9.-]+\.[A-Z]{2,4}', re.IGNORECASE)
|
|
self.password_pattern["alldigits"] = re.compile(r'^(\d+)$', re.IGNORECASE)
|
|
self.password_pattern["allspecial"]= re.compile(r'^([^a-z0-9]+)$', re.IGNORECASE)
|
|
|
|
########################################################################
|
|
# Hashcat Rules Engine
|
|
self.hashcat_rule = dict()
|
|
|
|
# Dummy rule
|
|
self.hashcat_rule[':'] = lambda x: x # Do nothing
|
|
|
|
# Case rules
|
|
self.hashcat_rule["l"] = lambda x: x.lower() # Lowercase all letters
|
|
self.hashcat_rule["u"] = lambda x: x.upper() # Capitalize all letters
|
|
self.hashcat_rule["c"] = lambda x: x.capitalize() # Capitalize the first letter
|
|
self.hashcat_rule["C"] = lambda x: x[0].lower() + x[1:].upper() # Lowercase the first found character, uppercase the rest
|
|
self.hashcat_rule["t"] = lambda x: x.swapcase() # Toggle the case of all characters in word
|
|
self.hashcat_rule["T"] = lambda x,y: x[:y] + x[y].swapcase() + x[y+1:] # Toggle the case of characters at position N
|
|
self.hashcat_rule["E"] = lambda x: " ".join([i[0].upper()+i[1:] for i in x.split(" ")]) # Upper case the first letter and every letter after a space
|
|
|
|
# Rotation rules
|
|
self.hashcat_rule["r"] = lambda x: x[::-1] # Reverse the entire word
|
|
self.hashcat_rule["{"] = lambda x: x[1:]+x[0] # Rotate the word left
|
|
self.hashcat_rule["}"] = lambda x: x[-1]+x[:-1] # Rotate the word right
|
|
|
|
# Duplication rules
|
|
self.hashcat_rule["d"] = lambda x: x+x # Duplicate entire word
|
|
self.hashcat_rule["p"] = lambda x,y: x*y # Duplicate entire word N times
|
|
self.hashcat_rule["f"] = lambda x: x+x[::-1] # Duplicate word reversed
|
|
self.hashcat_rule["z"] = lambda x,y: x[0]*y+x # Duplicate first character N times
|
|
self.hashcat_rule["Z"] = lambda x,y: x+x[-1]*y # Duplicate last character N times
|
|
self.hashcat_rule["q"] = lambda x: "".join([i+i for i in x]) # Duplicate every character
|
|
self.hashcat_rule["y"] = lambda x,y: x[:y]+x # Duplicate first N characters
|
|
self.hashcat_rule["Y"] = lambda x,y: x+x[-y:] # Duplicate last N characters
|
|
|
|
# Cutting rules
|
|
self.hashcat_rule["["] = lambda x: x[1:] # Delete first character
|
|
self.hashcat_rule["]"] = lambda x: x[:-1] # Delete last character
|
|
self.hashcat_rule["D"] = lambda x,y: x[:y]+x[y+1:] # Deletes character at position N
|
|
self.hashcat_rule["'"] = lambda x,y: x[:y] # Truncate word at position N
|
|
self.hashcat_rule["x"] = lambda x,y,z: x[:y]+x[y+z:] # Delete M characters, starting at position N
|
|
self.hashcat_rule["@"] = lambda x,y: x.replace(y,'') # Purge all instances of X
|
|
|
|
# Insertion rules
|
|
self.hashcat_rule["$"] = lambda x,y: x+y # Append character to end
|
|
self.hashcat_rule["^"] = lambda x,y: y+x # Prepend character to front
|
|
self.hashcat_rule["i"] = lambda x,y,z: x[:y]+z+x[y:] # Insert character X at position N
|
|
|
|
# Replacement rules
|
|
self.hashcat_rule["o"] = lambda x,y,z: x[:y]+z+x[y+1:] # Overwrite character at position N with X
|
|
self.hashcat_rule["s"] = lambda x,y,z: x.replace(y,z) # Replace all instances of X with Y
|
|
self.hashcat_rule["L"] = lambda x,y: x[:y]+chr(ord(x[y])<<1)+x[y+1:] # Bitwise shift left character @ N
|
|
self.hashcat_rule["R"] = lambda x,y: x[:y]+chr(ord(x[y])>>1)+x[y+1:] # Bitwise shift right character @ N
|
|
self.hashcat_rule["+"] = lambda x,y: x[:y]+chr(ord(x[y])+1)+x[y+1:] # Increment character @ N by 1 ascii value
|
|
self.hashcat_rule["-"] = lambda x,y: x[:y]+chr(ord(x[y])-1)+x[y+1:] # Decrement character @ N by 1 ascii value
|
|
self.hashcat_rule["."] = lambda x,y: x[:y]+x[y+1]+x[y+1:] # Replace character @ N with value at @ N plus 1
|
|
self.hashcat_rule[","] = lambda x,y: x[:y]+x[y-1]+x[y+1:] # Replace character @ N with value at @ N minus 1
|
|
|
|
# Swappping rules
|
|
self.hashcat_rule["k"] = lambda x: x[1]+x[0]+x[2:] # Swap first two characters
|
|
self.hashcat_rule["K"] = lambda x: x[:-2]+x[-1]+x[-2] # Swap last two characters
|
|
self.hashcat_rule["*"] = lambda x,y,z: x[:y]+x[z]+x[y+1:z]+x[y]+x[z+1:] if z > y else x[:z]+x[y]+x[z+1:y]+x[z]+x[y+1:] # Swap character X with Y
|
|
|
|
########################################################################
|
|
# Common numeric and special character substitutions (1337 5p34k)
|
|
self.leet = dict()
|
|
self.leet["1"] = "i"
|
|
self.leet["2"] = "z"
|
|
self.leet["3"] = "e"
|
|
self.leet["4"] = "a"
|
|
self.leet["5"] = "s"
|
|
self.leet["6"] = "b"
|
|
self.leet["7"] = "t"
|
|
self.leet["8"] = "b"
|
|
self.leet["9"] = "g"
|
|
self.leet["0"] = "o"
|
|
self.leet["!"] = "i"
|
|
self.leet["|"] = "i"
|
|
self.leet["@"] = "a"
|
|
self.leet["$"] = "s"
|
|
self.leet["+"] = "t"
|
|
|
|
########################################################################
|
|
# Preanalysis rules to bruteforce for each word
|
|
self.preanalysis_rules = []
|
|
self.preanalysis_rules.append(([],self.hashcat_rule[':'])) # Blank rule
|
|
self.preanalysis_rules.append((['r'],self.hashcat_rule['r'])) # Reverse rule
|
|
#self.preanalysis_rules.append((['{'],self.hashcat_rule['}'])) # Rotate left
|
|
#self.preanalysis_rules.append((['}'],self.hashcat_rule['{'])) # Rotate right
|
|
|
|
############################################################################
|
|
# Calculate Levenshtein edit path matrix
|
|
def levenshtein(self,word,password):
|
|
matrix = []
|
|
|
|
# Generate and populate the initial matrix
|
|
for i in range(len(password) + 1):
|
|
matrix.append([])
|
|
for j in range(len(word) + 1):
|
|
if i == 0:
|
|
matrix[i].append(j)
|
|
elif j == 0:
|
|
matrix[i].append(i)
|
|
else:
|
|
matrix[i].append(0)
|
|
|
|
# Calculate edit distance for each substring
|
|
for i in range(1,len(password) + 1):
|
|
for j in range(1,len(word) + 1):
|
|
if password[i-1] == word[j-1]:
|
|
matrix[i][j] = matrix[i-1][j-1]
|
|
else:
|
|
insertion = matrix[i-1][j] + 1
|
|
deletion = matrix[i][j-1] + 1
|
|
substitution = matrix[i-1][j-1] + 1
|
|
matrix[i][j] = min(insertion, deletion, substitution)
|
|
|
|
return matrix
|
|
|
|
def levenshtein_distance(self, s1, s2):
|
|
"""Calculate the Levenshtein distance between two strings.
|
|
|
|
This is straight from Wikipedia.
|
|
"""
|
|
if len(s1) < len(s2):
|
|
return self.levenshtein_distance(s2, s1)
|
|
if not s1:
|
|
return len(s2)
|
|
|
|
previous_row = range(len(s2) + 1)
|
|
for i, c1 in enumerate(s1):
|
|
current_row = [i + 1]
|
|
for j, c2 in enumerate(s2):
|
|
insertions = previous_row[j + 1] + 1
|
|
deletions = current_row[j] + 1
|
|
substitutions = previous_row[j] + (c1 != c2)
|
|
current_row.append(min(insertions, deletions, substitutions))
|
|
previous_row = current_row
|
|
|
|
return previous_row[-1]
|
|
|
|
def levenshtein_print(self,matrix,word,password):
|
|
""" Print word X password matrix """
|
|
print(" %s" % " ".join(list(word)))
|
|
for i,row in enumerate(matrix):
|
|
if i == 0: print(" "),
|
|
else: print(password[i-1]),
|
|
print(" ".join("%2d" % col for col in row))
|
|
|
|
def generate_levenshtein_rules(self, word, password):
|
|
""" Generates levenshtein rules. Returns a list of lists of levenshtein rules. """
|
|
|
|
# 1) Generate Levenshtein matrix
|
|
matrix = self.levenshtein(word, password)
|
|
|
|
# 2) Trace reverse paths through the matrix.
|
|
paths = self.levenshtein_reverse_recursive(matrix,len(matrix)-1,len(matrix[0])-1,0)
|
|
|
|
# 3) Return a collection of reverse paths.
|
|
return [path for path in paths if len(path) <= matrix[-1][-1]]
|
|
|
|
def levenshtein_reverse_recursive(self,matrix,i,j,path_len):
|
|
""" Calculate reverse Levenshtein paths.
|
|
Recursive, Depth First, Short-circuited algorithm by Peter Kacherginsky
|
|
Generates a list of edit operations necessary to transform a source word
|
|
into a password. Edit operations are recorded in the form:
|
|
(operation, password_offset, word_offset)
|
|
Where an operation can be either insertion, deletion or replacement.
|
|
"""
|
|
|
|
if i == 0 and j == 0 or path_len > matrix[-1][-1]:
|
|
return [[]]
|
|
else:
|
|
paths = list()
|
|
|
|
cost = matrix[i][j]
|
|
|
|
# Calculate minimum cost of each operation
|
|
cost_delete = cost_insert = cost_equal_or_replace = sys.maxsize
|
|
if i > 0: cost_insert = matrix[i-1][j]
|
|
if j > 0: cost_delete = matrix[i][j-1]
|
|
if i > 0 and j > 0: cost_equal_or_replace = matrix[i-1][j-1]
|
|
cost_min = min(cost_delete, cost_insert, cost_equal_or_replace)
|
|
|
|
# Recurse through reverse path for each operation
|
|
if cost_insert == cost_min:
|
|
insert_paths = self.levenshtein_reverse_recursive(matrix,i-1,j,path_len+1)
|
|
for insert_path in insert_paths: paths.append(insert_path + [('insert',i-1,j)])
|
|
|
|
if cost_delete == cost_min:
|
|
delete_paths = self.levenshtein_reverse_recursive(matrix,i,j-1,path_len+1)
|
|
for delete_path in delete_paths: paths.append(delete_path + [('delete',i,j-1)])
|
|
|
|
if cost_equal_or_replace == cost_min:
|
|
if cost_equal_or_replace == cost:
|
|
equal_paths = self.levenshtein_reverse_recursive(matrix,i-1,j-1,path_len)
|
|
for equal_path in equal_paths: paths.append(equal_path)
|
|
else:
|
|
replace_paths = self.levenshtein_reverse_recursive(matrix,i-1,j-1,path_len+1)
|
|
for replace_path in replace_paths: paths.append(replace_path + [('replace',i-1,j-1)])
|
|
|
|
return paths
|
|
|
|
def load_custom_wordlist(self,wordlist_file):
|
|
self.enchant = enchant.request_pwl_dict(wordlist_file)
|
|
|
|
def generate_words(self,password):
|
|
""" Generate source word candidates."""
|
|
|
|
LOGGER.debug("[*] Generating source words for %s" % password)
|
|
|
|
words = list()
|
|
words_collection = list()
|
|
|
|
# Let's collect best edit distance as soon as possible to prevent
|
|
# less efficient pre_rules like reversal and rotation from slowing
|
|
# us down with garbage
|
|
best_found_distance = 9999
|
|
|
|
#######################################################################
|
|
# Generate words for each preanalysis rule
|
|
if not self.brute_rules:
|
|
self.preanalysis_rules = self.preanalysis_rules[:1]
|
|
|
|
for pre_rule, pre_rule_lambda in self.preanalysis_rules:
|
|
|
|
pre_password = pre_rule_lambda(password)
|
|
|
|
# Generate word suggestions
|
|
if self.word: suggestions = [self.word]
|
|
elif self.simple_words: suggestions = self.generate_simple_words(pre_password)
|
|
else: suggestions = self.generate_advanced_words(pre_password)
|
|
|
|
# HACK: Perform some additional expansion on multi-word suggestions
|
|
# TODO: May be I should split these two and see if I can generate
|
|
# rules for each of the suggestions
|
|
for suggestion in suggestions[:self.max_words]:
|
|
suggestion = suggestion.replace(' ','')
|
|
suggestion = suggestion.replace('-','')
|
|
if not suggestion in suggestions:
|
|
suggestions.append(suggestion)
|
|
|
|
if len(suggestions) != len(set(suggestions)):
|
|
print(sorted(suggestions))
|
|
print(sorted(set(suggestions)))
|
|
|
|
|
|
for suggestion in suggestions:
|
|
|
|
distance = self.levenshtein_distance(suggestion,pre_password)
|
|
|
|
word = dict()
|
|
word["suggestion"] = suggestion
|
|
word["distance"] = distance
|
|
word["password"] = pre_password
|
|
word["pre_rule"] = pre_rule
|
|
word["best_rule_length"] = 9999
|
|
|
|
words.append(word)
|
|
|
|
#######################################################################
|
|
# Perform Optimization
|
|
for word in sorted(words, key=lambda word: word["distance"], reverse=False):
|
|
|
|
# Optimize for best distance
|
|
if not self.more_words:
|
|
if word["distance"] < best_found_distance:
|
|
best_found_distance = word["distance"]
|
|
|
|
elif word["distance"] > best_found_distance:
|
|
LOGGER.verbose("[-] %s => {edit distance suboptimal: %d (%d)} => %s" % \
|
|
(word["suggestion"], word["distance"], best_found_distance, word["password"]))
|
|
break
|
|
|
|
# Filter words with too big edit distance
|
|
if word["distance"] <= self.max_word_dist:
|
|
LOGGER.debug("[+] %s => {edit distance: %d (%d)} = > %s" % \
|
|
(word["suggestion"], word["distance"],best_found_distance, word["password"]))
|
|
|
|
words_collection.append(word)
|
|
|
|
else:
|
|
LOGGER.verbose("[-] %s => {max distance exceeded: %d (%d)} => %s" % \
|
|
(word["suggestion"], word["distance"], self.max_word_dist, word["password"]))
|
|
|
|
if self.max_words:
|
|
words_collection = words_collection[:self.max_words]
|
|
|
|
return words_collection
|
|
|
|
def generate_simple_words(self,password):
|
|
""" Generate simple words. A simple spellcheck."""
|
|
|
|
return self.enchant.suggest(password)
|
|
|
|
def generate_advanced_words(self,password):
|
|
""" Generate advanced words.
|
|
Perform some additional non-destructive cleaning to help spell-checkers:
|
|
1) Remove non-alpha prefixes and appendixes.
|
|
2) Perform common pattern matches (e.g. email).
|
|
3) Replace non-alpha character substitutions (1337 5p34k)
|
|
"""
|
|
|
|
# Remove non-alpha prefix and/or appendix
|
|
insertion_matches = self.password_pattern["insertion"].match(password)
|
|
if insertion_matches:
|
|
password = insertion_matches.group('password')
|
|
|
|
# Pattern matches
|
|
email_matches = self.password_pattern["email"].match(password)
|
|
if email_matches:
|
|
password = email_matches.group('password')
|
|
|
|
# Replace common special character replacements (1337 5p34k)
|
|
preanalysis_password = ''
|
|
for c in password:
|
|
if c in self.leet: preanalysis_password += self.leet[c]
|
|
else: preanalysis_password += c
|
|
password = preanalysis_password
|
|
|
|
LOGGER.debug("[*] Preanalysis Password: %s" % password)
|
|
|
|
return self.enchant.suggest(password)
|
|
|
|
############################################################################
|
|
# Hashcat specific offset definition 0-9,A-Z
|
|
def int_to_hashcat(self,N):
|
|
if N < 10: return N
|
|
else: return chr(65+N-10)
|
|
|
|
def hashcat_to_int(self,N):
|
|
if N.isdigit(): return int(N)
|
|
else: return ord(N)-65+10
|
|
|
|
def generate_hashcat_rules(self, suggestion, password):
|
|
""" Generate hashcat rules. Returns a length sorted list of lists of hashcat rules."""
|
|
|
|
# 2) Generate Levenshtein Rules
|
|
lev_rules = self.generate_levenshtein_rules(suggestion, password)
|
|
|
|
# 3) Generate Hashcat Rules
|
|
hashcat_rules = []
|
|
hashcat_rules_collection = []
|
|
|
|
#######################################################################
|
|
# Generate hashcat rule for each levenshtein rule
|
|
for lev_rule in lev_rules:
|
|
|
|
if self.simple_rules:
|
|
hashcat_rule = self.generate_simple_hashcat_rules(suggestion, lev_rule, password)
|
|
else:
|
|
hashcat_rule = self.generate_advanced_hashcat_rules(suggestion, lev_rule, password)
|
|
|
|
if hashcat_rule == None:
|
|
LOGGER.warning("[!] Processing FAILED: %s => ;( => %s" % (suggestion,password))
|
|
LOGGER.warning(" Sorry about that, please report this failure to")
|
|
LOGGER.warning(" the developer: iphelix [at] thesprawl.org")
|
|
|
|
else:
|
|
hashcat_rules.append(hashcat_rule)
|
|
|
|
best_found_rule_length = 9999
|
|
|
|
#######################################################################
|
|
# Perform Optimization
|
|
for hashcat_rule in sorted(hashcat_rules, key=lambda hashcat_rule: len(hashcat_rule)):
|
|
|
|
rule_length = len(hashcat_rule)
|
|
|
|
if not self.more_rules:
|
|
if rule_length < best_found_rule_length:
|
|
best_found_rule_length = rule_length
|
|
|
|
elif rule_length > best_found_rule_length:
|
|
LOGGER.verbose("[-] %s => {best rule length exceeded: %d (%d)} => %s" % \
|
|
(suggestion, rule_length, best_found_rule_length, password))
|
|
break
|
|
|
|
if rule_length <= self.max_rule_len:
|
|
hashcat_rules_collection.append(hashcat_rule)
|
|
|
|
return hashcat_rules_collection
|
|
|
|
def generate_simple_hashcat_rules(self,word,rules,password):
|
|
""" Generate basic hashcat rules using only basic insert,delete,replace rules. """
|
|
hashcat_rules = []
|
|
|
|
LOGGER.debug("[*] Simple Processing %s => %s" % (word,password))
|
|
|
|
# Dynamically apply rules to the source word
|
|
# NOTE: Special case were word == password this would work as well.
|
|
word_rules = word
|
|
|
|
for (op,p,w) in rules:
|
|
|
|
LOGGER.debug("\t[*] Simple Processing Started: %s - %s" % (word_rules, " ".join(hashcat_rules)))
|
|
|
|
if op == 'insert':
|
|
hashcat_rules.append("i%s%s" % (self.int_to_hashcat(p),password[p]))
|
|
word_rules = self.hashcat_rule['i'](word_rules,p,password[p])
|
|
|
|
elif op == 'delete':
|
|
hashcat_rules.append("D%s" % self.int_to_hashcat(p))
|
|
word_rules = self.hashcat_rule['D'](word_rules,p)
|
|
|
|
elif op == 'replace':
|
|
hashcat_rules.append("o%s%s" % (self.int_to_hashcat(p),password[p]))
|
|
word_rules = self.hashcat_rule['o'](word_rules,p,password[p])
|
|
|
|
LOGGER.debug("\t[*] Simple Processing Ended: %s => %s => %s" % (word_rules, " ".join(hashcat_rules),password))
|
|
|
|
# Check if rules result in the correct password
|
|
if word_rules == password:
|
|
return hashcat_rules
|
|
else:
|
|
LOGGER.debug("[!] Simple Processing FAILED: %s => %s => %s (%s)" % (word," ".join(hashcat_rules),password,word_rules))
|
|
return None
|
|
|
|
def generate_advanced_hashcat_rules(self,word,rules,password):
|
|
""" Generate advanced hashcat rules using full range of available rules. """
|
|
hashcat_rules = []
|
|
|
|
LOGGER.debug("[*] Advanced Processing %s => %s" % (word,password))
|
|
|
|
# Dynamically apply and store rules in word_rules variable.
|
|
# NOTE: Special case where word == password this would work as well.
|
|
word_rules = word
|
|
|
|
# Generate case statistics
|
|
password_lower = len([c for c in password if c.islower()])
|
|
password_upper = len([c for c in password if c.isupper()])
|
|
|
|
for i,(op,p,w) in enumerate(rules):
|
|
|
|
LOGGER.debug("\t[*] Advanced Processing Started: %s - %s" % (word_rules, " ".join(hashcat_rules)))
|
|
|
|
if op == 'insert':
|
|
hashcat_rules.append("i%s%s" % (self.int_to_hashcat(p),password[p]))
|
|
word_rules = self.hashcat_rule['i'](word_rules,p,password[p])
|
|
|
|
elif op == 'delete':
|
|
hashcat_rules.append("D%s" % self.int_to_hashcat(p))
|
|
word_rules = self.hashcat_rule['D'](word_rules,p)
|
|
|
|
elif op == 'replace':
|
|
|
|
# Detecting global replacement such as sXY, l, u, C, c is a non
|
|
# trivial problem because different characters may be added or
|
|
# removed from the word by other rules. A reliable way to solve
|
|
# this problem is to apply all of the rules the source word
|
|
# and keep track of its state at any given time. At the same
|
|
# time, global replacement rules can be tested by completing
|
|
# the rest of the rules using a simplified engine.
|
|
|
|
# The sequence of if statements determines the priority of rules
|
|
|
|
# This rule was made obsolete by a prior global replacement
|
|
if word_rules[p] == password[p]:
|
|
LOGGER.debug("\t[*] Advanced Processing Obsolete Rule: %s - %s" % (word_rules, " ".join(hashcat_rules)))
|
|
|
|
# Swapping rules
|
|
elif p < len(password)-1 and p < len(word_rules)-1 and word_rules[p] == password[p+1] and word_rules[p+1] == password[p]:
|
|
# Swap first two characters
|
|
if p == 0 and self.generate_simple_hashcat_rules( self.hashcat_rule['k'](word_rules), rules[i+1:],password):
|
|
hashcat_rules.append("k")
|
|
word_rules = self.hashcat_rule['k'](word_rules)
|
|
# Swap last two characters
|
|
elif p == len(word_rules)-2 and self.generate_simple_hashcat_rules( self.hashcat_rule['K'](word_rules), rules[i+1:],password):
|
|
hashcat_rules.append("K")
|
|
word_rules = self.hashcat_rule['K'](word_rules)
|
|
# Swap any two characters (only adjacent swapping is supported)
|
|
elif self.generate_simple_hashcat_rules( self.hashcat_rule['*'](word_rules,p,p+1), rules[i+1:],password):
|
|
hashcat_rules.append("*%s%s" % (self.int_to_hashcat(p),self.int_to_hashcat(p+1)))
|
|
word_rules = self.hashcat_rule['*'](word_rules,p,p+1)
|
|
else:
|
|
hashcat_rules.append("o%s%s" % (self.int_to_hashcat(p),password[p]))
|
|
word_rules = self.hashcat_rule['o'](word_rules,p,password[p])
|
|
|
|
# Case Toggle: Uppercased a letter
|
|
elif word_rules[p].islower() and word_rules[p].upper() == password[p]:
|
|
|
|
# Toggle the case of all characters in word (mixed cases)
|
|
if password_upper and password_lower and self.generate_simple_hashcat_rules( self.hashcat_rule['t'](word_rules), rules[i+1:],password):
|
|
hashcat_rules.append("t")
|
|
word_rules = self.hashcat_rule['t'](word_rules)
|
|
|
|
# Capitalize all letters
|
|
elif self.generate_simple_hashcat_rules( self.hashcat_rule['u'](word_rules), rules[i+1:],password):
|
|
hashcat_rules.append("u")
|
|
word_rules = self.hashcat_rule['u'](word_rules)
|
|
|
|
# Capitalize the first letter
|
|
elif p == 0 and self.generate_simple_hashcat_rules( self.hashcat_rule['c'](word_rules), rules[i+1:],password):
|
|
hashcat_rules.append("c")
|
|
word_rules = self.hashcat_rule['c'](word_rules)
|
|
|
|
# Toggle the case of characters at position N
|
|
else:
|
|
hashcat_rules.append("T%s" % self.int_to_hashcat(p))
|
|
word_rules = self.hashcat_rule['T'](word_rules,p)
|
|
|
|
# Case Toggle: Lowercased a letter
|
|
elif word_rules[p].isupper() and word_rules[p].lower() == password[p]:
|
|
|
|
# Toggle the case of all characters in word (mixed cases)
|
|
if password_upper and password_lower and self.generate_simple_hashcat_rules( self.hashcat_rule['t'](word_rules), rules[i+1:],password):
|
|
hashcat_rules.append("t")
|
|
word_rules = self.hashcat_rule['t'](word_rules)
|
|
|
|
# Lowercase all letters
|
|
elif self.generate_simple_hashcat_rules( self.hashcat_rule['l'](word_rules), rules[i+1:],password):
|
|
hashcat_rules.append("l")
|
|
word_rules = self.hashcat_rule['l'](word_rules)
|
|
|
|
# Lowercase the first found character, uppercase the rest
|
|
elif p == 0 and self.generate_simple_hashcat_rules( self.hashcat_rule['C'](word_rules), rules[i+1:],password):
|
|
hashcat_rules.append("C")
|
|
word_rules = self.hashcat_rule['C'](word_rules)
|
|
|
|
# Toggle the case of characters at position N
|
|
else:
|
|
hashcat_rules.append("T%s" % self.int_to_hashcat(p))
|
|
word_rules = self.hashcat_rule['T'](word_rules,p)
|
|
|
|
# Special case substitution of 'all' instances (1337 $p34k)
|
|
elif word_rules[p].isalpha() and not password[p].isalpha() and self.generate_simple_hashcat_rules( self.hashcat_rule['s'](word_rules,word_rules[p],password[p]), rules[i+1:],password):
|
|
|
|
# If we have already detected this rule, then skip it thus
|
|
# reducing total rule count.
|
|
# BUG: Elisabeth => sE3 sl1 u o3Z sE3 => 31IZAB3TH
|
|
#if not "s%s%s" % (word_rules[p],password[p]) in hashcat_rules:
|
|
hashcat_rules.append("s%s%s" % (word_rules[p],password[p]))
|
|
word_rules = self.hashcat_rule['s'](word_rules,word_rules[p],password[p])
|
|
|
|
# Replace next character with current
|
|
elif p < len(password)-1 and p < len(word_rules)-1 and password[p] == password[p+1] and password[p] == word_rules[p+1]:
|
|
hashcat_rules.append(".%s" % self.int_to_hashcat(p))
|
|
word_rules = self.hashcat_rule['.'](word_rules,p)
|
|
|
|
# Replace previous character with current
|
|
elif p > 0 and w > 0 and password[p] == password[p-1] and password[p] == word_rules[p-1]:
|
|
hashcat_rules.append(",%s" % self.int_to_hashcat(p))
|
|
word_rules = self.hashcat_rule[','](word_rules,p)
|
|
|
|
# ASCII increment
|
|
elif ord(word_rules[p]) + 1 == ord(password[p]):
|
|
hashcat_rules.append("+%s" % self.int_to_hashcat(p))
|
|
word_rules = self.hashcat_rule['+'](word_rules,p)
|
|
|
|
# ASCII decrement
|
|
elif ord(word_rules[p]) - 1 == ord(password[p]):
|
|
hashcat_rules.append("-%s" % self.int_to_hashcat(p))
|
|
word_rules = self.hashcat_rule['-'](word_rules,p)
|
|
|
|
# SHIFT left
|
|
elif ord(word_rules[p]) << 1 == ord(password[p]):
|
|
hashcat_rules.append("L%s" % self.int_to_hashcat(p))
|
|
word_rules = self.hashcat_rule['L'](word_rules,p)
|
|
|
|
# SHIFT right
|
|
elif ord(word_rules[p]) >> 1 == ord(password[p]):
|
|
hashcat_rules.append("R%s" % self.int_to_hashcat(p))
|
|
word_rules = self.hashcat_rule['R'](word_rules,p)
|
|
|
|
# Position based replacements.
|
|
else:
|
|
hashcat_rules.append("o%s%s" % (self.int_to_hashcat(p),password[p]))
|
|
word_rules = self.hashcat_rule['o'](word_rules,p,password[p])
|
|
|
|
LOGGER.debug("\t[*] Advanced Processing Ended: %s %s" % (word_rules, " ".join(hashcat_rules)))
|
|
|
|
########################################################################
|
|
# Prefix rules
|
|
last_prefix = 0
|
|
prefix_rules = list()
|
|
for hashcat_rule in hashcat_rules:
|
|
if hashcat_rule[0] == "i" and self.hashcat_to_int(hashcat_rule[1]) == last_prefix:
|
|
prefix_rules.append("^%s" % hashcat_rule[2])
|
|
last_prefix += 1
|
|
elif len(prefix_rules):
|
|
hashcat_rules = prefix_rules[::-1]+hashcat_rules[len(prefix_rules):]
|
|
break
|
|
else:
|
|
break
|
|
else:
|
|
hashcat_rules = prefix_rules[::-1]+hashcat_rules[len(prefix_rules):]
|
|
|
|
####################################################################
|
|
# Appendix rules
|
|
last_appendix = len(password) - 1
|
|
appendix_rules = list()
|
|
for hashcat_rule in hashcat_rules[::-1]:
|
|
if hashcat_rule[0] == "i" and self.hashcat_to_int(hashcat_rule[1]) == last_appendix:
|
|
appendix_rules.append("$%s" % hashcat_rule[2])
|
|
last_appendix-= 1
|
|
elif len(appendix_rules):
|
|
hashcat_rules = hashcat_rules[:-len(appendix_rules)]+appendix_rules[::-1]
|
|
break
|
|
else:
|
|
break
|
|
else:
|
|
hashcat_rules = hashcat_rules[:-len(appendix_rules)]+appendix_rules[::-1]
|
|
|
|
####################################################################
|
|
# Truncate left rules
|
|
last_precut = 0
|
|
precut_rules = list()
|
|
for hashcat_rule in hashcat_rules:
|
|
if hashcat_rule[0] == "D" and self.hashcat_to_int(hashcat_rule[1]) == last_precut:
|
|
precut_rules.append("[")
|
|
elif len(precut_rules):
|
|
hashcat_rules = precut_rules[::-1]+hashcat_rules[len(precut_rules):]
|
|
break
|
|
else:
|
|
break
|
|
else:
|
|
hashcat_rules = precut_rules[::-1]+hashcat_rules[len(precut_rules):]
|
|
|
|
####################################################################
|
|
# Truncate right rules
|
|
last_postcut = len(password)
|
|
postcut_rules = list()
|
|
for hashcat_rule in hashcat_rules[::-1]:
|
|
|
|
if hashcat_rule[0] == "D" and self.hashcat_to_int(hashcat_rule[1]) >= last_postcut:
|
|
postcut_rules.append("]")
|
|
elif len(postcut_rules):
|
|
hashcat_rules = hashcat_rules[:-len(postcut_rules)]+postcut_rules[::-1]
|
|
break
|
|
else:
|
|
break
|
|
else:
|
|
hashcat_rules = hashcat_rules[:-len(postcut_rules)]+postcut_rules[::-1]
|
|
|
|
# Check if rules result in the correct password
|
|
if word_rules == password:
|
|
return hashcat_rules
|
|
else:
|
|
LOGGER.debug("[!] Advanced Processing FAILED: %s => %s => %s (%s)" % (word," ".join(hashcat_rules),password,word_rules))
|
|
return None
|
|
|
|
|
|
def check_reversible_password(self, password):
|
|
""" Check whether the password is likely to be reversed successfuly. """
|
|
|
|
# Skip all numeric passwords
|
|
if password.isdigit():
|
|
if not self.quiet: LOGGER.verbose("[!] %s => {skipping numeric} => %s" % (password,password))
|
|
self.numeric_stats_total += 1
|
|
return False
|
|
|
|
# Skip passwords with less than 25% of alpha character
|
|
# TODO: Make random word detection more reliable based on word entropy.
|
|
elif len([c for c in password if c.isalpha()]) < len(password)/4:
|
|
if not self.quiet: LOGGER.verbose("[!] %s => {skipping alpha less than 25%%} => %s" % (password,password))
|
|
self.special_stats_total += 1
|
|
return False
|
|
|
|
# Only check english ascii passwords for now
|
|
# TODO: Add support for more languages.
|
|
elif [c for c in password if ord(c) < 32 or ord(c) > 126]:
|
|
if not self.quiet: LOGGER.verbose("[!] %s => {skipping non ascii english} => %s" % (password,password))
|
|
self.foreign_stats_total += 1
|
|
return False
|
|
|
|
else:
|
|
return True
|
|
|
|
def analyze_password(self,password, index=0, rules_queue=multiprocessing.Queue(), words_queue=multiprocessing.Queue()):
|
|
""" Analyze a single password. """
|
|
|
|
LOGGER.verbose("[*] Analyzing password: %s" % password)
|
|
|
|
words = []
|
|
|
|
# Short-cut words in the dictionary
|
|
if self.enchant.check(password) and not self.word:
|
|
|
|
word = dict()
|
|
word["password"] = password
|
|
word["suggestion"] = password
|
|
word["hashcat_rules"] = [[],]
|
|
word["pre_rule"] = []
|
|
word["best_rule_length"] = 9999
|
|
word["index"] = index
|
|
|
|
words.append(word)
|
|
|
|
# Generate rules for words not in the dictionary
|
|
else:
|
|
|
|
# Generate source words list
|
|
words = self.generate_words(password)
|
|
|
|
# Generate levenshtein reverse paths for each suggestion
|
|
for word in words:
|
|
|
|
# Generate a collection of hashcat_rules lists
|
|
word["hashcat_rules"] = self.generate_hashcat_rules(word["suggestion"],word["password"])
|
|
word["index"] = index
|
|
|
|
self.print_hashcat_rules(words, password, rules_queue, words_queue)
|
|
|
|
def print_hashcat_rules(self, words, password, rules_queue, words_queue):
|
|
|
|
best_found_rule_length = 9999
|
|
|
|
# Sorted list based on rule length
|
|
for word in sorted(words, key=lambda word: len(word["hashcat_rules"][0])):
|
|
|
|
words_queue.put((word["index"], word["suggestion"]))
|
|
|
|
for hashcat_rule in word["hashcat_rules"]:
|
|
|
|
rule_length = len(hashcat_rule)
|
|
|
|
if not self.more_rules:
|
|
if rule_length < best_found_rule_length:
|
|
best_found_rule_length = rule_length
|
|
|
|
elif rule_length > best_found_rule_length:
|
|
LOGGER.verbose("[-] %s => {best rule length exceeded: %d (%d)} => %s" % \
|
|
(word["suggestion"], rule_length, best_found_rule_length, password))
|
|
break
|
|
|
|
if rule_length <= self.max_rule_len:
|
|
|
|
hashcat_rule_str = " ".join(hashcat_rule + word["pre_rule"] or [':'])
|
|
LOGGER.verbose("[+] %s => %s => %s" % (word["suggestion"], hashcat_rule_str, password))
|
|
|
|
rules_queue.put((word["index"], hashcat_rule_str))
|
|
|
|
|
|
def password_worker(self,i, passwords_queue, rules_queue, words_queue):
|
|
LOGGER.debug("[*] Password analysis worker [%d] started." % i)
|
|
try:
|
|
while True:
|
|
password = passwords_queue.get()
|
|
|
|
# Interrupted by a Death Pill
|
|
if password == None: break
|
|
|
|
index, password = password
|
|
self.analyze_password(password, index, rules_queue, words_queue)
|
|
except (KeyboardInterrupt, SystemExit):
|
|
LOGGER.debug("[*] Password analysis worker [%d] terminated." % i)
|
|
|
|
LOGGER.debug("[*] Password analysis worker [%d] stopped." % i)
|
|
|
|
def rule_worker(self, rules_queue, output_rules_filename):
|
|
""" Worker to store generated rules. """
|
|
LOGGER.info("[*] Saving rules to %s" % output_rules_filename)
|
|
|
|
f = open(output_rules_filename, 'w')
|
|
LOGGER.debug("[*] Rule worker started.")
|
|
try:
|
|
while True:
|
|
rule = rules_queue.get()
|
|
|
|
# Interrupted by a Death Pill
|
|
if rule == None: break
|
|
|
|
f.write("%d\t%s\n" % rule)
|
|
f.flush()
|
|
|
|
except (KeyboardInterrupt, SystemExit):
|
|
LOGGER.debug("[*] Rule worker terminated.")
|
|
|
|
f.close()
|
|
LOGGER.debug("[*] Rule worker stopped.")
|
|
|
|
def word_worker(self, words_queue, output_words_filename):
|
|
""" Worker to store generated rules. """
|
|
LOGGER.info("[*] Saving words to %s" % output_words_filename)
|
|
|
|
f = open(output_words_filename, 'w')
|
|
LOGGER.debug("[*] Word worker started.")
|
|
try:
|
|
while True:
|
|
word = words_queue.get()
|
|
|
|
# Interrupted by a Death Pill
|
|
if word == None: break
|
|
|
|
f.write("%d\t%s\n" % word)
|
|
f.flush()
|
|
|
|
except (KeyboardInterrupt, SystemExit):
|
|
LOGGER.debug("[*] Word worker terminated.")
|
|
|
|
f.close()
|
|
LOGGER.debug("[*] Word worker stopped.")
|
|
|
|
# Analyze passwords file
|
|
def analyze_passwords_file(self,passwords_file, encoding):
|
|
""" Analyze provided passwords file. """
|
|
|
|
LOGGER.info("[*] Analyzing passwords file: %s:" % passwords_file)
|
|
LOGGER.info("[*] Press Ctrl-C to end execution and generate statistical analysis.")
|
|
|
|
# Setup queues
|
|
passwords_queue = multiprocessing.Queue(self.threads)
|
|
rules_queue = multiprocessing.Queue()
|
|
words_queue = multiprocessing.Queue()
|
|
|
|
# Start workers
|
|
for i in range(self.threads):
|
|
multiprocessing.Process(target=self.password_worker, args=(i, passwords_queue, rules_queue, words_queue)).start()
|
|
multiprocessing.Process(target=self.rule_worker, args=(rules_queue, "%s.rule" % self.basename)).start()
|
|
multiprocessing.Process(target=self.word_worker, args=(words_queue, "%s.word" % self.basename)).start()
|
|
|
|
# Continue with the main thread
|
|
|
|
f = open(passwords_file,'r', encoding=encoding)
|
|
|
|
output_pwd_filename = "%s.password" % self.basename
|
|
LOGGER.info("[*] Saving analysed passwords to to %s" % output_pwd_filename)
|
|
o = open(output_pwd_filename, "w")
|
|
|
|
password_count = 0
|
|
analysis_start = time.time()
|
|
segment_start = analysis_start
|
|
try:
|
|
for password in f:
|
|
password = password.rstrip('\r\n')
|
|
if len(password) > 0:
|
|
|
|
# Provide analysis time feedback to the user
|
|
if not self.quiet and password_count != 0 and password_count % 5000 == 0:
|
|
segment_time = time.time() - segment_start
|
|
LOGGER.info("[*] Processed %d passwords in %.2f seconds at the rate of %.2f p/sec" % \
|
|
(password_count, segment_start - analysis_start, 5000/segment_time ))
|
|
segment_start = time.time()
|
|
|
|
password_count += 1
|
|
|
|
# Perform preliminary checks and add password to the queue
|
|
if self.check_reversible_password(password):
|
|
passwords_queue.put((password_count, password))
|
|
o.write("%d\t%s\n" % (password_count, password))
|
|
o.flush()
|
|
|
|
except (KeyboardInterrupt, SystemExit):
|
|
LOGGER.warning("\n[!] Rulegen was interrupted.")
|
|
|
|
else:
|
|
# Signal workers to stop.
|
|
for i in range(self.threads):
|
|
passwords_queue.put(None)
|
|
|
|
# Wait for all of the queued passwords to finish.
|
|
while not passwords_queue.empty():
|
|
time.sleep(1)
|
|
|
|
# Signal writers to stop.
|
|
rules_queue.put(None)
|
|
words_queue.put(None)
|
|
|
|
o.close()
|
|
f.close()
|
|
|
|
analysis_time = time.time() - analysis_start
|
|
LOGGER.info("[*] Finished processing %d passwords in %.2f seconds at the rate of %.2f p/sec" % (password_count, analysis_time, float(password_count)/analysis_time ))
|
|
|
|
return
|
|
|
|
LOGGER.info("[*] Generating statistics for [%s] rules and words." % self.basename)
|
|
LOGGER.info("[-] Skipped %d all numeric passwords (%0.2f%%)" % \
|
|
(self.numeric_stats_total, float(self.numeric_stats_total)*100.0/float(password_count)))
|
|
LOGGER.info("[-] Skipped %d passwords with less than 25%% alpha characters (%0.2f%%)" % \
|
|
(self.special_stats_total, float(self.special_stats_total)*100.0/float(password_count)))
|
|
LOGGER.info("[-] Skipped %d passwords with non ascii characters (%0.2f%%)" % \
|
|
(self.foreign_stats_total, float(self.foreign_stats_total)*100.0/float(password_count)))
|
|
|
|
# TODO: Counter breaks on large files. uniq -c | sort -rn is still the most
|
|
# optimal way.
|
|
rules_file = open("%s.rule" % self.basename,'r')
|
|
rules_sorted_file = open("%s-sorted.rule" % self.basename, 'w')
|
|
rules_counter = Counter(rules_file)
|
|
rule_counter_total = sum(rules_counter.values())
|
|
|
|
LOGGER.info("[*] Top 10 rules")
|
|
rules_i = 0
|
|
for (rule, count) in rules_counter.most_common():
|
|
rules_sorted_file.write(rule)
|
|
if rules_i < 10: LOGGER.info("[+] %s - %d (%0.2f%%)" % (rule.rstrip('\r\n'), count, count*100/rule_counter_total))
|
|
rules_i += 1
|
|
|
|
rules_file.close()
|
|
rules_sorted_file.close()
|
|
|
|
|
|
words_file = open("%s.word" % self.basename,'r')
|
|
words_sorted_file = open("%s-sorted.word" % self.basename,'w')
|
|
words_counter = Counter(words_file)
|
|
word_counter_total = sum(rules_counter.values())
|
|
|
|
LOGGER.info("[*] Top 10 words")
|
|
words_i = 0
|
|
for (word, count) in words_counter.most_common():
|
|
words_sorted_file.write(word)
|
|
if words_i < 10: LOGGER.info("[+] %s - %d (%0.2f%%)" % (word.rstrip('\r\n'), count, count*100/word_counter_total))
|
|
words_i += 1
|
|
|
|
words_file.close()
|
|
words_sorted_file.close()
|
|
|
|
############################################################################
|
|
def verify_hashcat_rules(self,word, rules, password):
|
|
|
|
f = open("%s/test.rule" % HASHCAT_PATH,'w')
|
|
f.write(" ".join(rules))
|
|
f.close()
|
|
|
|
f = open("%s/test.word" % HASHCAT_PATH,'w')
|
|
f.write(word)
|
|
f.close()
|
|
|
|
p = subprocess.Popen(["%s/hashcat-cli64.bin" % HASHCAT_PATH,"-r","%s/test.rule" % HASHCAT_PATH,"--stdout","%s/test.word" % HASHCAT_PATH], stdout=subprocess.PIPE)
|
|
out, err = p.communicate()
|
|
out = out.strip()
|
|
|
|
if out == password:
|
|
hashcat_rules_str = " ".join(rules or [':'])
|
|
LOGGER.verbose("[+] %s => %s => %s" % (word, hashcat_rules_str, password))
|
|
|
|
else:
|
|
LOGGER.error("[!] Hashcat Verification FAILED: %s => %s => %s (%s)" % (word," ".join(rules or [':']),password,out))
|
|
|
|
if __name__ == "__main__":
|
|
|
|
header = " _ \n"
|
|
header += " RuleGen %s | |\n" % VERSION
|
|
header += " _ __ __ _ ___| | _\n"
|
|
header += " | '_ \\ / _` |/ __| |/ /\n"
|
|
header += " | |_) | (_| | (__| < \n"
|
|
header += " | .__/ \\__,_|\\___|_|\\_\\\n"
|
|
header += " | | \n"
|
|
header += " |_| iphelix@thesprawl.org\n"
|
|
header += "\n"
|
|
|
|
|
|
parser = OptionParser("%prog [options] passwords.txt", version="%prog "+VERSION)
|
|
|
|
parser.add_option("-b","--basename", help="Output base name. The following files will be generated: basename.words, basename.rules and basename.stats", default="analysis",metavar="rockyou")
|
|
parser.add_option("-w","--wordlist", help="Use a custom wordlist for rule analysis.", metavar="wiki.dict")
|
|
parser.add_option("-q", "--quiet", action="store_true", dest="quiet", default=False, help="Don't show headers.")
|
|
parser.add_option("--threads", type="int", default=multiprocessing.cpu_count(), help="Parallel threads to use for processing.")
|
|
parser.add_option("-e", "--encoding", help="Input file encoding.", default="utf-8")
|
|
|
|
wordtune = OptionGroup(parser, "Fine tune source word generation:")
|
|
wordtune.add_option("--maxworddist", help="Maximum word edit distance (Levenshtein)", type="int", default=10, metavar="10")
|
|
wordtune.add_option("--maxwords", help="Maximum number of source word candidates to consider", type="int", default=5, metavar="5")
|
|
wordtune.add_option("--morewords", help="Consider suboptimal source word candidates", action="store_true", default=False)
|
|
wordtune.add_option("--simplewords", help="Generate simple source words for given passwords", action="store_true", default=False)
|
|
parser.add_option_group(wordtune)
|
|
|
|
ruletune = OptionGroup(parser, "Fine tune rule generation:")
|
|
ruletune.add_option("--maxrulelen", help="Maximum number of operations in a single rule", type="int", default=10, metavar="10")
|
|
ruletune.add_option("--maxrules", help="Maximum number of rules to consider", type="int", default=5, metavar="5")
|
|
ruletune.add_option("--morerules", help="Generate suboptimal rules", action="store_true", default=False)
|
|
ruletune.add_option("--simplerules", help="Generate simple rules insert,delete,replace",action="store_true", default=False)
|
|
ruletune.add_option("--bruterules", help="Bruteforce reversal and rotation rules (slow)",action="store_true", default=False)
|
|
parser.add_option_group(ruletune)
|
|
|
|
spelltune = OptionGroup(parser, "Fine tune spell checker engine:")
|
|
spelltune.add_option("--providers", help="Comma-separated list of provider engines", default="aspell,myspell", metavar="aspell,myspell")
|
|
parser.add_option_group(spelltune)
|
|
|
|
debug = OptionGroup(parser, "Debugging options:")
|
|
debug.add_option("-v","--verbose", help="Show verbose information.", action="store_true", default=False)
|
|
debug.add_option("-d","--debug", help="Debug rules.", action="store_true", default=False)
|
|
debug.add_option("--password", help="Process the last argument as a password not a file.", action="store_true", default=False)
|
|
debug.add_option("--word", help="Use a custom word for rule analysis", metavar="Password")
|
|
debug.add_option("--hashcat", help="Test generated rules with hashcat-cli", action="store_true", default=False)
|
|
parser.add_option_group(debug)
|
|
|
|
(options, args) = parser.parse_args()
|
|
|
|
# Print program header
|
|
if not options.quiet:
|
|
print (header)
|
|
|
|
if len(args) < 1:
|
|
parser.error("no passwords file specified")
|
|
exit(1)
|
|
|
|
rulegen = RuleGen(language="en", providers=options.providers, basename=options.basename, threads=options.threads)
|
|
|
|
# Finetuning word generation
|
|
rulegen.max_word_dist=options.maxworddist
|
|
rulegen.max_words=options.maxwords
|
|
rulegen.more_words=options.morewords
|
|
rulegen.simple_words=options.simplewords
|
|
|
|
# Finetuning rule generation
|
|
rulegen.max_rule_len=options.maxrulelen
|
|
rulegen.max_rules=options.maxrules
|
|
rulegen.more_rules=options.morerules
|
|
rulegen.simple_rules=options.simplerules
|
|
rulegen.brute_rules=options.bruterules
|
|
if rulegen.brute_rules: LOGGER.info("[!] Bruteforcing reversal and rotation rules. (slower)")
|
|
|
|
# Debugging options
|
|
rulegen.word = options.word
|
|
rulegen.hashcat = options.hashcat
|
|
rulegen.quiet = options.quiet
|
|
|
|
logging.basicConfig(level=logging.DEBUG if options.debug
|
|
else logging.VERBOSE if options.verbose
|
|
else logging.INFO)
|
|
|
|
# Custom wordlist
|
|
if not options.word:
|
|
if options.wordlist: rulegen.load_custom_wordlist(options.wordlist)
|
|
LOGGER.info("[*] Using Enchant '%s' module. For best results please install" % rulegen.enchant.provider.name)
|
|
LOGGER.info(" '%s' module language dictionaries." % rulegen.enchant.provider.name)
|
|
|
|
# Analyze a single password or several passwords in a file
|
|
if options.password:
|
|
rulegen.analyze_password(args[0])
|
|
else:
|
|
rulegen.analyze_passwords_file(args[0], options.encoding)
|