Markov: port IRC bot module to django models
work in progress, not everything works yet, but the important stuff should
This commit is contained in:
parent
6a50db0e3d
commit
f1ff281da2
|
@ -24,7 +24,9 @@ import thread
|
||||||
import time
|
import time
|
||||||
|
|
||||||
from dateutil.relativedelta import relativedelta
|
from dateutil.relativedelta import relativedelta
|
||||||
import MySQLdb as mdb
|
|
||||||
|
from markov.models import MarkovContext, MarkovState, MarkovTarget
|
||||||
|
from markov.views import _generate_sentence, _learn_line
|
||||||
|
|
||||||
from extlib import irclib
|
from extlib import irclib
|
||||||
|
|
||||||
|
@ -48,12 +50,6 @@ class Markov(Module):
|
||||||
|
|
||||||
"""
|
"""
|
||||||
|
|
||||||
# set up some keywords for use in the chains --- don't change these
|
|
||||||
# once you've created a brain
|
|
||||||
self.start1 = '__start1'
|
|
||||||
self.start2 = '__start2'
|
|
||||||
self.stop = '__stop'
|
|
||||||
|
|
||||||
# set up regexes, for replying to specific stuff
|
# set up regexes, for replying to specific stuff
|
||||||
learnpattern = '^!markov\s+learn\s+(.*)$'
|
learnpattern = '^!markov\s+learn\s+(.*)$'
|
||||||
replypattern = '^!markov\s+reply(\s+min=(\d+))?(\s+max=(\d+))?(\s+(.*)$|$)'
|
replypattern = '^!markov\s+reply(\s+min=(\d+))?(\s+max=(\d+))?(\s+(.*)$|$)'
|
||||||
|
@ -70,66 +66,9 @@ class Markov(Module):
|
||||||
self.next_chatter_check = 0
|
self.next_chatter_check = 0
|
||||||
thread.start_new_thread(self.thread_do, ())
|
thread.start_new_thread(self.thread_do, ())
|
||||||
|
|
||||||
irc.xmlrpc_register_function(self._generate_line,
|
# TODO: bring this back somehow
|
||||||
"markov_generate_line")
|
#irc.xmlrpc_register_function(self._generate_line,
|
||||||
|
# "markov_generate_line")
|
||||||
def db_init(self):
|
|
||||||
"""Create the markov chain table."""
|
|
||||||
|
|
||||||
version = self.db_module_registered(self.__class__.__name__)
|
|
||||||
if version == None:
|
|
||||||
db = self.get_db()
|
|
||||||
try:
|
|
||||||
version = 1
|
|
||||||
cur = db.cursor(mdb.cursors.DictCursor)
|
|
||||||
cur.execute('''
|
|
||||||
CREATE TABLE markov_chatter_target (
|
|
||||||
id SERIAL,
|
|
||||||
target VARCHAR(256) NOT NULL,
|
|
||||||
chance INTEGER NOT NULL DEFAULT 99999
|
|
||||||
) ENGINE=InnoDB CHARACTER SET utf8 COLLATE utf8_bin
|
|
||||||
''')
|
|
||||||
cur.execute('''
|
|
||||||
CREATE TABLE markov_context (
|
|
||||||
id SERIAL,
|
|
||||||
context VARCHAR(256) NOT NULL
|
|
||||||
) ENGINE=InnoDB CHARACTER SET utf8 COLLATE utf8_bin
|
|
||||||
''')
|
|
||||||
cur.execute('''
|
|
||||||
CREATE TABLE markov_target_to_context_map (
|
|
||||||
id SERIAL,
|
|
||||||
target VARCHAR(256) NOT NULL,
|
|
||||||
context_id BIGINT(20) UNSIGNED NOT NULL,
|
|
||||||
FOREIGN KEY(context_id) REFERENCES markov_context(id)
|
|
||||||
) ENGINE=InnoDB CHARACTER SET utf8 COLLATE utf8_bin
|
|
||||||
''')
|
|
||||||
cur.execute('''
|
|
||||||
CREATE TABLE markov_chain (
|
|
||||||
id SERIAL,
|
|
||||||
k1 VARCHAR(128) NOT NULL,
|
|
||||||
k2 VARCHAR(128) NOT NULL,
|
|
||||||
v VARCHAR(128) NOT NULL,
|
|
||||||
context_id BIGINT(20) UNSIGNED NOT NULL,
|
|
||||||
FOREIGN KEY(context_id) REFERENCES markov_context(id)
|
|
||||||
) ENGINE=InnoDB CHARACTER SET utf8 COLLATE utf8_bin
|
|
||||||
''')
|
|
||||||
cur.execute('''
|
|
||||||
CREATE INDEX markov_chain_keys_and_context_id_index
|
|
||||||
ON markov_chain (k1, k2, context_id)''')
|
|
||||||
|
|
||||||
cur.execute('''
|
|
||||||
CREATE INDEX markov_chain_value_and_context_id_index
|
|
||||||
ON markov_chain (v, context_id)''')
|
|
||||||
|
|
||||||
db.commit()
|
|
||||||
self.db_register_module_version(self.__class__.__name__,
|
|
||||||
version)
|
|
||||||
except mdb.Error as e:
|
|
||||||
db.rollback()
|
|
||||||
self.log.error("database error trying to create tables")
|
|
||||||
self.log.exception(e)
|
|
||||||
raise
|
|
||||||
finally: cur.close()
|
|
||||||
|
|
||||||
def register_handlers(self):
|
def register_handlers(self):
|
||||||
"""Handle pubmsg/privmsg, to learn and/or reply to IRC events."""
|
"""Handle pubmsg/privmsg, to learn and/or reply to IRC events."""
|
||||||
|
@ -171,7 +110,9 @@ class Markov(Module):
|
||||||
if self.learnre.search(what) or self.replyre.search(what):
|
if self.learnre.search(what) or self.replyre.search(what):
|
||||||
return
|
return
|
||||||
|
|
||||||
self._learn_line(what, target, event)
|
if not event._recursing:
|
||||||
|
context = _get_or_create_target_context(target)
|
||||||
|
_learn_line(what, context)
|
||||||
|
|
||||||
def do(self, connection, event, nick, userhost, what, admin_unlocked):
|
def do(self, connection, event, nick, userhost, what, admin_unlocked):
|
||||||
"""Handle commands and inputs."""
|
"""Handle commands and inputs."""
|
||||||
|
@ -188,18 +129,25 @@ class Markov(Module):
|
||||||
if not self.shut_up:
|
if not self.shut_up:
|
||||||
# not a command, so see if i'm being mentioned
|
# not a command, so see if i'm being mentioned
|
||||||
if re.search(connection.get_nickname(), what, re.IGNORECASE) is not None:
|
if re.search(connection.get_nickname(), what, re.IGNORECASE) is not None:
|
||||||
|
context = _get_or_create_target_context(target)
|
||||||
|
|
||||||
addressed_pattern = '^' + connection.get_nickname() + '[:,]\s+(.*)'
|
addressed_pattern = '^' + connection.get_nickname() + '[:,]\s+(.*)'
|
||||||
addressed_re = re.compile(addressed_pattern)
|
addressed_re = re.compile(addressed_pattern)
|
||||||
if addressed_re.match(what):
|
if addressed_re.match(what):
|
||||||
# i was addressed directly, so respond, addressing
|
# i was addressed directly, so respond, addressing
|
||||||
# the speaker
|
# the speaker
|
||||||
|
topics = [x for x in addressed_re.match(what).group(1).split(' ') if len(x) >= 3]
|
||||||
|
|
||||||
self.lines_seen.append(('.self.said.', datetime.now()))
|
self.lines_seen.append(('.self.said.', datetime.now()))
|
||||||
return self.irc.reply(event, '{0:s}: {1:s}'.format(nick,
|
return self.irc.reply(event, '{0:s}: {1:s}'.format(nick,
|
||||||
self._generate_line(target, line=addressed_re.match(what).group(1))))
|
' '.join(_generate_sentence(context, topics=topics))))
|
||||||
else:
|
else:
|
||||||
# i wasn't addressed directly, so just respond
|
# i wasn't addressed directly, so just respond
|
||||||
|
topics = [x for x in what.split(' ') if len(x) >= 3]
|
||||||
self.lines_seen.append(('.self.said.', datetime.now()))
|
self.lines_seen.append(('.self.said.', datetime.now()))
|
||||||
return self.irc.reply(event, '{0:s}'.format(self._generate_line(target, line=what)))
|
|
||||||
|
return self.irc.reply(event, '{0:s}'.format(' '.join(_generate_sentence(context,
|
||||||
|
topics=topics))))
|
||||||
|
|
||||||
def markov_learn(self, event, nick, userhost, what, admin_unlocked):
|
def markov_learn(self, event, nick, userhost, what, admin_unlocked):
|
||||||
"""Learn one line, as provided to the command."""
|
"""Learn one line, as provided to the command."""
|
||||||
|
@ -212,7 +160,8 @@ class Markov(Module):
|
||||||
match = self.learnre.search(what)
|
match = self.learnre.search(what)
|
||||||
if match:
|
if match:
|
||||||
line = match.group(1)
|
line = match.group(1)
|
||||||
self._learn_line(line, target, event)
|
context = _get_or_create_target_context(target)
|
||||||
|
_learn_line(line, context)
|
||||||
|
|
||||||
# return what was learned, for weird chaining purposes
|
# return what was learned, for weird chaining purposes
|
||||||
return line
|
return line
|
||||||
|
@ -229,6 +178,7 @@ class Markov(Module):
|
||||||
if match:
|
if match:
|
||||||
min_size = 15
|
min_size = 15
|
||||||
max_size = 30
|
max_size = 30
|
||||||
|
context = _get_or_create_target_context(target)
|
||||||
|
|
||||||
if match.group(2):
|
if match.group(2):
|
||||||
min_size = int(match.group(2))
|
min_size = int(match.group(2))
|
||||||
|
@ -237,11 +187,13 @@ class Markov(Module):
|
||||||
|
|
||||||
if match.group(5) != '':
|
if match.group(5) != '':
|
||||||
line = match.group(6)
|
line = match.group(6)
|
||||||
|
topics = [x for x in line.split(' ') if len(x) >= 3]
|
||||||
|
|
||||||
self.lines_seen.append(('.self.said.', datetime.now()))
|
self.lines_seen.append(('.self.said.', datetime.now()))
|
||||||
return self._generate_line(target, line=line, min_size=min_size, max_size=max_size)
|
return ' '.join(_generate_sentence(context, topics=topics, min_words=min_size, max_words=max_size))
|
||||||
else:
|
else:
|
||||||
self.lines_seen.append(('.self.said.', datetime.now()))
|
self.lines_seen.append(('.self.said.', datetime.now()))
|
||||||
return self._generate_line(target, min_size=min_size, max_size=max_size)
|
return ' '.join(_generate_sentence(context, min_words=min_size, max_words=max_size))
|
||||||
|
|
||||||
def thread_do(self):
|
def thread_do(self):
|
||||||
"""Do various things."""
|
"""Do various things."""
|
||||||
|
@ -254,20 +206,8 @@ class Markov(Module):
|
||||||
def _do_random_chatter_check(self):
|
def _do_random_chatter_check(self):
|
||||||
"""Randomly say something to a channel."""
|
"""Randomly say something to a channel."""
|
||||||
|
|
||||||
# don't immediately potentially chatter, let the bot
|
# TODO: make this do stuff again
|
||||||
# join channels first
|
return
|
||||||
if self.next_chatter_check == 0:
|
|
||||||
self.next_chatter_check = time.time() + 600
|
|
||||||
|
|
||||||
if self.next_chatter_check < time.time():
|
|
||||||
self.next_chatter_check = time.time() + 600
|
|
||||||
|
|
||||||
targets = self._get_chatter_targets()
|
|
||||||
for t in targets:
|
|
||||||
if t['chance'] > 0:
|
|
||||||
a = random.randint(1, t['chance'])
|
|
||||||
if a == 1:
|
|
||||||
self.sendmsg(t['target'], self._generate_line(t['target']))
|
|
||||||
|
|
||||||
def _do_shut_up_checks(self):
|
def _do_shut_up_checks(self):
|
||||||
"""Check to see if we've been talking too much, and shut up if so."""
|
"""Check to see if we've been talking too much, and shut up if so."""
|
||||||
|
@ -293,426 +233,34 @@ class Markov(Module):
|
||||||
self.sendmsg(t['target'],
|
self.sendmsg(t['target'],
|
||||||
'shutting up for 30 seconds due to last 30 seconds of activity')
|
'shutting up for 30 seconds due to last 30 seconds of activity')
|
||||||
|
|
||||||
def _learn_line(self, line, target, event):
|
def _get_or_create_target_context(target_name):
|
||||||
"""Create Markov chains from the provided line."""
|
"""Return the context for a provided nick/channel, creating missing ones."""
|
||||||
|
|
||||||
# set up the head of the chain
|
# find the stuff, or create it
|
||||||
k1 = self.start1
|
try:
|
||||||
k2 = self.start2
|
target = MarkovTarget.objects.get(name=target_name)
|
||||||
|
return target.context
|
||||||
|
except MarkovContext.DoesNotExist:
|
||||||
|
# make a context
|
||||||
|
context = MarkovContext()
|
||||||
|
context.name = target_name
|
||||||
|
context.save()
|
||||||
|
|
||||||
context_id = self._get_context_id_for_target(target)
|
target.context = context
|
||||||
|
target.save()
|
||||||
|
|
||||||
# don't learn recursion
|
return target.context
|
||||||
if not event._recursing:
|
except MarkovTarget.DoesNotExist:
|
||||||
words = line.split()
|
# first we need to make a context for this
|
||||||
if len(words) == 0:
|
context = MarkovContext()
|
||||||
return line
|
context.name = target_name
|
||||||
|
context.save()
|
||||||
|
|
||||||
db = self.get_db()
|
target = MarkovTarget()
|
||||||
try:
|
target.name = target_name
|
||||||
cur = db.cursor(mdb.cursors.DictCursor)
|
target.context = context
|
||||||
statement = 'INSERT INTO markov_chain (k1, k2, v, context_id) VALUES (%s, %s, %s, %s)'
|
target.save()
|
||||||
for word in words:
|
|
||||||
cur.execute(statement, (k1, k2, word, context_id))
|
|
||||||
k1, k2 = k2, word
|
|
||||||
cur.execute(statement, (k1, k2, self.stop, context_id))
|
|
||||||
|
|
||||||
db.commit()
|
return target.context
|
||||||
except mdb.Error as e:
|
|
||||||
db.rollback()
|
|
||||||
self.log.error("database error learning line")
|
|
||||||
self.log.exception(e)
|
|
||||||
raise
|
|
||||||
finally: cur.close()
|
|
||||||
|
|
||||||
def _generate_line(self, target, line='', min_size=15, max_size=30):
|
|
||||||
"""Create a line, optionally using some text in a seed as a point in
|
|
||||||
the chain.
|
|
||||||
|
|
||||||
Keyword arguments:
|
|
||||||
target - the target to retrieve the context for (i.e. a channel or nick)
|
|
||||||
line - the line to reply to, by picking a random word and seeding with it
|
|
||||||
min_size - the minimum desired size in words. not guaranteed
|
|
||||||
max_size - the maximum desired size in words. not guaranteed
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
# if the limit is too low, there's nothing to do
|
|
||||||
if (max_size <= 3):
|
|
||||||
raise Exception("max_size is too small: %d" % max_size)
|
|
||||||
|
|
||||||
# if the min is too large, abort
|
|
||||||
if (min_size > 20):
|
|
||||||
raise Exception("min_size is too large: %d" % min_size)
|
|
||||||
|
|
||||||
seed_words = []
|
|
||||||
# shuffle the words in the input
|
|
||||||
seed_words = line.split()
|
|
||||||
random.shuffle(seed_words)
|
|
||||||
self.log.debug("seed words: {0:s}".format(seed_words))
|
|
||||||
|
|
||||||
# hit to generate a new seed word immediately if possible
|
|
||||||
seed_word = None
|
|
||||||
hit_word = None
|
|
||||||
|
|
||||||
context_id = self._get_context_id_for_target(target)
|
|
||||||
|
|
||||||
# start with an empty chain, and work from there
|
|
||||||
gen_words = [self.start1, self.start2]
|
|
||||||
|
|
||||||
# build a response by creating multiple sentences
|
|
||||||
while len(gen_words) < max_size + 2:
|
|
||||||
# if we're past the min and on a stop, we can end
|
|
||||||
if len(gen_words) > min_size + 2:
|
|
||||||
if gen_words[-1] == self.stop:
|
|
||||||
break
|
|
||||||
|
|
||||||
# pick a word from the shuffled seed words, if we need a new one
|
|
||||||
if seed_word == hit_word:
|
|
||||||
if len(seed_words) > 0:
|
|
||||||
seed_word = seed_words.pop()
|
|
||||||
self.log.debug("picked new seed word: "
|
|
||||||
"{0:s}".format(seed_word))
|
|
||||||
else:
|
|
||||||
seed_word = None
|
|
||||||
self.log.debug("ran out of seed words")
|
|
||||||
|
|
||||||
# if we have a stop, the word before it might need to be
|
|
||||||
# made to look like a sentence end
|
|
||||||
if gen_words[-1] == self.stop:
|
|
||||||
# chop off the stop, temporarily
|
|
||||||
gen_words = gen_words[:-1]
|
|
||||||
|
|
||||||
# we should have a real word, make it look like a
|
|
||||||
# sentence end
|
|
||||||
sentence_end = gen_words[-1]
|
|
||||||
eos_punctuation = ['!', '?', ',', '.']
|
|
||||||
if sentence_end[-1] not in eos_punctuation:
|
|
||||||
random.shuffle(eos_punctuation)
|
|
||||||
gen_words[-1] = sentence_end + eos_punctuation.pop()
|
|
||||||
self.log.debug("monkeyed with end of sentence, it's "
|
|
||||||
"now: {0:s}".format(gen_words[-1]))
|
|
||||||
|
|
||||||
# put the stop back on
|
|
||||||
gen_words.append(self.stop)
|
|
||||||
self.log.debug("gen_words: {0:s}".format(" ".join(gen_words)))
|
|
||||||
|
|
||||||
# first, see if we should start a new sentence. if so,
|
|
||||||
# work backwards
|
|
||||||
if gen_words[-1] in (self.start2, self.stop) and seed_word is not None and 0 == 1:
|
|
||||||
# drop a stop, since we're starting another sentence
|
|
||||||
if gen_words[-1] == self.stop:
|
|
||||||
gen_words = gen_words[:-1]
|
|
||||||
|
|
||||||
# work backwards from seed_word
|
|
||||||
working_backwards = []
|
|
||||||
back_k2 = self._retrieve_random_k2_for_value(seed_word, context_id)
|
|
||||||
if back_k2:
|
|
||||||
found_word = seed_word
|
|
||||||
if back_k2 == self.start2:
|
|
||||||
self.log.debug("random further back was start2, swallowing")
|
|
||||||
else:
|
|
||||||
working_backwards.append(back_k2)
|
|
||||||
working_backwards.append(found_word)
|
|
||||||
self.log.debug("started working backwards with: {0:s}".format(found_word))
|
|
||||||
self.log.debug("working_backwards: {0:s}".format(" ".join(working_backwards)))
|
|
||||||
|
|
||||||
# now work backwards until we randomly bump into a start
|
|
||||||
# to steer the chainer away from spending too much time on
|
|
||||||
# the weaker-context reverse chaining, we make max_size
|
|
||||||
# a non-linear distribution, making it more likely that
|
|
||||||
# some time is spent on better forward chains
|
|
||||||
max_back = min(random.randint(1, max_size/2) + random.randint(1, max_size/2),
|
|
||||||
max_size/4)
|
|
||||||
self.log.debug("max_back: {0:d}".format(max_back))
|
|
||||||
while len(working_backwards) < max_back:
|
|
||||||
back_k2 = self._retrieve_random_k2_for_value(working_backwards[0], context_id)
|
|
||||||
if back_k2 == self.start2:
|
|
||||||
self.log.debug("random further back was start2, finishing")
|
|
||||||
break
|
|
||||||
elif back_k2:
|
|
||||||
working_backwards.insert(0, back_k2)
|
|
||||||
self.log.debug("added '{0:s}' to working_backwards".format(back_k2))
|
|
||||||
self.log.debug("working_backwards: {0:s}".format(" ".join(working_backwards)))
|
|
||||||
else:
|
|
||||||
self.log.debug("nothing (at all!?) further back, finishing")
|
|
||||||
break
|
|
||||||
|
|
||||||
gen_words += working_backwards
|
|
||||||
self.log.debug("gen_words: {0:s}".format(" ".join(gen_words)))
|
|
||||||
hit_word = gen_words[-1]
|
|
||||||
else:
|
|
||||||
# we are working forward, with either:
|
|
||||||
# * a pair of words (normal path, filling out a sentence)
|
|
||||||
# * start1, start2 (completely new chain, no seed words)
|
|
||||||
# * stop (new sentence in existing chain, no seed words)
|
|
||||||
self.log.debug("working forwards")
|
|
||||||
forw_v = None
|
|
||||||
if gen_words[-1] in (self.start2, self.stop):
|
|
||||||
# case 2 or 3 above, need to work forward on a beginning
|
|
||||||
# of a sentence (this is slow)
|
|
||||||
if gen_words[-1] == self.stop:
|
|
||||||
# remove the stop if it's there
|
|
||||||
gen_words = gen_words[:-1]
|
|
||||||
|
|
||||||
new_sentence = self._create_chain_with_k1_k2(self.start1,
|
|
||||||
self.start2,
|
|
||||||
3, context_id,
|
|
||||||
avoid_address=True)
|
|
||||||
|
|
||||||
if len(new_sentence) > 0:
|
|
||||||
self.log.debug("started new sentence "
|
|
||||||
"'{0:s}'".format(" ".join(new_sentence)))
|
|
||||||
gen_words += new_sentence
|
|
||||||
self.log.debug("gen_words: {0:s}".format(" ".join(gen_words)))
|
|
||||||
else:
|
|
||||||
# this is a problem. we started a sentence on
|
|
||||||
# start1,start2, and still didn't find anything. to
|
|
||||||
# avoid endlessly looping we need to abort here
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
if seed_word:
|
|
||||||
self.log.debug("preferring: '{0:s}'".format(seed_word))
|
|
||||||
forw_v = self._retrieve_random_v_for_k1_and_k2_with_pref(gen_words[-2],
|
|
||||||
gen_words[-1],
|
|
||||||
seed_word,
|
|
||||||
context_id)
|
|
||||||
else:
|
|
||||||
forw_v = self._retrieve_random_v_for_k1_and_k2(gen_words[-2],
|
|
||||||
gen_words[-1],
|
|
||||||
context_id)
|
|
||||||
|
|
||||||
if forw_v:
|
|
||||||
gen_words.append(forw_v)
|
|
||||||
self.log.debug("added random word '{0:s}' to gen_words".format(forw_v))
|
|
||||||
self.log.debug("gen_words: {0:s}".format(" ".join(gen_words)))
|
|
||||||
hit_word = gen_words[-1]
|
|
||||||
else:
|
|
||||||
# append stop. this is an end to a sentence (since
|
|
||||||
# we had non-start words to begin with)
|
|
||||||
gen_words.append(self.stop)
|
|
||||||
self.log.debug("nothing found, added stop")
|
|
||||||
self.log.debug("gen_words: {0:s}".format(" ".join(gen_words)))
|
|
||||||
|
|
||||||
# chop off the seed data at the start
|
|
||||||
gen_words = gen_words[2:]
|
|
||||||
|
|
||||||
if len(gen_words):
|
|
||||||
# chop off the end text, if it was the keyword indicating an end of chain
|
|
||||||
if gen_words[-1] == self.stop:
|
|
||||||
gen_words = gen_words[:-1]
|
|
||||||
else:
|
|
||||||
self.log.warning("after all this we have an empty list of words. "
|
|
||||||
"there probably isn't any data for this context")
|
|
||||||
|
|
||||||
return ' '.join(gen_words)
|
|
||||||
|
|
||||||
def _retrieve_random_v_for_k1_and_k2(self, k1, k2, context_id):
|
|
||||||
"""Get one v for a given k1,k2."""
|
|
||||||
|
|
||||||
self.log.debug("searching with '{0:s}','{1:s}'".format(k1, k2))
|
|
||||||
values = []
|
|
||||||
db = self.get_db()
|
|
||||||
try:
|
|
||||||
query = '''
|
|
||||||
SELECT v FROM markov_chain AS r1
|
|
||||||
JOIN (
|
|
||||||
SELECT (RAND() * (SELECT MAX(id) FROM markov_chain)) AS id
|
|
||||||
) AS r2
|
|
||||||
WHERE r1.k1 = %s
|
|
||||||
AND r1.k2 = %s
|
|
||||||
AND r1.context_id = %s
|
|
||||||
ORDER BY r1.id >= r2.id DESC, r1.id ASC
|
|
||||||
LIMIT 1
|
|
||||||
'''
|
|
||||||
cur = db.cursor(mdb.cursors.DictCursor)
|
|
||||||
cur.execute(query, (k1, k2, context_id))
|
|
||||||
result = cur.fetchone()
|
|
||||||
if result:
|
|
||||||
self.log.debug("found '{0:s}'".format(result['v']))
|
|
||||||
return result['v']
|
|
||||||
except mdb.Error as e:
|
|
||||||
self.log.error("database error in _retrieve_random_v_for_k1_and_k2")
|
|
||||||
self.log.exception(e)
|
|
||||||
raise
|
|
||||||
finally: cur.close()
|
|
||||||
|
|
||||||
def _retrieve_random_v_for_k1_and_k2_with_pref(self, k1, k2, prefer, context_id):
|
|
||||||
"""Get one v for a given k1,k2.
|
|
||||||
|
|
||||||
Prefer that the result be prefer, if it's found.
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
self.log.debug("searching with '{0:s}','{1:s}', prefer "
|
|
||||||
"'{2:s}'".format(k1, k2, prefer))
|
|
||||||
values = []
|
|
||||||
db = self.get_db()
|
|
||||||
try:
|
|
||||||
query = '''
|
|
||||||
SELECT v FROM markov_chain AS r1
|
|
||||||
JOIN (
|
|
||||||
SELECT (RAND() * (SELECT MAX(id) FROM markov_chain)) AS id
|
|
||||||
) AS r2
|
|
||||||
WHERE r1.k1 = %s
|
|
||||||
AND r1.k2 = %s
|
|
||||||
AND r1.context_id = %s
|
|
||||||
ORDER BY r1.id >= r2.id DESC, r1.v = %s DESC, r1.id ASC
|
|
||||||
LIMIT 1
|
|
||||||
'''
|
|
||||||
cur = db.cursor(mdb.cursors.DictCursor)
|
|
||||||
cur.execute(query, (k1, k2, context_id, prefer))
|
|
||||||
result = cur.fetchone()
|
|
||||||
if result:
|
|
||||||
self.log.debug("found '{0:s}'".format(result['v']))
|
|
||||||
return result['v']
|
|
||||||
except mdb.Error as e:
|
|
||||||
self.log.error("database error in _retrieve_random_v_for_k1_and_k2_with_pref")
|
|
||||||
self.log.exception(e)
|
|
||||||
raise
|
|
||||||
finally: cur.close()
|
|
||||||
|
|
||||||
def _retrieve_random_k2_for_value(self, v, context_id):
|
|
||||||
"""Get one k2 for a given value."""
|
|
||||||
|
|
||||||
values = []
|
|
||||||
db = self.get_db()
|
|
||||||
try:
|
|
||||||
query = '''
|
|
||||||
SELECT k2 FROM markov_chain AS r1
|
|
||||||
JOIN (
|
|
||||||
SELECT (RAND() * (SELECT MAX(id) FROM markov_chain)) AS id
|
|
||||||
) AS r2
|
|
||||||
WHERE r1.v = %s
|
|
||||||
AND r1.context_id = %s
|
|
||||||
ORDER BY r1.id >= r2.id DESC, r1.id ASC
|
|
||||||
LIMIT 1
|
|
||||||
'''
|
|
||||||
cur = db.cursor(mdb.cursors.DictCursor)
|
|
||||||
cur.execute(query, (v, context_id))
|
|
||||||
result = cur.fetchone()
|
|
||||||
if result:
|
|
||||||
return result['k2']
|
|
||||||
except mdb.Error as e:
|
|
||||||
self.log.error("database error in _retrieve_random_k2_for_value")
|
|
||||||
self.log.exception(e)
|
|
||||||
raise
|
|
||||||
finally: cur.close()
|
|
||||||
|
|
||||||
def _create_chain_with_k1_k2(self, k1, k2, length, context_id,
|
|
||||||
avoid_address=False):
|
|
||||||
"""Create a chain of the given length, using k1,k2.
|
|
||||||
|
|
||||||
k1,k2 does not appear in the resulting chain.
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
chain = [k1, k2]
|
|
||||||
self.log.debug("creating chain for {0:s},{1:s}".format(k1, k2))
|
|
||||||
|
|
||||||
for _ in range(length):
|
|
||||||
v = self._retrieve_random_v_for_k1_and_k2(chain[-2],
|
|
||||||
chain[-1],
|
|
||||||
context_id)
|
|
||||||
if v:
|
|
||||||
chain.append(v)
|
|
||||||
|
|
||||||
# check for addresses (the "whoever:" in
|
|
||||||
# __start1 __start2 whoever: some words)
|
|
||||||
addressing_suffixes = [':', ',']
|
|
||||||
if len(chain) > 2 and chain[2][-1] in addressing_suffixes and avoid_address:
|
|
||||||
return chain[3:]
|
|
||||||
elif len(chain) > 2:
|
|
||||||
return chain[2:]
|
|
||||||
else:
|
|
||||||
return []
|
|
||||||
|
|
||||||
def _get_chatter_targets(self):
|
|
||||||
"""Get all possible chatter targets."""
|
|
||||||
|
|
||||||
db = self.get_db()
|
|
||||||
try:
|
|
||||||
# need to create our own db object, since this is likely going to be in a new thread
|
|
||||||
query = 'SELECT target, chance FROM markov_chatter_target'
|
|
||||||
cur = db.cursor(mdb.cursors.DictCursor)
|
|
||||||
cur.execute(query)
|
|
||||||
results = cur.fetchall()
|
|
||||||
return results
|
|
||||||
except mdb.Error as e:
|
|
||||||
self.log.error("database error in _get_chatter_targets")
|
|
||||||
self.log.exception(e)
|
|
||||||
raise
|
|
||||||
finally: cur.close()
|
|
||||||
|
|
||||||
def _get_context_id_for_target(self, target):
|
|
||||||
"""Get the context ID for the desired/input target."""
|
|
||||||
|
|
||||||
db = self.get_db()
|
|
||||||
try:
|
|
||||||
query = '''
|
|
||||||
SELECT mc.id FROM markov_context mc
|
|
||||||
INNER JOIN markov_target_to_context_map mt
|
|
||||||
ON mt.context_id = mc.id
|
|
||||||
WHERE mt.target = %s
|
|
||||||
'''
|
|
||||||
cur = db.cursor(mdb.cursors.DictCursor)
|
|
||||||
cur.execute(query, (target,))
|
|
||||||
result = cur.fetchone()
|
|
||||||
db.close()
|
|
||||||
if result:
|
|
||||||
return result['id']
|
|
||||||
else:
|
|
||||||
# auto-generate a context to keep things private
|
|
||||||
self._add_context_for_target(target)
|
|
||||||
return self._get_context_id_for_target(target)
|
|
||||||
except mdb.Error as e:
|
|
||||||
self.log.error("database error in _get_context_id_for_target")
|
|
||||||
self.log.exception(e)
|
|
||||||
raise
|
|
||||||
finally: cur.close()
|
|
||||||
|
|
||||||
def _add_context_for_target(self, target):
|
|
||||||
"""Create a new context for the desired/input target."""
|
|
||||||
|
|
||||||
db = self.get_db()
|
|
||||||
try:
|
|
||||||
statement = 'INSERT INTO markov_context (context) VALUES (%s)'
|
|
||||||
cur = db.cursor(mdb.cursors.DictCursor)
|
|
||||||
cur.execute(statement, (target,))
|
|
||||||
statement = '''
|
|
||||||
INSERT INTO markov_target_to_context_map (target, context_id)
|
|
||||||
VALUES (%s, (SELECT id FROM markov_context WHERE context = %s))
|
|
||||||
'''
|
|
||||||
cur.execute(statement, (target, target))
|
|
||||||
db.commit()
|
|
||||||
except mdb.Error as e:
|
|
||||||
db.rollback()
|
|
||||||
self.log.error("database error in _add_context_for_target")
|
|
||||||
self.log.exception(e)
|
|
||||||
raise
|
|
||||||
finally: cur.close()
|
|
||||||
|
|
||||||
try:
|
|
||||||
query = '''
|
|
||||||
SELECT mc.id FROM markov_context mc
|
|
||||||
INNER JOIN markov_target_to_context_map mt
|
|
||||||
ON mt.context_id = mc.id
|
|
||||||
WHERE mt.target = %s
|
|
||||||
'''
|
|
||||||
cur = db.cursor(mdb.cursors.DictCursor)
|
|
||||||
cur.execute(query, (target,))
|
|
||||||
result = cur.fetchone()
|
|
||||||
if result:
|
|
||||||
return result['id']
|
|
||||||
else:
|
|
||||||
# auto-generate a context to keep things private
|
|
||||||
self._add_context_for_target(target)
|
|
||||||
return self._get_context_id_for_target(target)
|
|
||||||
except mdb.Error as e:
|
|
||||||
self.log.error("database error in _get_context_id_for_target")
|
|
||||||
self.log.exception(e)
|
|
||||||
raise
|
|
||||||
finally: cur.close()
|
|
||||||
|
|
||||||
# vi:tabstop=4:expandtab:autoindent
|
# vi:tabstop=4:expandtab:autoindent
|
||||||
# kate: indent-mode python;indent-width 4;replace-tabs on;
|
|
||||||
|
|
Loading…
Reference in New Issue