| from dreamcoder.task import Task, EvaluationTimeout |
| import gc |
| from dreamcoder.utilities import * |
| from collections import Counter |
| import math |
|
|
| from dreamcoder.domains.regex.groundtruthRegexes import gt_dict |
|
|
| gt_dict = {"Data column no. "+str(num): r_str for num, r_str in gt_dict.items()} |
|
|
| class AllOrNothingLikelihoodModel: |
| def __init__(self, timeout=None): |
| self.timeout = timeout |
|
|
| def score(self, program, task): |
| logLikelihood = task.logLikelihood(program, self.timeout) |
| return valid(logLikelihood), logLikelihood |
|
|
|
|
| class EuclideanLikelihoodModel: |
| """Likelihood is based on Euclidean distance between features""" |
|
|
| def __init__(self, featureExtractor, successCutoff=0.9): |
| self.extract = featureExtractor |
| self.successCutoff = successCutoff |
|
|
| def score(self, program, task): |
| taskFeat = self.extract.featuresOfTask(task) |
| progFeat = self.extract.featuresOfProgram(program, task.request) |
| assert len(taskFeat) == len(progFeat) |
| distance = sum((x1 - x2)**2 for x1, x2 in zip(taskFeat, progFeat)) |
| logLikelihood = float(-distance) |
| return exp(logLikelihood) > self.successCutoff, logLikelihood |
|
|
| def longest_common_substr(arr): |
| |
|
|
| |
| |
| |
| |
| |
| n = len(arr) |
|
|
| |
| |
| s = arr[0] |
| l = len(s) |
| res = "" |
| for i in range(l) : |
| for j in range( i + 1, l + 1) : |
| |
| |
| stem = s[i:j] |
| k = 1 |
| for k in range(1, n): |
|
|
| |
| |
| if stem not in arr[k]: |
| break |
|
|
| |
| |
| |
| if (k + 1 == n and len(res) < len(stem)): res = stem |
| return res |
|
|
| def add_string_constants(tasks): |
| for task in tasks: |
| task.str_const = longest_common_substr([example[1] for example in task.examples]) |
| return tasks |
|
|
| def get_gt_ll(name, examples): |
| |
| import pregex as pre |
| r_str = gt_dict[name] |
| preg = pre.create(r_str) |
|
|
| if type(examples[0]) == list: |
| examples = [ "".join(example) for example in examples] |
|
|
| s = sum( preg.match(example) for example in examples) |
| if s == float("-inf"): |
| print("bad for ", name) |
| print('preg:', preg) |
| print('preg sample:', [preg.sample() for i in range(3)]) |
| print("exs", examples) |
| |
| return s |
|
|
|
|
| def add_cutoff_values(tasks, ll_cutoff): |
| from dreamcoder.domains.regex.makeRegexTasks import makeNewTasks |
| if ll_cutoff is None or ll_cutoff == "None": |
| for task in tasks: |
| task.ll_cutoff = None |
| return tasks |
| if ll_cutoff == "gt": |
| from dreamcoder.domains.regex.makeRegexTasks import regexHeldOutExamples |
| for task in tasks: |
| task.ll_cutoff = None |
| task.gt = get_gt_ll(task.name, [example[1] for example in task.examples]) |
| task.gt_test = get_gt_ll(task.name, |
| [example[1] for example in regexHeldOutExamples(task) ]) |
| return tasks |
| elif ll_cutoff == "plus": |
| for task in tasks: |
| task.ll_cutoff = regex_plus_bound([example[1] for example in task.examples]) |
| return tasks |
| elif ll_cutoff == "bigram": |
| eprint("WARNING: using entire corpus to make bigram model") |
| |
| model = make_corpus_bigram(show_tasks(makeNewTasks())) |
| for task in tasks: |
| task.ll_cutoff = bigram_corpus_score([example[1] for example in task.examples], model) |
| return tasks |
| elif ll_cutoff =="unigram": |
| eprint("WARNING: using entire corpus to make unigram model") |
| |
| model = make_corpus_unigram(show_tasks(makeNewTasks())) |
| for task in tasks: |
| task.ll_cutoff = unigram_corpus_score([example[1] for example in task.examples], model) |
| return tasks |
| elif ll_cutoff =="mix": |
| eprint("WARNING: using entire corpus to make bigram model") |
| eprint("WARNING: using entire corpus to make unigram model") |
| |
| unigram = make_corpus_unigram(show_tasks(makeNewTasks())) |
| bigram = make_corpus_bigram(show_tasks(makeNewTasks())) |
| for task in tasks: |
| uniscore = unigram_corpus_score([example[1] for example in task.examples], unigram) |
| biscore = bigram_corpus_score([example[1] for example in task.examples], bigram) |
| task.ll_cutoff = math.log(0.75*math.exp(biscore) + 0.25*math.exp(uniscore)) |
| return tasks |
| else: |
| eprint("not implemented") |
| eprint("cutoff val:") |
| eprint(ll_cutoff) |
| assert False |
|
|
| def show_tasks(dataset): |
| task_list = [] |
| for task in dataset: |
| task_list.append([example[1] for example in task.examples]) |
| return task_list |
|
|
| def regex_plus_bound(X): |
| from pregex import pregex |
| c = Counter(X) |
| regexes = [ |
| pregex.create(".+"), |
| pregex.create("\d+"), |
| pregex.create("\w+"), |
| pregex.create("\s+"), |
| pregex.create("\\u+"), |
| pregex.create("\l+")] |
| regex_scores = [] |
| for r in regexes: |
| regex_scores.append(sum(c[x] * r.match(x) for x in c)/float(sum([len(x) for x in X])) ) |
| return max(regex_scores) |
|
|
|
|
| def make_corpus_unigram(C): |
| str_list = [example + '\n' for task in C for example in task] |
| c = Counter(char for example in str_list for char in example ) |
| n = sum(c.values()) |
|
|
| logp = {x:math.log(c[x]/n) for x in c} |
| return logp |
|
|
| def unigram_corpus_score(X, logp): |
| task_ll = 0 |
| for x in X: |
| x = x + '\n' |
| task_ll += sum( logp.get(c, float('-inf')) for c in x)/len(x) |
|
|
| ll = task_ll/len(X) |
| return ll |
|
|
| def unigram_task_score(X): |
| """ |
| Given a list of strings, X, calculate the maximum log-likelihood per character for a unigram model over characters (including STOP symbol) |
| """ |
| c = Counter(x for s in X for x in s) |
| c.update("end" for s in X) |
| n = sum(c.values()) |
| logp = {x:math.log(c[x]/n) for x in c} |
| return sum(c[x]*logp[x] for x in c)/n |
|
|
| def make_corpus_bigram(C): |
| |
| |
|
|
| |
| str_list = [example + '\n' for task in C for example in task] |
|
|
| |
| head_count = Counter(element[0] for element in str_list) |
| head_n = sum(head_count.values()) |
| head_logp = {x:math.log(head_count[x]/head_n) for x in head_count} |
|
|
| body_count = Counter(element[i:i+2] for element in str_list for i in range(len(element)-1)) |
| body_bigram_n = sum(body_count.values()) |
| |
| body_character_n = Counter(char for element in str_list for char in element) |
| body_unigram_n = sum(body_character_n.values()) |
|
|
| body_logp = {x:math.log(body_count[x] / body_bigram_n / body_character_n[x[0]] * body_unigram_n) for x in body_count} |
|
|
| return {**head_logp, **body_logp} |
|
|
| def bigram_corpus_score(X, logp): |
| |
| task_ll = 0 |
| for x in X: |
| bigram_list = [x[0]] + [x[i:i+2] for i in range(len(x)-1)] + [x[-1] + '\n'] |
| bigram_list = [ ''.join(b) if isinstance(b,list) else b |
| for b in bigram_list ] |
|
|
| string_ll = sum(logp.get(bigram, float('-inf')) for bigram in bigram_list) |
|
|
| task_ll += string_ll |
|
|
| ll = task_ll |
| return ll |
|
|
|
|
| class ProbabilisticLikelihoodModel: |
|
|
| def __init__(self, timeout): |
| self.timeout = timeout |
| |
|
|
| def score(self, program, task): |
| |
| |
| try: |
| def timeoutCallBack(_1, _2): raise EvaluationTimeout() |
| signal.signal(signal.SIGVTALRM, timeoutCallBack) |
| signal.setitimer(signal.ITIMER_VIRTUAL, self.timeout) |
| try: |
| string_pregex = program.evaluate([]) |
| |
| |
| |
| preg = string_pregex |
| except IndexError: |
| |
| return False, NEGATIVEINFINITY |
| except Exception as e: |
| eprint("Exception during evaluation:", e) |
| if "Attempt to evaluate fragment variable" in e: |
| eprint("program (bc fragment error)", program) |
| return False, NEGATIVEINFINITY |
|
|
| |
|
|
| |
| |
| |
|
|
| cum_ll = 0 |
|
|
| example_list = [example[1] for example in task.examples] |
| c_example_list = Counter(example_list) |
|
|
| for c_example in c_example_list: |
| |
|
|
| try: |
| |
| |
| ll = preg.match(c_example) |
| |
| except ValueError as e: |
| eprint("ValueError:", e) |
| ll = float('-inf') |
| |
| |
| |
|
|
| if ll == float('-inf'): |
| return False, NEGATIVEINFINITY |
| else: |
| |
| |
|
|
| cum_ll += c_example_list[c_example] * ll |
| |
| |
| |
| |
| |
|
|
| normalized_cum_ll = cum_ll/ float(sum([len(example) for example in example_list])) |
|
|
|
|
|
|
| |
| |
| |
| success = normalized_cum_ll > task.ll_cutoff |
|
|
|
|
|
|
| |
|
|
| return success, normalized_cum_ll |
|
|
| except EvaluationTimeout: |
| eprint("Timed out while evaluating", program) |
| return False, NEGATIVEINFINITY |
| finally: |
| signal.signal(signal.SIGVTALRM, lambda *_: None) |
| signal.setitimer(signal.ITIMER_VIRTUAL, 0) |
|
|
|
|
| try: |
| import torch |
| import torch.nn as nn |
| import torch.nn.functional as F |
| from torch.nn.utils.rnn import pack_padded_sequence, pad_packed_sequence |
| from torch.autograd import Variable |
|
|
| class FeatureDiscriminatorLikelihoodModel(nn.Module): |
| def __init__(self, tasks, featureExtractor, |
| successCutoff=0.6, H=8, trainingSuccessRatio=0.5): |
| super(FeatureDiscriminatorLikelihoodModel, self).__init__() |
| self.extract = featureExtractor |
| self.successCutoff = successCutoff |
| self.trainingSuccessRatio = trainingSuccessRatio |
|
|
| self.W = nn.Linear(featureExtractor.outputDimensionality, H) |
| self.output = nn.Linear(H, 1) |
|
|
| |
| self.train(tasks) |
|
|
| def forward(self, examples): |
| """ |
| Examples is a list of feature sets corresponding to a particular example. |
| Output in [0,1] whether all examples correspond to the same program |
| """ |
| assert all( |
| len(x) == self.extract.outputDimensionality for x in examples) |
| examples = [F.tanh(self.W(ex)) for ex in examples] |
| maxed, _ = torch.max(torch.stack(examples), dim=0) |
| return F.sigmoid(self.output(maxed)) |
|
|
| def train(self, tasks, steps=400): |
| |
| optimizer = torch.optim.Adam(self.parameters()) |
| with timing("Trained discriminator"): |
| losses = [] |
| for i in range(steps): |
| self.zero_grad() |
| if random.random() <= self.trainingSuccessRatio: |
| |
| t = random.choice(tasks) |
| features = [self.extract.featuresOfTask( |
| Task(t.name, t.request, [ex], t.features)) |
| for ex in t.examples] |
| loss = (self(features) - 1.0)**2 |
| else: |
| |
| t1, t2 = random.sample(tasks, 2) |
| features1 = [self.extract.featuresOfTask( |
| Task(t1.name, t1.request, [ex], t1.features)) |
| for ex in t1.examples[:len(t1.examples) / 2]] |
| features2 = [self.extract.featuresOfTask( |
| Task(t2.name, t2.request, [ex], t2.features)) |
| for ex in t2.examples[len(t2.examples) / 2:]] |
| features = features1 + features2 |
| loss = self(features)**2 |
|
|
| loss.backward() |
| optimizer.step() |
| losses.append(loss.data[0]) |
| if not i % 50: |
| eprint( |
| "Discriminator Epoch", |
| i, |
| "Loss", |
| sum(losses) / |
| len(losses)) |
| gc.collect() |
|
|
| def score(self, program, task): |
| taskFeatures = self.extract.featuresOfTask(task) |
| progFeatures = self.extract.featuresOfProgram( |
| program, task.request) |
| likelihood = self([taskFeatures] + [progFeatures]) |
| likelihood = float(likelihood) |
| return likelihood > self.successCutoff, log(likelihood) |
| except ImportError: |
| pass |
|
|
|
|
| if __name__=="__main__": |
|
|
| arr = ['MAM.OSBS.2014.06', 'MAM.OSBS.2013.07', 'MAM.OSBS.2013.09', 'MAM.OSBS.2014.05', 'MAM.OSBS.2014.11'] |
| stems = longest_common_substr(arr) |
| print(stems) |
|
|
|
|
|
|
|
|