|
import datetime |
|
import json |
|
import os |
|
import pickle |
|
import subprocess |
|
import sys |
|
|
|
from dreamcoder.fragmentGrammar import FragmentGrammar |
|
from dreamcoder.frontier import Frontier, FrontierEntry |
|
from dreamcoder.grammar import Grammar |
|
from dreamcoder.task import Task |
|
from dreamcoder.program import Program, Invented |
|
from dreamcoder.utilities import eprint, timing, callCompiled, get_root_dir |
|
from dreamcoder.vs import induceGrammar_Beta |
|
|
|
|
|
def induceGrammar(*args, **kwargs): |
|
if sum(not f.empty for f in args[1]) == 0: |
|
eprint("No nonempty frontiers, exiting grammar induction early.") |
|
return args[0], args[1] |
|
backend = kwargs.pop("backend", "pypy") |
|
if 'pypy' in backend: |
|
|
|
|
|
|
|
|
|
g0,frontiers = args[0].strip_primitive_values(), \ |
|
[front.strip_primitive_values() for front in args[1]] |
|
original_tasks = {f.task.name: f.task for f in frontiers} |
|
frontiers = [Frontier(f.entries, Task(f.task.name,f.task.request,[])) |
|
for f in frontiers ] |
|
args = [g0,frontiers] |
|
|
|
|
|
with timing("Induced a grammar"): |
|
if backend == "pypy": |
|
g, newFrontiers = callCompiled(pypyInduce, *args, **kwargs) |
|
elif backend == "rust": |
|
g, newFrontiers = rustInduce(*args, **kwargs) |
|
elif backend == "vs": |
|
g, newFrontiers = rustInduce(*args, vs=True, **kwargs) |
|
elif backend == "pypy_vs": |
|
kwargs.pop('iteration') |
|
kwargs.pop('topk_use_only_likelihood') |
|
fn = '/tmp/vs.pickle' |
|
with open(fn, 'wb') as handle: |
|
pickle.dump((args, kwargs), handle) |
|
eprint("For debugging purposes, the version space compression invocation has been saved to", fn) |
|
g, newFrontiers = callCompiled(induceGrammar_Beta, *args, **kwargs) |
|
elif backend == "ocaml": |
|
kwargs.pop('iteration') |
|
kwargs.pop('topk_use_only_likelihood') |
|
kwargs['topI'] = 300 |
|
kwargs['bs'] = 1000000 |
|
g, newFrontiers = ocamlInduce(*args, **kwargs) |
|
elif backend == "memorize": |
|
g, newFrontiers = memorizeInduce(*args, **kwargs) |
|
else: |
|
assert False, "unknown compressor" |
|
|
|
if 'pypy' in backend: |
|
g, newFrontiers = g.unstrip_primitive_values(), \ |
|
[front.unstrip_primitive_values() for front in newFrontiers] |
|
newFrontiers = [Frontier(f.entries, original_tasks[f.task.name]) |
|
for f in newFrontiers] |
|
|
|
|
|
return g, newFrontiers |
|
|
|
def memorizeInduce(g, frontiers, **kwargs): |
|
existingInventions = {p.uncurry() |
|
for p in g.primitives } |
|
programs = {f.bestPosterior.program for f in frontiers if not f.empty} |
|
newInventions = programs - existingInventions |
|
newGrammar = Grammar.uniform([p for p in g.primitives] + \ |
|
[Invented(ni) for ni in newInventions]) |
|
|
|
|
|
def substitute(p): |
|
nonlocal newInventions |
|
if p in newInventions: return Invented(p).uncurry() |
|
return p |
|
newFrontiers = [Frontier([FrontierEntry(program=np, |
|
logPrior=newGrammar.logLikelihood(f.task.request, np), |
|
logLikelihood=e.logLikelihood) |
|
for e in f |
|
for np in [substitute(e.program)] ], |
|
task=f.task) |
|
for f in frontiers ] |
|
return newGrammar, newFrontiers |
|
|
|
|
|
|
|
|
|
|
|
def pypyInduce(*args, **kwargs): |
|
kwargs.pop('iteration') |
|
return FragmentGrammar.induceFromFrontiers(*args, **kwargs) |
|
|
|
|
|
def ocamlInduce(g, frontiers, _=None, |
|
topK=1, pseudoCounts=1.0, aic=1.0, |
|
structurePenalty=0.001, a=0, CPUs=1, |
|
bs=1000000, topI=300): |
|
|
|
|
|
|
|
|
|
if all(not p.isInvented for p in g.primitives): |
|
if a > 3: |
|
CPUs = max(1, int(CPUs / 6)) |
|
else: |
|
CPUs = max(1, int(CPUs / 3)) |
|
else: |
|
CPUs = max(1, int(CPUs / 2)) |
|
CPUs = 2 |
|
|
|
|
|
|
|
iterations = 99 |
|
|
|
while True: |
|
g0 = g |
|
|
|
originalFrontiers = frontiers |
|
t2f = {f.task: f for f in frontiers} |
|
frontiers = [f for f in frontiers if not f.empty] |
|
message = {"arity": a, |
|
"topK": topK, |
|
"pseudoCounts": float(pseudoCounts), |
|
"aic": aic, |
|
"bs": bs, |
|
"topI": topI, |
|
"structurePenalty": float(structurePenalty), |
|
"CPUs": CPUs, |
|
"DSL": g.json(), |
|
"iterations": iterations, |
|
"frontiers": [f.json() |
|
for f in frontiers]} |
|
|
|
message = json.dumps(message) |
|
if True: |
|
timestamp = datetime.datetime.now().isoformat() |
|
os.system("mkdir -p compressionMessages") |
|
fn = "compressionMessages/%s" % timestamp |
|
with open(fn, "w") as f: |
|
f.write(message) |
|
eprint("Compression message saved to:", fn) |
|
|
|
try: |
|
|
|
compressor_file = os.path.join(get_root_dir(), 'compression') |
|
process = subprocess.Popen(compressor_file, |
|
stdin=subprocess.PIPE, |
|
stdout=subprocess.PIPE) |
|
response, error = process.communicate(bytes(message, encoding="utf-8")) |
|
response = json.loads(response.decode("utf-8")) |
|
except OSError as exc: |
|
raise exc |
|
|
|
g = response["DSL"] |
|
g = Grammar(g["logVariable"], |
|
[(l, p.infer(), p) |
|
for production in g["productions"] |
|
for l in [production["logProbability"]] |
|
for p in [Program.parse(production["expression"])]], |
|
continuationType=g0.continuationType) |
|
|
|
frontiers = {original.task: |
|
Frontier([FrontierEntry(p, |
|
logLikelihood=e["logLikelihood"], |
|
logPrior=g.logLikelihood(original.task.request, p)) |
|
for e in new["programs"] |
|
for p in [Program.parse(e["program"])]], |
|
task=original.task) |
|
for original, new in zip(frontiers, response["frontiers"])} |
|
frontiers = [frontiers.get(f.task, t2f[f.task]) |
|
for f in originalFrontiers] |
|
if iterations == 1 and len(g) > len(g0): |
|
eprint("Grammar changed - running another round of consolidation.") |
|
continue |
|
else: |
|
eprint("Finished consolidation.") |
|
return g, frontiers |
|
|
|
|
|
def rustInduce(g0, frontiers, _=None, |
|
topK=1, pseudoCounts=1.0, aic=1.0, |
|
structurePenalty=0.001, a=0, CPUs=1, iteration=-1, |
|
topk_use_only_likelihood=False, |
|
vs=False): |
|
def finite_logp(l): |
|
return l if l != float("-inf") else -1000 |
|
|
|
message = { |
|
"strategy": {"version-spaces": {"top_i": 50}} |
|
if vs else |
|
{"fragment-grammars": {}}, |
|
"params": { |
|
"structure_penalty": structurePenalty, |
|
"pseudocounts": int(pseudoCounts + 0.5), |
|
"topk": topK, |
|
"topk_use_only_likelihood": topk_use_only_likelihood, |
|
"aic": aic if aic != float("inf") else None, |
|
"arity": a, |
|
}, |
|
"primitives": [{"name": p.name, "tp": str(t), "logp": finite_logp(l)} |
|
for l, t, p in g0.productions if p.isPrimitive], |
|
"inventions": [{"expression": str(p.body), |
|
"logp": finite_logp(l)} |
|
for l, t, p in g0.productions if p.isInvented], |
|
"variable_logprob": finite_logp(g0.logVariable), |
|
"frontiers": [{ |
|
"task_tp": str(f.task.request), |
|
"solutions": [{ |
|
"expression": str(e.program), |
|
"logprior": finite_logp(e.logPrior), |
|
"loglikelihood": e.logLikelihood, |
|
} for e in f], |
|
} for f in frontiers], |
|
} |
|
|
|
eprint("running rust compressor") |
|
|
|
messageJson = json.dumps(message) |
|
|
|
with open("jsonDebug", "w") as f: |
|
f.write(messageJson) |
|
|
|
|
|
|
|
if sys.version_info[1] >= 6: |
|
p = subprocess.Popen( |
|
['./rust_compressor/rust_compressor'], |
|
encoding='utf-8', |
|
stdin=subprocess.PIPE, |
|
stdout=subprocess.PIPE) |
|
elif sys.version_info[1] == 5: |
|
p = subprocess.Popen( |
|
['./rust_compressor/rust_compressor'], |
|
stdin=subprocess.PIPE, |
|
stdout=subprocess.PIPE) |
|
|
|
messageJson = bytearray(messageJson, encoding='utf-8') |
|
|
|
else: |
|
eprint("must be python 3.5 or 3.6") |
|
assert False |
|
|
|
p.stdin.write(messageJson) |
|
p.stdin.flush() |
|
p.stdin.close() |
|
|
|
if p.returncode is not None: |
|
raise ValueError("rust compressor failed") |
|
|
|
if sys.version_info[1] >= 6: |
|
resp = json.load(p.stdout) |
|
elif sys.version_info[1] == 5: |
|
import codecs |
|
resp = json.load(codecs.getreader('utf-8')(p.stdout)) |
|
|
|
productions = [(x["logp"], p) for p, x in |
|
zip((p for (_, _, p) in g0.productions if p.isPrimitive), resp["primitives"])] + \ |
|
[(i["logp"], Invented(Program.parse(i["expression"]))) |
|
for i in resp["inventions"]] |
|
productions = [(l if l is not None else float("-inf"), p) |
|
for l, p in productions] |
|
g = Grammar.fromProductions(productions, resp["variable_logprob"], continuationType=g0.continuationType) |
|
newFrontiers = [ |
|
Frontier( |
|
[ |
|
FrontierEntry( |
|
Program.parse( |
|
s["expression"]), |
|
logPrior=s["logprior"], |
|
logLikelihood=s["loglikelihood"]) for s in r["solutions"]], |
|
f.task) for f, |
|
r in zip( |
|
frontiers, |
|
resp["frontiers"])] |
|
return g, newFrontiers |
|
|