| | import datetime |
| |
|
| | import dill |
| |
|
| | from dreamcoder.compression import induceGrammar |
| | from dreamcoder.utilities import * |
| | try: |
| | from dreamcoder.recognition import * |
| | except: |
| | eprint("Failure loading recognition - only acceptable if using pypy ") |
| | from dreamcoder.enumeration import * |
| | from dreamcoder.fragmentGrammar import * |
| | from dreamcoder.taskBatcher import * |
| | from dreamcoder.primitiveGraph import graphPrimitives |
| | from dreamcoder.dreaming import backgroundHelmholtzEnumeration |
| |
|
| |
|
| | class ECResult(): |
| | def __init__(self, _=None, |
| | frontiersOverTime=None, |
| | testingSearchTime=None, |
| | learningCurve=None, |
| | grammars=None, |
| | taskSolutions=None, |
| | averageDescriptionLength=None, |
| | parameters=None, |
| | recognitionModel=None, |
| | searchTimes=None, |
| | recognitionTaskMetrics=None, |
| | numTestingTasks=None, |
| | sumMaxll=None, |
| | testingSumMaxll=None, |
| | hitsAtEachWake=None, |
| | timesAtEachWake=None, |
| | allFrontiers=None): |
| | self.frontiersOverTime = {} |
| | self.hitsAtEachWake = hitsAtEachWake or [] |
| | self.timesAtEachWake = timesAtEachWake or [] |
| | self.testingSearchTime = testingSearchTime or [] |
| | self.searchTimes = searchTimes or [] |
| | self.trainSearchTime = {} |
| | self.testSearchTime = {} |
| | self.recognitionTaskMetrics = recognitionTaskMetrics or {} |
| | self.recognitionModel = recognitionModel |
| | self.averageDescriptionLength = averageDescriptionLength or [] |
| | self.parameters = parameters |
| | self.learningCurve = learningCurve or [] |
| | self.grammars = grammars or [] |
| | self.taskSolutions = taskSolutions or {} |
| | self.numTestingTasks = numTestingTasks |
| | self.sumMaxll = sumMaxll or [] |
| | self.testingSumMaxll = testingSumMaxll or [] |
| | self.allFrontiers = allFrontiers or {} |
| |
|
| | def __repr__(self): |
| | attrs = ["{}={}".format(k, v) for k, v in self.__dict__.items()] |
| | return "ECResult({})".format(", ".join(attrs)) |
| |
|
| | def getTestingTasks(self): |
| | testing = [] |
| | training = self.taskSolutions.keys() |
| | for t in self.recognitionTaskMetrics: |
| | if isinstance(t, Task) and t not in training: testing.append(t) |
| | return testing |
| |
|
| |
|
| | def recordFrontier(self, frontier): |
| | t = frontier.task |
| | if t not in self.frontiersOverTime: self.frontiersOverTime[t] = [] |
| | self.frontiersOverTime[t].append(frontier) |
| |
|
| | |
| | |
| | abbreviations = {"frontierSize": "fs", |
| | "useDSL": "DSL", |
| | "taskReranker": "TRR", |
| | "matrixRank": "MR", |
| | "reuseRecognition": "RR", |
| | "ensembleSize": "ES", |
| | "recognitionTimeout": "RT", |
| | "recognitionSteps": "RS", |
| | "iterations": "it", |
| | "maximumFrontier": "MF", |
| | "pseudoCounts": "pc", |
| | "auxiliaryLoss": "aux", |
| | "structurePenalty": "L", |
| | "helmholtzRatio": "HR", |
| | "biasOptimal": "BO", |
| | "contextual": "CO", |
| | "topK": "K", |
| | "enumerationTimeout": "ET", |
| | "useRecognitionModel": "rec", |
| | "use_ll_cutoff": "llcut", |
| | "topk_use_only_likelihood": "topkNotMAP", |
| | "activation": "act", |
| | "storeTaskMetrics": 'STM', |
| | "topkNotMAP": "tknm", |
| | "rewriteTaskMetrics": "RW", |
| | 'taskBatchSize': 'batch'} |
| |
|
| | @staticmethod |
| | def abbreviate(parameter): return ECResult.abbreviations.get(parameter, parameter) |
| |
|
| | @staticmethod |
| | def parameterOfAbbreviation(abbreviation): |
| | return ECResult.abbreviationToParameter.get(abbreviation, abbreviation) |
| |
|
| | @staticmethod |
| | def clearRecognitionModel(path): |
| | SUFFIX = '.pickle' |
| | assert path.endswith(SUFFIX) |
| | |
| | with open(path,'rb') as handle: |
| | result = dill.load(handle) |
| | |
| | result.recognitionModel = None |
| | |
| | clearedPath = path[:-len(SUFFIX)] + "_graph=True" + SUFFIX |
| | with open(clearedPath,'wb') as handle: |
| | result = dill.dump(result, handle) |
| | eprint(" [+] Cleared recognition model from:") |
| | eprint(" %s"%path) |
| | eprint(" and exported to:") |
| | eprint(" %s"%clearedPath) |
| | eprint(" Use this one for graphing.") |
| |
|
| |
|
| | ECResult.abbreviationToParameter = { |
| | v: k for k, v in ECResult.abbreviations.items()} |
| |
|
| |
|
| | def explorationCompression(*arguments, **keywords): |
| | for r in ecIterator(*arguments, **keywords): |
| | pass |
| | return r |
| |
|
| |
|
| | def ecIterator(grammar, tasks, |
| | _=None, |
| | useDSL=True, |
| | noConsolidation=False, |
| | mask=False, |
| | seed=0, |
| | addFullTaskMetrics=False, |
| | matrixRank=None, |
| | solver='ocaml', |
| | compressor="rust", |
| | biasOptimal=False, |
| | contextual=False, |
| | testingTasks=[], |
| | iterations=None, |
| | resume=None, |
| | enumerationTimeout=None, |
| | testingTimeout=None, |
| | testEvery=1, |
| | reuseRecognition=False, |
| | ensembleSize=1, |
| | useRecognitionModel=True, |
| | recognitionTimeout=None, |
| | recognitionSteps=None, |
| | helmholtzRatio=0., |
| | featureExtractor=None, |
| | activation='relu', |
| | topK=1, |
| | topk_use_only_likelihood=False, |
| | use_map_search_times=True, |
| | maximumFrontier=None, |
| | pseudoCounts=1.0, aic=1.0, |
| | structurePenalty=0.001, arity=0, |
| | evaluationTimeout=1.0, |
| | taskBatchSize=None, |
| | taskReranker='default', |
| | CPUs=1, |
| | cuda=False, |
| | message="", |
| | outputPrefix=None, |
| | storeTaskMetrics=False, |
| | rewriteTaskMetrics=True, |
| | auxiliaryLoss=False, |
| | custom_wake_generative=None): |
| | if enumerationTimeout is None: |
| | eprint( |
| | "Please specify an enumeration timeout:", |
| | "explorationCompression(..., enumerationTimeout = ..., ...)") |
| | assert False |
| | if iterations is None: |
| | eprint( |
| | "Please specify a iteration count: explorationCompression(..., iterations = ...)") |
| | assert False |
| | if useRecognitionModel and featureExtractor is None: |
| | eprint("Warning: Recognition model needs feature extractor.", |
| | "Ignoring recognition model.") |
| | useRecognitionModel = False |
| | if ensembleSize > 1 and not useRecognitionModel: |
| | eprint("Warning: ensemble size requires using the recognition model, aborting.") |
| | assert False |
| | if biasOptimal and not useRecognitionModel: |
| | eprint("Bias optimality only applies to recognition models, aborting.") |
| | assert False |
| | if contextual and not useRecognitionModel: |
| | eprint("Contextual only applies to recognition models, aborting") |
| | assert False |
| | if reuseRecognition and not useRecognitionModel: |
| | eprint("Reuse of recognition model weights at successive iteration only applies to recognition models, aborting") |
| | assert False |
| | if matrixRank is not None and not contextual: |
| | eprint("Matrix rank only applies to contextual recognition models, aborting") |
| | assert False |
| | assert useDSL or useRecognitionModel, "You specified that you didn't want to use the DSL AND you don't want to use the recognition model. Figure out what you want to use." |
| | if testingTimeout > 0 and len(testingTasks) == 0: |
| | eprint("You specified a testingTimeout, but did not provide any held out testing tasks, aborting.") |
| | assert False |
| |
|
| | |
| | |
| | parameters = { |
| | k: v for k, |
| | v in locals().items() if k not in { |
| | "tasks", |
| | "use_map_search_times", |
| | "seed", |
| | "activation", |
| | "grammar", |
| | "cuda", |
| | "_", |
| | "testingTimeout", |
| | "testEvery", |
| | "message", |
| | "CPUs", |
| | "outputPrefix", |
| | "resume", |
| | "resumeFrontierSize", |
| | "addFullTaskMetrics", |
| | "featureExtractor", |
| | "evaluationTimeout", |
| | "testingTasks", |
| | "compressor", |
| | "custom_wake_generative"} and v is not None} |
| | if not useRecognitionModel: |
| | for k in {"helmholtzRatio", "recognitionTimeout", "biasOptimal", "mask", |
| | "contextual", "matrixRank", "reuseRecognition", "auxiliaryLoss", "ensembleSize"}: |
| | if k in parameters: del parameters[k] |
| | else: del parameters["useRecognitionModel"]; |
| | if useRecognitionModel and not contextual: |
| | if "matrixRank" in parameters: |
| | del parameters["matrixRank"] |
| | if "mask" in parameters: |
| | del parameters["mask"] |
| | if not mask and 'mask' in parameters: del parameters["mask"] |
| | if not auxiliaryLoss and 'auxiliaryLoss' in parameters: del parameters['auxiliaryLoss'] |
| | if not useDSL: |
| | for k in {"structurePenalty", "pseudoCounts", "aic"}: |
| | del parameters[k] |
| | else: del parameters["useDSL"] |
| | |
| | |
| | def checkpointPath(iteration, extra=""): |
| | parameters["iterations"] = iteration |
| | kvs = [ |
| | "{}={}".format( |
| | ECResult.abbreviate(k), |
| | parameters[k]) for k in sorted( |
| | parameters.keys())] |
| | return "{}_{}{}.pickle".format(outputPrefix, "_".join(kvs), extra) |
| |
|
| | if message: |
| | message = " (" + message + ")" |
| | eprint("Running EC%s on %s @ %s with %d CPUs and parameters:" % |
| | (message, os.uname()[1], datetime.datetime.now(), CPUs)) |
| | for k, v in parameters.items(): |
| | eprint("\t", k, " = ", v) |
| | eprint("\t", "evaluationTimeout", " = ", evaluationTimeout) |
| | eprint("\t", "cuda", " = ", cuda) |
| | eprint() |
| |
|
| | if addFullTaskMetrics: |
| | assert resume is not None, "--addFullTaskMetrics requires --resume" |
| |
|
| | def reportMemory(): |
| | eprint(f"Currently using this much memory: {getThisMemoryUsage()}") |
| | |
| | |
| | if resume is not None: |
| | try: |
| | resume = int(resume) |
| | path = checkpointPath(resume) |
| | except ValueError: |
| | path = resume |
| | with open(path, "rb") as handle: |
| | result = dill.load(handle) |
| | resume = len(result.grammars) - 1 |
| | eprint("Loaded checkpoint from", path) |
| | grammar = result.grammars[-1] if result.grammars else grammar |
| | else: |
| | |
| | numTestingTasks = len(testingTasks) if len(testingTasks) != 0 else None |
| |
|
| | result = ECResult(parameters=parameters, |
| | grammars=[grammar], |
| | taskSolutions={ |
| | t: Frontier([], |
| | task=t) for t in tasks}, |
| | recognitionModel=None, numTestingTasks=numTestingTasks, |
| | allFrontiers={ |
| | t: Frontier([], |
| | task=t) for t in tasks}) |
| |
|
| |
|
| | |
| | if taskReranker == 'default': |
| | taskBatcher = DefaultTaskBatcher() |
| | elif taskReranker == 'random': |
| | taskBatcher = RandomTaskBatcher() |
| | elif taskReranker == 'randomShuffle': |
| | taskBatcher = RandomShuffleTaskBatcher(seed) |
| | elif taskReranker == 'unsolved': |
| | taskBatcher = UnsolvedTaskBatcher() |
| | elif taskReranker == 'unsolvedEntropy': |
| | taskBatcher = UnsolvedEntropyTaskBatcher() |
| | elif taskReranker == 'unsolvedRandomEntropy': |
| | taskBatcher = UnsolvedRandomEntropyTaskBatcher() |
| | elif taskReranker == 'randomkNN': |
| | taskBatcher = RandomkNNTaskBatcher() |
| | elif taskReranker == 'randomLowEntropykNN': |
| | taskBatcher = RandomLowEntropykNNTaskBatcher() |
| | else: |
| | eprint("Invalid task reranker: " + taskReranker + ", aborting.") |
| | assert False |
| |
|
| | |
| | if addFullTaskMetrics: |
| | if testingTimeout is not None and testingTimeout > enumerationTimeout: |
| | enumerationTimeout = testingTimeout |
| | if result.recognitionModel is not None: |
| | _enumerator = lambda *args, **kw: result.recognitionModel.enumerateFrontiers(*args, **kw) |
| | else: _enumerator = lambda *args, **kw: multicoreEnumeration(result.grammars[-1], *args, **kw) |
| | enumerator = lambda *args, **kw: _enumerator(*args, |
| | maximumFrontier=maximumFrontier, |
| | CPUs=CPUs, evaluationTimeout=evaluationTimeout, |
| | solver=solver, |
| | **kw) |
| | trainFrontiers, _, trainingTimes = enumerator(tasks, enumerationTimeout=enumerationTimeout) |
| | testFrontiers, _, testingTimes = enumerator(testingTasks, enumerationTimeout=testingTimeout, testing=True) |
| |
|
| | recognizer = result.recognitionModel |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, trainingTimes, 'recognitionBestTimes') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, recognizer.taskGrammarLogProductions(tasks), 'taskLogProductions') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, recognizer.taskGrammarEntropies(tasks), 'taskGrammarEntropies') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskAuxiliaryLossLayer(tasks), 'taskAuxiliaryLossLayer') |
| | |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, testingTimes, 'heldoutTestingTimes') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, recognizer.taskGrammarLogProductions(testingTasks), 'heldoutTaskLogProductions') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, recognizer.taskGrammarEntropies(testingTasks), 'heldoutTaskGrammarEntropies') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskAuxiliaryLossLayer(testingTasks), 'heldoutAuxiliaryLossLayer') |
| |
|
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, {f.task: f |
| | for f in trainFrontiers + testFrontiers |
| | if len(f) > 0}, |
| | 'frontier') |
| | SUFFIX = ".pickle" |
| | assert path.endswith(SUFFIX) |
| | path = path[:-len(SUFFIX)] + "_FTM=True" + SUFFIX |
| | with open(path, "wb") as handle: dill.dump(result, handle) |
| | if useRecognitionModel: ECResult.clearRecognitionModel(path) |
| | |
| | sys.exit(0) |
| | |
| | |
| | for j in range(resume or 0, iterations): |
| | if storeTaskMetrics and rewriteTaskMetrics: |
| | eprint("Resetting task metrics for next iteration.") |
| | result.recognitionTaskMetrics = {} |
| |
|
| | reportMemory() |
| |
|
| | |
| | if testingTimeout > 0 and ((j % testEvery == 0) or (j == iterations - 1)): |
| | eprint("Evaluating on held out testing tasks for iteration: %d" % (j)) |
| | evaluateOnTestingTasks(result, testingTasks, grammar, |
| | CPUs=CPUs, maximumFrontier=maximumFrontier, |
| | solver=solver, |
| | enumerationTimeout=testingTimeout, evaluationTimeout=evaluationTimeout) |
| | |
| | |
| | if useRecognitionModel and biasOptimal and helmholtzRatio > 0 and \ |
| | all( str(p) != "REAL" for p in grammar.primitives ): |
| | |
| | if useDSL or 'helmholtzFrontiers' not in locals(): |
| | helmholtzFrontiers = backgroundHelmholtzEnumeration(tasks, grammar, enumerationTimeout, |
| | evaluationTimeout=evaluationTimeout, |
| | special=featureExtractor.special) |
| | else: |
| | print("Reusing dreams from previous iteration.") |
| | else: |
| | helmholtzFrontiers = lambda: [] |
| |
|
| | reportMemory() |
| |
|
| | |
| | wakingTaskBatch = taskBatcher.getTaskBatch(result, tasks, taskBatchSize, j) |
| | eprint("Using a waking task batch of size: " + str(len(wakingTaskBatch))) |
| |
|
| | |
| | if useDSL: |
| | wake_generative = custom_wake_generative if custom_wake_generative is not None else default_wake_generative |
| | topDownFrontiers, times = wake_generative(grammar, wakingTaskBatch, |
| | solver=solver, |
| | maximumFrontier=maximumFrontier, |
| | enumerationTimeout=enumerationTimeout, |
| | CPUs=CPUs, |
| | evaluationTimeout=evaluationTimeout) |
| | result.trainSearchTime = {t: tm for t, tm in times.items() if tm is not None} |
| | else: |
| | eprint("Skipping top-down enumeration because we are not using the generative model") |
| | topDownFrontiers, times = [], {t: None for t in wakingTaskBatch } |
| |
|
| | tasksHitTopDown = {f.task for f in topDownFrontiers if not f.empty} |
| | result.hitsAtEachWake.append(len(tasksHitTopDown)) |
| |
|
| | reportMemory() |
| |
|
| | |
| | for f in topDownFrontiers: |
| | if f.task not in result.allFrontiers: continue |
| | result.allFrontiers[f.task] = result.allFrontiers[f.task].combine(f).topK(maximumFrontier) |
| |
|
| | eprint("Frontiers discovered top down: " + str(len(tasksHitTopDown))) |
| | eprint("Total frontiers: " + str(len([f for f in result.allFrontiers.values() if not f.empty]))) |
| |
|
| | |
| | if useRecognitionModel: |
| | |
| | previousRecognitionModel = None |
| | if reuseRecognition and result.recognitionModel is not None: |
| | previousRecognitionModel = result.recognitionModel |
| |
|
| | thisRatio = helmholtzRatio |
| | |
| | if all( f.empty for f in result.allFrontiers.values() ): thisRatio = 1. |
| |
|
| | tasksHitBottomUp = \ |
| | sleep_recognition(result, grammar, wakingTaskBatch, tasks, testingTasks, result.allFrontiers.values(), |
| | ensembleSize=ensembleSize, featureExtractor=featureExtractor, mask=mask, |
| | activation=activation, contextual=contextual, biasOptimal=biasOptimal, |
| | previousRecognitionModel=previousRecognitionModel, matrixRank=matrixRank, |
| | timeout=recognitionTimeout, evaluationTimeout=evaluationTimeout, |
| | enumerationTimeout=enumerationTimeout, |
| | helmholtzRatio=thisRatio, helmholtzFrontiers=helmholtzFrontiers(), |
| | auxiliaryLoss=auxiliaryLoss, cuda=cuda, CPUs=CPUs, solver=solver, |
| | recognitionSteps=recognitionSteps, maximumFrontier=maximumFrontier) |
| |
|
| | showHitMatrix(tasksHitTopDown, tasksHitBottomUp, wakingTaskBatch) |
| | |
| | |
| | result.taskSolutions = {f.task: f.topK(topK) |
| | for f in result.allFrontiers.values()} |
| | for f in result.allFrontiers.values(): result.recordFrontier(f) |
| | result.learningCurve += [ |
| | sum(f is not None and not f.empty for f in result.taskSolutions.values())] |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, {f.task: f |
| | for f in result.allFrontiers.values() |
| | if len(f) > 0}, |
| | 'frontier') |
| | |
| | |
| | if useDSL and not(noConsolidation): |
| | eprint(f"Currently using this much memory: {getThisMemoryUsage()}") |
| | grammar = consolidate(result, grammar, topK=topK, pseudoCounts=pseudoCounts, arity=arity, aic=aic, |
| | structurePenalty=structurePenalty, compressor=compressor, CPUs=CPUs, |
| | iteration=j) |
| | eprint(f"Currently using this much memory: {getThisMemoryUsage()}") |
| | else: |
| | eprint("Skipping consolidation.") |
| | result.grammars.append(grammar) |
| | |
| | if outputPrefix is not None: |
| | path = checkpointPath(j + 1) |
| | with open(path, "wb") as handle: |
| | try: |
| | dill.dump(result, handle) |
| | except TypeError as e: |
| | eprint(result) |
| | assert(False) |
| | eprint("Exported checkpoint to", path) |
| | if useRecognitionModel: |
| | ECResult.clearRecognitionModel(path) |
| |
|
| | graphPrimitives(result, "%s_primitives_%d_"%(outputPrefix,j)) |
| | |
| |
|
| | yield result |
| |
|
| |
|
| | def showHitMatrix(top, bottom, tasks): |
| | tasks = set(tasks) |
| |
|
| | total = bottom | top |
| | eprint(len(total), "/", len(tasks), "total hit tasks") |
| | bottomMiss = tasks - bottom |
| | topMiss = tasks - top |
| |
|
| | eprint("{: <13s}{: ^13s}{: ^13s}".format("", "bottom miss", "bottom hit")) |
| | eprint("{: <13s}{: ^13d}{: ^13d}".format("top miss", |
| | len(bottomMiss & topMiss), |
| | len(bottom & topMiss))) |
| | eprint("{: <13s}{: ^13d}{: ^13d}".format("top hit", |
| | len(top & bottomMiss), |
| | len(top & bottom))) |
| |
|
| | def evaluateOnTestingTasks(result, testingTasks, grammar, _=None, |
| | CPUs=None, solver=None, maximumFrontier=None, enumerationTimeout=None, evaluationTimeout=None): |
| | if result.recognitionModel is not None: |
| | recognizer = result.recognitionModel |
| | testingFrontiers, times = \ |
| | recognizer.enumerateFrontiers(testingTasks, |
| | CPUs=CPUs, |
| | solver=solver, |
| | maximumFrontier=maximumFrontier, |
| | enumerationTimeout=enumerationTimeout, |
| | evaluationTimeout=evaluationTimeout, |
| | testing=True) |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, recognizer.taskGrammarLogProductions(testingTasks), 'heldoutTaskLogProductions') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, recognizer.taskGrammarEntropies(testingTasks), 'heldoutTaskGrammarEntropies') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, recognizer.taskGrammarEntropies(testingTasks), 'heldoutTaskGrammarEntropies') |
| | else: |
| | testingFrontiers, times = multicoreEnumeration(grammar, testingTasks, |
| | solver=solver, |
| | maximumFrontier=maximumFrontier, |
| | enumerationTimeout=enumerationTimeout, |
| | CPUs=CPUs, |
| | evaluationTimeout=evaluationTimeout, |
| | testing=True) |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, times, 'heldoutTestingTimes') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, |
| | {f.task: f for f in testingFrontiers if len(f) > 0 }, |
| | 'frontier') |
| | for f in testingFrontiers: result.recordFrontier(f) |
| | result.testSearchTime = {t: tm for t, tm in times.items() if tm is not None} |
| | times = [t for t in times.values() if t is not None ] |
| | eprint("\n".join(f.summarize() for f in testingFrontiers)) |
| | summaryStatistics("Testing tasks", times) |
| | eprint("Hits %d/%d testing tasks" % (len(times), len(testingTasks))) |
| | result.testingSearchTime.append(times) |
| |
|
| | |
| | def default_wake_generative(grammar, tasks, |
| | maximumFrontier=None, |
| | enumerationTimeout=None, |
| | CPUs=None, |
| | solver=None, |
| | evaluationTimeout=None): |
| | topDownFrontiers, times = multicoreEnumeration(grammar, tasks, |
| | maximumFrontier=maximumFrontier, |
| | enumerationTimeout=enumerationTimeout, |
| | CPUs=CPUs, |
| | solver=solver, |
| | evaluationTimeout=evaluationTimeout) |
| | eprint("Generative model enumeration results:") |
| | eprint(Frontier.describe(topDownFrontiers)) |
| | summaryStatistics("Generative model", [t for t in times.values() if t is not None]) |
| | return topDownFrontiers, times |
| |
|
| | def sleep_recognition(result, grammar, taskBatch, tasks, testingTasks, allFrontiers, _=None, |
| | ensembleSize=1, featureExtractor=None, matrixRank=None, mask=False, |
| | activation=None, contextual=True, biasOptimal=True, |
| | previousRecognitionModel=None, recognitionSteps=None, |
| | timeout=None, enumerationTimeout=None, evaluationTimeout=None, |
| | helmholtzRatio=None, helmholtzFrontiers=None, maximumFrontier=None, |
| | auxiliaryLoss=None, cuda=None, CPUs=None, solver=None): |
| | eprint("Using an ensemble size of %d. Note that we will only store and test on the best recognition model." % ensembleSize) |
| |
|
| | featureExtractorObjects = [featureExtractor(tasks, testingTasks=testingTasks, cuda=cuda) for i in range(ensembleSize)] |
| | recognizers = [RecognitionModel(featureExtractorObjects[i], |
| | grammar, |
| | mask=mask, |
| | rank=matrixRank, |
| | activation=activation, |
| | cuda=cuda, |
| | contextual=contextual, |
| | previousRecognitionModel=previousRecognitionModel, |
| | id=i) for i in range(ensembleSize)] |
| | eprint(f"Currently using this much memory: {getThisMemoryUsage()}") |
| | trainedRecognizers = parallelMap(min(CPUs,len(recognizers)), |
| | lambda recognizer: recognizer.train(allFrontiers, |
| | biasOptimal=biasOptimal, |
| | helmholtzFrontiers=helmholtzFrontiers, |
| | CPUs=CPUs, |
| | evaluationTimeout=evaluationTimeout, |
| | timeout=timeout, |
| | steps=recognitionSteps, |
| | helmholtzRatio=helmholtzRatio, |
| | auxLoss=auxiliaryLoss, |
| | vectorized=True), |
| | recognizers, |
| | seedRandom=True) |
| | eprint(f"Currently using this much memory: {getThisMemoryUsage()}") |
| | |
| | eprint("Trained an ensemble of %d recognition models, now enumerating." % len(trainedRecognizers)) |
| | ensembleFrontiers, ensembleTimes, ensembleRecognitionTimes = [], [], [] |
| | mostTasks = 0 |
| | bestRecognizer = None |
| | totalTasksHitBottomUp = set() |
| | for recIndex, recognizer in enumerate(trainedRecognizers): |
| | eprint("Enumerating from recognizer %d of %d" % (recIndex, len(trainedRecognizers))) |
| | bottomupFrontiers, allRecognitionTimes = \ |
| | recognizer.enumerateFrontiers(taskBatch, |
| | CPUs=CPUs, |
| | maximumFrontier=maximumFrontier, |
| | enumerationTimeout=enumerationTimeout, |
| | evaluationTimeout=evaluationTimeout, |
| | solver=solver) |
| | ensembleFrontiers.append(bottomupFrontiers) |
| | ensembleTimes.append([t for t in allRecognitionTimes.values() if t is not None]) |
| | ensembleRecognitionTimes.append(allRecognitionTimes) |
| |
|
| | recognizerTasksHitBottomUp = {f.task for f in bottomupFrontiers if not f.empty} |
| | totalTasksHitBottomUp.update(recognizerTasksHitBottomUp) |
| | eprint("Recognizer %d solved %d/%d tasks; total tasks solved is now %d." % (recIndex, len(recognizerTasksHitBottomUp), len(tasks), len(totalTasksHitBottomUp))) |
| | if len(recognizerTasksHitBottomUp) >= mostTasks: |
| | |
| | bestRecognizer = recIndex |
| |
|
| | |
| | eprint("Best recognizer: %d." % bestRecognizer) |
| | result.recognitionModel = trainedRecognizers[bestRecognizer] |
| | result.trainSearchTime = {tk: tm for tk, tm in ensembleRecognitionTimes[bestRecognizer].items() |
| | if tm is not None} |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, ensembleRecognitionTimes[bestRecognizer], 'recognitionBestTimes') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskHiddenStates(tasks), 'hiddenState') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskGrammarLogProductions(tasks), 'taskLogProductions') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskGrammarEntropies(tasks), 'taskGrammarEntropies') |
| | if contextual: |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, |
| | result.recognitionModel.taskGrammarStartProductions(tasks), |
| | 'startProductions') |
| |
|
| | result.hitsAtEachWake.append(len(totalTasksHitBottomUp)) |
| | eprint(f"Currently using this much memory: {getThisMemoryUsage()}") |
| |
|
| | """ Rescore and combine the frontiers across the ensemble of recognition models.""" |
| | eprint("Recognition model enumeration results for the best recognizer.") |
| | eprint(Frontier.describe(ensembleFrontiers[bestRecognizer])) |
| | summaryStatistics("Recognition model", ensembleTimes[bestRecognizer]) |
| |
|
| | eprint("Cumulative results for the full ensemble of %d recognizers: " % len(trainedRecognizers)) |
| | |
| | |
| | for bottomupFrontiers in ensembleFrontiers: |
| | for b in bottomupFrontiers: |
| | if b.task not in result.allFrontiers: continue |
| | result.allFrontiers[b.task] = result.allFrontiers[b.task].\ |
| | combine(grammar.rescoreFrontier(b)).\ |
| | topK(maximumFrontier) |
| |
|
| | eprint("Frontiers discovered bottom up: " + str(len(totalTasksHitBottomUp))) |
| | eprint("Total frontiers: " + str(len([f for f in result.allFrontiers.values() if not f.empty]))) |
| |
|
| | result.searchTimes.append(ensembleTimes[bestRecognizer]) |
| | if len(ensembleTimes[bestRecognizer]) > 0: |
| | eprint("Average search time: ", int(mean(ensembleTimes[bestRecognizer]) + 0.5), |
| | "sec.\tmedian:", int(median(ensembleTimes[bestRecognizer]) + 0.5), |
| | "\tmax:", int(max(ensembleTimes[bestRecognizer]) + 0.5), |
| | "\tstandard deviation", int(standardDeviation(ensembleTimes[bestRecognizer]) + 0.5)) |
| | return totalTasksHitBottomUp |
| |
|
| | def consolidate(result, grammar, _=None, topK=None, arity=None, pseudoCounts=None, aic=None, |
| | structurePenalty=None, compressor=None, CPUs=None, iteration=None): |
| | eprint("Showing the top 5 programs in each frontier being sent to the compressor:") |
| | for f in result.allFrontiers.values(): |
| | if f.empty: |
| | continue |
| | eprint(f.task) |
| | for e in f.normalize().topK(5): |
| | eprint("%.02f\t%s" % (e.logPosterior, e.program)) |
| | eprint() |
| |
|
| | |
| | needToSupervise = {f.task for f in result.allFrontiers.values() |
| | if f.task.supervision is not None and f.empty} |
| | compressionFrontiers = [f.replaceWithSupervised(grammar) if f.task in needToSupervise else f |
| | for f in result.allFrontiers.values() ] |
| |
|
| | if len([f for f in compressionFrontiers if not f.empty]) == 0: |
| | eprint("No compression frontiers; not inducing a grammar this iteration.") |
| | else: |
| | grammar, compressionFrontiers = induceGrammar(grammar, compressionFrontiers, |
| | topK=topK, |
| | pseudoCounts=pseudoCounts, a=arity, |
| | aic=aic, structurePenalty=structurePenalty, |
| | topk_use_only_likelihood=False, |
| | backend=compressor, CPUs=CPUs, iteration=iteration) |
| | |
| | for c in compressionFrontiers: |
| | result.allFrontiers[c.task] = c.topK(0) if c in needToSupervise else c |
| |
|
| |
|
| | result.grammars.append(grammar) |
| | eprint("Grammar after iteration %d:" % (iteration + 1)) |
| | eprint(grammar) |
| | |
| | return grammar |
| | |
| |
|
| |
|
| | def commandlineArguments(_=None, |
| | iterations=None, |
| | enumerationTimeout=None, |
| | testEvery=1, |
| | topK=1, |
| | reuseRecognition=False, |
| | CPUs=1, |
| | solver='ocaml', |
| | compressor="ocaml", |
| | useRecognitionModel=True, |
| | recognitionTimeout=None, |
| | activation='relu', |
| | helmholtzRatio=1., |
| | featureExtractor=None, |
| | cuda=None, |
| | maximumFrontier=None, |
| | pseudoCounts=1.0, aic=1.0, |
| | structurePenalty=0.001, a=0, |
| | taskBatchSize=None, taskReranker="default", |
| | extras=None, |
| | storeTaskMetrics=False, |
| | rewriteTaskMetrics=True): |
| | if cuda is None: |
| | cuda = torch.cuda.is_available() |
| | print("CUDA is available?:", torch.cuda.is_available()) |
| | print("using cuda?:", cuda) |
| | import argparse |
| | parser = argparse.ArgumentParser(description="") |
| | parser.add_argument("--resume", |
| | help="Resumes EC algorithm from checkpoint. You can either pass in the path of a checkpoint, or you can pass in the iteration to resume from, in which case it will try to figure out the path.", |
| | default=None, |
| | type=str) |
| | parser.add_argument("-i", "--iterations", |
| | help="default: %d" % iterations, |
| | default=iterations, |
| | type=int) |
| | parser.add_argument("-t", "--enumerationTimeout", |
| | default=enumerationTimeout, |
| | help="In seconds. default: %s" % enumerationTimeout, |
| | type=int) |
| | parser.add_argument("-R", "--recognitionTimeout", |
| | default=recognitionTimeout, |
| | help="In seconds. Amount of time to train the recognition model on each iteration. Defaults to enumeration timeout.", |
| | type=int) |
| | parser.add_argument("-RS", "--recognitionSteps", |
| | default=None, |
| | help="Number of gradient steps to train the recognition model. Can be specified instead of train time.", |
| | type=int) |
| | parser.add_argument( |
| | "-k", |
| | "--topK", |
| | default=topK, |
| | help="When training generative and discriminative models, we train them to fit the top K programs. Ideally we would train them to fit the entire frontier, but this is often intractable. default: %d" % |
| | topK, |
| | type=int) |
| | parser.add_argument("-p", "--pseudoCounts", |
| | default=pseudoCounts, |
| | help="default: %f" % pseudoCounts, |
| | type=float) |
| | parser.add_argument("-b", "--aic", |
| | default=aic, |
| | help="default: %f" % aic, |
| | type=float) |
| | parser.add_argument("-l", "--structurePenalty", |
| | default=structurePenalty, |
| | help="default: %f" % structurePenalty, |
| | type=float) |
| | parser.add_argument("-a", "--arity", |
| | default=a, |
| | help="default: %d" % a, |
| | type=int) |
| | parser.add_argument("-c", "--CPUs", |
| | default=CPUs, |
| | help="default: %d" % CPUs, |
| | type=int) |
| | parser.add_argument("--no-cuda", |
| | action="store_false", |
| | dest="cuda", |
| | help="""cuda will be used if available (which it %s), |
| | unless this is set""" % ("IS" if cuda else "ISN'T")) |
| | parser.add_argument("-m", "--maximumFrontier", |
| | help="""Even though we enumerate --frontierSize |
| | programs, we might want to only keep around the very |
| | best for performance reasons. This is a cut off on the |
| | maximum size of the frontier that is kept around. |
| | Default: %s""" % maximumFrontier, |
| | type=int) |
| | parser.add_argument("--reuseRecognition", |
| | help="""Should we initialize recognition model weights to be what they were at the previous DreamCoder iteration? Default: %s""" % reuseRecognition, |
| | default=reuseRecognition, |
| | action="store_true") |
| | parser.add_argument("--recognition", |
| | dest="useRecognitionModel", |
| | action="store_true", |
| | help="""Enable bottom-up neural recognition model. |
| | Default: %s""" % useRecognitionModel) |
| | parser.add_argument("--ensembleSize", |
| | dest="ensembleSize", |
| | default=1, |
| | help="Number of recognition models to train and enumerate from at each iteration.", |
| | type=int) |
| | parser.add_argument("-g", "--no-recognition", |
| | dest="useRecognitionModel", |
| | action="store_false", |
| | help="""Disable bottom-up neural recognition model. |
| | Default: %s""" % (not useRecognitionModel)) |
| | parser.add_argument("-d", "--no-dsl", |
| | dest="useDSL", |
| | action="store_false", |
| | help="""Disable DSL enumeration and updating.""") |
| | parser.add_argument("--no-consolidation", |
| | dest="noConsolidation", |
| | action="store_true", |
| | help="""Disable DSL updating.""") |
| | parser.add_argument( |
| | "--testingTimeout", |
| | type=int, |
| | dest="testingTimeout", |
| | default=0, |
| | help="Number of seconds we should spend evaluating on each held out testing task.") |
| | parser.add_argument( |
| | "--testEvery", |
| | type=int, |
| | dest="testEvery", |
| | default=1, |
| | help="Run heldout testing every X iterations." |
| | ) |
| | parser.add_argument( |
| | "--seed", |
| | type=int, |
| | default=0, |
| | help="Random seed. Currently this only matters for random batching strategies.") |
| | parser.add_argument( |
| | "--activation", |
| | choices=[ |
| | "relu", |
| | "sigmoid", |
| | "tanh"], |
| | default=activation, |
| | help="""Activation function for neural recognition model. |
| | Default: %s""" % |
| | activation) |
| | parser.add_argument( |
| | "--solver", |
| | choices=[ |
| | "ocaml", |
| | "pypy", |
| | "python"], |
| | default=solver, |
| | help="""Solver for enumeration. |
| | Default: %s""" % |
| | solver) |
| | parser.add_argument( |
| | "-r", |
| | "--Helmholtz", |
| | dest="helmholtzRatio", |
| | help="""When training recognition models, what fraction of the training data should be samples from the generative model? Default %f""" % |
| | helmholtzRatio, |
| | default=helmholtzRatio, |
| | type=float) |
| | parser.add_argument( |
| | "--compressor", |
| | default=compressor, |
| | choices=["pypy","rust","vs","pypy_vs","ocaml","memorize"]) |
| | parser.add_argument( |
| | "--matrixRank", |
| | help="Maximum rank of bigram transition matrix for contextual recognition model. Defaults to full rank.", |
| | default=None, |
| | type=int) |
| | parser.add_argument( |
| | "--mask", |
| | help="Unconditional bigram masking", |
| | default=False, action="store_true") |
| | parser.add_argument("--biasOptimal", |
| | help="Enumerate dreams rather than sample them & bias-optimal recognition objective", |
| | default=False, action="store_true") |
| | parser.add_argument("--contextual", |
| | help="bigram recognition model (default is unigram model)", |
| | default=False, action="store_true") |
| | parser.add_argument("--clear-recognition", |
| | dest="clear-recognition", |
| | help="Clears the recognition model from a checkpoint. Necessary for graphing results with recognition models, because pickle is kind of stupid sometimes.", |
| | default=None, |
| | type=str) |
| | parser.add_argument("--primitive-graph", |
| | dest="primitive-graph", |
| | nargs='+', |
| | help="Displays a dependency graph of the learned primitives", |
| | default=None, |
| | type=str) |
| | parser.add_argument( |
| | "--taskBatchSize", |
| | dest="taskBatchSize", |
| | help="Number of tasks to train on during wake. Defaults to all tasks if None.", |
| | default=None, |
| | type=int) |
| | parser.add_argument( |
| | "--taskReranker", |
| | dest="taskReranker", |
| | help="Reranking function used to order the tasks we train on during waking.", |
| | choices=[ |
| | "default", |
| | "random", |
| | "randomShuffle", |
| | "unsolved", |
| | "unsolvedEntropy", |
| | "unsolvedRandomEntropy", |
| | "randomkNN", |
| | "randomLowEntropykNN"], |
| | default=taskReranker, |
| | type=str) |
| | parser.add_argument( |
| | "--storeTaskMetrics", |
| | dest="storeTaskMetrics", |
| | default=True, |
| | help="Whether to store task metrics directly in the ECResults.", |
| | action="store_true" |
| | ) |
| | parser.add_argument( |
| | "--rewriteTaskMetrics", |
| | dest="rewriteTaskMetrics", |
| | help="Whether to rewrite a new task metrics dictionary at each iteration, rather than retaining the old.", |
| | action="store_true" |
| | ) |
| | parser.add_argument("--addTaskMetrics", |
| | dest="addTaskMetrics", |
| | help="Creates a checkpoint with task metrics and no recognition model for graphing.", |
| | default=None, |
| | nargs='+', |
| | type=str) |
| | parser.add_argument("--auxiliary", |
| | action="store_true", default=False, |
| | help="Add auxiliary classification loss to recognition network training", |
| | dest="auxiliaryLoss") |
| | parser.add_argument("--addFullTaskMetrics", |
| | help="Only to be used in conjunction with --resume. Loads checkpoint, solves both testing and training tasks, stores frontiers, solve times, and task metrics, and then dies.", |
| | default=False, |
| | action="store_true") |
| | parser.add_argument("--countParameters", |
| | help="Load a checkpoint then report how many parameters are in the recognition model.", |
| | default=None, type=str) |
| | parser.set_defaults(useRecognitionModel=useRecognitionModel, |
| | useDSL=True, |
| | featureExtractor=featureExtractor, |
| | maximumFrontier=maximumFrontier, |
| | cuda=cuda) |
| | if extras is not None: |
| | extras(parser) |
| | v = vars(parser.parse_args()) |
| | if v["clear-recognition"] is not None: |
| | ECResult.clearRecognitionModel(v["clear-recognition"]) |
| | sys.exit(0) |
| | else: |
| | del v["clear-recognition"] |
| | |
| | if v["primitive-graph"] is not None: |
| | |
| | for n,pg in enumerate(v["primitive-graph"]): |
| | with open(pg,'rb') as handle: |
| | result = dill.load(handle) |
| | graphPrimitives(result,f"figures/deepProgramLearning/{sys.argv[0]}{n}",view=True) |
| | sys.exit(0) |
| | else: |
| | del v["primitive-graph"] |
| |
|
| | if v["addTaskMetrics"] is not None: |
| | for path in v["addTaskMetrics"]: |
| | with open(path,'rb') as handle: |
| | result = dill.load(handle) |
| | addTaskMetrics(result, path) |
| | sys.exit(0) |
| | else: |
| | del v["addTaskMetrics"] |
| |
|
| | if v["useRecognitionModel"] and v["recognitionTimeout"] is None: |
| | v["recognitionTimeout"] = v["enumerationTimeout"] |
| |
|
| | if v["countParameters"]: |
| | with open(v["countParameters"], "rb") as handle: |
| | result = dill.load(handle) |
| | eprint("The recognition model has", |
| | sum(p.numel() for p in result.recognitionModel.parameters() if p.requires_grad), |
| | "trainable parameters and", |
| | sum(p.numel() for p in result.recognitionModel.parameters() ), |
| | "total parameters.\n", |
| | "The feature extractor accounts for", |
| | sum(p.numel() for p in result.recognitionModel.featureExtractor.parameters() ), |
| | "of those parameters.\n", |
| | "The grammar builder accounts for", |
| | sum(p.numel() for p in result.recognitionModel.grammarBuilder.parameters() ), |
| | "of those parameters.\n") |
| | sys.exit(0) |
| | del v["countParameters"] |
| | |
| | |
| | return v |
| |
|
| | def addTaskMetrics(result, path): |
| | """Adds a task metrics to ECResults that were pickled without them.""" |
| | with torch.no_grad(): return addTaskMetrics_(result, path) |
| | def addTaskMetrics_(result, path): |
| | SUFFIX = '.pickle' |
| | assert path.endswith(SUFFIX) |
| |
|
| | tasks = result.taskSolutions.keys() |
| | everyTask = set(tasks) |
| | for t in result.recognitionTaskMetrics: |
| | if isinstance(t, Task) and t not in everyTask: everyTask.add(t) |
| |
|
| | eprint(f"Found {len(tasks)} training tasks.") |
| | eprint(f"Scrounged up {len(everyTask) - len(tasks)} testing tasks.") |
| | if not hasattr(result, "recognitionTaskMetrics") or result.recognitionTaskMetrics is None: |
| | result.recognitionTaskMetrics = {} |
| |
|
| | |
| | if hasattr(list(tasks)[0], 'getImage'): |
| | images = {t: t.getImage(pretty=True) for t in tasks} |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, images, 'taskImages') |
| |
|
| | if hasattr(list(tasks)[0], 'highresolution'): |
| | images = {t: t.highresolution for t in tasks} |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, images, 'taskImages') |
| |
|
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.auxiliaryPrimitiveEmbeddings(), 'auxiliaryPrimitiveEmbeddings') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskAuxiliaryLossLayer(tasks), 'taskAuxiliaryLossLayer') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskAuxiliaryLossLayer(everyTask), 'every_auxiliaryLossLayer') |
| |
|
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskGrammarFeatureLogProductions(tasks), 'grammarFeatureLogProductions') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskGrammarFeatureLogProductions(everyTask), 'every_grammarFeatureLogProductions') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskGrammarLogProductions(tasks), 'contextualLogProductions') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskGrammarLogProductions(everyTask), 'every_contextualLogProductions') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskHiddenStates(tasks), 'hiddenState') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, result.recognitionModel.taskHiddenStates(everyTask), 'every_hiddenState') |
| | g = result.grammars[-2] |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, {f.task: f.expectedProductionUses(g) |
| | for f in result.taskSolutions.values() |
| | if len(f) > 0}, |
| | 'expectedProductionUses') |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, {f.task: f.expectedProductionUses(g) |
| | for t, metrics in result.recognitionTaskMetrics.items() |
| | if "frontier" in metrics |
| | for f in [metrics["frontier"]] |
| | if len(f) > 0}, |
| | 'every_expectedProductionUses') |
| | if False: |
| | eprint(f"About to do an expensive Monte Carlo simulation w/ {len(tasks)} tasks") |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, |
| | {task: result.recognitionModel.grammarOfTask(task).untorch().expectedUsesMonteCarlo(task.request, debug=False) |
| | for task in tasks }, |
| | 'expectedProductionUsesMonteCarlo') |
| | try: |
| | updateTaskSummaryMetrics(result.recognitionTaskMetrics, |
| | result.recognitionModel.taskGrammarStartProductions(tasks), |
| | 'startProductions') |
| | except: pass |
| |
|
| | |
| | |
| |
|
| | result.recognitionModel = None |
| | |
| | clearedPath = path[:-len(SUFFIX)] + "_graph=True" + SUFFIX |
| | with open(clearedPath,'wb') as handle: |
| | result = dill.dump(result, handle) |
| | eprint(" [+] Cleared recognition model from:") |
| | eprint(" %s"%path) |
| | eprint(" and exported to:") |
| | eprint(" %s"%clearedPath) |
| | eprint(" Use this one for graphing.") |
| |
|
| |
|