| import os |
| from enum import Enum |
| import pexpect |
| import json |
| import re |
| from tqdm import tqdm |
| import json |
| import git |
| import pickle |
| import requests |
| from collections import defaultdict |
|
|
| from utils import make_lean_repl, send_tactic, send_command_icanon, send_command_zsh, BASE_PATH |
|
|
| class ParseState(Enum): |
| defn = 0 |
| inductive = 1 |
| block_comment = 2 |
| prop = 3 |
| mutual = 4 |
|
|
| def loc_tag(fname, line_ind): |
| return f'{fname}:{line_ind}' |
|
|
| def parse_file(fname): |
| lines = None |
| with open(fname, 'r') as f: |
| lines = f.readlines() |
| imports = [] |
| defns = [] |
| props = [] |
| comments = [] |
|
|
| prev_state = None |
| state = None |
| data = [] |
|
|
| def change_state(new_state, line_ind, line): |
| nonlocal data, state, prev_state |
| route_map = {ParseState.defn: defns, ParseState.inductive: defns, |
| ParseState.prop: props, ParseState.block_comment: comments, |
| ParseState.mutual: defns} |
| if state in route_map: |
| route_map[state].append(data) |
| data = [] |
| if new_state in route_map: |
| data = [(loc_tag(fname, line_ind), line)] |
| prev_state = state |
| state = new_state |
|
|
| for line_ind, line in enumerate(lines): |
| line_str = line.strip() |
| if state == ParseState.block_comment: |
| if line_str.endswith('-/'): |
| state = prev_state |
| prev_state = None |
| continue |
| elif line_str.startswith('--'): |
| comments.append((loc_tag(fname, line_ind), line)) |
| continue |
| elif line_str.startswith('/-'): |
| change_state(ParseState.block_comment, line_ind, line) |
| continue |
| elif line_str.startswith('mutual'): |
| change_state(ParseState.mutual, line_ind, line) |
| continue |
| elif line_str.startswith('end') and state == ParseState.mutual: |
| |
| data.append((loc_tag(fname, line_ind), line)) |
| change_state(None, line_ind, line) |
| continue |
| elif state == ParseState.mutual: |
| data.append((loc_tag(fname,line_ind), line)) |
| continue |
| elif line.startswith('import'): |
| assert state is None |
| imports.append(line) |
| continue |
| elif line_str.startswith('def prop'): |
| change_state(ParseState.prop, line_ind, line) |
| elif line_str.startswith('def') or line_str.startswith('lemma') or line_str.startswith('theorem'): |
| change_state(ParseState.defn, line_ind, line) |
| elif line_str.startswith('inductive'): |
| change_state(ParseState.inductive, line_ind, line) |
| elif len(line_str) == 0: |
| change_state(None, line_ind, line) |
| else: |
| data.append((loc_tag(fname,line_ind), line)) |
| |
| change_state(None, -1, '') |
|
|
|
|
| return imports, defns, props, comments |
|
|
|
|
| def process_defns(defns): |
| new_defns = [] |
| for defn in defns: |
| inds, lines = zip(*defn) |
| prop_text = ''.join(lines) |
| ind = min(inds) |
| max_ind = max(inds) |
|
|
| if lines[0].strip().startswith('mutual'): |
| |
| names = [] |
| for line in lines: |
| if line.strip().startswith('def'): |
| inner_name = [s for s in line.strip().split(' ') if len(s) > 0][1] |
| names.append(inner_name) |
| |
| else: |
|
|
| names = [[s for s in prop_text.split(' ') if len(s) > 0][1].strip()] |
|
|
| for name in names: |
| if name.endswith(':'): |
| name = name[:-1] |
| new_defns.append(((ind, max_ind), name, prop_text)) |
| return new_defns |
|
|
| |
| def process_prop(prop, default_proof=':= by sorry'): |
| inds, lines = zip(*prop) |
| prop_text = ''.join(lines) |
| ind = min(inds) |
| max_ind = max(inds) |
| name = prop_text.split(' ')[1] |
| assert prop_text[:3] == 'def' |
| prop2 = 'theorem' + prop_text[3:] |
| |
| |
| |
| prop2 = prop2.strip().replace(':=', ':') + f'{default_proof}' |
| return ((ind, max_ind), name, prop2) |
|
|
| |
| |
| |
| |
| |
| |
| def collect_import_defns(import_names): |
| import_names = import_names[:] |
| defns = [] |
| seen = set() |
| while len(import_names) > 0: |
| imp = import_names.pop() |
| if imp in seen: |
| continue |
| seen.add(imp) |
| i, d, p, c = parse_file(f'{BASE_PATH}/{imp}.lean') |
| import_names += [imp_str.split('.')[-1].strip() for imp_str in i if 'LeanSrc' in imp_str] |
| defns += d |
| return defns |
|
|
| |
| |
| |
| def match_error(err_str): |
| m1 = re.search('expected at\n\s+(.+)\nterm has', err_str) |
| if m1 is not None: |
| return m1.group(1) |
| m2 = re.search("unknown (identifier|constant) '(.+)'", err_str) |
| if m2 is not None: |
| return m2.group(2) |
| if 'invalid dotted identifier notation' in err_str: |
| return err_str.strip().split(' ')[-1] |
| print(f'ERROR: err string <<{err_str}>> is not a recognized error pattern') |
| exit() |
| return None |
|
|
| |
| |
| |
| ignore_errs = ['equality or iff proof expected', 'invalid occurrence of universe level', |
| 'function is not recursive', 'failed to prove termination', 'unsolved goals', |
| 'invalid field notation', 'result is not type correct', |
| 'invalid argument, variable is not a proposition', 'tactic failed. Possible reasons'] |
| def collect_relevant_defns(prop, defns, lean_repl, env, import_order): |
| """ |
| Collect the functions and type definitions used in a prop from a list of defns sourced from the |
| current file and potentially from other libraries, although for now I'm not handling mathlib. |
| This will rely on names being unique, so please don't shadow any names in the files you're importing. |
| """ |
| |
| outp, _env = send_command(lean_repl, prop, env=env) |
| errors = [m for m in outp['messages'] if m['severity'] == 'error'] |
| |
| seen = set() |
| seen_locs = set() |
| all_deps = [] |
|
|
| while True: |
| |
| env2 = env |
| all_deps = order_deps(all_deps, import_order) |
| errors = [] |
| seen_err = set() |
| for defn in all_deps: |
| |
| |
| outp, env2 = send_command(lean_repl, defn[1], env=env2) |
| tmp = [(m, defn[1]) for m in outp.get('messages', []) if m['severity'] == 'error' and m['data'] not in seen_err] |
| errors += tmp |
| for m, _ in tmp: |
| seen_err.add(m['data']) |
| |
| |
| |
| |
|
|
| |
| if len(errors) == 0: |
| outp, env2 = send_command(lean_repl, prop, env=env2) |
| errors = [(m, prop) for m in outp.get('messages', []) if m['severity'] == 'error'] |
|
|
| if len(errors) == 0: |
| break |
|
|
| while len(errors) > 0: |
| err, err_cause = errors.pop() |
| if any([uerr in err['data'] for uerr in ignore_errs]): |
| continue |
| if 'invalid pattern variable, must be atomic' in err['data']: |
| found_ind = False |
| defn_line = err_cause.split('\n')[0] |
| for ident in defn_line.strip().split(' '): |
| if ident in defns and ident not in seen: |
| found_ind = True |
| cp = err.copy() |
| |
| cp['data'] = f"unknown identifier '{ident}'" |
| |
| errors.append((cp, err_cause)) |
| if not found_ind: |
| print('ERROR: failed to resolve inductive type pattern var problem') |
| exit() |
| continue |
|
|
| ident_str = match_error(err['data']) |
| ident_str = ident_str.replace('_root_.','') |
| |
| if ident_str not in defns: |
| print(f'ERROR: couldnt find identifier {ident_str}') |
| print(err) |
| exit() |
| continue |
| if ident_str in seen: |
| continue |
| |
| |
| seen.add(ident_str) |
|
|
| if defns[ident_str][0] in seen_locs: |
| continue |
| seen_locs.add(defns[ident_str][0]) |
| all_deps.append(defns[ident_str]) |
|
|
| return all_deps |
|
|
| def order_deps(defns, import_order): |
| if len(defns) == 0: |
| return defns |
| order_map = {fname: i for i, fname in enumerate(import_order)} |
| line_nums = [int(defn[0][0].split(':')[-1]) for defn in defns] |
| max_line_num = max(line_nums) |
| def import_rank(defn): |
| fpath, line_ind = defn[0][0].split(':') |
| fname = re.search(BASE_PATH + '/(\S+)\.lean', fpath).group(1) |
| return order_map[fname]*max_line_num + int(line_ind) |
| return sorted(defns, key=import_rank) |
|
|
| def extract_file_props(fname, full_path, send_command, default_proof=':= by sorry', repl_type='zsh'): |
| |
| |
| |
| i, d, p, c = parse_file(full_path) |
|
|
| imp_names = [imp_str.split('.')[-1].strip() for imp_str in i if 'LeanSrc' in imp_str] |
| imp_d = collect_import_defns(imp_names) |
|
|
| all_d = imp_d + d |
| import_order = imp_names + [fname] |
| all_d = process_defns(all_d) |
| defns_by_name = {name: (ind, defn) for ind, name, defn in all_d} |
| |
|
|
| props = [process_prop(prop, default_proof=default_proof) for prop in p] |
| |
| lemma_props = [(ind, name, defn.split('\n')[0].strip().replace('lemma', 'theorem').replace(':= by', default_proof)) |
| for ind, name, defn in all_d if defn.strip().startswith('lemma')] |
| props = lemma_props |
| |
| lean_repl = make_lean_repl(repl_type=repl_type) |
|
|
| props_with_deps = {} |
|
|
| outp, mathlib_env = send_command(lean_repl, 'import Mathlib', env=None, first=True) |
| ct = 0 |
| for prop_loc, prop_name, prop in tqdm(props, desc='analyzing and loading lean code + properties'): |
| ct += 1 |
| env = mathlib_env |
|
|
| all_deps = collect_relevant_defns(prop, defns_by_name, lean_repl, env, import_order) |
|
|
| for defn in all_deps: |
| print(defn[-1]) |
| outp, env = send_command(lean_repl, defn[-1], env=env) |
| print('final output of deps', outp) |
|
|
| outp, env = send_command(lean_repl, prop, env=env) |
| for message in outp['messages']: |
| if message['severity'] == 'error': |
| print(f'error at prop {prop_name}') |
| print(message) |
| print() |
| exit() |
| props_with_deps[prop_name] = all_deps + [(prop_loc, prop)] |
|
|
| lean_repl.close() |
| |
| return props_with_deps, c |
|
|
| def output_prop_with_deps(prop, prop_name, folder='LeanSrc/benchmark'): |
| lines = '\n'.join(['import Mathlib'] + [code_lines for _loc, code_lines in prop]) |
| with open(os.path.join(folder, prop_name + '.lean'), 'w') as f: |
| f.write(lines) |
|
|
| def convert_file_props(fname, new_fname): |
| i, d, p, c = parse_file(f'{BASE_PATH}/{fname}.lean') |
| imp_names = [imp_str.split('.')[-1].strip() for imp_str in i if 'LeanSrc' in imp_str] |
| imp_d = collect_import_defns(imp_names) |
|
|
| all_d = imp_d + d |
| import_order = imp_names + [fname] |
| all_d = process_defns(all_d) |
| defns_by_name = {name: (ind, defn) for ind, name, defn in all_d} |
|
|
| props = [process_prop(prop) for prop in p] |
|
|
| with open(new_fname, 'w') as f: |
| defn_lines = '\n'.join([defn for _, _, defn in all_d]) |
| f.write(defn_lines + '\n') |
|
|
| prop_lines = '\n'.join([prop for _, _, prop in props]) |
| f.write(prop_lines + '\n') |
|
|
| def format_llema_input(pwd, lean_url, lean_sha): |
| dcts = [] |
| for prop_name in pwd: |
| |
| lines = '\n'.join(['import Mathlib'] + [code_lines for _loc, code_lines in pwd[prop_name]]) |
| lines = lines.replace(':= by sorry', '') |
| loc, _ = pwd[prop_name][-1] |
| fpath = loc.split(':')[0] |
| dct = {'full_name': prop_name, |
| 'statement': lines, |
| 'url': lean_url, |
| 'commit': lean_sha, |
| 'file_path': fpath, |
| 'split': 'valid'} |
| |
| dcts.append(json.dumps(dct) + '\n') |
| with open('leancb_lemma_inp.jsonl', 'w') as f: |
| f.writelines(dcts) |
|
|
| def pwd_to_json(pwd, send_command, loc2comm, repl_type='zsh'): |
|
|
| lean_repl = make_lean_repl(repl_type=repl_type) |
|
|
| outp, mathlib_env = send_command(lean_repl, 'import Mathlib', env=None, first=True) |
| assert len([m for m in outp.get('messages', []) if m['severity'] == 'error']) == 0, str(outp) |
| |
| dcts = [] |
| for prop_name in pwd: |
| deps = '\n\n'.join(['import Mathlib'] + [code_lines for _loc, code_lines in pwd[prop_name][:-1]]) |
| prop_loc, prop_defn = pwd[prop_name][-1] |
| fpath = prop_loc[0].split(':')[0] |
| cline = int(prop_loc[0].split(':')[1]) - 1 |
| score = 5 |
| if cline in loc2comm: |
| comm = loc2comm[cline] |
| if 'core: ' in comm: |
| score = int(comm.split('core:')[1].strip().split('/')[0].strip()) |
|
|
| env = mathlib_env |
| for _loc, code_lines in pwd[prop_name]: |
| outp, env = send_command(lean_repl, code_lines, env=env) |
| ps = outp['sorries'][0]['goal'] |
|
|
| locs = [loc for loc, _code_lines in pwd[prop_name]] |
| fname2line = defaultdict(lambda: 0) |
| for loc in locs: |
| fpath, line_num = loc[1].split(':') |
| fname2line[fpath] = max(fname2line[fpath], int(line_num)) |
|
|
| dct = {'full_name': prop_name, |
| 'prop_defn': prop_defn, |
| 'prop_loc': prop_loc[0], |
| 'score': score, |
| 'deps': deps, |
| 'proof_state': ps, |
| 'file_locs': [(fpath, fname2line[fpath]) for fpath in fname2line]} |
| |
| dcts.append(json.dumps(dct) + '\n') |
| with open('codeprops_bench.jsonl', 'w') as f: |
| f.writelines(dcts) |
|
|
|
|
| if __name__ == '__main__': |
| |
| main_fname = 'Sorts' |
|
|
| |
|
|
| main_full_path = f'{BASE_PATH}/{main_fname}.lean' |
| """ |
| pwd = extract_file_props(main_fname, main_full_path) # props with deps |
| for prop_name in pwd: |
| output_prop_with_deps(pwd[prop_name], prop_name, folder=folder) |
| """ |
| |
| use_icanon = True |
|
|
| if use_icanon: |
| send_command = send_command_icanon |
| repl_type = 'icanon' |
| else: |
| send_command = send_command_zsh |
| repl_type = 'zsh' |
|
|
|
|
| rerun = True |
| if rerun: |
| pwd, comments = extract_file_props(main_fname, main_full_path, send_command, repl_type=repl_type) |
| with open(f'comm_{main_fname}.pkl', 'wb') as f: |
| pickle.dump(comments, f) |
| with open(f'pwd_{main_fname}.pkl', 'wb') as f: |
| pickle.dump(pwd, f) |
| else: |
| with open(f'pwd_{main_fname}.pkl', 'rb') as f: |
| pwd = pickle.load(f) |
| with open(f'comm_{main_fname}.pkl', 'rb') as f: |
| comments = pickle.load(f) |
|
|
| loc2comm = {} |
| for loc, comm in comments: |
| fname, line_str = loc.strip().split(':') |
| if fname != main_full_path: |
| continue |
| loc2comm[int(line_str.strip())] = comm |
|
|
| |
| |
| |
| |
| |
|
|
| pwd_to_json(pwd, send_command, loc2comm, repl_type=repl_type) |
| |
| |
|
|
| |
| |
| |
| |
| |
| |