-
Notifications
You must be signed in to change notification settings - Fork 19
/
Copy pathcount_proofs.py
executable file
·185 lines (161 loc) · 7.28 KB
/
count_proofs.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
#!/usr/bin/env python3
##########################################################################
#
# This file is part of Proverbot9001.
#
# Proverbot9001 is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Proverbot9001 is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Proverbot9001. If not, see <https://www.gnu.org/licenses/>.
#
# Copyright 2019 Alex Sanchez-Stern and Yousef Alhessi
#
##########################################################################
import argparse
import multiprocessing
import functools
import itertools
import coq_serapy
from coq_serapy.contexts import (ScrapedCommand, ScrapedTactic,
strip_scraped_output, TacticContext)
from context_filter import get_context_filter
from util import eprint, stringified_percent
from data import (read_all_text_data, read_all_text_data_worker__,
MixedDataset, file_chunks)
from pathlib import Path
from typing import List, Optional, Tuple, cast
def main() -> None:
args, parser = parse_arguments()
# Set up --all and --some flags
if (not args.all) and (not args.some):
args.all = True
# Do the counting
total_proofs = 0
total_match = 0
with multiprocessing.Pool(args.num_threads) as pool:
results = pool.imap(functools.partial(count_proofs, args),
args.filenames)
for (matches, num_proofs), filename in zip(results, args.filenames):
if args.print_stmt:
for (prefix, stmt) in matches:
print(stmt)
elif args.print_name:
for (prefix, stmt) in matches:
print(prefix + coq_serapy.lemma_name_from_statement(stmt))
else:
print(f"{filename}: "
f"{len(matches)}/{num_proofs} "
f"{stringified_percent(len(matches),num_proofs)}%")
total_proofs += num_proofs
total_match += len(matches)
if not args.print_name and not args.print_stmt:
print(f"Total: {total_match}/{total_proofs} "
f"({stringified_percent(total_match, total_proofs)}%)")
def parse_arguments() -> Tuple[argparse.Namespace, argparse.ArgumentParser]:
parser = argparse.ArgumentParser(
description="Count the number of proofs matching criteria")
parser.add_argument('--prelude', default=".")
parser.add_argument('--debug', action='store_true')
parser.add_argument("--verbose", "-v", help="verbose output",
action='count')
parser.add_argument('--context-filter', dest="context_filter", type=str,
default="count-default")
parser.add_argument('--num-threads', "-j", default=None, type=int)
g = parser.add_mutually_exclusive_group()
g.add_argument(
'--print-name', dest="print_name",
help="Don't print counts just print the names of matching lemmas",
action='store_true')
g.add_argument(
'--print-stmt', dest="print_stmt",
help="Don't print counts just print the matching lemma statements",
action='store_true')
parser.add_argument("--max-length", dest="max_length", type=int,
default=120)
g = parser.add_mutually_exclusive_group()
g.add_argument("--all", "-a", action='store_true')
g.add_argument("--some", "-s", action='store_true')
parser.add_argument('filenames', nargs="+", help="proof file name (*.v)")
return parser.parse_args(), parser
def count_proofs(args: argparse.Namespace, filename: str) \
-> Tuple[List[Tuple[str, str]], int]:
eprint(f"Counting {filename}", guard=args.debug)
scrapefile = args.prelude + "/" + filename + ".scrape"
interactions = list(read_all_text_data_singlethreaded(Path(scrapefile)))
filter_func = get_context_filter(args.context_filter)
sm_stack = coq_serapy.initial_sm_stack(filename)
matches = []
total_count = 0
cur_proof_counts = False
cur_lemma_stmt = ""
extended_interactions: List[Optional[ScrapedCommand]] = \
cast(List[Optional[ScrapedCommand]], interactions[1:]) + [None]
for inter, next_inter in zip(interactions, extended_interactions):
if isinstance(inter, ScrapedTactic):
context_before = strip_scraped_output(inter)
command = inter.tactic
else:
context_before = TacticContext([], [], [], "")
command = inter
sm_stack = coq_serapy.update_sm_stack(sm_stack, command)
if next_inter and isinstance(next_inter, ScrapedTactic):
context_after = strip_scraped_output(next_inter)
else:
context_after = TacticContext([], [], [], "")
entering_proof = bool(isinstance(inter, str) and
isinstance(next_inter, ScrapedTactic))
exiting_proof = bool(isinstance(inter, ScrapedTactic) and
isinstance(next_inter, str))
if entering_proof:
assert isinstance(next_inter, ScrapedTactic)
cur_lemma_stmt = next_inter.prev_tactics[0]
cur_proof_counts = False if args.some else True
continue
if cur_lemma_stmt:
if filter_func(context_before, command,
context_after, args):
if args.some and not cur_proof_counts:
cur_proof_counts = True
else:
if args.all and cur_proof_counts:
cur_lemma_name = coq_serapy.lemma_name_from_statement(
cur_lemma_stmt)
if cur_proof_counts:
eprint(f"Eliminating proof {cur_lemma_name} "
f"because tactic {command.strip()} "
"doesn't match",
guard=args.debug)
cur_proof_counts = False
if exiting_proof:
cur_lemma_name = coq_serapy.lemma_name_from_statement(
cur_lemma_stmt)
if not cur_lemma_name:
cur_lemma_stmt = ""
continue
if cur_proof_counts:
eprint(f"Proof of {cur_lemma_name} counts",
guard=args.debug)
matches.append((coq_serapy.sm_prefix_from_stack(sm_stack), cur_lemma_stmt))
total_count += 1
cur_lemma_stmt = ""
return matches, total_count
def read_all_text_data_singlethreaded(data_path: Path,
num_threads: Optional[int] = None) \
-> MixedDataset:
line_chunks = file_chunks(data_path, 32768)
try:
yield from itertools.chain.from_iterable((
read_all_text_data_worker__(chunk) for chunk in line_chunks))
except AssertionError:
print(f"Couldn't parse data in {str(data_path)}")
raise
if __name__ == "__main__":
main()