| | |
| | |
| | |
| | |
| | import json |
| | import re |
| | from pathlib import Path |
| |
|
| | from rapidfuzz import fuzz |
| |
|
| |
|
| | def clean_text(text: str): |
| | text = re.sub( |
| | r"\nAvailable online at.+?\.\s\s$", "", text, flags=re.DOTALL | re.MULTILINE |
| | ) |
| | return text |
| |
|
| |
|
| | def extract_problems(markdown_text: str): |
| | problems = {} |
| |
|
| | |
| | |
| | |
| | problems_section = re.search( |
| | r"^## Problems\s*$(.*?)^##.*Solutions to Day 1", |
| | markdown_text, |
| | re.DOTALL | re.MULTILINE, |
| | ) |
| |
|
| | if not problems_section: |
| | print(" - Not found '## Problems' section.") |
| | return problems |
| |
|
| | problems_block = problems_section.group(1) |
| |
|
| | |
| | |
| | matchs = list(re.finditer(r"^(\d+)\.\s+", problems_block, flags=re.MULTILINE)) |
| |
|
| | |
| | for i, m in enumerate(matchs): |
| | problem_label = m.group(1) |
| | problem = problems_block[ |
| | m.end() : matchs[i + 1].start() |
| | if i + 1 < len(matchs) |
| | else len(problems_block) |
| | ] |
| |
|
| | problems[problem_label] = (problem.strip(), m.group()) |
| |
|
| | print(f" - Extracted {len(problems)} problems.") |
| |
|
| | return problems |
| |
|
| |
|
| | def extract_solutions(markdown_text): |
| | solutions = {} |
| |
|
| | |
| | |
| | solutions_section = re.search( |
| | r"^##.*?Solutions to Day 1\s*$(.*)", |
| | markdown_text, |
| | re.DOTALL | re.MULTILINE, |
| | ) |
| |
|
| | if not solutions_section: |
| | print(" - Not found Solutions section.") |
| | return solutions |
| |
|
| | solutions_block = solutions_section.group(1) |
| |
|
| | |
| | |
| | matchs = list( |
| | re.finditer(r"^##.*?IMO\s\d+\/(\d+).*?\n$", solutions_block, flags=re.MULTILINE) |
| | ) |
| |
|
| | |
| | for i, m in enumerate(matchs): |
| | problem_label = m.group(1) |
| | solution = solutions_block[ |
| | m.end() : matchs[i + 1].start() |
| | if i + 1 < len(matchs) |
| | else len(solutions_block) |
| | ] |
| |
|
| | solutions[problem_label] = (solution.strip(), m.group()) |
| |
|
| | print(f" - Extracted {len(solutions)} solutions.") |
| |
|
| | return solutions |
| |
|
| |
|
| | def join(problems: dict, solutions: dict): |
| | pairs = [] |
| |
|
| | for problem_label, (problem, p_match) in problems.items(): |
| | solution, s_match = solutions.get(problem_label) |
| |
|
| | |
| | problem_align = fuzz.partial_ratio_alignment(solution, problem) |
| | solution = solution.replace( |
| | solution[problem_align.src_start : problem_align.src_end], "" |
| | ) |
| | solution = re.sub( |
| | r"^\s*## Problem statement", "", solution, flags=re.IGNORECASE |
| | ).strip() |
| |
|
| | if not solution: |
| | print(f" - Warning: No solution found for problem {problem_label}.") |
| | pairs.append((problem, solution, problem_label, p_match, s_match)) |
| |
|
| | return pairs |
| |
|
| |
|
| | def write_pairs(output_file: Path, pairs: list, year: str, project_root: Path): |
| | output_jsonl_text = "" |
| | for problem, solution, problem_label, p_match, s_match in pairs: |
| | output_jsonl_text += ( |
| | json.dumps( |
| | { |
| | "year": year, |
| | "tier": "T0", |
| | "problem_label": problem_label, |
| | "problem_type": None, |
| | "exam": "IMO", |
| | "problem": problem, |
| | "solution": solution, |
| | "metadata": { |
| | "resource_path": output_file.relative_to( |
| | project_root |
| | ).as_posix(), |
| | "problem_match": p_match, |
| | "solution_match": s_match, |
| | }, |
| | }, |
| | ensure_ascii=False, |
| | ) |
| | + "\n" |
| | ) |
| |
|
| | output_file.write_text(output_jsonl_text, encoding="utf-8") |
| |
|
| |
|
| | if __name__ == "__main__": |
| | compet_base_path = Path(__file__).resolve().parent.parent |
| | compet_md_path = compet_base_path / "md" |
| | seg_output_path = compet_base_path / "segmented" |
| | project_root = compet_base_path.parent |
| |
|
| | num_problems = 0 |
| | num_solutions = 0 |
| |
|
| | for md_file in list(compet_md_path.glob("**/*notes.md")): |
| | print(f"Processing {md_file}...") |
| | output_file = seg_output_path / md_file.relative_to(compet_md_path).with_suffix( |
| | ".jsonl" |
| | ) |
| | output_file.parent.mkdir(parents=True, exist_ok=True) |
| |
|
| | |
| | markdown_text = "\n" + md_file.read_text(encoding="utf-8") |
| | markdown_text = clean_text(markdown_text) |
| |
|
| | problems = extract_problems(markdown_text) |
| | solutions = extract_solutions(markdown_text) |
| |
|
| | num_problems += len(problems) |
| | num_solutions += len(solutions) |
| |
|
| | pairs = join(problems, solutions) |
| |
|
| | year = re.search(r"\d{4}", output_file.stem).group() |
| |
|
| | write_pairs(output_file, pairs, year, project_root) |
| |
|
| | print() |
| |
|
| | print(f"Total problems extracted: {num_problems}") |
| | print(f"Total solutions extracted: {num_solutions}") |
| |
|