| | from __future__ import annotations |
| |
|
| | import importlib |
| | import re |
| | import unicodedata |
| | from dataclasses import dataclass, field |
| | from pathlib import Path |
| | from typing import Any, Dict, Iterable, Optional |
| |
|
| | try: |
| | etree = importlib.import_module("lxml.etree") |
| | except Exception as exc: |
| | raise RuntimeError( |
| | "Missing lxml dependency. Install with `pip install lxml`.", |
| | ) from exc |
| |
|
| | NS = {"tei": "http://www.tei-c.org/ns/1.0"} |
| |
|
| | BLOCK_TAGS = { |
| | "head", |
| | "p", |
| | "item", |
| | "title", |
| | "label", |
| | "figDesc", |
| | "caption", |
| | "cell", |
| | "ab", |
| | "note", |
| | "quote", |
| | } |
| |
|
| | HASH_BASE = 1_000_003 |
| | HASH_MASK = (1 << 64) - 1 |
| | HASH_VERSION = "tokhash_v2" |
| | ANCHOR_WINDOW = 12 |
| | ANCHOR_WINDOW_ALT = 6 |
| |
|
| |
|
| | def normalize_paper_id(paper_id: str) -> str: |
| | raw = paper_id.strip() |
| | if raw.lower().startswith("doi:"): |
| | raw = raw[4:] |
| | if raw.startswith("dx.doi.org/"): |
| | raw = raw[len("dx.doi.org/") :] |
| | if "doi.org/" in raw: |
| | raw = raw.split("doi.org/", 1)[1] |
| | if raw.startswith("http://") or raw.startswith("https://"): |
| | match = re.search(r"arxiv\.org/(abs|pdf)/([^?#]+)", raw) |
| | if match: |
| | return match.group(2).replace(".pdf", "") |
| | match = re.search(r"/pii/([^/?#]+)", raw) |
| | if match: |
| | return match.group(1) |
| | match = re.search(r"10\.[0-9]+/([^?#]+)", raw) |
| | if match: |
| | return match.group(1) |
| | parts = [p for p in re.split(r"[/?#]", raw) if p] |
| | if parts: |
| | return parts[-1] |
| | match = re.search(r"10\.[0-9]+/([^\s]+)", raw) |
| | if match: |
| | return match.group(1) |
| | return raw |
| |
|
| |
|
| | def normalize_arxiv(value: str) -> str: |
| | cleaned = value.strip() |
| | match = re.search( |
| | r"(\d{4}\.\d{4,5}v\d+|[a-z-]+/\d{7}v\d+)", |
| | cleaned, |
| | re.I, |
| | ) |
| | if match: |
| | return match.group(1) |
| | return cleaned.replace("arXiv:", "").strip() |
| |
|
| |
|
| | def normalize_doi(value: str) -> str: |
| | cleaned = value.strip() |
| | if cleaned.startswith("https://doi.org/"): |
| | cleaned = cleaned[len("https://doi.org/") :] |
| | if cleaned.startswith("http://doi.org/"): |
| | cleaned = cleaned[len("http://doi.org/") :] |
| | return cleaned.lower() |
| |
|
| |
|
| | def doi_suffix(value: str) -> str: |
| | cleaned = normalize_doi(value) |
| | match = re.search(r"10\.[0-9]+/(.+)", cleaned) |
| | if match: |
| | return match.group(1) |
| | return cleaned |
| |
|
| |
|
| | def extract_ids_from_tei(path: Path) -> tuple[Optional[str], Optional[str]]: |
| | try: |
| | root = etree.parse(str(path)).getroot() |
| | except (OSError, etree.XMLSyntaxError): |
| | return None, None |
| | bibl = root.find( |
| | ".//tei:teiHeader/tei:fileDesc/tei:sourceDesc/tei:biblStruct", |
| | namespaces=NS, |
| | ) |
| | if bibl is None: |
| | return None, None |
| | idnos = bibl.findall(".//tei:idno", namespaces=NS) |
| | doi = arxiv = None |
| | for idno in idnos: |
| | if idno.text is None: |
| | continue |
| | value = idno.text.strip() |
| | if not value: |
| | continue |
| | id_type = (idno.get("type") or "").lower() |
| | if doi is None and ( |
| | id_type == "doi" or value.lower().startswith("10.") |
| | ): |
| | doi = value |
| | if arxiv is None and ( |
| | "arxiv" in id_type or value.lower().startswith("arxiv") |
| | ): |
| | arxiv = normalize_arxiv(value) |
| | return doi, arxiv |
| |
|
| |
|
| | def extract_title_from_tei(path: Path) -> Optional[str]: |
| | try: |
| | root = etree.parse(str(path)).getroot() |
| | except (OSError, etree.XMLSyntaxError): |
| | return None |
| | for path_expr in ( |
| | ".//tei:teiHeader/tei:fileDesc/tei:titleStmt/tei:title", |
| | ".//tei:teiHeader/tei:fileDesc/tei:sourceDesc/tei:biblStruct/tei:analytic/tei:title", |
| | ".//tei:teiHeader/tei:fileDesc/tei:sourceDesc/tei:biblStruct/tei:monogr/tei:title", |
| | ): |
| | node = root.find(path_expr, namespaces=NS) |
| | if node is None: |
| | continue |
| | text = "".join(node.itertext()).strip() |
| | if text: |
| | return text |
| | return None |
| |
|
| |
|
| | def extract_text_from_pdf(path: Path) -> str: |
| | pdfplumber_module: Any | None |
| | try: |
| | pdfplumber_module = importlib.import_module("pdfplumber") |
| | except Exception: |
| | pdfplumber_module = None |
| | pdfplumber_any: Any = pdfplumber_module |
| | if pdfplumber_any is not None: |
| | try: |
| | texts = [] |
| | with pdfplumber_any.open(str(path)) as pdf: |
| | for page in pdf.pages: |
| | page_text = page.extract_text() or "" |
| | words_text = "" |
| | try: |
| | words = page.extract_words( |
| | use_text_flow=True, |
| | keep_blank_chars=False, |
| | ) |
| | if words: |
| | words_text = " ".join( |
| | w.get("text", "") for w in words |
| | ) |
| | except Exception: |
| | words_text = "" |
| | if words_text and words_text not in page_text: |
| | if page_text: |
| | page_text = f"{page_text}\n{words_text}" |
| | else: |
| | page_text = words_text |
| | texts.append(page_text) |
| | return "\n\n".join(texts) |
| | except Exception: |
| | |
| | pass |
| | try: |
| | from pdfminer.high_level import extract_text |
| | except Exception: |
| | try: |
| | from PyPDF2 import PdfReader |
| | except Exception as py_exc: |
| | raise RuntimeError( |
| | "PDF fallback extraction requires pdfminer.six " |
| | "or PyPDF2. Install with `pip install pdfminer.six`.", |
| | ) from py_exc |
| | reader = PdfReader(str(path)) |
| | texts = [] |
| | for page in reader.pages: |
| | texts.append(page.extract_text() or "") |
| | return "\n\n".join(texts) |
| | return extract_text(str(path)) |
| |
|
| |
|
| | def build_tei_index(tei_dirs: Iterable[Path]) -> Dict[str, Path]: |
| | index: Dict[str, Path] = {} |
| | for tei_dir in tei_dirs: |
| | if not tei_dir.exists(): |
| | continue |
| | for path in tei_dir.glob("*.grobid.tei.xml"): |
| | stem = path.name[: -len(".grobid.tei.xml")] |
| | if stem.startswith("paper_"): |
| | stem = stem[len("paper_") :] |
| | index.setdefault(stem, path) |
| | return index |
| |
|
| |
|
| | def _local_name(tag: str) -> str: |
| | return tag.split("}", 1)[-1] if "}" in tag else tag |
| |
|
| |
|
| | def extract_blocks_from_tei(path: Path) -> list[str]: |
| | root = etree.parse(str(path)).getroot() |
| | blocks: list[str] = [] |
| |
|
| | def add_blocks(elem) -> None: |
| | tag = _local_name(elem.tag) |
| | if tag in BLOCK_TAGS: |
| | text = "".join(elem.itertext()).strip() |
| | if text: |
| | number = elem.get("n") |
| | if number and tag in {"head", "label"}: |
| | text = f"{number} {text}".strip() |
| | blocks.append(text) |
| | return |
| | for child in elem: |
| | add_blocks(child) |
| |
|
| | for abstract in root.xpath( |
| | ".//tei:teiHeader//tei:abstract", |
| | namespaces=NS, |
| | ): |
| | add_blocks(abstract) |
| |
|
| | text_root = root.find(".//tei:text", namespaces=NS) |
| | if text_root is not None: |
| | add_blocks(text_root) |
| | return blocks |
| |
|
| |
|
| | def _normalize_token(token: str) -> str: |
| | return unicodedata.normalize("NFKC", token).lower() |
| |
|
| |
|
| | HYPHEN_CHARS = { |
| | "-", |
| | "\u2010", |
| | "\u2011", |
| | "\u2012", |
| | "\u2013", |
| | "\u2014", |
| | "\u2212", |
| | } |
| | SOFT_HYPHEN = "\u00ad" |
| |
|
| | CITATION_BRACKET_RE = re.compile(r"\[[^\]]{0,120}\]") |
| | CITATION_PAREN_RE = re.compile(r"\([^\)]{0,120}\)") |
| |
|
| |
|
| | def _looks_like_bracket_citation(text: str) -> bool: |
| | return any(ch.isdigit() for ch in text) |
| |
|
| |
|
| | def _looks_like_paren_citation(text: str) -> bool: |
| | if not any(ch.isdigit() for ch in text): |
| | return False |
| | lowered = text.lower() |
| | if "et al" in lowered: |
| | return True |
| | if re.search(r"\b(19|20)\d{2}\b", text): |
| | return True |
| | return False |
| |
|
| |
|
| | def strip_citations( |
| | text: str, |
| | *, |
| | strip_brackets: bool = True, |
| | strip_parens: bool = False, |
| | ) -> str: |
| | if not text: |
| | return text |
| | spans: list[tuple[int, int]] = [] |
| | if strip_brackets: |
| | for match in CITATION_BRACKET_RE.finditer(text): |
| | if _looks_like_bracket_citation(match.group(0)): |
| | spans.append((match.start(), match.end())) |
| | if strip_parens: |
| | for match in CITATION_PAREN_RE.finditer(text): |
| | if _looks_like_paren_citation(match.group(0)): |
| | spans.append((match.start(), match.end())) |
| | if not spans: |
| | return text |
| | spans.sort() |
| | merged: list[tuple[int, int]] = [] |
| | for start, end in spans: |
| | if not merged or start > merged[-1][1]: |
| | merged.append((start, end)) |
| | else: |
| | merged[-1] = (merged[-1][0], max(merged[-1][1], end)) |
| | parts = [] |
| | cursor = 0 |
| | for start, end in merged: |
| | if cursor < start: |
| | parts.append(text[cursor:start]) |
| | parts.append(" ") |
| | cursor = end |
| | if cursor < len(text): |
| | parts.append(text[cursor:]) |
| | return "".join(parts) |
| |
|
| |
|
| | def tokenize_text( |
| | text: str, |
| | *, |
| | return_spans: bool = False, |
| | ) -> tuple[list[str], Optional[list[tuple[int, int]]]]: |
| | tokens: list[str] = [] |
| | spans: list[tuple[int, int]] = [] |
| | i = 0 |
| | while i < len(text): |
| | ch = text[i] |
| | if ch == SOFT_HYPHEN: |
| | i += 1 |
| | continue |
| | if ch.isalnum(): |
| | start = i |
| | last_idx = i |
| | last_alpha = ch.isalpha() |
| | token_chars = [ch] |
| | i += 1 |
| | while i < len(text): |
| | ch = text[i] |
| | if ch == SOFT_HYPHEN: |
| | i += 1 |
| | continue |
| | if ch.isalnum(): |
| | is_alpha = ch.isalpha() |
| | if is_alpha != last_alpha: |
| | break |
| | token_chars.append(ch) |
| | last_idx = i |
| | last_alpha = is_alpha |
| | i += 1 |
| | continue |
| | if ch in HYPHEN_CHARS and last_alpha: |
| | j = i + 1 |
| | while j < len(text) and text[j].isspace(): |
| | j += 1 |
| | if j < len(text) and text[j].isalpha(): |
| | i = j |
| | continue |
| | break |
| | tokens.append(_normalize_token("".join(token_chars))) |
| | if return_spans: |
| | spans.append((start, last_idx + 1)) |
| | else: |
| | i += 1 |
| | return tokens, spans if return_spans else None |
| |
|
| |
|
| | def hash_token(token: str) -> int: |
| | import hashlib |
| |
|
| | digest = hashlib.blake2b(token.encode("utf-8"), digest_size=8).digest() |
| | return int.from_bytes(digest, "big") |
| |
|
| |
|
| | def hash_token_sequence(tokens: list[str]) -> tuple[int, str, int]: |
| | import hashlib |
| |
|
| | rolling = 0 |
| | normalized = [_normalize_token(token) for token in tokens] |
| | for token in normalized: |
| | rolling = ((rolling * HASH_BASE) + hash_token(token)) & HASH_MASK |
| | joined = " ".join(normalized).encode("utf-8") |
| | sha = hashlib.sha256(joined).hexdigest() |
| | return rolling, sha, len(normalized) |
| |
|
| |
|
| | @dataclass |
| | class TokenIndex: |
| | doc_text: str |
| | tokens: list[str] |
| | spans: list[tuple[int, int]] |
| | token_hashes: list[int] |
| | rolling_cache: dict[int, dict[int, list[int]]] = field( |
| | default_factory=dict, |
| | ) |
| |
|
| | @classmethod |
| | def from_text(cls, doc_text: str) -> "TokenIndex": |
| | tokens, spans = tokenize_text(doc_text, return_spans=True) |
| | token_hashes = [hash_token(t) for t in tokens] |
| | return cls( |
| | doc_text=doc_text, |
| | tokens=tokens, |
| | spans=spans or [], |
| | token_hashes=token_hashes, |
| | ) |
| |
|
| | def _build_rolling_index(self, window: int) -> dict[int, list[int]]: |
| | if window in self.rolling_cache: |
| | return self.rolling_cache[window] |
| | index: dict[int, list[int]] = {} |
| | if window <= 0 or window > len(self.tokens): |
| | self.rolling_cache[window] = index |
| | return index |
| |
|
| | pow_base = 1 |
| | for _ in range(window - 1): |
| | pow_base = (pow_base * HASH_BASE) & HASH_MASK |
| |
|
| | rolling = 0 |
| | for i in range(window): |
| | rolling = ( |
| | (rolling * HASH_BASE) + self.token_hashes[i] |
| | ) & HASH_MASK |
| | index.setdefault(rolling, []).append(0) |
| |
|
| | for i in range(1, len(self.tokens) - window + 1): |
| | remove = (self.token_hashes[i - 1] * pow_base) & HASH_MASK |
| | rolling = (rolling - remove) & HASH_MASK |
| | rolling = ( |
| | (rolling * HASH_BASE) + self.token_hashes[i + window - 1] |
| | ) & HASH_MASK |
| | index.setdefault(rolling, []).append(i) |
| |
|
| | self.rolling_cache[window] = index |
| | return index |
| |
|
| | def _positions_for_hash( |
| | self, |
| | window: int, |
| | target_hash: int, |
| | target_sha: str, |
| | ) -> list[int]: |
| | index = self._build_rolling_index(window) |
| | candidates = index.get(target_hash, []) |
| | if not candidates: |
| | return [] |
| | import hashlib |
| |
|
| | positions: list[int] = [] |
| | for start_idx in candidates: |
| | end_idx = start_idx + window - 1 |
| | if end_idx >= len(self.tokens): |
| | continue |
| | token_slice = self.tokens[start_idx : start_idx + window] |
| | sha = hashlib.sha256( |
| | " ".join(token_slice).encode("utf-8"), |
| | ).hexdigest() |
| | if sha == target_sha: |
| | positions.append(start_idx) |
| | return positions |
| |
|
| | def find_token_span_by_hash( |
| | self, |
| | window: int, |
| | target_hash: int, |
| | target_sha: str, |
| | ) -> Optional[tuple[int, int]]: |
| | positions = self._positions_for_hash(window, target_hash, target_sha) |
| | if not positions: |
| | return None |
| | start_idx = positions[0] |
| | end_idx = start_idx + window - 1 |
| | return start_idx, end_idx |
| |
|
| | def find_token_positions_by_hash( |
| | self, |
| | window: int, |
| | target_hash: int, |
| | target_sha: str, |
| | ) -> list[int]: |
| | return self._positions_for_hash(window, target_hash, target_sha) |
| |
|
| | def find_span_by_hash( |
| | self, |
| | window: int, |
| | target_hash: int, |
| | target_sha: str, |
| | ) -> Optional[tuple[int, int]]: |
| | span = self.find_token_span_by_hash(window, target_hash, target_sha) |
| | if span is None: |
| | return None |
| | start_idx, end_idx = span |
| | start_char = self.spans[start_idx][0] |
| | end_char = self.spans[end_idx][1] |
| | return start_char, end_char |
| | return None |
| |
|
| |
|
| | @dataclass |
| | class DocIndex: |
| | doc_text: str |
| | norm_space: str |
| | norm_space_map: list[int] |
| | norm_nospace: str |
| | norm_nospace_map: list[int] |
| |
|
| | @classmethod |
| | def from_tei(cls, tei_path: Path) -> "DocIndex": |
| | blocks = extract_blocks_from_tei(tei_path) |
| | doc_text = " ".join(blocks) |
| | return cls.from_text(doc_text) |
| |
|
| | @classmethod |
| | def from_text(cls, doc_text: str) -> "DocIndex": |
| | norm_space: list[str] = [] |
| | norm_space_map: list[int] = [] |
| | norm_nospace: list[str] = [] |
| | norm_nospace_map: list[int] = [] |
| | prev_space = False |
| | i = 0 |
| | while i < len(doc_text): |
| | ch = doc_text[i] |
| | if ch == "-" and i > 0 and doc_text[i - 1].isalpha(): |
| | j = i + 1 |
| | while j < len(doc_text) and doc_text[j].isspace(): |
| | j += 1 |
| | if j < len(doc_text) and doc_text[j].isalpha(): |
| | i = j |
| | continue |
| | lower = ch.lower() |
| | if lower.isalnum(): |
| | norm_space.append(lower) |
| | norm_space_map.append(i) |
| | norm_nospace.append(lower) |
| | norm_nospace_map.append(i) |
| | prev_space = False |
| | else: |
| | if not prev_space: |
| | norm_space.append(" ") |
| | norm_space_map.append(i) |
| | prev_space = True |
| | i += 1 |
| |
|
| | while norm_space and norm_space[0] == " ": |
| | norm_space.pop(0) |
| | norm_space_map.pop(0) |
| | while norm_space and norm_space[-1] == " ": |
| | norm_space.pop() |
| | norm_space_map.pop() |
| |
|
| | return cls( |
| | doc_text=doc_text, |
| | norm_space="".join(norm_space), |
| | norm_space_map=norm_space_map, |
| | norm_nospace="".join(norm_nospace), |
| | norm_nospace_map=norm_nospace_map, |
| | ) |
| |
|
| | def find_span(self, query: str) -> Optional[tuple[int, int, str]]: |
| | if not query: |
| | return None |
| | n_q, n_q_ns = _normalize_query(query) |
| | idx = self.norm_space.find(n_q) |
| | if idx != -1: |
| | start = self.norm_space_map[idx] |
| | end = self.norm_space_map[idx + len(n_q) - 1] + 1 |
| | return start, end, "space" |
| |
|
| | trimmed = re.sub(r"^\s*\d+(?:\.\d+)*\s+", "", query) |
| | if trimmed != query: |
| | n_q_trim, n_q_trim_ns = _normalize_query(trimmed) |
| | idx = self.norm_space.find(n_q_trim) |
| | if idx != -1: |
| | start = self.norm_space_map[idx] |
| | end = self.norm_space_map[idx + len(n_q_trim) - 1] + 1 |
| | return start, end, "space_trim" |
| | n_q_ns = n_q_trim_ns |
| |
|
| | idx = self.norm_nospace.find(n_q_ns) |
| | if idx != -1: |
| | start = self.norm_nospace_map[idx] |
| | end = self.norm_nospace_map[idx + len(n_q_ns) - 1] + 1 |
| | return start, end, "nospace" |
| | return None |
| |
|
| | def extract_span(self, start: Optional[int], end: Optional[int]) -> str: |
| | if start is None or end is None: |
| | return "" |
| | if start < 0 or end > len(self.doc_text) or start >= end: |
| | return "" |
| | return self.doc_text[start:end] |
| |
|
| |
|
| | def _normalize_query(text: str) -> tuple[str, str]: |
| | norm_space: list[str] = [] |
| | norm_nospace: list[str] = [] |
| | prev_space = False |
| | i = 0 |
| | while i < len(text): |
| | ch = text[i] |
| | if ch == "-" and i > 0 and text[i - 1].isalpha(): |
| | j = i + 1 |
| | while j < len(text) and text[j].isspace(): |
| | j += 1 |
| | if j < len(text) and text[j].isalpha(): |
| | i = j |
| | continue |
| | lower = ch.lower() |
| | if lower.isalnum(): |
| | norm_space.append(lower) |
| | norm_nospace.append(lower) |
| | prev_space = False |
| | else: |
| | if not prev_space: |
| | norm_space.append(" ") |
| | prev_space = True |
| | i += 1 |
| |
|
| | while norm_space and norm_space[0] == " ": |
| | norm_space.pop(0) |
| | while norm_space and norm_space[-1] == " ": |
| | norm_space.pop() |
| | return "".join(norm_space), "".join(norm_nospace) |
| |
|