|
|
|
|
|
""" |
|
|
GitHub API utilities for scraping and metadata collection. |
|
|
Separated from scrape_gh_docs.py to keep the main script slimmer. |
|
|
""" |
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import os |
|
|
import time |
|
|
import logging |
|
|
import threading |
|
|
from urllib.parse import quote_plus |
|
|
from typing import Optional, Dict, Any, List |
|
|
import json |
|
|
|
|
|
import requests |
|
|
import aiohttp |
|
|
|
|
|
GITHUB_API = "https://api.github.com" |
|
|
|
|
|
|
|
|
logger = logging.getLogger("scrape_gh_docs") |
|
|
|
|
|
_thread_local = threading.local() |
|
|
|
|
|
|
|
|
def github_headers() -> Dict[str, str]: |
|
|
token = os.getenv("GITHUB_TOKEN") |
|
|
assert token is not None, "Set the Github token to run this script!" |
|
|
h = {"Accept": "application/vnd.github.v3+json", "User-Agent": "docs-scraper/1.0"} |
|
|
if token: |
|
|
h["Authorization"] = f"token {token}" |
|
|
return h |
|
|
|
|
|
|
|
|
def get_session() -> requests.Session: |
|
|
sess = getattr(_thread_local, "session", None) |
|
|
if sess is None: |
|
|
sess = requests.Session() |
|
|
_thread_local.session = sess |
|
|
return sess |
|
|
|
|
|
|
|
|
def request_json( |
|
|
url: str, params: Optional[dict] = None, accept_status=(200,), max_retries: int = 3 |
|
|
): |
|
|
for attempt in range(max_retries): |
|
|
resp = get_session().get( |
|
|
url, headers=github_headers(), params=params, timeout=30 |
|
|
) |
|
|
if resp.status_code in accept_status: |
|
|
|
|
|
try: |
|
|
return resp.json() |
|
|
except Exception: |
|
|
return None |
|
|
if resp.status_code in (403, 529): |
|
|
|
|
|
reset = resp.headers.get("X-RateLimit-Reset") |
|
|
ra = resp.headers.get("Retry-After") |
|
|
if ra: |
|
|
wait = int(ra) |
|
|
elif reset: |
|
|
wait = max(5, int(reset) - int(time.time())) |
|
|
else: |
|
|
wait = 30 |
|
|
logger.warning( |
|
|
f"{resp.status_code} from {url}. Sleeping {wait}s (attempt {attempt + 1}/{max_retries})" |
|
|
) |
|
|
time.sleep(wait) |
|
|
continue |
|
|
if 500 <= resp.status_code < 600: |
|
|
backoff = (attempt + 1) * 5 |
|
|
logger.warning(f"{resp.status_code} from {url}. Backing off {backoff}s") |
|
|
time.sleep(backoff) |
|
|
continue |
|
|
logger.error(f"Request to {url} returned {resp.status_code}: {resp.text}") |
|
|
return None |
|
|
logger.error(f"Exhausted retries for {url}") |
|
|
return None |
|
|
|
|
|
|
|
|
def find_readme_file( |
|
|
owner: str, repo: str, ref: Optional[str] = None |
|
|
) -> Optional[Dict[str, str]]: |
|
|
""" |
|
|
Find a README file located at the repository root using the Contents API. |
|
|
|
|
|
Returns a dict with keys: 'name', 'path', 'download_url' if found, else None. |
|
|
README detection is case-insensitive and prioritizes extensions: .md < .rst < .org < others, |
|
|
and shorter names if tie-break is needed. |
|
|
""" |
|
|
|
|
|
root_list = get_contents(owner, repo, "", ref=ref) |
|
|
if not isinstance(root_list, list): |
|
|
return None |
|
|
candidates: List[tuple[tuple[int, int], Dict[str, Any]]] = [] |
|
|
prio_map = {".md": 0, ".rst": 1, ".org": 2} |
|
|
for it in root_list: |
|
|
try: |
|
|
if it.get("type") != "file": |
|
|
continue |
|
|
name = it.get("name") or "" |
|
|
if not name: |
|
|
continue |
|
|
if not name.lower().startswith("readme"): |
|
|
continue |
|
|
ext = "" |
|
|
if "." in name: |
|
|
ext = name[name.rfind(".") :].lower() |
|
|
prio = (prio_map.get(ext, 3), len(name)) |
|
|
candidates.append((prio, it)) |
|
|
except Exception: |
|
|
continue |
|
|
if not candidates: |
|
|
return None |
|
|
candidates.sort(key=lambda x: x[0]) |
|
|
best = candidates[0][1] |
|
|
|
|
|
name = best.get("name") |
|
|
path = best.get("path") |
|
|
download_url = best.get("download_url") |
|
|
if ( |
|
|
isinstance(name, str) |
|
|
and isinstance(path, str) |
|
|
and isinstance(download_url, str) |
|
|
): |
|
|
return {"name": name, "path": path, "download_url": download_url} |
|
|
return None |
|
|
|
|
|
|
|
|
def download_readme_to( |
|
|
owner: str, repo: str, dest_root, ref: Optional[str] = None |
|
|
) -> Optional[Dict[str, str]]: |
|
|
""" |
|
|
Locate README at repo root and download it into dest_root/NAME. |
|
|
Returns the README info dict if downloaded, else None. |
|
|
""" |
|
|
info = find_readme_file(owner, repo, ref=ref) |
|
|
if not info: |
|
|
return None |
|
|
dest_root.mkdir(parents=True, exist_ok=True) |
|
|
dest = dest_root / info["name"] |
|
|
try: |
|
|
download_file(info["download_url"], dest) |
|
|
return info |
|
|
except Exception: |
|
|
logger.warning(f"Failed to download README for {owner}/{repo}: {info['name']}") |
|
|
return None |
|
|
|
|
|
|
|
|
def download_file(url: str, dest_path): |
|
|
dest_path.parent.mkdir(parents=True, exist_ok=True) |
|
|
with get_session().get(url, headers=github_headers(), stream=True, timeout=60) as r: |
|
|
r.raise_for_status() |
|
|
with open(dest_path, "wb") as f: |
|
|
for chunk in r.iter_content(chunk_size=8192): |
|
|
if chunk: |
|
|
f.write(chunk) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def get_repo_info(owner: str, repo: str) -> Optional[Dict[str, Any]]: |
|
|
url = f"{GITHUB_API}/repos/{owner}/{repo}" |
|
|
return request_json(url) |
|
|
|
|
|
|
|
|
def get_default_branch( |
|
|
owner: str, repo: str, repo_json: Optional[Dict[str, Any]] = None |
|
|
) -> Optional[str]: |
|
|
if repo_json and "default_branch" in repo_json: |
|
|
return repo_json["default_branch"] |
|
|
info = get_repo_info(owner, repo) |
|
|
if not info: |
|
|
return None |
|
|
return info.get("default_branch") |
|
|
|
|
|
|
|
|
def get_latest_commit_date( |
|
|
owner: str, |
|
|
repo: str, |
|
|
ref: Optional[str], |
|
|
repo_json: Optional[Dict[str, Any]] = None, |
|
|
) -> Optional[str]: |
|
|
""" |
|
|
Return ISO8601 date string of the latest commit on the given ref (branch or SHA). |
|
|
Falls back to repo's pushed_at if commits endpoint returns nothing. |
|
|
""" |
|
|
branch = ref or (repo_json.get("default_branch") if repo_json else None) or "main" |
|
|
commits = request_json( |
|
|
f"{GITHUB_API}/repos/{owner}/{repo}/commits", |
|
|
params={"sha": branch, "per_page": 1}, |
|
|
accept_status=(200,), |
|
|
) |
|
|
if isinstance(commits, list) and commits: |
|
|
try: |
|
|
return commits[0]["commit"]["author"]["date"] |
|
|
except Exception: |
|
|
pass |
|
|
if repo_json is None: |
|
|
repo_json = get_repo_info(owner, repo) or {} |
|
|
return repo_json.get("pushed_at") |
|
|
|
|
|
|
|
|
def get_contents(owner: str, repo: str, path: str, ref: Optional[str] = None): |
|
|
url = f"{GITHUB_API}/repos/{owner}/{repo}/contents/{quote_plus(path)}" |
|
|
params = {"ref": ref} if ref else None |
|
|
return request_json(url, params=params, accept_status=(200, 404)) |
|
|
|
|
|
|
|
|
def get_owner_type(owner: str) -> Optional[str]: |
|
|
info = request_json(f"{GITHUB_API}/users/{owner}", accept_status=(200, 404)) |
|
|
if not info: |
|
|
return None |
|
|
return info.get("type") |
|
|
|
|
|
|
|
|
def get_org_repos(owner: str, per_page: int = 100) -> List[Dict[str, Any]]: |
|
|
owner_type = get_owner_type(owner) |
|
|
base = "orgs" if owner_type == "Organization" else "users" |
|
|
repos: List[Dict[str, Any]] = [] |
|
|
page = 1 |
|
|
while True: |
|
|
url = f"{GITHUB_API}/{base}/{owner}/repos" |
|
|
params = {"per_page": per_page, "page": page} |
|
|
data = request_json(url, params=params) |
|
|
if not data: |
|
|
if page == 1 and base == "orgs": |
|
|
base = "users" |
|
|
continue |
|
|
break |
|
|
repos.extend(data) |
|
|
if len(data) < per_page: |
|
|
break |
|
|
page += 1 |
|
|
return repos |
|
|
|
|
|
|
|
|
def search_repos(query: str, per_page: int = 5) -> List[Dict[str, Any]]: |
|
|
url = f"{GITHUB_API}/search/repositories" |
|
|
params = {"q": query, "per_page": per_page} |
|
|
res = request_json(url, params=params, accept_status=(200,)) |
|
|
if not res: |
|
|
return [] |
|
|
return res.get("items", []) |
|
|
|
|
|
|
|
|
def get_repo_tree_paths(owner: str, repo: str, ref: Optional[str]) -> List[str]: |
|
|
ref = ref or "main" |
|
|
url = f"{GITHUB_API}/repos/{owner}/{repo}/git/trees/{quote_plus(ref)}" |
|
|
params = {"recursive": 1} |
|
|
data = request_json(url, params=params, accept_status=(200,)) |
|
|
if not data or "tree" not in data: |
|
|
return [] |
|
|
paths: List[str] = [] |
|
|
for entry in data["tree"]: |
|
|
if entry.get("type") == "blob" and "path" in entry: |
|
|
paths.append(entry["path"]) |
|
|
return paths |
|
|
|
|
|
|
|
|
def get_repo_tree_md_paths(owner: str, repo: str, ref: Optional[str]) -> List[str]: |
|
|
""" |
|
|
Return only Markdown file paths from the repository tree on the given ref |
|
|
using the Git Trees API (recursive=1). |
|
|
|
|
|
This is a convenience wrapper over get_repo_tree_paths() that filters to |
|
|
.md files, case-insensitive. |
|
|
""" |
|
|
all_paths = get_repo_tree_paths(owner, repo, ref) |
|
|
return [p for p in all_paths if p.lower().endswith(".md")] |
|
|
|
|
|
|
|
|
async def fetch_repo_readme_markdown( |
|
|
session: aiohttp.ClientSession, owner: str, repo: str |
|
|
) -> Optional[str]: |
|
|
""" |
|
|
Fetch README markdown using the contents API, trying README.md and readme.md. |
|
|
Returns the markdown text or None if not found. |
|
|
""" |
|
|
headers = github_headers() |
|
|
for name in ("README.md", "readme.md"): |
|
|
url = f"{GITHUB_API}/repos/{owner}/{repo}/contents/{name}" |
|
|
try: |
|
|
async with session.get(url, headers=headers) as resp: |
|
|
if resp.status == 200: |
|
|
data = await resp.json() |
|
|
if isinstance(data, dict) and "download_url" in data: |
|
|
download_url = data["download_url"] |
|
|
async with session.get(download_url, headers=headers) as d: |
|
|
if d.status == 200: |
|
|
return await d.text() |
|
|
except Exception: |
|
|
continue |
|
|
|
|
|
try: |
|
|
|
|
|
repo_url = f"{GITHUB_API}/repos/{owner}/{repo}" |
|
|
default_branch = "main" |
|
|
async with session.get(repo_url, headers=headers) as rinfo: |
|
|
if rinfo.status == 200: |
|
|
info = await rinfo.json() |
|
|
if isinstance(info, dict) and info.get("default_branch"): |
|
|
default_branch = info["default_branch"] |
|
|
|
|
|
|
|
|
tree_url = ( |
|
|
f"{GITHUB_API}/repos/{owner}/{repo}/git/trees/{quote_plus(default_branch)}" |
|
|
) |
|
|
async with session.get(tree_url, headers=headers) as rtree: |
|
|
if rtree.status != 200: |
|
|
return None |
|
|
tree = await rtree.json() |
|
|
if not isinstance(tree, dict) or "tree" not in tree: |
|
|
return None |
|
|
|
|
|
entries = tree["tree"] |
|
|
|
|
|
candidates = [] |
|
|
for e in entries: |
|
|
if e.get("type") != "blob": |
|
|
continue |
|
|
path = e.get("path") |
|
|
if not path: |
|
|
continue |
|
|
name_lower = path.lower() |
|
|
if name_lower.startswith("readme"): |
|
|
|
|
|
prio_map = {".md": 0, ".rst": 1, ".org": 2} |
|
|
ext = "" |
|
|
if "." in path: |
|
|
ext = path[path.rfind(".") :].lower() |
|
|
prio = (prio_map.get(ext, 3), len(path)) |
|
|
candidates.append((prio, path)) |
|
|
|
|
|
if not candidates: |
|
|
return None |
|
|
candidates.sort() |
|
|
chosen_path = candidates[0][1] |
|
|
|
|
|
|
|
|
contents_url = ( |
|
|
f"{GITHUB_API}/repos/{owner}/{repo}/contents/{quote_plus(chosen_path)}" |
|
|
) |
|
|
async with session.get(contents_url, headers=headers) as rc: |
|
|
if rc.status != 200: |
|
|
return None |
|
|
cdata = await rc.json() |
|
|
if isinstance(cdata, dict) and "download_url" in cdata: |
|
|
download_url = cdata["download_url"] |
|
|
async with session.get(download_url, headers=headers) as rd: |
|
|
if rd.status == 200: |
|
|
return await rd.text() |
|
|
except Exception: |
|
|
return None |
|
|
|
|
|
return None |
|
|
|
|
|
|
|
|
async def fetch_repo_description( |
|
|
session: aiohttp.ClientSession, owner: str, repo: str |
|
|
) -> Optional[str]: |
|
|
url = f"https://api.github.com/repos/{owner}/{repo}" |
|
|
try: |
|
|
async with session.get(url, headers=github_headers()) as resp: |
|
|
if resp.status == 200: |
|
|
data = await resp.json() |
|
|
if isinstance(data, dict) and "description" in data: |
|
|
desc = data["description"] |
|
|
if isinstance(desc, str): |
|
|
return desc |
|
|
except Exception: |
|
|
return None |
|
|
return None |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
GITHUB_GRAPHQL = "https://api.github.com/graphql" |
|
|
|
|
|
|
|
|
def _graphql_headers() -> Dict[str, str]: |
|
|
token = os.getenv("GITHUB_TOKEN") |
|
|
h = {"Content-Type": "application/json", "User-Agent": "docs-scraper/1.0"} |
|
|
if token: |
|
|
|
|
|
h["Authorization"] = f"bearer {token}" |
|
|
return h |
|
|
|
|
|
|
|
|
def fetch_repos_metadata_graphql( |
|
|
pairs: List[tuple[str, str]], |
|
|
topics_limit: int = 20, |
|
|
) -> Dict[str, Dict[str, Any]]: |
|
|
""" |
|
|
Batch-fetch repository metadata via GitHub GraphQL for a list of (owner, repo) pairs. |
|
|
|
|
|
Returns a mapping from "owner/repo" -> { |
|
|
"name": str | None, |
|
|
"description": str | None, |
|
|
"stars": int | None, |
|
|
"default_branch": str | None, |
|
|
"last_commit_date": str | None, |
|
|
"language": str | None, |
|
|
"topics": List[str], |
|
|
"is_fork": bool | None, |
|
|
"parent_url": str | None, |
|
|
} |
|
|
|
|
|
Notes: |
|
|
- last_commit_date is taken from the latest commit on the default branch when available; |
|
|
falls back to pushedAt otherwise. |
|
|
- Callers should keep batch sizes modest (e.g., <= 30) to stay under GraphQL cost limits. |
|
|
""" |
|
|
if not pairs: |
|
|
return {} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
vars: Dict[str, Any] = {} |
|
|
fields: List[str] = [] |
|
|
for i, (owner, name) in enumerate(pairs): |
|
|
vars[f"owner{i}"] = owner |
|
|
vars[f"name{i}"] = name |
|
|
alias = f"repo_{i}" |
|
|
header = f"{alias}: repository(owner: $owner{i}, name: $name{i}) " |
|
|
topics_fragment = ( |
|
|
" repositoryTopics(first: " |
|
|
+ str(topics_limit) |
|
|
+ ") { nodes { topic { name } } }\n" |
|
|
if topics_limit and topics_limit > 0 |
|
|
else "" |
|
|
) |
|
|
body = ( |
|
|
"{\n" |
|
|
" name\n" |
|
|
" description\n" |
|
|
" stargazerCount\n" |
|
|
" pushedAt\n" |
|
|
" isFork\n" |
|
|
" parent { url nameWithOwner }\n" |
|
|
" primaryLanguage { name }\n" + topics_fragment + " defaultBranchRef {\n" |
|
|
" name\n" |
|
|
" target {\n" |
|
|
" ... on Commit {\n" |
|
|
" history(first: 1) {\n" |
|
|
" nodes {\n" |
|
|
" committedDate\n" |
|
|
" }\n" |
|
|
" }\n" |
|
|
" }\n" |
|
|
" }\n" |
|
|
" }\n" |
|
|
"}\n" |
|
|
) |
|
|
fields.append(header + body) |
|
|
|
|
|
var_decls = " ".join( |
|
|
[f"$owner{i}: String!, $name{i}: String!" for i in range(len(pairs))] |
|
|
) |
|
|
query = "query(" + var_decls + ") {\n" + "\n".join(fields) + "\n}" |
|
|
|
|
|
payload = {"query": query, "variables": vars} |
|
|
max_retries = 3 |
|
|
for attempt in range(max_retries): |
|
|
resp = get_session().post( |
|
|
GITHUB_GRAPHQL, |
|
|
headers=_graphql_headers(), |
|
|
data=json.dumps(payload), |
|
|
timeout=60, |
|
|
) |
|
|
if resp.status_code == 200: |
|
|
try: |
|
|
data = resp.json() |
|
|
except Exception: |
|
|
logger.error("Failed to decode GraphQL JSON response") |
|
|
return {} |
|
|
if "errors" in data and data["errors"]: |
|
|
|
|
|
logger.warning(f"GraphQL response has errors: {data['errors'][0]}") |
|
|
repos = data.get("data", {}) or {} |
|
|
break |
|
|
if resp.status_code in (403, 529): |
|
|
reset = resp.headers.get("X-RateLimit-Reset") |
|
|
ra = resp.headers.get("Retry-After") |
|
|
if ra: |
|
|
wait = int(ra) |
|
|
elif reset: |
|
|
wait = max(5, int(reset) - int(time.time())) |
|
|
else: |
|
|
wait = 30 |
|
|
logger.warning( |
|
|
f"GraphQL {resp.status_code}. Sleeping {wait}s (attempt {attempt + 1}/{max_retries})" |
|
|
) |
|
|
time.sleep(wait) |
|
|
continue |
|
|
if 500 <= resp.status_code < 600: |
|
|
backoff = (attempt + 1) * 5 |
|
|
logger.warning(f"GraphQL {resp.status_code}. Backing off {backoff}s") |
|
|
time.sleep(backoff) |
|
|
continue |
|
|
logger.error(f"GraphQL error {resp.status_code}: {resp.text}") |
|
|
return {} |
|
|
else: |
|
|
logger.error("Exhausted retries for GraphQL request") |
|
|
return {} |
|
|
|
|
|
out: Dict[str, Dict[str, Any]] = {} |
|
|
for i, (owner, name) in enumerate(pairs): |
|
|
alias = f"repo_{i}" |
|
|
r = repos.get(alias) |
|
|
key = f"{owner}/{name}" |
|
|
if not isinstance(r, dict): |
|
|
out[key] = { |
|
|
"name": None, |
|
|
"description": None, |
|
|
"stars": None, |
|
|
"default_branch": None, |
|
|
"last_commit_date": None, |
|
|
"language": None, |
|
|
"topics": [], |
|
|
"is_fork": None, |
|
|
"parent_url": None, |
|
|
} |
|
|
continue |
|
|
repo_name = r.get("name") |
|
|
desc = r.get("description") |
|
|
stars = r.get("stargazerCount") |
|
|
pushed_at = r.get("pushedAt") |
|
|
is_fork = r.get("isFork") |
|
|
parent = r.get("parent") or {} |
|
|
parent_url = parent.get("url") if isinstance(parent, dict) else None |
|
|
lang_obj = r.get("primaryLanguage") or {} |
|
|
language = lang_obj.get("name") if isinstance(lang_obj, dict) else None |
|
|
topics_list: List[str] = [] |
|
|
rt = r.get("repositoryTopics") or {} |
|
|
nodes = rt.get("nodes") if isinstance(rt, dict) else None |
|
|
if isinstance(nodes, list): |
|
|
for node in nodes: |
|
|
t = None |
|
|
if isinstance(node, dict): |
|
|
topic = node.get("topic") |
|
|
if isinstance(topic, dict): |
|
|
t = topic.get("name") |
|
|
if isinstance(t, str): |
|
|
topics_list.append(t) |
|
|
dbr = r.get("defaultBranchRef") or {} |
|
|
default_branch = dbr.get("name") if isinstance(dbr, dict) else None |
|
|
last_commit_date = pushed_at |
|
|
if isinstance(dbr, dict): |
|
|
tgt = dbr.get("target") or {} |
|
|
hist = tgt.get("history") or {} |
|
|
nodes = hist.get("nodes") or [] |
|
|
if isinstance(nodes, list) and nodes: |
|
|
lcd = nodes[0].get("committedDate") |
|
|
if isinstance(lcd, str): |
|
|
last_commit_date = lcd |
|
|
out[key] = { |
|
|
"name": repo_name if isinstance(repo_name, str) else None, |
|
|
"description": desc if isinstance(desc, str) else None, |
|
|
"stars": stars, |
|
|
"default_branch": default_branch, |
|
|
"last_commit_date": last_commit_date, |
|
|
"language": language if isinstance(language, str) else None, |
|
|
"topics": topics_list, |
|
|
"is_fork": is_fork if isinstance(is_fork, bool) else None, |
|
|
"parent_url": parent_url if isinstance(parent_url, str) else None, |
|
|
} |
|
|
return out |
|
|
|