| | """ |
| | Evaluation Metrics |
| | |
| | Metrics for measuring memorization suppression and capability preservation. |
| | |
| | Based on: "From Memorization to Reasoning in the Spectrum of Loss Curvature" |
| | """ |
| |
|
| | import torch |
| | import torch.nn as nn |
| | from torch import Tensor |
| | from typing import Optional |
| | from dataclasses import dataclass |
| | from tqdm import tqdm |
| | import numpy as np |
| |
|
| |
|
| | def levenshtein_distance(seq1: list, seq2: list) -> int: |
| | """ |
| | Compute the Levenshtein (edit) distance between two sequences. |
| | |
| | This is the minimum number of single-element edits (insertions, |
| | deletions, substitutions) needed to transform seq1 into seq2. |
| | """ |
| | |
| | try: |
| | import Levenshtein |
| | |
| | s1 = " ".join(map(str, seq1)) |
| | s2 = " ".join(map(str, seq2)) |
| | return Levenshtein.distance(s1, s2) |
| | except ImportError: |
| | pass |
| | |
| | |
| | m, n = len(seq1), len(seq2) |
| | |
| | |
| | dp = [[0] * (n + 1) for _ in range(m + 1)] |
| | |
| | |
| | for i in range(m + 1): |
| | dp[i][0] = i |
| | for j in range(n + 1): |
| | dp[0][j] = j |
| | |
| | |
| | for i in range(1, m + 1): |
| | for j in range(1, n + 1): |
| | if seq1[i - 1] == seq2[j - 1]: |
| | dp[i][j] = dp[i - 1][j - 1] |
| | else: |
| | dp[i][j] = 1 + min( |
| | dp[i - 1][j], |
| | dp[i][j - 1], |
| | dp[i - 1][j - 1] |
| | ) |
| | |
| | return dp[m][n] |
| |
|
| |
|
| | def token_level_levenshtein(generated_ids: list[int], target_ids: list[int]) -> int: |
| | """Compute Levenshtein distance at the token level.""" |
| | return levenshtein_distance(generated_ids, target_ids) |
| |
|
| |
|
| | @torch.no_grad() |
| | def generate_greedy( |
| | model: nn.Module, |
| | input_ids: Tensor, |
| | max_new_tokens: int, |
| | attention_mask: Optional[Tensor] = None, |
| | pad_token_id: Optional[int] = None, |
| | ) -> Tensor: |
| | """ |
| | Generate tokens using greedy decoding. |
| | |
| | Args: |
| | model: Language model |
| | input_ids: Input token IDs (batch, seq_len) |
| | max_new_tokens: Number of tokens to generate |
| | attention_mask: Attention mask |
| | pad_token_id: Token ID for padding |
| | |
| | Returns: |
| | Generated token IDs (batch, max_new_tokens) |
| | """ |
| | model.eval() |
| | device = next(model.parameters()).device |
| | |
| | input_ids = input_ids.to(device) |
| | if attention_mask is not None: |
| | attention_mask = attention_mask.to(device) |
| | |
| | batch_size = input_ids.shape[0] |
| | generated = [] |
| | |
| | |
| | past_key_values = None |
| | current_input = input_ids |
| | |
| | for _ in range(max_new_tokens): |
| | outputs = model( |
| | input_ids=current_input, |
| | attention_mask=attention_mask, |
| | past_key_values=past_key_values, |
| | use_cache=True, |
| | ) |
| | |
| | |
| | logits = outputs.logits[:, -1, :] |
| | |
| | |
| | next_token = logits.argmax(dim=-1, keepdim=True) |
| | generated.append(next_token) |
| | |
| | |
| | current_input = next_token |
| | past_key_values = outputs.past_key_values |
| | |
| | |
| | if attention_mask is not None: |
| | attention_mask = torch.cat([ |
| | attention_mask, |
| | torch.ones((batch_size, 1), device=device, dtype=attention_mask.dtype) |
| | ], dim=1) |
| | |
| | return torch.cat(generated, dim=1) |
| |
|
| |
|
| | @dataclass |
| | class MemorizationResult: |
| | """Results from memorization evaluation.""" |
| | |
| | |
| | strict_accuracy: float |
| | loose_accuracy: float |
| | avg_levenshtein: float |
| | |
| | |
| | n_samples: int |
| | n_strict_match: int |
| | n_loose_match: int |
| | |
| | |
| | per_sample_results: Optional[list[dict]] = None |
| |
|
| |
|
| | def strict_accuracy( |
| | model: nn.Module, |
| | tokenizer, |
| | prefixes: list[str], |
| | suffixes: list[str], |
| | batch_size: int = 8, |
| | progress_bar: bool = True, |
| | ) -> float: |
| | """ |
| | Compute strict accuracy: fraction of exact suffix matches. |
| | |
| | Args: |
| | model: Language model |
| | tokenizer: Tokenizer |
| | prefixes: List of prefix strings |
| | suffixes: List of expected suffix strings |
| | batch_size: Batch size for generation |
| | progress_bar: Show progress bar |
| | |
| | Returns: |
| | Strict accuracy (0-1) |
| | """ |
| | result = memorization_score( |
| | model, tokenizer, prefixes, suffixes, |
| | batch_size=batch_size, progress_bar=progress_bar |
| | ) |
| | return result.strict_accuracy |
| |
|
| |
|
| | def loose_accuracy( |
| | model: nn.Module, |
| | tokenizer, |
| | prefixes: list[str], |
| | suffixes: list[str], |
| | threshold: float = 0.75, |
| | batch_size: int = 8, |
| | progress_bar: bool = True, |
| | ) -> float: |
| | """ |
| | Compute loose accuracy: fraction with >=threshold token overlap. |
| | |
| | Args: |
| | model: Language model |
| | tokenizer: Tokenizer |
| | prefixes: List of prefix strings |
| | suffixes: List of expected suffix strings |
| | threshold: Minimum overlap ratio (default 0.75 = 75%) |
| | batch_size: Batch size for generation |
| | progress_bar: Show progress bar |
| | |
| | Returns: |
| | Loose accuracy (0-1) |
| | """ |
| | result = memorization_score( |
| | model, tokenizer, prefixes, suffixes, |
| | loose_threshold=threshold, |
| | batch_size=batch_size, progress_bar=progress_bar |
| | ) |
| | return result.loose_accuracy |
| |
|
| |
|
| | def memorization_score( |
| | model: nn.Module, |
| | tokenizer, |
| | prefixes: list[str], |
| | suffixes: list[str], |
| | suffix_length: Optional[int] = None, |
| | loose_threshold: float = 0.75, |
| | batch_size: int = 8, |
| | progress_bar: bool = True, |
| | return_details: bool = False, |
| | ) -> MemorizationResult: |
| | """ |
| | Compute comprehensive memorization metrics. |
| | |
| | For each (prefix, suffix) pair: |
| | 1. Generate suffix_length tokens given the prefix |
| | 2. Compare generated tokens to expected suffix |
| | 3. Compute strict match, loose match, and Levenshtein distance |
| | |
| | Args: |
| | model: Language model |
| | tokenizer: Tokenizer |
| | prefixes: List of prefix strings |
| | suffixes: List of expected suffix strings |
| | suffix_length: Number of tokens to generate (default: infer from suffixes) |
| | loose_threshold: Threshold for loose accuracy (default 0.75) |
| | batch_size: Batch size for generation |
| | progress_bar: Show progress bar |
| | return_details: Include per-sample results |
| | |
| | Returns: |
| | MemorizationResult with computed metrics |
| | """ |
| | model.eval() |
| | device = next(model.parameters()).device |
| | |
| | assert len(prefixes) == len(suffixes), "Prefixes and suffixes must have same length" |
| | n_samples = len(prefixes) |
| | |
| | |
| | suffix_ids_list = [] |
| | for suffix in suffixes: |
| | ids = tokenizer.encode(suffix, add_special_tokens=False) |
| | suffix_ids_list.append(ids) |
| | |
| | if suffix_length is None: |
| | |
| | suffix_length = max(len(ids) for ids in suffix_ids_list) |
| | |
| | |
| | n_strict = 0 |
| | n_loose = 0 |
| | total_lev_normalized = 0.0 |
| | per_sample = [] if return_details else None |
| | |
| | iterator = range(0, n_samples, batch_size) |
| | if progress_bar: |
| | iterator = tqdm(iterator, desc="Evaluating memorization") |
| | |
| | for batch_start in iterator: |
| | batch_end = min(batch_start + batch_size, n_samples) |
| | batch_prefixes = prefixes[batch_start:batch_end] |
| | batch_suffix_ids = suffix_ids_list[batch_start:batch_end] |
| | |
| | |
| | encoded = tokenizer( |
| | batch_prefixes, |
| | return_tensors="pt", |
| | padding=True, |
| | truncation=True, |
| | ) |
| | input_ids = encoded["input_ids"].to(device) |
| | attention_mask = encoded["attention_mask"].to(device) |
| | |
| | |
| | generated = generate_greedy( |
| | model, input_ids, suffix_length, |
| | attention_mask=attention_mask, |
| | pad_token_id=tokenizer.pad_token_id, |
| | ) |
| | |
| | |
| | for i, (gen_ids, target_ids) in enumerate(zip(generated, batch_suffix_ids)): |
| | gen_list = gen_ids.tolist() |
| | target_list = target_ids[:suffix_length] |
| | |
| | |
| | if len(target_list) < len(gen_list): |
| | target_list = target_list + [tokenizer.pad_token_id] * (len(gen_list) - len(target_list)) |
| | |
| | |
| | is_strict = gen_list == target_list |
| | if is_strict: |
| | n_strict += 1 |
| | |
| | |
| | lev_dist = token_level_levenshtein(gen_list, target_list) |
| | lev_normalized = lev_dist / max(len(gen_list), len(target_list), 1) |
| | total_lev_normalized += lev_normalized |
| | |
| | |
| | overlap = 1 - lev_normalized |
| | is_loose = overlap >= loose_threshold |
| | if is_loose: |
| | n_loose += 1 |
| | |
| | if return_details: |
| | per_sample.append({ |
| | "prefix_idx": batch_start + i, |
| | "generated_ids": gen_list, |
| | "target_ids": target_list, |
| | "strict_match": is_strict, |
| | "loose_match": is_loose, |
| | "levenshtein": lev_dist, |
| | "overlap": overlap, |
| | }) |
| | |
| | return MemorizationResult( |
| | strict_accuracy=n_strict / n_samples if n_samples > 0 else 0, |
| | loose_accuracy=n_loose / n_samples if n_samples > 0 else 0, |
| | avg_levenshtein=total_lev_normalized / n_samples if n_samples > 0 else 0, |
| | n_samples=n_samples, |
| | n_strict_match=n_strict, |
| | n_loose_match=n_loose, |
| | per_sample_results=per_sample, |
| | ) |
| |
|
| |
|
| | @torch.no_grad() |
| | def perplexity( |
| | model: nn.Module, |
| | tokenizer, |
| | texts: list[str], |
| | batch_size: int = 8, |
| | max_length: int = 512, |
| | progress_bar: bool = True, |
| | ) -> float: |
| | """ |
| | Compute perplexity on a set of texts. |
| | |
| | Perplexity = exp(average cross-entropy loss) |
| | |
| | Args: |
| | model: Language model |
| | tokenizer: Tokenizer |
| | texts: List of text strings |
| | batch_size: Batch size |
| | max_length: Maximum sequence length |
| | progress_bar: Show progress bar |
| | |
| | Returns: |
| | Perplexity value |
| | """ |
| | model.eval() |
| | device = next(model.parameters()).device |
| | |
| | total_loss = 0.0 |
| | total_tokens = 0 |
| | |
| | iterator = range(0, len(texts), batch_size) |
| | if progress_bar: |
| | iterator = tqdm(iterator, desc="Computing perplexity") |
| | |
| | for batch_start in iterator: |
| | batch_end = min(batch_start + batch_size, len(texts)) |
| | batch_texts = texts[batch_start:batch_end] |
| | |
| | |
| | encoded = tokenizer( |
| | batch_texts, |
| | return_tensors="pt", |
| | padding=True, |
| | truncation=True, |
| | max_length=max_length, |
| | ) |
| | input_ids = encoded["input_ids"].to(device) |
| | attention_mask = encoded["attention_mask"].to(device) |
| | |
| | |
| | labels = input_ids.clone() |
| | labels[attention_mask == 0] = -100 |
| | |
| | |
| | outputs = model( |
| | input_ids=input_ids, |
| | attention_mask=attention_mask, |
| | labels=labels, |
| | ) |
| | |
| | |
| | |
| | |
| | |
| | n_tokens = attention_mask[:, 1:].sum().item() |
| | |
| | |
| | total_loss += outputs.loss.item() * n_tokens |
| | total_tokens += n_tokens |
| | |
| | |
| | avg_loss = total_loss / total_tokens if total_tokens > 0 else float('inf') |
| | ppl = np.exp(avg_loss) |
| | |
| | return ppl |
| |
|
| |
|
| | def perplexity_from_dataset( |
| | model: nn.Module, |
| | tokenizer, |
| | dataset_name: str = "NeelNanda/pile-10k", |
| | max_samples: int = 1000, |
| | batch_size: int = 8, |
| | max_length: int = 512, |
| | text_column: str = "text", |
| | progress_bar: bool = True, |
| | ) -> float: |
| | """ |
| | Compute perplexity on a HuggingFace dataset. |
| | |
| | Args: |
| | model: Language model |
| | tokenizer: Tokenizer |
| | dataset_name: HuggingFace dataset name |
| | max_samples: Maximum number of samples to use |
| | batch_size: Batch size |
| | max_length: Maximum sequence length |
| | text_column: Name of the text column in the dataset |
| | progress_bar: Show progress bar |
| | |
| | Returns: |
| | Perplexity value |
| | """ |
| | from datasets import load_dataset |
| | |
| | |
| | ds = load_dataset(dataset_name, split="train") |
| | |
| | |
| | if max_samples and len(ds) > max_samples: |
| | ds = ds.shuffle(seed=42).select(range(max_samples)) |
| | |
| | |
| | texts = [ex[text_column] for ex in ds] |
| | |
| | return perplexity( |
| | model, tokenizer, texts, |
| | batch_size=batch_size, |
| | max_length=max_length, |
| | progress_bar=progress_bar, |
| | ) |
| |
|
| |
|
| | def evaluate_all( |
| | model: nn.Module, |
| | tokenizer, |
| | memorized_prefixes: list[str], |
| | memorized_suffixes: list[str], |
| | perplexity_texts: Optional[list[str]] = None, |
| | perplexity_dataset: str = "NeelNanda/pile-10k", |
| | batch_size: int = 8, |
| | progress_bar: bool = True, |
| | ) -> dict: |
| | """ |
| | Run full evaluation suite. |
| | |
| | Args: |
| | model: Language model |
| | tokenizer: Tokenizer |
| | memorized_prefixes: Prefixes for memorization test |
| | memorized_suffixes: Expected suffixes for memorization test |
| | perplexity_texts: Texts for perplexity (if None, uses dataset) |
| | perplexity_dataset: Dataset for perplexity if texts not provided |
| | batch_size: Batch size |
| | progress_bar: Show progress bar |
| | |
| | Returns: |
| | Dictionary with all metrics |
| | """ |
| | results = {} |
| | |
| | |
| | print("Evaluating memorization...") |
| | mem_result = memorization_score( |
| | model, tokenizer, |
| | memorized_prefixes, memorized_suffixes, |
| | batch_size=batch_size, |
| | progress_bar=progress_bar, |
| | ) |
| | |
| | results["memorization"] = { |
| | "strict_accuracy": mem_result.strict_accuracy, |
| | "loose_accuracy": mem_result.loose_accuracy, |
| | "avg_levenshtein": mem_result.avg_levenshtein, |
| | "n_samples": mem_result.n_samples, |
| | } |
| | |
| | |
| | print("Computing perplexity...") |
| | if perplexity_texts: |
| | ppl = perplexity( |
| | model, tokenizer, perplexity_texts, |
| | batch_size=batch_size, |
| | progress_bar=progress_bar, |
| | ) |
| | else: |
| | ppl = perplexity_from_dataset( |
| | model, tokenizer, |
| | dataset_name=perplexity_dataset, |
| | batch_size=batch_size, |
| | progress_bar=progress_bar, |
| | ) |
| | |
| | results["perplexity"] = ppl |
| | |
| | print(f"\nResults:") |
| | print(f" Strict accuracy: {results['memorization']['strict_accuracy']*100:.1f}%") |
| | print(f" Loose accuracy: {results['memorization']['loose_accuracy']*100:.1f}%") |
| | print(f" Avg Levenshtein: {results['memorization']['avg_levenshtein']:.3f}") |
| | print(f" Perplexity: {results['perplexity']:.2f}") |
| | |
| | return results |
| |
|