| | """ |
| | K-FAC Weight Editor |
| | |
| | Applies K-FAC-based weight editing to suppress memorization by |
| | removing low-curvature weight components. |
| | |
| | Based on: "From Memorization to Reasoning in the Spectrum of Loss Curvature" |
| | """ |
| |
|
| | import torch |
| | import torch.nn as nn |
| | from torch import Tensor |
| | from typing import Optional, Literal |
| | from dataclasses import dataclass |
| | import numpy as np |
| |
|
| |
|
| | @dataclass |
| | class EditConfig: |
| | """Configuration for K-FAC weight editing.""" |
| | |
| | |
| | energy_threshold: float = 0.6 |
| | |
| | |
| | |
| | |
| | formula: Literal["original", "modified"] = "original" |
| | |
| | |
| | device: str = "cuda" |
| | |
| | |
| | inplace: bool = True |
| |
|
| |
|
| | class KFACEditor: |
| | """ |
| | Edits model weights using K-FAC-based compression to suppress memorization. |
| | |
| | The editing procedure: |
| | 1. Eigendecompose A and G matrices from K-FAC |
| | 2. Transform weights to curvature basis: C = U_G^T @ W @ U_A |
| | 3. Compute importance scores Π_ij for each component |
| | 4. Select top components by cumulative energy threshold |
| | 5. Reconstruct weights: W_edited = U_G @ (C ⊙ M) @ U_A^T |
| | |
| | The key insight is that high-curvature components (high Π) correspond |
| | to generalizing directions, while low-curvature components are used |
| | for memorization. |
| | """ |
| | |
| | def __init__( |
| | self, |
| | model: nn.Module, |
| | kfac_stats: dict[str, tuple[Tensor, Tensor]], |
| | config: Optional[EditConfig] = None, |
| | ): |
| | """ |
| | Initialize the editor. |
| | |
| | Args: |
| | model: The model to edit |
| | kfac_stats: Dictionary mapping layer names to (A, G) tuples |
| | config: Edit configuration |
| | """ |
| | self.model = model |
| | self.kfac_stats = kfac_stats |
| | self.config = config or EditConfig() |
| | |
| | |
| | self._eigen_cache: dict[str, dict] = {} |
| | |
| | |
| | self.edit_stats: dict[str, dict] = {} |
| | |
| | def eigendecompose( |
| | self, |
| | A: Tensor, |
| | G: Tensor, |
| | regularization: float = 1e-6, |
| | ) -> dict: |
| | """ |
| | Compute eigendecomposition of K-FAC factors. |
| | |
| | Args: |
| | A: Activation covariance matrix (d_in x d_in) |
| | G: Gradient covariance matrix (d_out x d_out) |
| | regularization: Small value added to diagonal for numerical stability |
| | |
| | Returns: |
| | Dictionary with eigenvalues and eigenvectors for both A and G |
| | """ |
| | |
| | |
| | original_device = A.device |
| | original_dtype = A.dtype |
| | |
| | |
| | A = A.to(device="cpu", dtype=torch.float32) |
| | G = G.to(device="cpu", dtype=torch.float32) |
| | |
| | |
| | A = (A + A.T) / 2 |
| | G = (G + G.T) / 2 |
| | |
| | |
| | A = A + regularization * torch.eye(A.shape[0], device=A.device, dtype=A.dtype) |
| | G = G + regularization * torch.eye(G.shape[0], device=G.device, dtype=G.dtype) |
| | |
| | |
| | |
| | lambda_A, U_A = torch.linalg.eigh(A) |
| | lambda_G, U_G = torch.linalg.eigh(G) |
| | |
| | |
| | idx_A = torch.argsort(lambda_A, descending=True) |
| | idx_G = torch.argsort(lambda_G, descending=True) |
| | |
| | lambda_A = lambda_A[idx_A] |
| | U_A = U_A[:, idx_A] |
| | lambda_G = lambda_G[idx_G] |
| | U_G = U_G[:, idx_G] |
| | |
| | |
| | lambda_A = torch.clamp(lambda_A, min=0) |
| | lambda_G = torch.clamp(lambda_G, min=0) |
| | |
| | return { |
| | "lambda_A": lambda_A, |
| | "U_A": U_A, |
| | "lambda_G": lambda_G, |
| | "U_G": U_G, |
| | } |
| | |
| | def transform_to_curvature_basis( |
| | self, |
| | W: Tensor, |
| | U_A: Tensor, |
| | U_G: Tensor, |
| | ) -> Tensor: |
| | """ |
| | Transform weights to curvature basis. |
| | |
| | C = U_G^T @ W @ U_A |
| | |
| | Each C_ij represents the component of W along the direction |
| | defined by the i-th gradient eigenvector and j-th activation eigenvector. |
| | """ |
| | return U_G.T @ W @ U_A |
| | |
| | def compute_importance_original( |
| | self, |
| | lambda_A: Tensor, |
| | lambda_G: Tensor, |
| | ) -> Tensor: |
| | """ |
| | Compute importance scores using the original formula. |
| | |
| | Π_ij = λ_i * μ_j |
| | |
| | This is the outer product of the eigenvalues. |
| | """ |
| | |
| | |
| | |
| | return torch.outer(lambda_G, lambda_A) |
| | |
| | def compute_importance_modified( |
| | self, |
| | lambda_A: Tensor, |
| | lambda_G: Tensor, |
| | C: Tensor, |
| | ) -> Tensor: |
| | """ |
| | Compute importance scores using the modified formula. |
| | |
| | Π_ij = λ_i * μ_j * |C_ij|² |
| | |
| | This weights the curvature by the actual magnitude of the |
| | weight component in that direction. |
| | """ |
| | |
| | Pi_base = torch.outer(lambda_G, lambda_A) |
| | |
| | |
| | return Pi_base * (C ** 2) |
| | |
| | def compute_energy_mask( |
| | self, |
| | importance: Tensor, |
| | threshold: float, |
| | ) -> tuple[Tensor, dict]: |
| | """ |
| | Compute binary mask keeping top components by cumulative energy. |
| | |
| | Args: |
| | importance: Importance scores (d_out, d_in) |
| | threshold: Fraction of total energy to retain (e.g., 0.6 = 60%) |
| | |
| | Returns: |
| | Tuple of (mask, statistics) |
| | """ |
| | |
| | flat_importance = importance.flatten() |
| | total_energy = flat_importance.sum() |
| | |
| | |
| | sorted_vals, sorted_indices = torch.sort(flat_importance, descending=True) |
| | |
| | |
| | cumsum = torch.cumsum(sorted_vals, dim=0) |
| | |
| | |
| | target_energy = threshold * total_energy |
| | cutoff_idx = torch.searchsorted(cumsum, target_energy).item() |
| | cutoff_idx = max(1, min(cutoff_idx + 1, len(flat_importance))) |
| | |
| | |
| | mask = torch.zeros_like(flat_importance, dtype=torch.bool) |
| | mask[sorted_indices[:cutoff_idx]] = True |
| | mask = mask.reshape(importance.shape) |
| | |
| | |
| | n_kept = mask.sum().item() |
| | n_total = mask.numel() |
| | actual_energy = flat_importance[mask.flatten()].sum().item() |
| | |
| | stats = { |
| | "n_kept": n_kept, |
| | "n_total": n_total, |
| | "fraction_kept": n_kept / n_total, |
| | "energy_retained": actual_energy / total_energy.item() if total_energy > 0 else 0, |
| | "threshold": threshold, |
| | } |
| | |
| | return mask, stats |
| | |
| | def reconstruct_weights( |
| | self, |
| | C: Tensor, |
| | mask: Tensor, |
| | U_A: Tensor, |
| | U_G: Tensor, |
| | ) -> Tensor: |
| | """ |
| | Reconstruct weights from masked curvature components. |
| | |
| | W_edited = U_G @ (C ⊙ M) @ U_A^T |
| | """ |
| | C_masked = C * mask.float() |
| | return U_G @ C_masked @ U_A.T |
| | |
| | def _get_weight_matrix(self, layer_name: str) -> Tensor: |
| | """Get the weight matrix for a given layer name.""" |
| | |
| | parts = layer_name.split(".") |
| | layer_idx = int(parts[0].replace("layer_", "")) |
| | proj_name = parts[1] |
| | |
| | |
| | layers = None |
| | if hasattr(self.model, "model") and hasattr(self.model.model, "layers"): |
| | layers = self.model.model.layers |
| | elif hasattr(self.model, "transformer") and hasattr(self.model.transformer, "blocks"): |
| | layers = self.model.transformer.blocks |
| | elif hasattr(self.model, "layers"): |
| | layers = self.model.layers |
| | |
| | if layers is None: |
| | raise ValueError(f"Could not find layers in model") |
| | |
| | layer = layers[layer_idx] |
| | |
| | |
| | mlp = None |
| | if hasattr(layer, "mlp"): |
| | mlp = layer.mlp |
| | elif hasattr(layer, "feed_forward"): |
| | mlp = layer.feed_forward |
| | elif hasattr(layer, "ff"): |
| | mlp = layer.ff |
| | |
| | if mlp is None: |
| | raise ValueError(f"Could not find MLP in layer {layer_idx}") |
| | |
| | |
| | proj = getattr(mlp, proj_name) |
| | return proj.weight |
| | |
| | def _set_weight_matrix(self, layer_name: str, new_weight: Tensor) -> None: |
| | """Set the weight matrix for a given layer name.""" |
| | parts = layer_name.split(".") |
| | layer_idx = int(parts[0].replace("layer_", "")) |
| | proj_name = parts[1] |
| | |
| | layers = None |
| | if hasattr(self.model, "model") and hasattr(self.model.model, "layers"): |
| | layers = self.model.model.layers |
| | elif hasattr(self.model, "transformer") and hasattr(self.model.transformer, "blocks"): |
| | layers = self.model.transformer.blocks |
| | elif hasattr(self.model, "layers"): |
| | layers = self.model.layers |
| | |
| | layer = layers[layer_idx] |
| | |
| | mlp = None |
| | if hasattr(layer, "mlp"): |
| | mlp = layer.mlp |
| | elif hasattr(layer, "feed_forward"): |
| | mlp = layer.feed_forward |
| | elif hasattr(layer, "ff"): |
| | mlp = layer.ff |
| | |
| | proj = getattr(mlp, proj_name) |
| | proj.weight.data = new_weight |
| | |
| | def edit_layer( |
| | self, |
| | layer_name: str, |
| | energy_threshold: Optional[float] = None, |
| | formula: Optional[str] = None, |
| | ) -> dict: |
| | """ |
| | Apply K-FAC editing to a single layer. |
| | |
| | Args: |
| | layer_name: Name of the layer to edit (e.g., "layer_11.gate_proj") |
| | energy_threshold: Override config threshold |
| | formula: Override config formula |
| | |
| | Returns: |
| | Statistics about the edit |
| | """ |
| | threshold = energy_threshold or self.config.energy_threshold |
| | edit_formula = formula or self.config.formula |
| | |
| | if layer_name not in self.kfac_stats: |
| | raise ValueError(f"No K-FAC statistics for layer {layer_name}") |
| | |
| | A, G = self.kfac_stats[layer_name] |
| | |
| | A = A.to(device="cpu", dtype=torch.float32) |
| | G = G.to(device="cpu", dtype=torch.float32) |
| | |
| | |
| | if layer_name not in self._eigen_cache: |
| | self._eigen_cache[layer_name] = self.eigendecompose(A, G) |
| | |
| | eigen = self._eigen_cache[layer_name] |
| | lambda_A = eigen["lambda_A"] |
| | lambda_G = eigen["lambda_G"] |
| | U_A = eigen["U_A"] |
| | U_G = eigen["U_G"] |
| | |
| | |
| | W_original = self._get_weight_matrix(layer_name) |
| | original_device = W_original.device |
| | original_dtype = W_original.dtype |
| | W = W_original.to(device="cpu", dtype=torch.float32) |
| | original_norm = torch.norm(W).item() |
| | |
| | |
| | C = self.transform_to_curvature_basis(W, U_A, U_G) |
| | |
| | |
| | if edit_formula == "original": |
| | importance = self.compute_importance_original(lambda_A, lambda_G) |
| | elif edit_formula == "modified": |
| | importance = self.compute_importance_modified(lambda_A, lambda_G, C) |
| | else: |
| | raise ValueError(f"Unknown formula: {edit_formula}") |
| | |
| | |
| | mask, mask_stats = self.compute_energy_mask(importance, threshold) |
| | |
| | |
| | W_edited = self.reconstruct_weights(C, mask, U_A, U_G) |
| | edited_norm = torch.norm(W_edited).item() |
| | |
| | |
| | W_edited = W_edited.to(device=original_device, dtype=original_dtype) |
| | |
| | |
| | if self.config.inplace: |
| | self._set_weight_matrix(layer_name, W_edited) |
| | |
| | |
| | stats = { |
| | "layer_name": layer_name, |
| | "formula": edit_formula, |
| | "threshold": threshold, |
| | "original_norm": original_norm, |
| | "edited_norm": edited_norm, |
| | "norm_change": (edited_norm - original_norm) / original_norm if original_norm > 0 else 0, |
| | **mask_stats, |
| | } |
| | |
| | self.edit_stats[layer_name] = stats |
| | return stats |
| | |
| | def edit_model( |
| | self, |
| | layers: Optional[list[str]] = None, |
| | energy_threshold: Optional[float] = None, |
| | formula: Optional[str] = None, |
| | verbose: bool = True, |
| | ) -> dict: |
| | """ |
| | Apply K-FAC editing to multiple layers. |
| | |
| | Args: |
| | layers: List of layer names to edit (default: all available) |
| | energy_threshold: Override config threshold |
| | formula: Override config formula |
| | verbose: Print progress |
| | |
| | Returns: |
| | Summary statistics |
| | """ |
| | if layers is None: |
| | layers = list(self.kfac_stats.keys()) |
| | |
| | if verbose: |
| | print(f"Editing {len(layers)} layers with {formula or self.config.formula} formula, " |
| | f"{(energy_threshold or self.config.energy_threshold)*100:.0f}% energy threshold") |
| | |
| | all_stats = [] |
| | for layer_name in layers: |
| | stats = self.edit_layer(layer_name, energy_threshold, formula) |
| | all_stats.append(stats) |
| | |
| | if verbose: |
| | print(f" {layer_name}: kept {stats['fraction_kept']*100:.1f}% components, " |
| | f"energy {stats['energy_retained']*100:.1f}%, " |
| | f"norm change {stats['norm_change']*100:+.1f}%") |
| | |
| | |
| | summary = { |
| | "n_layers_edited": len(all_stats), |
| | "avg_fraction_kept": np.mean([s["fraction_kept"] for s in all_stats]), |
| | "avg_energy_retained": np.mean([s["energy_retained"] for s in all_stats]), |
| | "avg_norm_change": np.mean([s["norm_change"] for s in all_stats]), |
| | "layers": all_stats, |
| | } |
| | |
| | return summary |
| | |
| | def restore_original(self, layer_name: str) -> None: |
| | """ |
| | Restore original weights for a layer. |
| | |
| | Note: This only works if we kept a copy of the original weights, |
| | which we don't by default. This method would need to be called |
| | before editing if restoration is desired. |
| | """ |
| | raise NotImplementedError( |
| | "Original weight restoration not implemented. " |
| | "Reload the model to restore original weights." |
| | ) |
| |
|
| |
|
| | def compare_formulas( |
| | model: nn.Module, |
| | kfac_stats: dict[str, tuple[Tensor, Tensor]], |
| | test_fn, |
| | layers: Optional[list[str]] = None, |
| | energy_thresholds: list[float] = [0.5, 0.6, 0.7, 0.8], |
| | device: str = "cuda", |
| | ) -> dict: |
| | """ |
| | Compare original and modified formulas across different thresholds. |
| | |
| | Args: |
| | model: Model to edit (will be modified!) |
| | kfac_stats: K-FAC statistics |
| | test_fn: Function that takes model and returns metrics dict |
| | layers: Layers to edit |
| | energy_thresholds: List of thresholds to test |
| | device: Device for computation |
| | |
| | Returns: |
| | Results dictionary |
| | """ |
| | import copy |
| | |
| | results = { |
| | "baseline": None, |
| | "original": {}, |
| | "modified": {}, |
| | } |
| | |
| | |
| | results["baseline"] = test_fn(model) |
| | print(f"Baseline: {results['baseline']}") |
| | |
| | |
| | original_state = copy.deepcopy(model.state_dict()) |
| | |
| | for formula in ["original", "modified"]: |
| | for threshold in energy_thresholds: |
| | |
| | model.load_state_dict(original_state) |
| | |
| | |
| | config = EditConfig( |
| | energy_threshold=threshold, |
| | formula=formula, |
| | device=device, |
| | ) |
| | editor = KFACEditor(model, kfac_stats, config) |
| | edit_stats = editor.edit_model(layers, verbose=False) |
| | |
| | |
| | metrics = test_fn(model) |
| | metrics["edit_stats"] = edit_stats |
| | |
| | results[formula][threshold] = metrics |
| | print(f"{formula} @ {threshold*100:.0f}%: {metrics}") |
| | |
| | |
| | model.load_state_dict(original_state) |
| | |
| | return results |
| |
|