LIME (Local Interpretable Model-Agnostic Explanations) is the explainability method that explains individual predictions of any black-box model by training a simple, interpretable surrogate model on locally perturbed samples around the input — providing human-readable feature importance explanations for any classifier or regressor regardless of architecture.
What Is LIME?
- Definition: An explanation method that approximates the complex decision boundary of a black-box model (neural network, random forest, SVM) near a specific input instance with a simple, interpretable model (linear regression, decision tree) trained on perturbed versions of that instance.
- Core Insight: Even if the global model is complex and non-linear, it may be locally approximately linear near any specific input — enabling simple explanation of local behavior without understanding the global model.
- Publication: "Why Should I Trust You? Explaining the Predictions of Any Classifier" — Ribeiro, Singh, Guestrin (UW, 2016).
- Model-Agnostic: Requires only the ability to query the model for predictions — works for image classifiers, text models, tabular models, or any other ML system.
Why LIME Matters
- Universal Applicability: Works for any model that can produce predictions — no access to gradients, weights, or model internals required. A single implementation explains neural networks, random forests, and commercial black-box APIs.
- Human-Interpretable Explanations: Produces simple, linear explanations ("The word Viagra contributed +0.3 to spam probability; Hello contributed -0.05") that non-experts can understand and act upon.
- Trust Calibration: Users can evaluate whether model explanations are sensible for their domain — if the explanation highlights irrelevant features, the model should not be trusted for that instance.
- Debugging: Identify specific inputs where the model learned incorrect features — find systematic bugs affecting classes of inputs.
- Regulatory Compliance: Produce explanations for individual automated decisions required by GDPR, ECOA, and similar regulations.
The LIME Procedure
Step 1 — Select Instance to Explain:
- Choose the specific input (one image, one text document, one row of tabular data) to explain.
Step 2 — Perturb the Input:
- Generate N perturbed versions of the instance (typically N=1,000–5,000):
- Images: Randomly hide/reveal "superpixels" (contiguous image regions).
- Text: Randomly remove words from the sentence.
- Tabular: Randomly sample feature values from the training distribution.
Step 3 — Query the Black Box:
- Run all N perturbed instances through the original model.
- Collect predictions (probabilities or class labels) for each.
Step 4 — Weight by Proximity:
- Assign higher weight to perturbed instances closer to the original input.
- Distance metric: cosine similarity for text, L2 for tabular.
- Weight function: W_i = exp(-D(x, x_i)² / σ²).
Step 5 — Train Surrogate Model:
- Fit a weighted linear regression (or decision tree) on the perturbed instances and their black-box predictions.
- The linear model coefficients become the explanation — each coefficient is the importance of the corresponding interpretable feature.
Step 6 — Present Explanation:
- Top positive/negative coefficients are the most important features for this prediction.
- For images: highlight/suppress superpixels by coefficient sign.
- For text: color-code words by positive (green) or negative (red) contribution.
LIME Examples
Text Spam Classification:
- Input: "URGENT: You have won $1,000,000! Call now!"
- LIME explanation: "Predicted SPAM because: 'URGENT' (+0.41), '$1,000,000' (+0.38), 'won' (+0.21). Despite: 'Call' (-0.05)."
Medical Diagnosis (Chest X-Ray):
- LIME highlights specific lung regions that contributed to "Pneumonia" classification.
- Clinician can verify: are the highlighted regions the actual areas of concern?
Credit Scoring:
- LIME explanation: "Loan denied primarily because: credit_score=580 (-0.32), payment_history=missed (-0.28). Income=$45k contributed slightly (+0.08)."
LIME Limitations
- Local Approximation Instability: Because LIME samples randomly and trains a new surrogate per explanation, running LIME twice on the same input may produce different explanations — reducing reliability.
- Superpixel Boundary Sensitivity: LIME for images depends heavily on how superpixels are segmented — different segmentation algorithms produce different explanations.
- Neighborhood Definition: The "local" region LIME optimizes is defined by the perturbation process — if the perturbation distribution is unrealistic, the local model is fit on out-of-distribution data.
- Kernel Width: The bandwidth parameter σ for proximity weighting significantly affects results — smaller σ produces very local (noisy) explanations; larger σ produces less local (potentially unfaithful) ones.
LIME vs. SHAP Comparison
| Property | LIME | SHAP |
|----------|------|------|
| Speed | Moderate | Slow (KernelSHAP) / Fast (TreeSHAP) |
| Stability | Low (random sampling) | Higher |
| Theoretical grounding | Heuristic | Game-theoretic axioms |
| Completeness | No | Yes |
| Model-agnostic | Yes | Yes |
| Ease of use | Simple | Moderate |
LIME is the practical, universal explanation tool that made black-box ML interpretability accessible — by requiring only the ability to query a model rather than model internals, LIME democratized explanation generation for any deployed ML system, making it the go-to explainability method for practitioners who need fast, readable explanations across heterogeneous model types and modalities.