From 819f727e219fe9436c1599073db4974bfd3ff4d9 Mon Sep 17 00:00:00 2001
From: pat-alt <altmeyerpat@gmail.com>
Date: Sun, 12 Feb 2023 00:25:01 -0500
Subject: [PATCH] proposal

---
 .../dev/proposal/execute-results/html.json    |  11 ++
 _quarto.yml                                   |   3 +
 build/dev/proposal.html                       |  39 +++-
 dev/proposal.ipynb                            | 174 ------------------
 4 files changed, 48 insertions(+), 179 deletions(-)
 create mode 100644 _freeze/dev/proposal/execute-results/html.json
 delete mode 100644 dev/proposal.ipynb

diff --git a/_freeze/dev/proposal/execute-results/html.json b/_freeze/dev/proposal/execute-results/html.json
new file mode 100644
index 00000000..7363fbbe
--- /dev/null
+++ b/_freeze/dev/proposal/execute-results/html.json
@@ -0,0 +1,11 @@
+{
+  "hash": "a2ac106a6b675eafee9a47455706943d",
+  "result": {
+    "markdown": "---\ntitle: Conformal Counterfactual Explanations\nsubtitle: Research Proposal\nabstract: |\n    We propose Conformal Counterfactual Explanations: an effortless and rigorous way to produce realistic and faithful Counterfactual Explanations using Conformal Prediction. To address the need for realistic counterfactuals, existing work has primarily relied on separate generative models to learn the data generating process. While this an effective way to produce plausible and model-agnostic counterfactual explanations, it not only introduces an significant engineering overhead, but also reallocates the task of creating realistic model explanations from the model itsel to the generative model. Recent work has shown that there is no need for any of this when working with probabilistic models that explicitly quantify their own uncertainty. Unfortunately, most models used in practice still do not fulfil that basic requirement, in which case we would like to have a way to quantify predictive uncertainty in a post-hoc fashion.\n---\n\n\n\n## Motivation\n\nCounterfactual Explanations are a powerful, flexible and intuitive way to not only explain black-box models, but also enable affected individuals to challenge them though the means of Algorithmic Recourse. \n\n### From Adversarial Examples to Counterfactual Explanations\n\nMost state-of-the-art approaches to generating Counterfactual Explanations (CE) rely on gradient descent in the feature space. The key idea is to perturb inputs $x\\in\\mathcal{X}$ into a black-box model $f: \\mathcal{X} \\mapsto \\mathcal{Y}$ in order to change the model output $f(x)$ to some pre-specified target value $t\\in\\mathcal{Y}$. Formally, this boils down to defining some loss function $\\ell(f(x),t)$ and taking gradient steps in the minimizing direction. The so generated counterfactuals are considered valid as soon as the predicted label matches the target label. A stripped down counterfactual explanation is therefore little different from an adversarial example.\n\n> You may not like it, but this is what counterfactuals look like\n\n\n\n\n\nThe crucial difference between adversarial examples and counterfactuals is one of intent. While adversarial examples are typically intened to go unnoticed, counterfactuals in the context of Explainable AI are generally sought to be \"plausible\" or \"realistic\". To fulfill this latter goal, researchers have come up with a myriad of ways. @joshi2019realistic were among the first to suggest that instead of searching counterfactuals in the feature space, we can instead traverse a latent embedding learned by a surrogate generative model. This ensures that the generated counterfactuals comply with the (learned) data-generating process (DGB). Similarly, @poyiadzi2020face use density ...\n\n- Show DiCE for weak MLP\n- Show Latent for same weak MLP\n- Latent can be manipulated: \n    - train biased model\n    - train VAE with biased variable removed/attacked (use Boston housing dataset)\n    - hypothesis: will generate bias-free explanations\n\n::: {#prp-surrogate}\n\n## Avoid Surrogates\n\nSince we are in the business of explaining a black-box model, the task of learning realistic representations of the data should not be reallocated from the model itself to some surrogate model.\n\n:::\n\n## Introduction to Conformal Prediction\n\n- distribution-free, model-agnostic and scalable approach to predictive uncertainty quantification\n\n### Post-hoc\n\n- Take any fitted model and turn it into a conformal model using calibration data.\n\n### Intrinsic --- Conformal Training [MAYBE]\n\n- Model explicitly trained for conformal prediction.\n\n## Conformal Counterfactuals\n\n- Realistic counterfactuals by minimizing predictive uncertainty [@schut2021generating].\n- Problem: restricted to Bayesian models.\n- Solution: post-hoc predictive uncertainty quantification. \n- Conformal prediction is instance-based. So is CE. \n- Does the coverage guarantee carry over to counterfactuals?\n\n### Research Questions\n\n- Is CP alone enough to ensure realistic counterfactuals?\n- Do counterfactuals improve further as the models get better?\n- Do counterfactuals get more realistic as coverage\n- What happens as we vary coverage and setsize?\n- What happens as we improve the model robustness?\n- What happens as we improve the model's ability to incorporate predictive uncertainty (deep ensemble, laplace)?\n\n## Experiments\n\n- Maybe: conformalised Laplace\n- Benchmarking:\n    - add PROBE into the mix\n    - compare travel costs to domain shits.\n\n## References\n\n",
+    "supporting": [
+      "proposal_files/figure-html"
+    ],
+    "filters": [],
+    "includes": {}
+  }
+}
\ No newline at end of file
diff --git a/_quarto.yml b/_quarto.yml
index c178089d..711a4fd2 100644
--- a/_quarto.yml
+++ b/_quarto.yml
@@ -10,6 +10,9 @@ execute:
   echo: false
   eval: false
   output: false
+  freeze: auto
+
+jupyter: julia-1.8
 
 
 
diff --git a/build/dev/proposal.html b/build/dev/proposal.html
index a37e9d2a..eea1c51f 100644
--- a/build/dev/proposal.html
+++ b/build/dev/proposal.html
@@ -7,7 +7,7 @@
 <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes">
 
 
-<title>Research Proposal</title>
+<title>Conformal Counterfactual Explanations</title>
 <style>
 code{white-space: pre-wrap;}
 span.smallcaps{font-variant: small-caps;}
@@ -53,6 +53,7 @@ div.csl-indent {
 <link href="proposal_files/libs/bootstrap/bootstrap-icons.css" rel="stylesheet">
 <link href="proposal_files/libs/bootstrap/bootstrap.min.css" rel="stylesheet" id="quarto-bootstrap" data-mode="light">
 
+  <script src="https://cdn.jsdelivr.net/npm/mathjax@3/es5/tex-chtml-full.js" type="text/javascript"></script>
 
 </head>
 
@@ -64,7 +65,10 @@ div.csl-indent {
     <h2 id="toc-title">Table of contents</h2>
    
   <ul>
-  <li><a href="#you-may-not-like-it-but-this-is-what-counterfactuals-look-like" id="toc-you-may-not-like-it-but-this-is-what-counterfactuals-look-like" class="nav-link active" data-scroll-target="#you-may-not-like-it-but-this-is-what-counterfactuals-look-like">You may not like it, but this is what counterfactuals look like</a></li>
+  <li><a href="#motivation" id="toc-motivation" class="nav-link active" data-scroll-target="#motivation">Motivation</a>
+  <ul class="collapse">
+  <li><a href="#from-adversarial-examples-to-counterfactual-explanations" id="toc-from-adversarial-examples-to-counterfactual-explanations" class="nav-link" data-scroll-target="#from-adversarial-examples-to-counterfactual-explanations">From Adversarial Examples to Counterfactual Explanations</a></li>
+  </ul></li>
   <li><a href="#introduction-to-conformal-prediction" id="toc-introduction-to-conformal-prediction" class="nav-link" data-scroll-target="#introduction-to-conformal-prediction">Introduction to Conformal Prediction</a>
   <ul class="collapse">
   <li><a href="#post-hoc" id="toc-post-hoc" class="nav-link" data-scroll-target="#post-hoc">Post-hoc</a></li>
@@ -83,7 +87,8 @@ div.csl-indent {
 
 <header id="title-block-header" class="quarto-title-block default">
 <div class="quarto-title">
-<h1 class="title">Research Proposal</h1>
+<h1 class="title">Conformal Counterfactual Explanations</h1>
+<p class="subtitle lead">Research Proposal</p>
 </div>
 
 
@@ -95,11 +100,25 @@ div.csl-indent {
     
   </div>
   
+<div>
+  <div class="abstract">
+    <div class="abstract-title">Abstract</div>
+    <p>We propose Conformal Counterfactual Explanations: an effortless and rigorous way to produce realistic and faithful Counterfactual Explanations using Conformal Prediction. To address the need for realistic counterfactuals, existing work has primarily relied on separate generative models to learn the data generating process. While this an effective way to produce plausible and model-agnostic counterfactual explanations, it not only introduces an significant engineering overhead, but also reallocates the task of creating realistic model explanations from the model itsel to the generative model. Recent work has shown that there is no need for any of this when working with probabilistic models that explicitly quantify their own uncertainty. Unfortunately, most models used in practice still do not fulfil that basic requirement, in which case we would like to have a way to quantify predictive uncertainty in a post-hoc fashion.</p>
+  </div>
+</div>
 
 </header>
 
-<section id="you-may-not-like-it-but-this-is-what-counterfactuals-look-like" class="level2">
-<h2 class="anchored" data-anchor-id="you-may-not-like-it-but-this-is-what-counterfactuals-look-like">You may not like it, but this is what counterfactuals look like</h2>
+<section id="motivation" class="level2">
+<h2 class="anchored" data-anchor-id="motivation">Motivation</h2>
+<p>Counterfactual Explanations are a powerful, flexible and intuitive way to not only explain black-box models, but also enable affected individuals to challenge them though the means of Algorithmic Recourse.</p>
+<section id="from-adversarial-examples-to-counterfactual-explanations" class="level3">
+<h3 class="anchored" data-anchor-id="from-adversarial-examples-to-counterfactual-explanations">From Adversarial Examples to Counterfactual Explanations</h3>
+<p>Most state-of-the-art approaches to generating Counterfactual Explanations (CE) rely on gradient descent in the feature space. The key idea is to perturb inputs <span class="math inline">\(x\in\mathcal{X}\)</span> into a black-box model <span class="math inline">\(f: \mathcal{X} \mapsto \mathcal{Y}\)</span> in order to change the model output <span class="math inline">\(f(x)\)</span> to some pre-specified target value <span class="math inline">\(t\in\mathcal{Y}\)</span>. Formally, this boils down to defining some loss function <span class="math inline">\(\ell(f(x),t)\)</span> and taking gradient steps in the minimizing direction. The so generated counterfactuals are considered valid as soon as the predicted label matches the target label. A stripped down counterfactual explanation is therefore little different from an adversarial example.</p>
+<blockquote class="blockquote">
+<p>You may not like it, but this is what counterfactuals look like</p>
+</blockquote>
+<p>The crucial difference between adversarial examples and counterfactuals is one of intent. While adversarial examples are typically intened to go unnoticed, counterfactuals in the context of Explainable AI are generally sought to be “plausible” or “realistic”. To fulfill this latter goal, researchers have come up with a myriad of ways. <span class="citation" data-cites="joshi2019realistic">Joshi et al. (<a href="#ref-joshi2019realistic" role="doc-biblioref">2019</a>)</span> were among the first to suggest that instead of searching counterfactuals in the feature space, we can instead traverse a latent embedding learned by a surrogate generative model. This ensures that the generated counterfactuals comply with the (learned) data-generating process (DGB). Similarly, <span class="citation" data-cites="poyiadzi2020face">Poyiadzi et al. (<a href="#ref-poyiadzi2020face" role="doc-biblioref">2020</a>)</span> use density …</p>
 <ul>
 <li>Show DiCE for weak MLP</li>
 <li>Show Latent for same weak MLP</li>
@@ -110,6 +129,10 @@ div.csl-indent {
 <li>hypothesis: will generate bias-free explanations</li>
 </ul></li>
 </ul>
+<div id="prp-surrogate" class="theorem proposition">
+<p><span class="theorem-title"><strong>Proposition 1 (Avoid Surrogates) </strong></span>Since we are in the business of explaining a black-box model, the task of learning realistic representations of the data should not be reallocated from the model itself to some surrogate model.</p>
+</div>
+</section>
 </section>
 <section id="introduction-to-conformal-prediction" class="level2">
 <h2 class="anchored" data-anchor-id="introduction-to-conformal-prediction">Introduction to Conformal Prediction</h2>
@@ -167,6 +190,12 @@ div.csl-indent {
 </section>
 
 <div id="quarto-appendix" class="default"><section class="quarto-appendix-contents" role="doc-bibliography"><h2 class="anchored quarto-appendix-heading">References</h2><div id="refs" class="references csl-bib-body hanging-indent" role="doc-bibliography">
+<div id="ref-joshi2019realistic" class="csl-entry" role="doc-biblioentry">
+Joshi, Shalmali, Oluwasanmi Koyejo, Warut Vijitbenjaronk, Been Kim, and Joydeep Ghosh. 2019. <span>“Towards Realistic Individual Recourse and Actionable Explanations in Black-Box Decision Making Systems.”</span> <a href="https://arxiv.org/abs/1907.09615">https://arxiv.org/abs/1907.09615</a>.
+</div>
+<div id="ref-poyiadzi2020face" class="csl-entry" role="doc-biblioentry">
+Poyiadzi, Rafael, Kacper Sokol, Raul Santos-Rodriguez, Tijl De Bie, and Peter Flach. 2020. <span>“<span>FACE</span>: <span>Feasible</span> and Actionable Counterfactual Explanations.”</span> In <em>Proceedings of the <span>AAAI</span>/<span>ACM Conference</span> on <span>AI</span>, <span>Ethics</span>, and <span>Society</span></em>, 344–50.
+</div>
 <div id="ref-schut2021generating" class="csl-entry" role="doc-biblioentry">
 Schut, Lisa, Oscar Key, Rory Mc Grath, Luca Costabello, Bogdan Sacaleanu, Yarin Gal, et al. 2021. <span>“Generating <span>Interpretable Counterfactual Explanations By Implicit Minimisation</span> of <span>Epistemic</span> and <span>Aleatoric Uncertainties</span>.”</span> In <em>International <span>Conference</span> on <span>Artificial Intelligence</span> and <span>Statistics</span></em>, 1756–64. <span>PMLR</span>.
 </div>
diff --git a/dev/proposal.ipynb b/dev/proposal.ipynb
deleted file mode 100644
index 1a03e4c4..00000000
--- a/dev/proposal.ipynb
+++ /dev/null
@@ -1,174 +0,0 @@
-{
-  "cells": [
-    {
-      "cell_type": "raw",
-      "metadata": {},
-      "source": [
-        "---\n",
-        "title: Conformal Counterfactual Explanations\n",
-        "subtitle: Research Proposal\n",
-        "abstract: |\n",
-        "    We propose Conformal Counterfactual Explanations: an effortless and rigorous way to produce realistic and faithful Counterfactual Explanations using Conformal Prediction. To address the need for realistic counterfactuals, existing work has primarily relied on separate generative models to learn the data generating process. While this an effective way to produce plausible and model-agnostic counterfactual explanations, it not only introduces an significant engineering overhead, but also reallocates the task of creating realistic model explanations from the model itsel to the generative model. Recent work has shown that there is no need for any of this when working with probabilistic models that explicitly quantify their own uncertainty. Unfortunately, most models used in practice still do not fulfil that basic requirement, in which case we would like to have a way to quantify predictive uncertainty in a post-hoc fashion.\n",
-        "---"
-      ],
-      "id": "74a21809"
-    },
-    {
-      "cell_type": "code",
-      "metadata": {},
-      "source": [
-        "using CounterfactualExplanations\n",
-        "using CounterfactualExplanations.Data: load_mnist\n",
-        "using CounterfactualExplanations.Models: load_mnist_mlp\n",
-        "using Images\n",
-        "using MLDatasets\n",
-        "using MLDatasets: convert2image\n",
-        "using Plots"
-      ],
-      "id": "274dc69a",
-      "execution_count": null,
-      "outputs": []
-    },
-    {
-      "cell_type": "markdown",
-      "metadata": {},
-      "source": [
-        "## Motivation\n",
-        "\n",
-        "Counterfactual Explanations are a powerful, flexible and intuitive way to not only explain black-box models, but also enable affected individuals to challenge them though the means of Algorithmic Recourse. \n",
-        "\n",
-        "### From Adversarial Examples to Counterfactual Explanations\n",
-        "\n",
-        "Most state-of-the-art approaches to generating Counterfactual Explanations (CE) rely on gradient descent in the feature space. The key idea is to perturb inputs $x\\in\\mathcal{X}$ into a black-box model $f: \\mathcal{X} \\mapsto \\mathcal{Y}$ in order to change the model output $f(x)$ to some pre-specified target value $t\\in\\mathcal{Y}$. Formally, this boils down to defining some loss function $\\ell(f(x),t)$ and taking gradient steps in the minimizing direction. The so generated counterfactuals are considered valid as soon as the predicted label matches the target label. A stripped down counterfactual explanation is therefore little different from an adversarial example.\n",
-        "\n",
-        "> You may not like it, but this is what counterfactuals look like\n"
-      ],
-      "id": "17241786"
-    },
-    {
-      "cell_type": "code",
-      "metadata": {},
-      "source": [
-        "# Data:\n",
-        "counterfactual_data = load_mnist()\n",
-        "X, y = CounterfactualExplanations.DataPreprocessing.unpack_data(counterfactual_data)\n",
-        "input_dim, n_obs = size(counterfactual_data.X)\n",
-        "M = load_mnist_mlp()\n",
-        "\n",
-        "# Target:\n",
-        "factual_label = 8\n",
-        "x = reshape(X[:,rand(findall(predict_label(M, counterfactual_data).==factual_label))],input_dim,1)\n",
-        "target = 3\n",
-        "factual = predict_label(M, counterfactual_data, x)[1]\n",
-        "\n",
-        "# Search:\n",
-        "n_ce = 3\n",
-        "generator = GenericGenerator()\n",
-        "ces = generate_counterfactual(x, target, counterfactual_data, M, generator; num_counterfactuals=n_ce)"
-      ],
-      "id": "7351f8c8",
-      "execution_count": null,
-      "outputs": []
-    },
-    {
-      "cell_type": "code",
-      "metadata": {},
-      "source": [
-        "image_size = 300\n",
-        "p1 = plot(\n",
-        "    convert2image(MNIST, reshape(x,28,28)),\n",
-        "    axis=nothing, \n",
-        "    size=(image_size, image_size),\n",
-        "    title=\"Factual\"\n",
-        ")\n",
-        "plts = [p1]\n",
-        "\n",
-        "for i in eachindex(ces)\n",
-        "    ce = ces[i]\n",
-        "    plt = plot(\n",
-        "        convert2image(MNIST, reshape(CounterfactualExplanations.counterfactual(ce),28,28)),\n",
-        "        axis=nothing, \n",
-        "        size=(image_size, image_size),\n",
-        "        title=\"Counterfactual $i\"\n",
-        "    )\n",
-        "    plts = [plts..., plt]\n",
-        "end\n",
-        "\n",
-        "plt = plot(plts...; size=(image_size * (n_ce + 1),image_size), layout=(1,(n_ce + 1)))"
-      ],
-      "id": "a1853fa9",
-      "execution_count": null,
-      "outputs": []
-    },
-    {
-      "cell_type": "markdown",
-      "metadata": {},
-      "source": [
-        "The crucial difference between adversarial examples and counterfactuals is one of intent. While adversarial examples are typically intened to go unnoticed, counterfactuals in the context of Explainable AI are generally sought to be \"plausible\" or \"realistic\". To fulfill this latter goal, researchers have come up with a myriad of ways. @joshi2019realistic were among the first to suggest that instead of searching counterfactuals in the feature space, we can instead traverse a latent embedding learned by a surrogate generative model. This ensures that the generated counterfactuals comply with the (learned) data-generating process (DGB). Similarly, @poyiadzi2020face use density ...\n",
-        "\n",
-        "- Show DiCE for weak MLP\n",
-        "- Show Latent for same weak MLP\n",
-        "- Latent can be manipulated: \n",
-        "    - train biased model\n",
-        "    - train VAE with biased variable removed/attacked (use Boston housing dataset)\n",
-        "    - hypothesis: will generate bias-free explanations\n",
-        "\n",
-        "::: {#prp-surrogate}\n",
-        "\n",
-        "## Avoid Surrogates\n",
-        "\n",
-        "Since we are in the business of explaining a black-box model, the task of learning realistic representations of the data should not be reallocated from the model itself to some surrogate model.\n",
-        "\n",
-        ":::\n",
-        "\n",
-        "## Introduction to Conformal Prediction\n",
-        "\n",
-        "- distribution-free, model-agnostic and scalable approach to predictive uncertainty quantification\n",
-        "\n",
-        "### Post-hoc\n",
-        "\n",
-        "- Take any fitted model and turn it into a conformal model using calibration data.\n",
-        "\n",
-        "### Intrinsic --- Conformal Training [MAYBE]\n",
-        "\n",
-        "- Model explicitly trained for conformal prediction.\n",
-        "\n",
-        "## Conformal Counterfactuals\n",
-        "\n",
-        "- Realistic counterfactuals by minimizing predictive uncertainty [@schut2021generating].\n",
-        "- Problem: restricted to Bayesian models.\n",
-        "- Solution: post-hoc predictive uncertainty quantification. \n",
-        "- Conformal prediction is instance-based. So is CE. \n",
-        "- Does the coverage guarantee carry over to counterfactuals?\n",
-        "\n",
-        "### Research Questions\n",
-        "\n",
-        "- Is CP alone enough to ensure realistic counterfactuals?\n",
-        "- Do counterfactuals improve further as the models get better?\n",
-        "- Do counterfactuals get more realistic as coverage\n",
-        "- What happens as we vary coverage and setsize?\n",
-        "- What happens as we improve the model robustness?\n",
-        "- What happens as we improve the model's ability to incorporate predictive uncertainty (deep ensemble, laplace)?\n",
-        "\n",
-        "## Experiments\n",
-        "\n",
-        "- Maybe: conformalised Laplace\n",
-        "- Benchmarking:\n",
-        "    - add PROBE into the mix\n",
-        "    - compare travel costs to domain shits.\n",
-        "\n",
-        "## References\n"
-      ],
-      "id": "8abba11d"
-    }
-  ],
-  "metadata": {
-    "kernelspec": {
-      "name": "julia-(4-threads)-1.8",
-      "language": "julia",
-      "display_name": "Julia (4 threads) 1.8.3"
-    }
-  },
-  "nbformat": 4,
-  "nbformat_minor": 5
-}
\ No newline at end of file
-- 
GitLab