I have a public Zotero library where I archive and catalogue the papers I read, contact me if you wish to be added to the library. Follows a catalogue of the tags that I use.
Tags
We have several classes of tags, each identified by its token cardinality, ranging from one (single-tag or 1-tag) to n (n-tag). Single-tags are comprised of a single phrase (possibly with spaces), such as survey
or use case
and have no tag terminator. Conversely, n-tags (2-tag, 3-tag, …) are a sequence of colon-separated tags comprised of a prefix and suffix:
1
prefix:suffix
prefix
is a general tag, e.g. data:
, scope:
, while suffix
is a more specific tag, e.g. local
, text
, feature importance
. They are defined in BNF form.
Example:
1
2
Prefix scope:
Suffix local | global | local to global
generates tags
1
2
3
4
scope:local
scope:global
scope:local to global
scope:sub-global
n-tags allow for varying levels of specificity and compositionality. Moreover, when dealing with unclear or overly general papers the prefix can be omitted. For instance, if we are dealing with a paper on explanatory interactive learning (prefix xil:
) and the type of explanation provided to the user is not specified we can just use the tag xil:
. On the other hand, if the paper specifies it, we can insert it as a suffix, e.g. xil:feature:importance
, xil:decision rules
.
Tag usage across collections: the alg
tag
Some tags belong to some collections, e.g. the gen:autoencoder
tag for the ml/generative models/autoencoder
collection, but they can (and are) used in other collections too, provided they are prefixed by a alg:
prefix. The alg:
prefix indicates that the given tag is not the focus of the paper, rather an accessory algorithm/technique/feature.
Say we have a paper on local explainability leveraging autoencoders (AE) like ABELE: an obvious tag to use would be gen:autoencoder
. Still, the paper merely leverages existing autoencoder literature to define a XAI algorithm. If we were to use gen:autoencoder
here, we could not tell apart papers on autoencoders (tag gen:autoencoder
) from papers merely leveraging autoencoders (tag alg:gen:autoencoder
). To separate the two classes of papers, we simply add a alg:
prefix to the gen:autoencoder
tag.
General tags
As seen in “Tag usage across collections: the alg
tag”, alg:
is a key tag used across collections to add accessory information. Some tags do not belong to any collection, thus we use them as 2-tags with an alg:
Very cool/important papers
star
Papers on software tools/frameworks
software
Theoretical papers
theoretical
Explanation derived/inspired from…?
alg:
1
shapley values | shap | lime
Algorithm based on…?
alg:
1
backpropagation | input perturbation | knowledge graph
XAI: must-have tags!
Papers in any XAI in collection. Should not use a alg:
prefix when using outside these collections.
What is the explanation scope?
scope:
1
local | global | local to global
Data: what’s the model input data?
data:
1
any | text | image | tabular | sequence | shapelet | graph | health
Model: is the XAI algorithm model-agnostic? Is it explaining a specific model? A model for a specific task?
explains:
1
2
3
4
5
6
7
8
9
10
11
model-agnostic | by design |
linear model |
nn | rnn | cnn | transformer | neurons |
svm |
latent space |
language model | qa |
clustering |
ensemble | forest |
kb:completion
information retriever | recommendation system |
reinforcement agent
kb:completion
explains a knowledge base completion model
What kind of explanation is provided?
Is it a counterfactual? Feature importance (saliency and word importance included)? Does it explain the interaction between features? Does it provide the importance of the neurons in a neural model? Does it provide decision rules (decision trees included)?
explanation:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
aut:grammar | aut:automaton |
caption |
counterfactual |
decision rules | oblique decision rules |
feature:importance | feature:interaction |
graph nodes | graph edges |
input importance |
knowledge graph nodes | knowledge graph relations |
multi-modal |
natural language |
neuron importance |
prototypes |
proof |
shapelet |
structural equations |
taxonomy |
visualization
aut:automaton
an automatonaut:grammar
a (first-order, context-free, etc.) grammarcaption
caption for an imagecounterfactual
a counterfactualdecision rules
feature:importance
what important are the input features?feature:interaction
how are sets of features correlated?prototypes
a (set of) prototypical example(s)graph nodes
nodes in a graphgraph edges
edges in a graphknowledge graph nodes
nodes in a knowledge graphknowledge graph relations
edges in a knowledge graphlatent dimension
multi-modal
multi-modal explanations for reinforcement agentsnatural language
a natural language explanationneuron importance
important or significant neuronsstructural equations
feature equations in causal modelstaxonomy
Task: what task is the black box solving?
Given that most tasks are n-ary classification, we omit the tag for classification problems.
task:
1
2
3
4
5
6
7
8
9
10
11
benchmarking |
dfa distillation |
entailment trees |
interactive xai |
masking |
multiclass classification |
probing | prompting | prompt tuning | knowledge editing |
natural language inference | qa | fact checking |
regression | ordinal regression |
st graphs |
kg-to-text | concepts-to-text
benchmarking
generic papers on benchmarking (XAI included)dfa distillation
distilling a DFA from a modelentailment trees
models who construct proof-like natural language trees to reach a goalinteractive xai
XAI with interactive human interventionprompting
probing masked language models for knowledgeprobing
generic papers on probing techniquesprompt tuning
papers on injecting knowledge in language models through promptsknowledge editing
papers on editing beliefs of language modelskg-to-text
papers on generating text from knowledge graphsconcepts-to-text
papers on generating text from given conceptsst graphs
structured graphs of belief
Explanations: definitions & desiderata and evaluations & benchmarking
exp:
1
definition | validation | benchmark | critique
definition
what is an explanation?validation
how to validate explanations?benchmark
best explanations?critique
a critique of explanation algorithms
Discrete spaces: automata, languages and symbols
Papers on grammars, deterministic finite state automata (DFA), epsilon deterministic finite state automata (eps-DFA), regular and context-free grammars
aut:
1
2
3
grammar | grammar:regular | grammar:context-free | grammar:temporal |
automata:dfa | automata:multiset | automata:symbolic |
symbolic learning
grammar
regular grammar
regular grammarscontext-free grammar
context free grammarstemporal grammar
temporal grammarsdfa
deterministic finite state automataeps-dfa
epsilon deterministic finite state automatasym-dfa
symbolic dfasymbolic learning
learning (from) discrete symbols
Logic
Papers on logic programming, logic-powered models and logic programs inductors.
log:
1
2
3
4
5
6
7
markov network | logic network |
answer set programming | logic programming | inductive logic programming |
inductive logic programming:meta |
theorem proving |
reasoning |
answer set programming |
adaptive | fuzzy
markov network
markov networkslogic network
neural models performing logic operationslogic programming
logic programminginductive logic programming
learning logic programs from datainductive logic programming
learning programs of logic programs from datatheorem proving
can we prove a theorem?reasoning
reasoning in neural modelsadaptive
adaptive logicfuzzy
fuzzy logic
Argumentation
Papers on argumentation.
arg:
Polyhedra and combinatorics
Papers on polyhedra and combinatorics: search algorithm, CSP, SAT, MAXSAT and Set Cover problem, Mixed Integer Programming and Polyhedral learning
cmb:
1
2
3
4
5
6
7
search |
csp |
mip | ilp |
optimal transport |
sat | maxsat |
set cover |
polyhedral learning | polytope | polytope approximation
search
search in a combinatorial spacecsp
constraint satisfaction problemsoptimal transport
optimal transport problemsmip
mixed integer programming problemsilp
integer linear programming problemsat
satisfiability problemsmaxsat
maximum satisfiability problemsset cover
set coverage problemspolyhedral learning
learning polyhedra
Computer vision
cv:
1
segmentation | object detection | style transfer
segmentation
finding meaningful parts of an imageobject detection
detect objects in imagesstyle transfer
applying stylistic patterns to images
Sequence
seq:
1
sax
sax
the SAX family of shapelet discretization algorithms
Clustering
clus:
1
clustering | pruning
clustering
learning a clusteringpruning
merging clusters (mainly in hierarchical clusterings)
Paper type: surveys, benchmarks and theoretical papers
Use tags survey
, benchmark
and theoretical
appropriately.
Adversarial
Papers on adversarial attack/defense in ml/adversarial
Attack or defense?
adv:
1
attack | defense
attack
adversarial attacksdefense
defense against adversarial attacks
Latent space
Papers on embedding, latent space representation, traversal and (dis)entanglement.
lat:
1
2
3
4
dimensionality reduction |
disentanglement |
embedding | representation |
traversal | interpolation
dimensionality reduction
dimensionality reductiondisentanglement
to each latent dimension its own meaningembedding
learning an unsupervised embedding (latent) representationrepresentation
learning a latent representationtraversal
percolating the latent spaceinterpolation
latent space interpolation
Generative models
Papers on generative models (AE, VAE, GAN, flow models, ecc) in /ml/generative models
.
What’s the model class?
gen:
1
2
3
4
5
6
autoencoder | variational autoencoder |
conditioned |
flow |
gan |
process |
other
autoencoder
autoencodersvariational autoencoder
variational autoencodersflow
flow modelsgan
generative adversarial networksprocess
stochastic processes (mainly gaussian)other
other ad hoc generative models
Knowledge bases
kb:
1
2
3
4
5
common sense |
completion |
embedding |
inject | linking
mining
common sense
knowledge bases on common sense reasoningcompletion
infer missing edges in a given knowledge graphembedding
creating embeddings for a knowledge graph nodes and edgesinject
enrich model training with a given knowledge baselinking
map tokens to entities in the knowledge base/graphmining
distilling a knowledge base from data/models
Explainable Interactive Learning (XIL)
Explainable Interactive Learning (XIL) inserts a human-in-the-loop in the training procedure of a machine learning model by periodically presenting the model to the human and prompting an explainable correction
xil:
1
2
3
4
5
counterfactuals |
decision rules | decision making |
feature:importance |
prototypes |
?
feature:importance
training black box models with explainable human intervention in the form of feature importance measuresdecision rules
training black box models with explainable human intervention in the form of decision rulesdecision making
humans and algorithms collaborating at prediction timeprototypes
training black box models with explainable human intervention in the form of prototypes selectioncounterfactuals
training black box models with explainable human intervention in the form of counterfactual decision rules?
unclear
Machine Learning & Pattern Recognition
mlpr:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
backpropagation |
boosting |
boundary tree |
differentiable tree |
capsule network |
concept learning |
decision tree | forest |
distillation |
program synthesis |
graphical model
regularization | constraints |
transfer learning |
som |
training
model training
trainingdistillation
distill models in other modelsconcept learning
learning higher level featuresprogram synthesis
papers on synthesizing programs from dataregularization
regularizationconstraints
learning the set of constraints inferred by a model/constraining a model behaviorsom
the self-organizing maptransfer learning
leverage existing models in different domains/datasets/etc
Computational Mathematics
Maths, duh
cm:
1
matrix factorization | piecewise linear models | voronoi diagram
matrix factorization
factorizing matricespiecewise linear models
on piecewise linear modelsvoronoi diagram
on voronoi diagrams
Optimization
opt:
1
submodular optimization | genetic programming
Statistics
Ususally prefixed by alg:
as the literature is not focused on statistics
stat:
1
bayes | shapley | bandits | mutual information
shapley
leveraging the Shapley valuesbandits
bandits selection algorithmmutual information
Natural Language Processing (NLP)
nlp:
1
2
3
attention | transformer |
language model | perplexity |
qa | summarization | text similarity
attention
defining and leveraging the attention mechanismtransformer
defining and leveraging the multi-head attention mechanism for Transformersqa
question answering
Fairness
Papers on fairness in /fairness
on defining fairness, learning fair models or remove unfair behavior from trained unfair models
fairness:
1
definition | learning | cleaning
definition
formally defining a fairness measurelearning
learning fair modelscleaning
removing unfair behavior from given models