I have a public Zotero library where I archive and catalogue the papers I read, contact me if you wish to be added to the library. Follows a catalogue of the tags that I use.
Tags
We have several classes of tags, each identified by its token cardinality, ranging from one (single-tag or 1-tag) to n (n-tag). Single-tags are comprised of a single phrase (possibly with spaces), such as survey or use case and have no tag terminator. Conversely, n-tags (2-tag, 3-tag, …) are a sequence of colon-separated tags comprised of a prefix and suffix:
1
prefix:suffix
prefix is a general tag, e.g. data:, scope:, while suffix is a more specific tag, e.g. local, text, feature importance.  They are defined in BNF form.
Example:
1
2
Prefix scope:
Suffix local | global | local to global
generates tags
1
2
3
4
scope:local
scope:global
scope:local to global
scope:sub-global
n-tags allow for varying levels of specificity and compositionality. Moreover, when dealing with unclear or overly general papers the prefix can be omitted. For instance, if we are dealing with a paper on explanatory interactive learning (prefix  xil:) and the type of explanation provided to the user is not specified we can just use the tag xil:. On the other hand, if the paper specifies it, we can insert it as a suffix, e.g. xil:feature:importance, xil:decision rules.
Tag usage across collections: the alg tag
Some tags belong to some collections, e.g. the gen:autoencoder tag for the ml/generative models/autoencoder collection, but they can (and are) used in other collections too, provided they are prefixed by a alg: prefix. The alg: prefix indicates that the given tag is not the focus of the paper, rather an accessory algorithm/technique/feature.
Say we have a paper on local explainability leveraging autoencoders (AE) like ABELE: an obvious tag to use would be gen:autoencoder. Still, the paper merely leverages existing autoencoder literature to define a XAI algorithm. If we were to use gen:autoencoder here, we could not tell apart papers on autoencoders (tag gen:autoencoder) from papers merely leveraging autoencoders (tag alg:gen:autoencoder). To separate the two classes of papers, we simply add a alg: prefix to the gen:autoencoder tag.
General tags
As seen in “Tag usage across collections: the alg tag”, alg: is a key tag used across collections to add accessory information. Some tags do not belong to any collection, thus we use them as 2-tags with an alg:
Very cool/important papers
star
Papers on software tools/frameworks
software
Theoretical papers
theoretical
Explanation derived/inspired from…?
alg:
1
shapley values | shap | lime
Algorithm based on…?
alg:
1
backpropagation | input perturbation | knowledge graph
XAI: must-have tags!
Papers in any XAI in collection. Should not use a alg: prefix when using outside these collections.
What is the explanation scope?
scope:
1
local | global | local to global
Data: what’s the model input data?
data:
1
any | text | image | tabular | sequence | shapelet | graph | health
Model: is the XAI algorithm model-agnostic? Is it explaining a specific model? A model for a specific task?
explains:
1
2
3
4
5
6
7
8
9
10
11
model-agnostic | by design |
linear model |
nn | rnn | cnn | transformer | neurons |
svm |
latent space |
language model | qa |
clustering |
ensemble | forest |
kb:completion
information retriever | recommendation system |
reinforcement agent
- kb:completionexplains a knowledge base completion model
What kind of explanation is provided?
Is it a counterfactual? Feature importance (saliency and word importance included)? Does it explain the interaction between features? Does it provide the importance of the neurons in a neural model? Does it provide decision rules (decision trees included)?
explanation:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
aut:grammar | aut:automaton |
caption |
counterfactual |
decision rules | oblique decision rules |
feature:importance | feature:interaction |
graph nodes | graph edges |
input importance |
knowledge graph nodes | knowledge graph relations |
multi-modal |
natural language |
neuron importance |
prototypes |
proof |
shapelet |
structural equations |
taxonomy |
visualization
- aut:automatonan automaton
- aut:grammara (first-order, context-free, etc.) grammar
- captioncaption for an image
- counterfactuala counterfactual
- decision rules
- feature:importancewhat important are the input features?
- feature:interactionhow are sets of features correlated?
- prototypesa (set of) prototypical example(s)
- graph nodesnodes in a graph
- graph edgesedges in a graph
- knowledge graph nodesnodes in a knowledge graph
- knowledge graph relationsedges in a knowledge graph
- latent dimension
- multi-modalmulti-modal explanations for reinforcement agents
- natural languagea natural language explanation
- neuron importanceimportant or significant neurons
- structural equationsfeature equations in causal models
- taxonomy
Task: what task is the black box solving?
Given that most tasks are n-ary classification, we omit the tag for classification problems.
task:
1
2
3
4
5
6
7
8
9
10
11
benchmarking |
dfa distillation |
entailment trees |
interactive xai |
masking |
multiclass classification |
probing | prompting | prompt tuning | knowledge editing |
natural language inference | qa | fact checking |
regression | ordinal regression |
st graphs |
kg-to-text | concepts-to-text
- benchmarkinggeneric papers on benchmarking (XAI included)
- dfa distillationdistilling a DFA from a model
- entailment treesmodels who construct proof-like natural language trees to reach a goal
- interactive xaiXAI with interactive human intervention
- promptingprobing masked language models for knowledge
- probinggeneric papers on probing techniques
- prompt tuningpapers on injecting knowledge in language models through prompts
- knowledge editingpapers on editing beliefs of language models
- kg-to-textpapers on generating text from knowledge graphs
- concepts-to-textpapers on generating text from given concepts
- st graphsstructured graphs of belief
Explanations: definitions & desiderata and evaluations & benchmarking
exp:
1
definition | validation | benchmark | critique
- definitionwhat is an explanation?
- validationhow to validate explanations?
- benchmarkbest explanations?
- critiquea critique of explanation algorithms
Discrete spaces: automata, languages and symbols
Papers on grammars, deterministic finite state automata (DFA), epsilon deterministic finite state automata (eps-DFA), regular and context-free grammars
aut:
1
2
3
grammar | grammar:regular | grammar:context-free | grammar:temporal |
automata:dfa | automata:multiset | automata:symbolic |
symbolic learning
- grammar
- regular grammarregular grammars
- context-free grammarcontext free grammars
- temporal grammartemporal grammars
- dfadeterministic finite state automata
- eps-dfaepsilon deterministic finite state automata
- sym-dfasymbolic dfa
- symbolic learninglearning (from) discrete symbols
Logic
Papers on logic programming, logic-powered models and logic programs inductors.
log:
1
2
3
4
5
6
7
markov network | logic network |
answer set programming | logic programming | inductive logic programming |
inductive logic programming:meta |
theorem proving |
reasoning |
answer set programming |
adaptive | fuzzy
- markov networkmarkov networks
- logic networkneural models performing logic operations
- logic programminglogic programming
- inductive logic programminglearning logic programs from data
- inductive logic programminglearning programs of logic programs from data
- theorem provingcan we prove a theorem?
- reasoningreasoning in neural models
- adaptiveadaptive logic
- fuzzyfuzzy logic
Argumentation
Papers on argumentation.
arg:
Polyhedra and combinatorics
Papers on polyhedra and combinatorics: search algorithm, CSP, SAT, MAXSAT and Set Cover problem, Mixed Integer Programming and Polyhedral learning
cmb:
1
2
3
4
5
6
7
search |
csp |
mip | ilp |
optimal transport |
sat | maxsat |
set cover |
polyhedral learning | polytope | polytope approximation
- searchsearch in a combinatorial space
- cspconstraint satisfaction problems
- optimal transportoptimal transport problems
- mipmixed integer programming problems
- ilpinteger linear programming problem
- satsatisfiability problems
- maxsatmaximum satisfiability problems
- set coverset coverage problems
- polyhedral learninglearning polyhedra
Computer vision
cv:
1
segmentation | object detection | style transfer
- segmentationfinding meaningful parts of an image
- object detectiondetect objects in images
- style transferapplying stylistic patterns to images
Sequence
seq:
1
sax
- saxthe SAX family of shapelet discretization algorithms
Clustering
clus:
1
clustering | pruning
- clusteringlearning a clustering
- pruningmerging clusters (mainly in hierarchical clusterings)
Paper type: surveys, benchmarks and theoretical papers
Use tags survey, benchmark and theoretical appropriately.
Adversarial
Papers on adversarial attack/defense in  ml/adversarial
Attack or defense?
adv:
1
attack | defense
- attackadversarial attacks
- defensedefense against adversarial attacks
Latent space
Papers on embedding, latent space representation, traversal and (dis)entanglement.
lat:
1
2
3
4
dimensionality reduction |
disentanglement |
embedding | representation |
traversal | interpolation
- dimensionality reductiondimensionality reduction
- disentanglementto each latent dimension its own meaning
- embeddinglearning an unsupervised embedding (latent) representation
- representationlearning a latent representation
- traversalpercolating the latent space
- interpolationlatent space interpolation
Generative models
Papers on generative models (AE, VAE, GAN, flow models, ecc) in /ml/generative models.
What’s the model class?
gen:
1
2
3
4
5
6
autoencoder | variational autoencoder |
conditioned |
flow |
gan |
process |
other
- autoencoderautoencoders
- variational autoencodervariational autoencoders
- flowflow models
- gangenerative adversarial networks
- processstochastic processes (mainly gaussian)
- otherother ad hoc generative models
Knowledge bases
kb:
1
2
3
4
5
common sense |
completion |
embedding |
inject | linking
mining
- common senseknowledge bases on common sense reasoning
- completioninfer missing edges in a given knowledge graph
- embeddingcreating embeddings for a knowledge graph nodes and edges
- injectenrich model training with a given knowledge base
- linkingmap tokens to entities in the knowledge base/graph
- miningdistilling a knowledge base from data/models
Explainable Interactive Learning (XIL)
Explainable Interactive Learning (XIL) inserts a human-in-the-loop in the training procedure of a machine learning model by periodically presenting the model to the human and prompting an explainable correction
xil:
1
2
3
4
5
counterfactuals |
decision rules | decision making |
feature:importance |
prototypes |
?
- feature:importancetraining black box models with explainable human intervention in the form of feature importance measures
- decision rulestraining black box models with explainable human intervention in the form of decision rules
- decision makinghumans and algorithms collaborating at prediction time
- prototypestraining black box models with explainable human intervention in the form of prototypes selection
- counterfactualstraining black box models with explainable human intervention in the form of counterfactual decision rules
- ?unclear
Machine Learning & Pattern Recognition
mlpr:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
backpropagation |
boosting |
boundary tree |
differentiable tree |
capsule network |
concept learning |
decision tree | forest |
distillation |
program synthesis |
graphical model
regularization | constraints |
transfer learning |
som |
training
- model trainingtraining
- distillationdistill models in other models
- concept learninglearning higher level features
- program synthesispapers on synthesizing programs from data
- regularizationregularization
- constraintslearning the set of constraints inferred by a model/constraining a model behavior
- somthe self-organizing map
- transfer learningleverage existing models in different domains/datasets/etc
Computational Mathematics
Maths, duh
cm:
1
matrix factorization | piecewise linear models | voronoi diagram
- matrix factorizationfactorizing matrices
- piecewise linear modelson piecewise linear models
- voronoi diagramon voronoi diagrams
Optimization
opt:
1
submodular optimization | genetic programming
Statistics
Ususally prefixed by alg: as the literature is not focused on statistics
stat:
1
bayes | shapley | bandits | mutual information
- shapleyleveraging the Shapley values
- banditsbandits selection algorithm
- mutual information
Natural Language Processing (NLP)
nlp:
1
2
3
attention | transformer |
language model | perplexity |
qa | summarization | text similarity
- attentiondefining and leveraging the attention mechanism
- transformerdefining and leveraging the multi-head attention mechanism for Transformers
- qaquestion answering
Fairness
Papers on fairness in /fairness on defining fairness, learning fair models or remove unfair behavior from trained unfair models
fairness:
1
definition | learning | cleaning
- definitionformally defining a fairness measure
- learninglearning fair models
- cleaningremoving unfair behavior from given models