ServiceNow AI Research

Disentanglement

Causal Differentiating Concepts: Interpreting LM Behavior via Causal Representation Learning
Language model activations entangle concepts that mediate their behavior, making it difficult to interpret these factors, which has …
Monotonicity Regularization: Improved Penalties and Novel Applications to Disentangled Representation Learning and Robust Classification
We study settings where gradient penalties are used alongside risk minimization with the goal of obtaining predictors satisfying …
Disentangling the independently controllable factors of variation by interacting with the world
It has been postulated that a good representation is one that disentangles the underlying explanatory factors of variation. However, it …