18-08-27 New Concrete Representation Learning Ideas
Category: Idea Lists (Upon Request)
<!-- gdoc-inlined -->
Let them be shitty. But go super concrete. Super concrete.
I like this idea of ‘new metrics over representations’. I want to learn abstract representations.
- Sensitivity score (maml based).
- Do close analysis of how the representation represents a single image.
- Credit assignment over particular filters.
- Look at the way the filters are recombined with one another - find simple examples of composition that models a particular part of the image.
- Some filters will be composed with more or less other filters, which is a different metric than their importance. Which filters activate over the most images? At each level of the network?
- Can we use this metric of generality as a heuristic for transfer? Say, only filters that are sufficiently general get used in transfer?
- Find ‘conflation’ in a representation. (this may be hard)
- Have a notion of which features should be recombined to create a higher level feature
- Look for overlapping activations where they should not exist (misclassified examples should be really good for this)
- Have a notion of which features should be recombined to create a higher level feature
- Get to a ‘why’ for misclassified examples
- Look at the ways that the representation couldn’t distinguish between particular parts of an input, look at the mistakes made over 4-5 examples and diagnose them
- Are you allowed to publish a paper titled ‘why our networks fail?’
- This may be hard to get causal on, but could be extremely useful.
- Do VQ-VAE, but with a forward predictive model. Generative model of future, rather than present. Auto-regressive generative model.
- I guess this is what the forward predictive lstm over VAE state is, in a way.
- Take Ben Poole’s Gaussian Mixture Model VQ-VAE and apply it to something like world models (where you want this ability to go discrete or continuous)
- Is this idea for generative future prediction a thing in general? VAE + LSTM to do it is awesome, but is it the best in its class for that task?
- Check manifold learning hypothesis
- Dog manifold on animal manifold, for example
- Causal representation learning - which filtermaps have counterfactual impact on the output?
- Use filter-level dropout to estimate this
Source: Original Google Doc