Curse of dimensionality occurs when learning structureless data in high dimension \(d\):
VS
\(\varepsilon\sim P^{-\beta}\)
1/10
[Kaplan 20]
[Hestness 17]
\(\Rightarrow\) Data must be structured and
Machine Learning should capture such structure.
Key questions motivating this thesis:
2/10
Reducing complexity with depth
Deep networks build increasingly abstract representations with depth (also in brain)
Intuition: reduces complexity of the task, ultimately beating curse of dimensionality.
Two ways for losing information
by learning invariances
Discrete
Continuous
[Zeiler and Fergus 14, Yosinski 15, Olah 17, Doimo 20,
Van Essen 83, Grill-Spector 04]
[Shwartz-Ziv and Tishby 17, Ansuini 19, Recanatesi 19, ]
[Bruna and Mallat 13, Mallat 16, Petrini 21]
3/10
Hierarchical structure
How many training points?
Quantitative predictions in a model of data
sofa
[Chomsky 1965]
[Grenander 1996]
4/10
5/10
\(P^*\)
\(P^*\sim n_c m^L\)
6/10
How many training points are needed to group synonyms?
7/10
Patch \(\mu\)
Label \(\alpha\)
8/10
At \(P^*\) the task and the synonyms are learnt
1.0
0.5
0.0
\(10^4\)
Training set size \(P\)
9/10
Future directions
Thank you!
sofa
28/28
Thank you!
10/10
Generative technique:
The problem (NL-CSP):
Our approach:
[U.S patent]
Focus on LLM part: