Jary Pomponi(@JaryPom) 's Twitter Profile Photo

Aside from achieving very good results, we also extensively studied the components of our approach to support our claims.

Aside from achieving very good results, we also extensively studied the components of our approach to support our claims.
account_circle
Jary Pomponi(@JaryPom) 's Twitter Profile Photo

The classifier head builds up the prediction in a cascaded way, by scaling past samples using multiple gating values. It simplifies the regularisation of the model and can be used with any approach, improving the performance, especially when the external memory size is contained

The classifier head builds up the prediction in a cascaded way, by scaling past samples using multiple gating values. It simplifies the regularisation of the model and can be used with any approach, improving the performance, especially when the external memory size is contained
account_circle
Jary Pomponi(@JaryPom) 's Twitter Profile Photo

The regularisation schema tries to constrain past probabilities to be lower than the ground truth one.

Combining it with a simple knowledge distillation approach removes the necessity of training on rehearsal samples, improving the plasticity and preserving the stability!

The regularisation schema tries to constrain past probabilities to be lower than the ground truth one.

Combining it with a simple knowledge distillation approach removes the necessity of training on rehearsal samples, improving the plasticity and preserving the stability!
account_circle
Jary Pomponi(@JaryPom) 's Twitter Profile Photo

Back in the Continual Learning game with Alessio Devoto and Simone Scardapane!

In this paper, we propose a regularization schema and a novel classifier head. Combining these two will improve the plasticity of your model while actively fighting catastrophic forgetting.

Back in the Continual Learning game with @devoto_alessio and @s_scardapane! 

In this paper, we propose a regularization schema and a novel classifier head. Combining these two will improve the plasticity of your model while actively fighting catastrophic forgetting.
account_circle
Simone Scardapane(@s_scardapane) 's Twitter Profile Photo

*Conditional computation in NNs: principles and research trends*
by Alessio Devoto Valerio Marsocci Jary Pomponi Pasquale Minervini πŸš€ looking for postdocs!

Our latest tutorial on increasing modularity in NNs with conditional computation, covering MoEs, token selection, & early exits.

arxiv.org/abs/2403.07965

*Conditional computation in NNs: principles and research trends*
by @devoto_alessio @valeriomarsocci @JaryPom @PMinervini 

Our latest tutorial on increasing modularity in NNs with conditional computation, covering MoEs, token selection, & early exits.

arxiv.org/abs/2403.07965
account_circle
Jary Pomponi(@JaryPom) 's Twitter Profile Photo

My first paper of 2024 is out!

With Matteo Gambella, Manuel Roveri, and Simone Scardapane, we developed NACHOS, the first Neural Architecture Search framework for automatically designing optimal Early Exits NNs.

You can find the paper here: arxiv.org/abs/2401.13330

account_circle
Jary Pomponi(@JaryPom) 's Twitter Profile Photo

You can find the paper here: proceedings.mlr.press/v158/qendro21a…

Which comes with a very easy-to-use code: github.com/ajrcampbell/ea…

account_circle
Jary Pomponi(@JaryPom) 's Twitter Profile Photo

The paper is currently under review, but we just released it on arXiv: arxiv.org/abs/2402.01262

Also, the source code, based on Avalanche, is available: github.com/jaryP/Cascaded…

account_circle
Jary Pomponi(@JaryPom) 's Twitter Profile Photo

Ferenc HuszΓ‘r ICML Conference I created a repo containing only the modified paper and then anonymized it.

I will share the latter link, and I hope it will work or the reviewer...

account_circle
Jary Pomponi(@JaryPom) 's Twitter Profile Photo

Martin Mundt Kristian Kersting Wolfgang Stammer Very interesting, thanks for posting it. I will try to use it in my next Continual Learning research!

Hope to get interesting results

account_circle