Desmond Elliott(@delliott) 's Twitter Profileg
Desmond Elliott

@delliott

Assistant Professor at the University of Copenhagen working on multimodal machine learning.

ID:8164582

linkhttps://elliottd.github.io/ calendar_today13-08-2007 18:37:04

3,1K Tweets

2,8K Followers

440 Following

Nikos Aletras(@nikaletras) 's Twitter Profile Photo

Really enjoyed showcasing research that we've been doing the past few years at Sheffield. The main premise of the lecture was that LLMs are idiot savants with remarkable abilities, performing well under extreme or non-intuitive settings (e.g. pre-training, info loss etc.)

account_circle
Shomir Wilson(@ShomirWilson) 's Twitter Profile Photo

Here's the diagram. This aggregate view provides some insights that are otherwise obscure to student researchers who are just getting started: nearly all of our manuscripts eventually get accepted somewhere, but the number of resubmissions necessary varies widely.

Here's the diagram. This aggregate view provides some insights that are otherwise obscure to student researchers who are just getting started: nearly all of our manuscripts eventually get accepted somewhere, but the number of resubmissions necessary varies widely.
account_circle
Desmond Elliott(@delliott) 's Twitter Profile Photo

I'm excited about this project on compositional understanding across multimodal sequences taken from egocentric kitchen activities

account_circle
ELLIS Unit Lisbon(@Lisbon_ELLIS) 's Twitter Profile Photo

πŸ“’ Only 10 days left to apply to the Lisbon Machine Learning School in July. Check out our fantastic lineup of speakers πŸ‘‡

πŸ”— for more information: lxmls.it.pt/2024/

πŸ”— to apply: tinyurl.com/288raakv

πŸ“’ Only 10 days left to apply to the Lisbon Machine Learning School in July. Check out our fantastic lineup of speakers πŸ‘‡ πŸ”— for more information: lxmls.it.pt/2024/ πŸ”— to apply: tinyurl.com/288raakv #LxMLS #MachineLearning #NLP
account_circle
Nathan Godey @ICLR(@nthngdy) 's Twitter Profile Photo

🀏 Why do small Language Models underperform?

We prove empirically and theoretically that the LM head on top of language models can limit performance through the softmax bottleneck phenomenon, especially when the hidden dimension <1000.

πŸ“„Paper: arxiv.org/pdf/2404.07647…
(1/10)

🀏 Why do small Language Models underperform? We prove empirically and theoretically that the LM head on top of language models can limit performance through the softmax bottleneck phenomenon, especially when the hidden dimension <1000. πŸ“„Paper: arxiv.org/pdf/2404.07647… (1/10)
account_circle
babyLM(@babyLMchallenge) 's Twitter Profile Photo

πŸ‘Ά BabyLM Challenge is back!
Can you improve pretraining with a small data budget?

BabyLMs for better LLMs
& for understanding how humans learn from 100M words

New:
How vision affects learning
Bring your own data
Paper track

babylm.github.io
🧡

account_circle
Hilde Kuehne(@HildeKuehne) 's Twitter Profile Photo

Ever asked yourself what's the best explainability method for ViT Transformer at the moment?

🚨 We present you LeGrad, a Layerwise Explainability GRADient method for large ViT transformer architectures. 🚨

So, grab your β˜•οΈ and πŸ₯, release your inner πŸ‡«πŸ‡·, and join us on our…

Ever asked yourself what's the best explainability method for ViT Transformer at the moment? 🚨 We present you LeGrad, a Layerwise Explainability GRADient method for large ViT transformer architectures. 🚨 So, grab your β˜•οΈ and πŸ₯, release your inner πŸ‡«πŸ‡·, and join us on our…
account_circle
Desmond Elliott(@delliott) 's Twitter Profile Photo

I am looking for an ultra-emergency reviewer for ARR to review a paper on speech multimodality. Please reach out if you can help

account_circle
Desmond Elliott(@delliott) 's Twitter Profile Photo

Sadly I won’t make it to in Malta this year but Wenyan Li will be there to talk about Data Curation for Image Captioning on Monday at 1400 in Marie Louise 2

account_circle
Stephanie Brandl(@StephanieBrandl) 's Twitter Profile Photo

I am thrilled to share that our paper 'Evaluating Webcam-based Gaze Data as an Alternative for Human Rationale Annotations' was accepted to LREC-COLING.
πŸ“œ arxiv.org/pdf/2402.19133…
❀️ with Oliver Eberle, Tiago Ribeiro, Anders Søgaard and Nora Hollenstein ML Group, TU Berlin CoAStAL NLP

account_circle
Desmond Elliott(@delliott) 's Twitter Profile Photo

Do you know of any high-quality language-specific pretraining datasets? I’d love to hear about them and why you think they are of high quality.

account_circle