Yann LeCun(@ylecun) 's Twitter Profileg
Yann LeCun

@ylecun

Professor at NYU. Chief AI Scientist at Meta.
Researcher in AI, Machine Learning, Robotics, etc.
ACM Turing Award Laureate.

ID:48008938

linkhttp://yann.lecun.com calendar_today17-06-2009 16:05:51

18,9K Tweets

707,9K Followers

716 Following

Philipp Schmid(@_philschmid) 's Twitter Profile Photo

Easily Fine-tune AI at Meta Llama 3 70B! 🦙 I am excited to share a new guide on how to fine-tune Llama 3 70B with PyTorch FSDP, Q-Lora, and Flash Attention 2 (SDPA) using Hugging Face build for consumer-size GPUs (4x 24GB). 🚀

Blog: philschmid.de/fsdp-qlora-lla…

The blog covers:
👨‍💻

account_circle
Jeremy Howard(@jeremyphoward) 's Twitter Profile Photo

Today at Answer.AI we've got something new for you: FSDP/QDoRA. We've tested it with AI at Meta Llama3 and the results blow away anything we've seen before.

I believe that this combination is likely to create better task-specific models than anything else at any cost. 🧵

Today at @answerdotai we've got something new for you: FSDP/QDoRA. We've tested it with @AIatMeta Llama3 and the results blow away anything we've seen before. I believe that this combination is likely to create better task-specific models than anything else at any cost. 🧵
account_circle
Laura Gwilliams(@GwilliamsL) 's Twitter Profile Photo

Delighted to share our new preprint! with Alec Marantz, @DavidPoeppel and Jean-Rémi King:
biorxiv.org/content/10.110…

'Hierarchical dynamic coding coordinates speech comprehension in the brain'
Summary below 👇

1/8

account_circle
The Technology Brother(@thetechbrother) 's Twitter Profile Photo

Is it possible to produce an AI system that is not biased?

As we saw with Google Gemini, it's clear that bias in our AI models will always be a huge problem. So what's the solution?

Here Yann LeCun nails it - It's the same solution we found with the press: freedom and diversity.

account_circle
TuringPost(@TheTuringPost) 's Twitter Profile Photo

Yann LeCun Yann LeCun delivered a lecture on Objective-Driven AI.

He began with a reality check: 'Machine Learning falls short compared to humans and animals!'

Here's his insight on constructing AI systems that learn, reason, plan, and prioritize safety:

1/5

Yann LeCun @ylecun delivered a lecture on Objective-Driven AI. He began with a reality check: 'Machine Learning falls short compared to humans and animals!' Here's his insight on constructing AI systems that learn, reason, plan, and prioritize safety: 1/5
account_circle
martin_casado(@martin_casado) 's Twitter Profile Photo

At this point I feel like we understand pretty well what's going on with LLMs:

- Outputs are roughly equivalent to kernel smoothing over positional embeddings (arxiv.org/pdf/1908.11775…)
- The learned computation model is *probably* bounded by RASP-L (arxiv.org/pdf/2310.16028…)
-

account_circle
Massimo(@Rainmaker1973) 's Twitter Profile Photo

Corvids are known to be smart problem solvers.

This one, for example, plays tic-tac-toe.

[📹voron_gosha_tv]

account_circle
Yann LeCun(@ylecun) 's Twitter Profile Photo

Agency (and planning) can't be a wart on top of Auto-Regressive LLMs.
It must be an intrinsic property of the architecture.

account_circle
Michael Bronstein(@mmbronstein) 's Twitter Profile Photo

We started releasing the first chapters of our Geometric Deep Learning book and the accompanying slides from the corresponding Oxford and Cambridge courses.

account_circle
Mike Schroepfer(@schrep) 's Twitter Profile Photo

True Story!

One of the many reasons I love open source is it doesn't give a damn about the org chart or 'managing up.' If people outside of FB/Meta didn't use or like our OSS then something was wrong with it.

PyTorch succeeded because of the hyper focus on developer

account_circle
Artificial Analysis(@ArtificialAnlys) 's Twitter Profile Photo

First throughput tokens/s benchmarks of Llama 3 🐎
Groq Inc is coming fast out of the gate with 284 tokens/s on Llama 3 70B and 876 tokens/s on Llama 3 8B.
Once-again showing the step-change speed benefits of their custom LPU chips.

Congratulations to other providers too,

First throughput tokens/s benchmarks of Llama 3 🐎 @GroqInc is coming fast out of the gate with 284 tokens/s on Llama 3 70B and 876 tokens/s on Llama 3 8B. Once-again showing the step-change speed benefits of their custom LPU chips. Congratulations to other providers too,
account_circle