Tim G. J. Rudner(@timrudner) 's Twitter Profileg
Tim G. J. Rudner

@timrudner

Faculty Fellow at NYU. Trustworthy (safe x transparent x fair) ML.

I'm on the faculty job market.

Prev: PhD Student & Rhodes Scholar at University of Oxford.

ID:4338555809

linkhttps://timrudner.com calendar_today01-12-2015 09:55:56

2,1K Tweets

1,8K Followers

695 Following

NYU Data Science(@NYUDataScience) 's Twitter Profile Photo

CDS’ Tim G. J. Rudner & Julia Stoyanovich (@stoyanoj) join the national effort to set standards for AI safety with the U.S. AI Safety Institute Consortium.

“The stakes are high. And it’s important to get this right,” said Rudner.

Learn more: nyudatascience.medium.com/cds-tim-g-j-ru…

account_circle
Soheil Feizi(@FeiziSoheil) 's Twitter Profile Photo

Just dropped a 4 hour lecture on 'Large Language Models': youtu.be/2yjzZfDQxy8

0:00 Basics of language models
2:30 Word2vec
16:27 Transfer Learning
19:23 BERT
1:00:39 T5
1:31:14 GPT1-3
1:53:05 ChatGPT
2:20:03 LLMs as Deep RL
2:53:00 Policy Gradient
3:32:50 Train your…

account_circle
Helen Toner(@hlntnr) 's Twitter Profile Photo

Seems like everyone wants to know more about how large language models work, but existing explainers often:
-Are too shallow or too technical
-Focus on how LLMs predict the next word, which is only part of the story

So we wrote 3 explainers of our own! 🧵
cset.georgetown.edu/article/the-su…

account_circle
Dan Hendrycks(@DanHendrycks) 's Twitter Profile Photo

Making a good benchmark may seem easy---just collect a dataset---but it requires getting multiple high-level design choices right.

Thomas Woodside and I wrote a post on how to design good ML benchmarks:
safe.ai/blog/devising-…

account_circle
Marc Finzi(@m_finzi) 's Twitter Profile Photo

In this work we construct the first nonvacuous generalization bounds for LLMs, helping to explain why these models generalize.
w/ Sanae Lotfi, Yilun Kuang, Tim G. J. Rudner Micah Goldblum, Andrew Gordon Wilson

arxiv.org/abs/2312.17173

A 🧵on how we make these bounds
1/9

account_circle
Tim G. J. Rudner(@timrudner) 's Twitter Profile Photo

📢 We obtained the **first non-vacuous generalization bounds** for pre-trained LLMs! 🚀🤖

Check out our paper (📄: arxiv.org/abs/2312.17173) and Micah Goldblum's excellent thread 🧵below:

📢 We obtained the **first non-vacuous generalization bounds** for pre-trained LLMs! 🚀🤖 Check out our paper (📄: arxiv.org/abs/2312.17173) and @micahgoldblum's excellent thread 🧵below:
account_circle
Tim G. J. Rudner(@timrudner) 's Twitter Profile Photo

📢 We obtained the **first non-vacuous generalization bounds** for pre-trained large language models! 🚀🤖

Check out our paper (arxiv.org/abs/2312.17173) and Sanae Lotfi's excellent thread 🧵 below:

account_circle
Cong Lu(@cong_ml) 's Twitter Profile Photo

🚨 Model-based methods for offline RL aren’t working for the reasons you think! 🚨

In our new work, led by Anya Sims, we uncover a hidden “edge-of-reach” pathology which we show is the actual reason why offline MBRL methods work or fail!

Let's dive in! 🧵

[1/N]

🚨 Model-based methods for offline RL aren’t working for the reasons you think! 🚨 In our new work, led by @anyaasims, we uncover a hidden “edge-of-reach” pathology which we show is the actual reason why offline MBRL methods work or fail! Let's dive in! 🧵 [1/N]
account_circle
Yee Whye Teh(@yeewhye) 's Twitter Profile Photo

Please apply if you are a UK undergraduate from under-represented backgrounds interested in exploring what a career in AI research is like. Deadline Feb 17!
Google DeepMind has kindly supported the AI internship projects UNIQ+, University of Oxford, thank you!

account_circle
Dan Hendrycks(@DanHendrycks) 's Twitter Profile Photo

To help make models more robust and defend against misuse, we created HarmBench, an evaluation framework for automated red teaming and testing the adversarial robustness of LLMs and multimodal models.

🌐 harmbench.org
📝 arxiv.org/abs/2402.04249

To help make models more robust and defend against misuse, we created HarmBench, an evaluation framework for automated red teaming and testing the adversarial robustness of LLMs and multimodal models. 🌐 harmbench.org 📝 arxiv.org/abs/2402.04249
account_circle
Ahmad Beirami(@abeirami) 's Twitter Profile Photo

[ paper]

TL;DR We introduce 𝗴𝗿𝗮𝗱𝗶𝗲𝗻𝘁-𝗯𝗮𝘀𝗲𝗱 𝗿𝗲𝗱 𝘁𝗲𝗮𝗺𝗶𝗻𝗴 (𝗚𝗕𝗥𝗧), an effective method for triggering language models to produce unsafe responses, even when the LM is finetuned to be safe through 𝑎𝑙𝑖𝑔𝑛𝑚𝑒𝑛𝑡.

[#eacl2024 paper] TL;DR We introduce 𝗴𝗿𝗮𝗱𝗶𝗲𝗻𝘁-𝗯𝗮𝘀𝗲𝗱 𝗿𝗲𝗱 𝘁𝗲𝗮𝗺𝗶𝗻𝗴 (𝗚𝗕𝗥𝗧), an effective method for triggering language models to produce unsafe responses, even when the LM is finetuned to be safe through 𝑎𝑙𝑖𝑔𝑛𝑚𝑒𝑛𝑡.
account_circle
Shalmali Joshi(@shalmali_joshi_) 's Twitter Profile Photo

MLHC 2024 Cfp is out! Submission deadline is April 9, 2024. We're switching to OpenReview this cycle and the submission site will open soon. Stay tuned!

account_circle
Alex Yanko 🇺🇦(@LeopolisDream) 's Twitter Profile Photo

State Space Models: A Modern Approach

This is an interactive textbook on state space models (SSM) using the JAX Python library.

probml.github.io/ssm-book/root.…

State Space Models: A Modern Approach This is an interactive textbook on state space models (SSM) using the JAX Python library. probml.github.io/ssm-book/root.…
account_circle
Pierre Alquier(@PierreAlquier) 's Twitter Profile Photo

'User-friendly Introduction to PAC-Bayes bounds' is now published: nowpublishers.com/article/Detail…

Thanks a lot for all the feedback on the first version I posted on arXiv some time ago 🙏

I'll be teaching a course on this topic at SSML 2024 OIST groups.oist.jp/mlss

'User-friendly Introduction to PAC-Bayes bounds' is now published: nowpublishers.com/article/Detail… Thanks a lot for all the feedback on the first version I posted on arXiv some time ago 🙏 I'll be teaching a course on this topic at SSML 2024 @OISTedu groups.oist.jp/mlss
account_circle
Yarin(@yaringal) 's Twitter Profile Photo

I'm hiring!

I'm building 4 research groups under me at AISI (formerly the UK's Taskforce on Frontier AI) to work on foundational AI safety research.
[1/5]
gov.uk/government/pub…

account_circle
Jeffrey Ding(@jjding99) 's Twitter Profile Photo

Excited to share that my book on technological revolutions and power transitions will be published in Princeton University Press @princetonupress.bsky.s in August.
Available to preorder ($30 paperback):
- amazon.com/Technology-Ris…
- bookshop.org/p/books/techno…

Excited to share that my book on technological revolutions and power transitions will be published in @PrincetonUPress in August. Available to preorder ($30 paperback): - amazon.com/Technology-Ris… - bookshop.org/p/books/techno…
account_circle
Robert Palgrave(@Robert_Palgrave) 's Twitter Profile Photo

We have now completed our analysis of new materials reported in the Google Deepmind / Berkeley autonomous lab paper. My own initial analysis is in the quote tweet.

Happy to have worked with Schoop Lab to jointly put together a comprehensive analysis, now available on ChemRxiv.…

account_circle
Ben Eysenbach(@ben_eysenbach) 's Twitter Profile Photo

I'm hiring visiting students/RAs for Summer 2024! ☀️

Focus: scaling unsupervised RL & understanding emergent properties. 🚀
Who: PhD, MS, rising undergrad seniors w/ prior research experience.
Apply by emailing me a resume.

Details below👇

account_circle
Jia-Bin Huang(@jbhuang0604) 's Twitter Profile Photo

New YouTube video!

Diffusion models are awesome! But how do they actually work? 🤔

Check out the video and learn with me!

youtube.com/watch?v=i2qSxM…

New YouTube video! Diffusion models are awesome! But how do they actually work? 🤔 Check out the video and learn with me! youtube.com/watch?v=i2qSxM…
account_circle