Gantavya Bhatt(@BhattGantavya) 's Twitter Profileg
Gantavya Bhatt

@BhattGantavya

Ph.D. Student @UW, MELODI Lab and @uw_wail at @uwcse Formerly @amazonscience, EE undergrad @iitdelhi. An active photographer and Alpinist!

ID:1011498496648548358

linkhttps://sites.google.com/view/gbhatt/ calendar_today26-06-2018 06:36:49

1,0K Tweets

545 Followers

1,4K Following

Raghav Somani(@SomaniRaghav) 's Twitter Profile Photo

There are too many problems with this. This is not like spelling bee or olympiads.

How about first taking a step to help students better learn calculus and probability? It saddens me to see even many senior undergraduates struggling with basic high school-level probability.

account_circle
Ashima Suvarna 🌻(@suvarna_ashima) 's Twitter Profile Photo

📢 We propose a new benchmark called PhonologyBench for testing how well LLMs perform on three tasks that require sound knowledge : phonemic transcription, counting syllables, and listing possible rhymes.
w/ Harshita Khandelwal, Violet Peng 1/3

📢 We propose a new benchmark called PhonologyBench for testing how well LLMs perform on three tasks that require sound knowledge : phonemic transcription, counting syllables, and listing possible rhymes. w/ Harshita Khandelwal, @VioletNPeng 1/3
account_circle
Ahmad Beirami(@abeirami) 's Twitter Profile Photo

Robustness methods
1) augment data with natural/synthetic perturbations and a consistency loss
2) reweight samples to improve generalization (like DRO)

We do it differently!
We show significant robustness with a simple tweak of the first layer and loss motivated by comms theory.

account_circle
Ashima Suvarna 🌻(@suvarna_ashima) 's Twitter Profile Photo

📢Our project website for DOVE 🕊️is up !
🌐 dove-alignment.github.io
📜 arxiv.org/abs/2404.00530
💻 github.com/Hritikbansal/d…
🤗 huggingface.co/jointpreferenc…

📢Our project website for DOVE 🕊️is up ! 🌐 dove-alignment.github.io 📜 arxiv.org/abs/2404.00530 💻 github.com/Hritikbansal/d… 🤗 huggingface.co/jointpreferenc…
account_circle
Gantavya Bhatt(@BhattGantavya) 's Twitter Profile Photo

Story of discussion phase in conferences: reviewers don’t respond even after addressing all of the raised concerns (this time they’re even ignoring ACs message about replying to the rebuttal). Oh well.

account_circle
Hritik Bansal(@hbXNov) 's Twitter Profile Photo

Our data, code, and checkpoints are available on huggingface 🤗 and github:
Paper: arxiv.org/abs/2404.00530
Data: huggingface.co/datasets/joint…
Code: github.com/Hritikbansal/d…

account_circle
Hritik Bansal(@hbXNov) 's Twitter Profile Photo

We find that the LLM trained with joint instruction-response preference data using DOVE outperforms the LLM trained with DPO by 5.2% and 3.3% win-rate for the TL;DR and Anthropic-Helpful datasets, respectively!

We find that the LLM trained with joint instruction-response preference data using DOVE outperforms the LLM trained with DPO by 5.2% and 3.3% win-rate for the TL;DR and Anthropic-Helpful datasets, respectively!
account_circle
Hritik Bansal(@hbXNov) 's Twitter Profile Photo

To learn from joint preferences, we introduce a new preference optimization objective. Intuitively, it upweights the joint probability of the preferred instruction-response pair over the rejected pair. If instructions are identical, then DOVE boils down to DPO!

To learn from joint preferences, we introduce a new preference optimization objective. Intuitively, it upweights the joint probability of the preferred instruction-response pair over the rejected pair. If instructions are identical, then DOVE boils down to DPO!
account_circle
Hritik Bansal(@hbXNov) 's Twitter Profile Photo

We find (a) in 71% of cases, humans/AI can make a decisive choice under joint setup, when both the responses are chosen/rejected under conditional setup, (b) in joint setup, humans/AI can favor a response that was rejected based on conditional setup over a preferred response.

We find (a) in 71% of cases, humans/AI can make a decisive choice under joint setup, when both the responses are chosen/rejected under conditional setup, (b) in joint setup, humans/AI can favor a response that was rejected based on conditional setup over a preferred response.
account_circle
Hritik Bansal(@hbXNov) 's Twitter Profile Photo

Let’s get humans to provide feedback ✍️! In the conditional setup, Resp. B and D are rejected against A and C. However, in the joint preference setup, an instruction-response (I1-Resp. B) is preferred over another instruction-response (I2-Resp. D) with a valid feedback reasoning.

Let’s get humans to provide feedback ✍️! In the conditional setup, Resp. B and D are rejected against A and C. However, in the joint preference setup, an instruction-response (I1-Resp. B) is preferred over another instruction-response (I2-Resp. D) with a valid feedback reasoning.
account_circle
Hritik Bansal(@hbXNov) 's Twitter Profile Photo

Traditional conditional feedback approaches are limited in capturing the multifaceted nature of human preferences! Thus, we collect human and AI preferences jointly over instruction-response pairs i.e., (I1,R1) vs (I2, R2). Joint preferences subsume conditional pref. when I1=I2.

account_circle
Hritik Bansal(@hbXNov) 's Twitter Profile Photo

Common LLM alignment protocols acquire ranking feedback from human/AI conditioned on an identical context. Is a fixed context necessary? Would you prefer a detailed, well-articulated product review ✍️😍over a vague, inaccurate movie review 📽🚫?

account_circle
Ben Recht(@beenwrekt) 's Twitter Profile Photo

Gergely is right. There is absolutely no reason we have to continue with this rebuttal process.

Set the accept threshold to 50% and work on inclusion in this mad time of too many papers.

account_circle