EdinburghNLP(@EdinburghNLP) 's Twitter Profileg
EdinburghNLP

@EdinburghNLP

The Natural Language Processing Group at the University of Edinburgh

ID:862649978170245120

linkhttp://edinburghnlp.inf.ed.ac.uk/ calendar_today11-05-2017 12:45:56

988 Tweets

11,3K Followers

140 Following

Giwon Hong(@GiwonHong413849) 's Twitter Profile Photo

Introducing The Hallucinations Leaderboard! 🚀 An open effort to measure the LLMs’ tendency to generate hallucinations across various tasks, like open-domain QA, instruction following, and summarisation! 🧵1/N

📄Paper:arxiv.org/abs/2404.05904
🤗Leaderboard:huggingface.co/spaces/halluci…

Introducing The Hallucinations Leaderboard! 🚀 An open effort to measure the LLMs’ tendency to generate hallucinations across various tasks, like open-domain QA, instruction following, and summarisation! 🧵1/N 📄Paper:arxiv.org/abs/2404.05904 🤗Leaderboard:huggingface.co/spaces/halluci…
account_circle
Pasquale Minervini 🚀 looking for postdocs!(@PMinervini) 's Twitter Profile Photo

Yi Tay Yu Zhao Yeah! However, other (open-source) pre-training code bases we analysed were doing plain causal masking, where the likelihood of each token was conditioned on all previous tokens in the pre-training chunk. This (sub-optimal, as we found in arxiv.org/abs/2402.13991) choice probably…

@YiTayML @yuzhaouoe Yeah! However, other (open-source) pre-training code bases we analysed were doing plain causal masking, where the likelihood of each token was conditioned on all previous tokens in the pre-training chunk. This (sub-optimal, as we found in arxiv.org/abs/2402.13991) choice probably…
account_circle
Aryo Pradipta Gema(@aryopg) 's Twitter Profile Photo

Fine-tune your LLM unless you have access to GPT-4!
In our SemEval 2024 NLI4CT solution, we evaluated LLMs with ICL, CoT, and a novel PEFT method. And yet, GPT-4 produces surprisingly good results, ranking joint-first in the leaderboard!
👉 arxiv.org/abs/2404.00484
🧵 1/9

Fine-tune your LLM unless you have access to GPT-4! In our SemEval 2024 @NLI4CT solution, we evaluated LLMs with ICL, CoT, and a novel PEFT method. And yet, GPT-4 produces surprisingly good results, ranking joint-first in the leaderboard! 👉 arxiv.org/abs/2404.00484 🧵 1/9
account_circle
Pasquale Minervini 🚀 looking for postdocs!(@PMinervini) 's Twitter Profile Photo

Llama 3 was trained using intra-document causal masking, as suggested by Yu Zhao's paper 'Analysing The Impact of Sequence Composition on Language Model Pre-Training'! 🚀🚀🚀 arxiv.org/abs/2402.13991

Llama 3 was trained using intra-document causal masking, as suggested by @yuzhaouoe's paper 'Analysing The Impact of Sequence Composition on Language Model Pre-Training'! 🚀🚀🚀 arxiv.org/abs/2402.13991
account_circle
Tom Hosking(@tomhosking) 's Twitter Profile Photo

HRQ-VAE is now available in Pythae!

Hierarchical Residual Quantization learns a recursive clustering of a dense vector space, end-to-end. And it learns more meaningful clusters than doing quantization post-hoc.

📝 Paper: aclanthology.org/2022.acl-long.…
🤖 Code: github.com/clementchadebe…

account_circle
Pasquale Minervini 🚀 looking for postdocs!(@PMinervini) 's Twitter Profile Photo

If you're working on LLM pre-training, you may want to take these findings into account! E.g., intra-document causal masking is a one-line drop-in replacement of 'classic' causal masking; it's fully supported by e.g., Flash Attention (github.com/Dao-AILab/flas…); and provides…

If you're working on LLM pre-training, you may want to take these findings into account! E.g., intra-document causal masking is a one-line drop-in replacement of 'classic' causal masking; it's fully supported by e.g., Flash Attention (github.com/Dao-AILab/flas…); and provides…
account_circle
Rohit Saxena(@rohit_saxena) 's Twitter Profile Photo

🚨New paper🚨
Excited to share that 'Select and Summarize: Scene Saliency for Movie Script Summarization' has been accepted at the Findings. 🎬
w. Frank Keller
arxiv.org/abs/2404.03561…

account_circle
Pasquale Minervini 🚀 looking for postdocs!(@PMinervini) 's Twitter Profile Photo

Excited to share our latest work on improving LLM pre-training! 🚀 The amazing Yu Zhao et al. found that focusing on how pre-training sequences are composed and attended over can significantly improve the generalisation properties of LLMs on a wide array of downstream tasks,…

Excited to share our latest work on improving LLM pre-training! 🚀 The amazing @yuzhaouoe et al. found that focusing on how pre-training sequences are composed and attended over can significantly improve the generalisation properties of LLMs on a wide array of downstream tasks,…
account_circle
fly51fly(@fly51fly) 's Twitter Profile Photo

[CL] Learning to Plan and Generate Text with Citations
arxiv.org/abs/2404.03381
- Long-form generation in information-seeking scenarios has seen increasing demand for verifiable systems that generate responses with supporting evidence like citations. Most approaches rely on…

[CL] Learning to Plan and Generate Text with Citations arxiv.org/abs/2404.03381 - Long-form generation in information-seeking scenarios has seen increasing demand for verifiable systems that generate responses with supporting evidence like citations. Most approaches rely on…
account_circle
Dongwei Jiang(@Dongwei__Jiang) 's Twitter Profile Photo

📢 New paper at NAACL:

🤔 > LLMs still struggle with complex reasoning problems. Can we use reasoning nuggets from mathematical proofs to help with it?

Introducing LeanReasoner, a framework that uses Lean to solve complex natural language logical reasoning problems 🧵

📢 New paper at NAACL: 🤔 > LLMs still struggle with complex reasoning problems. Can we use reasoning nuggets from mathematical proofs to help with it? Introducing LeanReasoner, a framework that uses Lean to solve complex natural language logical reasoning problems 🧵
account_circle
Antonio Valerio Miceli Barone(@AVMiceliBarone) 's Twitter Profile Photo

New paper:

Scaling Behavior of Machine Translation with Large Language Models under Prompt Injection Attacks
arxiv.org/abs/2403.09832

account_circle
Edoardo Ponti(@PontiEdoardo) 's Twitter Profile Photo

If you are curious to discover more about Dynamic Memory Compression, I will give a preview during my keynote talk at the MOOMIN workshop eaclmeeting

See you on Thursday, March 21st at 9:30 AM!

moomin-workshop.github.io/program

account_circle
Irina Saparina(@irisaparina) 's Twitter Profile Photo

Next week I’ll be in Malta 🇲🇹 to present our work on Improving Generalization in Semantic Parsing by Increasing Natural Language Variation at !

1/3

Next week I’ll be in Malta 🇲🇹 to present our work on Improving Generalization in Semantic Parsing by Increasing Natural Language Variation at #EACL2024! 1/3
account_circle
EdinburghNLP(@EdinburghNLP) 's Twitter Profile Photo

Dynamic Memory Compression (DMC) retrofits LLMs like Llama 2 7/13/70B into models with the same downstream performance but with +370% throughput (tok/s) during generation.

account_circle
Simone Scardapane(@s_scardapane) 's Twitter Profile Photo

*Conditional computation in NNs: principles and research trends*
by Alessio Devoto Valerio Marsocci Jary Pomponi Pasquale Minervini 🚀 looking for postdocs!

Our latest tutorial on increasing modularity in NNs with conditional computation, covering MoEs, token selection, & early exits.

arxiv.org/abs/2403.07965

*Conditional computation in NNs: principles and research trends* by @devoto_alessio @valeriomarsocci @JaryPom @PMinervini Our latest tutorial on increasing modularity in NNs with conditional computation, covering MoEs, token selection, & early exits. arxiv.org/abs/2403.07965
account_circle
Tom Sherborne(@tomsherborne) 's Twitter Profile Photo

CR for TRAM is now live! See you at in Vienna (as a spotlight poster)

now feat.
* Vision exps (Better Imagenet→CIFAR/Cars/Flowers transfer)
* +ablations (XL model, weird combos)
* Pictures (see below!)
w/ Naomi Saphra Pradeep Dasigi Hao Peng

openreview.net/forum?id=kxebD…

CR for TRAM is now live! See you at #ICLR2024 in Vienna (as a spotlight poster) now feat. * Vision exps (Better Imagenet→CIFAR/Cars/Flowers transfer) * +ablations (XL model, weird combos) * Pictures (see below!) w/ @nsaphra @pdasigi @haopeng_nlp openreview.net/forum?id=kxebD…
account_circle
Edoardo Ponti(@PontiEdoardo) 's Twitter Profile Photo

Can open-source LLMs execute *chains of instructions* in a single query? Not so well, we found.

However, they can learn this ability by:
- augmenting examples from public SFT mixtures with chains of instructions automatically
- performing *sequential instruction tuning* on them.…

Can open-source LLMs execute *chains of instructions* in a single query? Not so well, we found. However, they can learn this ability by: - augmenting examples from public SFT mixtures with chains of instructions automatically - performing *sequential instruction tuning* on them.…
account_circle