Anthony Chen(@_anthonychen) 's Twitter Profile Photo

Do you know there's more than one Michael Jordan?🤨 Of course you do. Do retrievers?🤔 Our paper evaluates the entity disambiguation capabilities of retrievers & how popularity bias affects retrieval

Idea: Construct sets of queries for all entities which share a name

Do you know there's more than one Michael Jordan?🤨 Of course you do. Do retrievers?🤔 Our #ACL2021NLP paper evaluates the entity disambiguation capabilities of retrievers & how popularity bias affects retrieval

Idea: Construct sets of queries for all entities which share a name
account_circle
Ori Ram(@ori__ram) 's Twitter Profile Photo

Ever wanted to train a QA model with only 100 examples?
Check out Splinter, our new extractive QA model, trained with a novel self-supervised pretraining task, which excels at the few-shot setting 🔥

at

Paper: arxiv.org/abs/2101.00438

1/N

Ever wanted to train a QA model with only 100 examples?
Check out Splinter, our new extractive QA model, trained with a novel self-supervised pretraining task, which excels at the few-shot setting 🔥

at #ACL2021NLP

Paper: arxiv.org/abs/2101.00438

1/N
account_circle
Belinda Li(@belindazli) 's Twitter Profile Photo

Do neural language models (trained on text alone!) construct representations of meaning? In a new paper, we find that LM representations implicitly model *entities and situations* as they evolve through a discourse. 1/
arxiv.org/abs/2106.00737

Do neural language models (trained on text alone!) construct representations of meaning? In a new #ACL2021NLP paper, we find that LM representations implicitly model *entities and situations* as they evolve through a discourse. 1/
arxiv.org/abs/2106.00737
account_circle
Ruiqi Zhong(@ZhongRuiqi) 's Twitter Profile Photo

On which datapoints are smaller models better? 🤔Surprisingly, even pinpointing them is challenging and requires new statistical tools!😯 Check out our paper:

Are Larger Pretrained Language Models Uniformly Better? Comparing Performance at the Instance Level

On which datapoints are smaller models better? 🤔Surprisingly, even pinpointing them is challenging and requires new statistical tools!😯 Check out our paper: 

Are Larger Pretrained Language Models Uniformly Better? Comparing Performance at the Instance Level #NLProc #ACL2021NLP
account_circle
Jun Yan(@jun_yannn) 's Twitter Profile Photo

Excited to introduce our paper “Learning Contextualized Knowledge Structures for Commonsense Reasoning”, which will appear in Findings of !

Paper: arxiv.org/abs/2010.12873
Code: github.com/INK-USC/HGN

🧵[1/n]

Excited to introduce our paper “Learning Contextualized Knowledge Structures for Commonsense Reasoning”, which will appear in Findings of #ACL2021NLP! 

Paper: arxiv.org/abs/2010.12873 
Code: github.com/INK-USC/HGN

🧵[1/n]
account_circle
Qinyuan Ye(@qinyuan_ye) 's Twitter Profile Photo

We’ll be presenting “Learning to Generate Task-Specific Adapters from Task Description” at Poster 2G (5-7pm PT). Please come and chat with us! 😊

TL;DR: We train a hypernetwork that takes in a task description and generates task-specific adapter parameters.

We’ll be presenting “Learning to Generate Task-Specific Adapters from Task Description” at Poster 2G (5-7pm PT). Please come and chat with us! 😊 #acl2021nlp

TL;DR: We train a hypernetwork that takes in a task description and generates task-specific adapter parameters.
account_circle
Jun Yan(@jun_yannn) 's Twitter Profile Photo

Happy to introduce my internship work at Amazon Science: “AdaTag: Multi-Attribute Value Extraction from Product Profiles with Adaptive Decoding”, which got accepted to !

Paper: arxiv.org/abs/2106.02318

🧵[1/n]

Happy to introduce my internship work at @AmazonScience: “AdaTag: Multi-Attribute Value Extraction from Product Profiles with Adaptive Decoding”, which got accepted to #ACL2021NLP!

Paper: arxiv.org/abs/2106.02318

🧵[1/n]
account_circle
Tiago Pimentel(@tpimentelms) 's Twitter Profile Photo

Have you ever wanted to get word token distributions, but had to throw away oov items because simply counting frequencies can’t handle them? This paper is for you!

Just in time for the party, here is a summary of our findings on Modeling the Unigram Distribution :)

Have you ever wanted to get word token distributions, but had to throw away oov items because simply counting frequencies can’t handle them? This paper is for you!

Just in time for the party, here is a summary of our #ACL2021NLP findings on Modeling the Unigram Distribution :)
account_circle
Rada Mihalcea(@radamihalcea) 's Twitter Profile Photo

We often ask: what is the difference between Computational Linguistics and Natural Language Processing?

A wonderful perspective from Prof. Junichi Tsuji, Lifetime Achievement Award winner, with the message:

It is time for us to reconnect NLP and CL!

We often ask: what is the difference between Computational Linguistics and Natural Language Processing? 

A wonderful perspective from Prof. Junichi Tsuji,  #ACL2021NLP Lifetime Achievement Award winner, with the message: 

It is time for us to reconnect NLP and CL!
account_circle
Valentin Hofmann(@vjhofmann) 's Twitter Profile Photo

Word meaning varies across linguistic and extralinguistic contexts, but so far there has been no attempt in to model both types of variation jointly. Our paper fills this gap by introducing dynamic contextualized word embeddings. arxiv.org/pdf/2010.12684… [1/4]

Word meaning varies across linguistic and extralinguistic contexts, but so far there has been no attempt in #NLProc to model both types of variation jointly. Our #ACL2021NLP paper fills this gap by introducing dynamic contextualized word embeddings. arxiv.org/pdf/2010.12684… [1/4]
account_circle
Bill Yuchen Lin 🤖(@billyuchenlin) 's Twitter Profile Photo

Wanna have some fun with your NLU models? Try out RiddleSense, our new QA dataset in . It consists of multi-choice questions featuring both linguistic creativity and common-sense knowledge.
Project website: inklab.usc.edu/RiddleSense/ Paper: arxiv.org/abs/2101.00376 [1/5]

Wanna have some fun with your NLU models? Try out RiddleSense, our new QA dataset in #ACL2021NLP. It consists of multi-choice questions featuring both linguistic creativity and common-sense knowledge. 
Project website: inklab.usc.edu/RiddleSense/ Paper: arxiv.org/abs/2101.00376 [1/5]
account_circle
Valentin Hofmann(@vjhofmann) 's Twitter Profile Photo

BERT's tokenizations are sometimes morphologically superb-iza-rre, but does that impact BERT's semantic representations of complex words? Our upcoming paper (w/ Janet Pierrehumbert & hinrich schuetze) takes a look at this question. arxiv.org/pdf/2101.00403… /1

BERT's tokenizations are sometimes morphologically superb-iza-rre, but does that impact BERT's semantic representations of complex words? Our upcoming #ACL2021NLP paper (w/ Janet Pierrehumbert & @HinrichSchuetze) takes a look at this question. arxiv.org/pdf/2101.00403… /1
account_circle
Kayo Yin(@kayo_yin) 's Twitter Profile Photo

Thank you everyone for attending my talk on 'Including Signed Languages' at the Best Paper Session! I am so grateful for all the enthusiasm 😊

We will upload ASL interpretations soon! In the meantime, here's my pre-recording with EN subtitles: youtu.be/AYEIcOsUyWs

Thank you everyone for attending my talk on 'Including Signed Languages' at the #ACL2021NLP Best Paper Session! I am so grateful for all the enthusiasm 😊

We will upload ASL interpretations soon! In the meantime, here's my pre-recording with EN subtitles: youtu.be/AYEIcOsUyWs
account_circle
Lei Li(@lileics) 's Twitter Profile Photo

Honored and flattered to receive the best paper award of . Thank to wonderful co-authors Jingjing, Hao, Chun and Zaixiang, to the extraordinary MLNLC team at Bytedance AI Lab, and zhenyuan, weiying, hang, and ACL committee for support!
to Olympic sprinter

Honored and flattered to receive the best paper award of #ACL2021NLP. Thank to wonderful co-authors Jingjing, Hao, Chun and Zaixiang, to the extraordinary MLNLC team at Bytedance AI Lab, and zhenyuan, weiying, hang, and ACL committee for support! 
to Olympic sprinter #SuBingtian
account_circle
Kawin Ethayarajh(@ethayarajh) 's Twitter Profile Photo

Is there a connection between Shapley Values and attention-based explanations in NLP?

Yes! Our paper proves that **attention flows** can be Shapley Value explanations, but regular attention and leave-one-out cannot.

arxiv.org/abs/2105.14652

w/ Dan Jurafsky Stanford NLP Group

Is there a connection between Shapley Values and attention-based explanations in NLP?

Yes! Our #ACL2021NLP paper proves that **attention flows** can be Shapley Value explanations, but regular attention and leave-one-out cannot.

arxiv.org/abs/2105.14652

w/ @jurafsky @stanfordnlp
account_circle
Bill Yuchen Lin 🤖(@billyuchenlin) 's Twitter Profile Photo

Happy to announce that 2 of our papers on commonsense reasoning got accepted to (1main+1findings). 🎉One for analyzing and improving multilingual LMs on CSR; the other presents RiddleSense, a QA dataset for answering ! Code, data & papers coming soon! USC NLP

Happy to announce that 2 of our papers on commonsense reasoning got accepted to #ACL2021NLP (1main+1findings). 🎉One for analyzing and improving multilingual LMs on CSR; the other presents RiddleSense, a QA dataset for answering #riddles! Code, data & papers coming soon! @nlp_usc
account_circle
Bill Yuchen Lin 🤖(@billyuchenlin) 's Twitter Profile Photo

Introducing our work on evaluating and improving multi-lingual language models (ML-LMs) for commonsense reasoning (CSR). We present resources on probing & benchmarking, and a method for improving ML-LMs. [1/6]
Paper, code, data, etc.: inklab.usc.edu/XCSR/

Introducing our #acl2021nlp work on evaluating and improving multi-lingual language models (ML-LMs) for commonsense reasoning (CSR). We present resources  on probing & benchmarking, and a method for improving ML-LMs. [1/6] 
Paper, code, data, etc.: inklab.usc.edu/XCSR/
account_circle
Kaiser Sun(@KaiserWhoLearns) 's Twitter Profile Photo

Our paper “Effective Attention Sheds Light On Interpretability”(w/ Ana Marasović) was accepted into Findings of ACL2021

Pre-print available at: arxiv.org/abs/2105.08855
Thread⬇️

Our paper “Effective Attention Sheds Light On Interpretability”(w/ @anmarasovic) was accepted into Findings of ACL2021  #ACL2021NLP #NLProc

Pre-print available at: arxiv.org/abs/2105.08855
Thread⬇️
account_circle
Tatsuki Kuribayashi(@ttk_kuribayashi) 's Twitter Profile Photo

Our paper 'Lower Perplexity is Not Always Human-Like' is now at arxiv.org/abs/2106.01229

The perplexity of language models is getting lower. Does this mean that they better simulate human incremental language processing? The answer is in our title.
(Tohoku NLP Group (migrated to @tohoku_nlp))

Our #ACL2021NLP paper 'Lower Perplexity is Not Always Human-Like' is now at arxiv.org/abs/2106.01229

The perplexity of language models is getting lower. Does this mean that they better simulate human incremental language processing? The answer is in our title. 
(@NlpTohoku)
account_circle