Hosein Mohebbi(@hmohebbi75) 's Twitter Profileg
Hosein Mohebbi

@hmohebbi75

PhD candidate @TilburgU, doing research on interpretability for text and speech. #NLProc

ID:1271141134857895936

linkhttps://hmohebbi.github.io calendar_today11-06-2020 18:06:16

87 Tweets

252 Followers

341 Following

Javier Ferrando(@javifer_96) 's Twitter Profile Photo

[1/4] Introducing “A Primer on the Inner Workings of Transformer-based Language Models”, a comprehensive survey on interpretability methods and the findings into the functioning of language models they have led to.

ArXiv: arxiv.org/pdf/2405.00208

[1/4] Introducing “A Primer on the Inner Workings of Transformer-based Language Models”, a comprehensive survey on interpretability methods and the findings into the functioning of language models they have led to. ArXiv: arxiv.org/pdf/2405.00208
account_circle
Parisa Kordjamshidi(@Kordjamshidi) 's Twitter Profile Photo

Hiring a Post-doctoral researcher in my lab, please consider applying if you see the position fits your expertise and interest. The research is at the intersection of NLP and CV leveraging neuro-symbolic modeling in V&L navigation; , , careers.msu.edu/en-us/job/5182… .

account_circle
Hosein Mohebbi(@hmohebbi75) 's Twitter Profile Photo

On April 20th, I’ll run for Spierfonds to raise money for research on muscular diseases. Your donation here would greatly please me :)
spieractie.nl/fundraisers/ho…

account_circle
Michael Hanna(@michaelwhanna) 's Twitter Profile Photo

Circuits are a hot topic in interpretability, but how do you find a circuit and guarantee it reflects how your model works?

We (Sandro Pezzelle, Yonatan Belinkov, and I) introduce a new circuit-finding method, EAP-IG, and show it finds more faithful circuits arxiv.org/abs/2403.17806 1/8

account_circle
Hosein Mohebbi(@hmohebbi75) 's Twitter Profile Photo

Still looking at self-attention weights to see how (token) representations are being mixed in a Transformer layer? Try more informative alternatives by simply enabling ‘output_context_mixings’ in your HuggingFace model call:
Notebook: colab.research.google.com/drive/114Yigbe…

account_circle
Hosein Mohebbi(@hmohebbi75) 's Twitter Profile Photo

The moment for “Transformer-specific Interpretability” tutorial is very approaching!
Check out our detailed program & an extensive suggested reading list here: projects.illc.uva.nl/indeep/tutoria…

🗓️ Thursday, 21 March
⏱️ 14:00-17:30 pm
📍 Radisson room

account_circle
Sara Rajaee(@rajaee_sara) 's Twitter Profile Photo

Today I’ll present my joint work with Christof Monz on “Analyzing the Evaluation of Cross-lingual Knowledge Transfer in Multilingual Language Models” at
Happy to see everyone there!
📍 Marie Louise 1, Multilingual session
🕟 4 p.m.
📖 arxiv.org/abs/2402.02099

account_circle
Gaofei Shen(@linguisticshen) 's Twitter Profile Photo

Sharing some great news today. My paper titled 'Encoding of lexical tone in self-supervised models of spoken language' has been accepted to the main NAACL 2024 conference!
Many thanks to my co-authors Michaela Watkins afra alishahi Arianna Bisazza Grzegorz Chrupała 🇪🇺🇺🇦

A brief 🧵below:

account_circle
Lisa Beinborn(@lisabeinborn) 's Twitter Profile Photo

We work on human-centered natural language processing and I am hiring a PhD candidate and a PostDoc. Please let me know if you are interested or share with someone who might be.

PostDoc description: uni-goettingen.de/en/644546.html…
PhD description: uni-goettingen.de/en/644546.html…

account_circle
Hosein Mohebbi(@hmohebbi75) 's Twitter Profile Photo

Over the coming two weeks, I’ll have a great opportunity to share my research through a series of talks at:
- CSTR, Edinburgh (Feb 26)
- CardiffNLP (Feb 29)
- TeIAS, Tehran (Mar 6)
- InDeep workshop, Amsterdam (Mar 8)
Excited to meet, discuss & learn from everyone there!

account_circle
Hosein Mohebbi(@hmohebbi75) 's Twitter Profile Photo

Arthur’s Seat view from Informatics’ terrace serves as a refresh button during my break times!
Very enjoying my visit w/ Ivan Titov and his team at EdinburghNLP!

Arthur’s Seat view from Informatics’ terrace serves as a refresh button during my break times! Very enjoying my visit w/ @iatitov and his team at @EdinburghNLP!
account_circle
Goro Kobayashi(@goro_koba) 's Twitter Profile Photo

Happy to share our paper on LM’s Feed-forward network (FF) analysis has been accepted as an spotlight!
💡FF boosts attention between words forming compound nouns, named entities, etc.
💡FF and LayerNorm cancel out each other’s effects

📄arxiv.org/abs/2302.00456

Happy to share our paper on LM’s Feed-forward network (FF) analysis has been accepted as an #ICLR2024 spotlight! 💡FF boosts attention between words forming compound nouns, named entities, etc. 💡FF and LayerNorm cancel out each other’s effects 📄arxiv.org/abs/2302.00456
account_circle
Zhijing Jin@ICLR/COLING/NAACL(@ZhijingJin) 's Twitter Profile Photo

For people interested in and Interpretability & Robustness of , I just shared my Keynote Talk at BlackboxNLP workshop: 'Causal NLP: A Path towards Opening the Black Box of NLP'
- 🎦Talk video: youtube.com/watch?v=Auls8a…
- 💻Slides: docs.google.com/presentation/d…

For people interested in #CausalNLP and Interpretability & Robustness of #LLMs, I just shared my Keynote Talk at #EMNLP2023 @BlackboxNLP workshop: 'Causal NLP: A Path towards Opening the Black Box of NLP' - 🎦Talk video: youtube.com/watch?v=Auls8a… - 💻Slides: docs.google.com/presentation/d…
account_circle
Arianna Bisazza(@AriannaBisazza) 's Twitter Profile Photo

Even more grateful to also receive an Outstanding Paper Award at for our paper on cross-lingual consistency of factual knowledge in LMs! with Jirui Qi and Raquel Fernández.
@GroNLP AmsterdamNLP

arxiv.org/abs/2310.10378

Even more grateful to also receive an Outstanding Paper Award at #EMNLP2023 for our paper on cross-lingual consistency of factual knowledge in LMs! with @Jirui_Qi and @raquel_dmg. @GroNLP @AmsterdamNLP arxiv.org/abs/2310.10378
account_circle
AmsterdamNLP(@AmsterdamNLP) 's Twitter Profile Photo

Fantastic news! Outstanding paper award for this paper on opening up the blackbox of neural speech recognition models at !

by Hosein Mohebbi (Tilburg & InDeep) and collaborators from Tilburg and Amsterdam.

projects.illc.uva.nl/indeep/

account_circle