Kyle Mahowald(@kmahowald) 's Twitter Profileg
Kyle Mahowald

@kmahowald

UT Austin linguist https://t.co/1GaRxR8rOu. cognition, psycholinguistics, data, NLP, crosswords. He/him.

ID:22515678

linkhttp://mahowak.github.io calendar_today02-03-2009 18:28:50

511 Tweets

1,6K Followers

722 Following

Jessy Li(@jessyjli) 's Twitter Profile Photo

A 🔥🔥🔥 event at UT Austin crossing AI/NLP, linguistics, philosophy, and media! This is virtual, free registration with link below 👇

account_circle
Eunsol Choi(@eunsolc) 's Twitter Profile Photo

Can LLM comprehensively capture information spread across multiple documents?
Can LLM distinguish confusing entity mentions?

Please check out our preprint on multi-document reasoning for LLM, focusing on entity disambiguation!

account_circle
Jessy Li(@jessyjli) 's Twitter Profile Photo

Congrats to Venkat for successfully defending his thesis today! David Beaver and I are so proud of his excellent work.
Venkat will join Ithaca College as a faculty member this fall 🎉
Huge thanks to friends, collaborators & committee members Kyle Mahowald and Malihe Alikhani

Congrats to @_venkatasg for successfully defending his thesis today! @David_Beaver and I are so proud of his excellent work. Venkat will join @IthacaCollege as a faculty member this fall 🎉 Huge thanks to friends, collaborators & committee members @kmahowald and @malihealikhani
account_circle
Tiago Pimentel(@tpimentelms) 's Twitter Profile Photo

Have you ever wondered about what's the impact of having so many near duplicate subwords in your model's vocabulary (e.g., _book vs book vs books vs Book 📚) on its performance?! If yes, checkout Anton Schäfer's new paper on it! :)

account_circle
Ted Gibson, Language Lab MIT(@LanguageMIT) 's Twitter Profile Photo

We are gruntled editors of Open Mind, a completely free open access cognitive science journal published by MIT Press: send us your papers!!

account_circle
Ethan Gotlieb Wilcox(@weGotlieb) 's Twitter Profile Photo

BabyLM is back for round two! This time we have: 1. Multimodal track 2. Bring your own data and 3. Paper track! See the CfP for all the juicy details!

account_circle
Kyle Mahowald(@kmahowald) 's Twitter Profile Photo

Super exciting to see resources like this being built for studying constructions more systematically in computational linguistics!

account_circle
Dan Roberts(@danintheory) 's Twitter Profile Photo

Do LLMs really need to be so L?

That's a rejected title for a new paper w/ Andrey Gromov, Kushal Tirumala, Hassan Shapourian, Paolo Glorioso on pruning open-weight LLMs: we can remove up to *half* the layers of Llama-2 70B w/ essentially no impact on performance on QA benchmarks.

1/

Do LLMs really need to be so L? That's a rejected title for a new paper w/ @Andr3yGR, @kushal_tirumala, @Hasan_Shap, @PaoloGlorioso1 on pruning open-weight LLMs: we can remove up to *half* the layers of Llama-2 70B w/ essentially no impact on performance on QA benchmarks. 1/
account_circle
Kyle Mahowald(@kmahowald) 's Twitter Profile Photo

Very cool survey on mental experience. A real surprise to me that some see a new name visually written in their mind when they first hear it…even more surprised it turns out you can live for years with someone who does that and have no idea that’s what’s been going on in there!

account_circle
Kanishka Misra 😶‍🌫️(@kanishkamisra) 's Twitter Profile Photo

☀️minicons 🌖 now supports sequence scoring with Vision-Language Models!!

Looking forward to see how ppl use it (if at all!) -- feedback always welcome!

🐧🐦

☀️minicons 🌖 now supports sequence scoring with Vision-Language Models!! Looking forward to see how ppl use it (if at all!) -- feedback always welcome! 🐧🐦
account_circle
Christopher Potts(@ChrisGPotts) 's Twitter Profile Photo

I know I am late in the project cycle for this, but I do have suggested edits for the team behind nature.com/articles/d4158… My overall comment is that the central claims in the original are lacking in empirical support.

I know I am late in the project cycle for this, but I do have suggested edits for the team behind nature.com/articles/d4158… My overall comment is that the central claims in the original are lacking in empirical support.
account_circle
Trends in Cognitive Sciences(@TrendsCognSci) 's Twitter Profile Photo

Dissociating language and thought in large language models

Feature Review by Kyle Mahowald (Kyle Mahowald), Anna Ivanova (@neuranna), Idan Blank (@IbanDlank), Nancy Kanwisher (@Nancy_Kanwisher), Joshua Tenenbaum, & Evelina Fedorenko (@ev_fedorenko)

doi.org/10.1016/j.tics…

Dissociating language and thought in large language models Feature Review by Kyle Mahowald (@kmahowald), Anna Ivanova (@neuranna), Idan Blank (@IbanDlank), Nancy Kanwisher (@Nancy_Kanwisher), Joshua Tenenbaum, & Evelina Fedorenko (@ev_fedorenko) doi.org/10.1016/j.tics…
account_circle