John Nay(@johnjnay) 's Twitter Profile Photo

LLMs Can Define Reward Params to Optimize for Robots

-LLMs define params to optimize for control policies to accomplish robotic tasks
-Bridges gap from high-level language instruct & low-level robot actions

-Reliably tackles 90% of tasks, baseline 50%

arxiv.org/abs/2306.08647

LLMs Can Define Reward Params to Optimize for Robots

-LLMs define params to optimize for control policies to accomplish robotic tasks
-Bridges gap from high-level language instruct & low-level robot actions

-Reliably tackles 90% of tasks, baseline 50%

arxiv.org/abs/2306.08647
account_circle
Jaya Gupta(@JayaGup10) 's Twitter Profile Photo

For the non AI-companies that want to train their own LLM similar to Bloomberg GPT:

the infrastructure and tooling is less of a constraint - it’s really only an option if you can hire sophisticated AI talent which is a subset of some of the roles on this slide

these are also…

For the non AI-companies that want to train their own LLM similar to Bloomberg GPT: 

the infrastructure and tooling is less of a constraint - it’s really only an option if you can hire sophisticated AI talent which is a subset of some of the roles on this slide 

these are also…
account_circle
Alec Stapp(@AlecStapp) 's Twitter Profile Photo

'The ‘Safeguarding the Future’ course at MIT tasked non-scientist students with investigating whether LLM chatbots could be prompted to assist non-experts in causing a pandemic.

In 1 hour, the chatbots:

- suggested 4 potential pandemic pathogens

- explained how they can be…

'The ‘Safeguarding the Future’ course at MIT tasked non-scientist students with investigating whether LLM chatbots could be prompted to assist non-experts in causing a pandemic.

In 1 hour, the chatbots:

- suggested 4 potential pandemic pathogens

- explained how they can be…
account_circle
AI Notkilleveryoneism Memes(@AISafetyMemes) 's Twitter Profile Photo

'These results suggest that LLMs will make pandemic-class agents widely accessible as soon as they are credibly identified, even to people with little or no laboratory training.'

MIT tasked non-scientist students to investigate: can LLM chatbots be used by non-experts to cause a…

'These results suggest that LLMs will make pandemic-class agents widely accessible as soon as they are credibly identified, even to people with little or no laboratory training.'

MIT tasked non-scientist students to investigate: can LLM chatbots be used by non-experts to cause a…
account_circle
Akshay 🚀(@akshay_pachaar) 's Twitter Profile Photo

Just a year ago, it would have seemed unbelievable!

Now, you can interact with any API in the world using Plain English! 🤯

Thanks to LLMs + LangChain, the possibilities are endless! 🔥

Check it out👇

Just a year ago, it would have seemed unbelievable!

Now, you can interact with any API in the world using Plain English! 🤯

Thanks to LLMs + LangChain, the possibilities are endless! 🔥

Check it out👇
account_circle
kinocow(@kinocow) 's Twitter Profile Photo

I love that pwning dumb incels can be now outsourced to LLMs. Here's they thought it will come to their rescue and end up getting smacked in the face with dry logic instead. This is a win.

I love that pwning dumb incels can be now outsourced to LLMs. Here's they thought it will come to their rescue and end up getting smacked in the face with dry logic instead. This is a win.
account_circle
Joey Gonzalez(@profjoeyg) 's Twitter Profile Photo

Serving LLMs? My students found a way to accelerate serving by over an order-of-magnitude just by changing the way memory is managed (spoiler alert): gpu memory fragmentation = slow. Introducing vLLM with PagedAttention:

Serving LLMs? My students found a way to accelerate serving by over an order-of-magnitude just by changing the way memory is managed (spoiler alert): gpu memory fragmentation = slow. Introducing vLLM with PagedAttention:
account_circle
Yam Peleg(@Yampeleg) 's Twitter Profile Photo

Cheatsheet:

Full Training vs. LoRA Adapters

VS. VS VS.

Smaller Models vs. Larger Models

👇 All Results 👇

---

LLMs Training TL;DR:

1. larger model = better model. [1]
2. larger model = more…

Cheatsheet:

Full Training             vs.       LoRA Adapters

           VS.                   VS                 VS.

Smaller Models      vs.      Larger Models

👇  All Results  👇

---

LLMs Training TL;DR:

1. larger model = better model. [1]
2. larger model = more…
account_circle
Jaya Gupta(@JayaGup10) 's Twitter Profile Photo

This great chart from BCV shows the limited number of companies that can construct their own LLM.

What’s interesting is the number of cos configuring LLMs for use cases specific to financial services, like risk modeling, fraud detection, policy underwriting or purchase order…

This great chart from BCV shows the limited number of companies that can construct their own LLM.

What’s interesting is the number of cos configuring LLMs for use cases specific to financial services, like risk modeling, fraud detection, policy underwriting or purchase order…
account_circle
Ben Zhao(@ravenben) 's Twitter Profile Photo

Had amusing exchange with someone defending LLMs as not memorizing content, particularly copyrighted content. So I jumped on chatGPT and tried a little experiment. Folks can decide for themselves.
12 prompts, each producing exact copy of a text snippet verbatim. There’s more…

Had amusing exchange with someone defending LLMs as not memorizing content, particularly copyrighted content. So I jumped on chatGPT and tried a little experiment. Folks can decide for themselves.
12 prompts, each producing exact copy of a text snippet verbatim. There’s more…
account_circle
Yash(@itsYash777) 's Twitter Profile Photo

In a study by OpenAI, it was found that for every 100 responses generated by an LLM, 40 of them were likely to be hallucinations.

Let's understand hallucination in LLMs 🧵

In a study by OpenAI, it was found that for every 100 responses generated by an LLM, 40 of them were likely to be hallucinations.

Let's understand hallucination in LLMs 🧵
account_circle
Cameron R. Wolfe, Ph.D.(@cwolferesearch) 's Twitter Profile Photo

Can large language models (LLMs) train themselves? The explosion of imitation-based open-source LLMs drew criticism due to cursory evaluation that covered up performance gaps. However, recent research shows powerful open-source LLMs can actually be created by imitating other…

Can large language models (LLMs) train themselves?  The explosion of imitation-based open-source LLMs drew criticism due to cursory evaluation that covered up performance gaps. However, recent research shows powerful open-source LLMs can actually be created by imitating other…
account_circle
Sam Ching(@samcwl) 's Twitter Profile Photo

Day 1 of the first ever multi-day legal LLM hackathon 🔥

Excited to see folks:

- finetune open-source LLMs on legal text
- improve on LegalBench
- Legal red-teaming on agent simulations
- build legal apps!

Day 1 of the first ever multi-day legal LLM hackathon 🔥

Excited to see folks:

- finetune open-source LLMs on legal text
- improve on LegalBench
- Legal red-teaming on agent simulations
- build legal apps!
account_circle
kenshinsamurai9000(@kenshinsamurai9) 's Twitter Profile Photo

𝐆𝐏𝐓𝟒 𝐜𝐚𝐧 𝐦𝐮𝐥𝐭𝐢𝐩𝐥𝐲 𝐚𝐫𝐛𝐢𝐭𝐫𝐚𝐫𝐲 𝐥𝐞𝐧𝐠𝐭𝐡 𝐢𝐧𝐭𝐞𝐠𝐞𝐫𝐬. 𝐓𝐡𝐢𝐬 𝐦𝐞𝐚𝐧𝐬 𝐫𝐞𝐚𝐬𝐨𝐧𝐢𝐧𝐠 𝐬𝐭𝐚𝐭𝐞𝐬 𝐚𝐫𝐞 𝐨𝐧𝐥𝐲 𝐥𝐢𝐦𝐢𝐭𝐞𝐝 𝐛𝐲 𝐜𝐨𝐧𝐭𝐞𝐱𝐭 𝐥𝐞𝐧𝐠𝐭𝐡.

One of the most intriguing aspects of GPT-based LLMs, is that while they are…

𝐆𝐏𝐓𝟒 𝐜𝐚𝐧 𝐦𝐮𝐥𝐭𝐢𝐩𝐥𝐲 𝐚𝐫𝐛𝐢𝐭𝐫𝐚𝐫𝐲 𝐥𝐞𝐧𝐠𝐭𝐡 𝐢𝐧𝐭𝐞𝐠𝐞𝐫𝐬. 𝐓𝐡𝐢𝐬 𝐦𝐞𝐚𝐧𝐬 𝐫𝐞𝐚𝐬𝐨𝐧𝐢𝐧𝐠 𝐬𝐭𝐚𝐭𝐞𝐬 𝐚𝐫𝐞 𝐨𝐧𝐥𝐲 𝐥𝐢𝐦𝐢𝐭𝐞𝐝 𝐛𝐲 𝐜𝐨𝐧𝐭𝐞𝐱𝐭 𝐥𝐞𝐧𝐠𝐭𝐡.

One of the most intriguing aspects of GPT-based  LLMs, is that while they are…
account_circle
Ben Remington(@RemingtonBen) 's Twitter Profile Photo

Yesterday I submitted the first completed version of my thesis on streamwise vorticity currents and their effect on LLMs and tornadoes. Today I felt like I was rewarded for those efforts.

photography

Yesterday I submitted the first completed version of my thesis on streamwise vorticity currents and their effect on LLMs and tornadoes. Today I felt like I was rewarded for those efforts. 

#ndwx #mnwx #nikon #nikonphotography
account_circle
Zhuohan Li(@zhuohan123) 's Twitter Profile Photo

🌟 Thrilled to introduce vLLM with Woosuk Kwon!

🚀 vLLM is an open-source LLM inference and serving library that accelerates HuggingFace Transformers by 24x and powers lmsys.org Vicuna and Chatbot Arena.

Github: github.com/vllm-project/v…
Blog: vllm.ai

account_circle
Sanyam Bhutani(@bhutanisanyam1) 's Twitter Profile Photo

Combining Knowledge Graphs and LLMs! 🙏

To my surprise, this paper is extremely detailed around training strategies of building such models and goes beyond “just prompting” ChatGPT and GPT-4.

In fact, I realised after reading-it doesn’t even mention these models in most of the…

Combining Knowledge Graphs and LLMs! 🙏

To my surprise, this paper is extremely detailed around training strategies of building such models and goes beyond “just prompting” ChatGPT and GPT-4. 

In fact, I realised after reading-it doesn’t even mention these models in most of the…
account_circle
Yuchen Zhuang(@yuchen_zhuang) 's Twitter Profile Photo

🔧Thrilled to introduce , a new dataset to evaluate the capabilities of in answering challenging questions with external tools. It offers two levels (easy/hard) across eight real-life scenarios. 🚀

More details below:
🧵(1/5)

🔧Thrilled to introduce #ToolQA, a new dataset to evaluate the capabilities of #LLMs in answering challenging questions with external tools. It offers two levels (easy/hard) across eight real-life scenarios. 🚀

More details below:
🧵(1/5)
account_circle