undefined cover
undefined cover
Taming Erratic Behavior in AI Agents cover
Taming Erratic Behavior in AI Agents cover
The Prompt Desk

Taming Erratic Behavior in AI Agents

Taming Erratic Behavior in AI Agents

24min |05/06/2024
Play
undefined cover
undefined cover
Taming Erratic Behavior in AI Agents cover
Taming Erratic Behavior in AI Agents cover
The Prompt Desk

Taming Erratic Behavior in AI Agents

Taming Erratic Behavior in AI Agents

24min |05/06/2024
Play

Description

As AI agents powered by large language models become more complex, developers often encounter erratic and unexpected behaviors during testing. From agents falling into infinite loops to models struggling with certain data formats, these issues can be tricky to diagnose and resolve. In this episode, Bradley Arsenault and Justin Macorin explore real-world examples of AI agents going off the rails. They discuss practical techniques like action governors, confusion matrix analysis, minimum task requirements, and targeted fine-tuning to create more robust and reliable agents. Tune in for valuable insights on taming unruly AI from two experienced practitioners at the forefront of prompt engineering and AI product development.


Continue listening to The Prompt Desk Podcast for everything LLM & GPT, Prompt Engineering, Generative AI, and LLM Security.
Check out PromptDesk.ai for an open-source prompt management tool.
Check out Brad’s AI Consultancy at bradleyarsenault.me
Add Justin Macorin and Bradley Arsenault on LinkedIn.
Please fill out our listener survey here to help us create a better podcast: https://docs.google.com/forms/d/e/1FAIpQLSfNjWlWyg8zROYmGX745a56AtagX_7cS16jyhjV2u_ebgc-tw/viewform?usp=sf_link


Hosted by Ausha. See ausha.co/privacy-policy for more information.

Description

As AI agents powered by large language models become more complex, developers often encounter erratic and unexpected behaviors during testing. From agents falling into infinite loops to models struggling with certain data formats, these issues can be tricky to diagnose and resolve. In this episode, Bradley Arsenault and Justin Macorin explore real-world examples of AI agents going off the rails. They discuss practical techniques like action governors, confusion matrix analysis, minimum task requirements, and targeted fine-tuning to create more robust and reliable agents. Tune in for valuable insights on taming unruly AI from two experienced practitioners at the forefront of prompt engineering and AI product development.


Continue listening to The Prompt Desk Podcast for everything LLM & GPT, Prompt Engineering, Generative AI, and LLM Security.
Check out PromptDesk.ai for an open-source prompt management tool.
Check out Brad’s AI Consultancy at bradleyarsenault.me
Add Justin Macorin and Bradley Arsenault on LinkedIn.
Please fill out our listener survey here to help us create a better podcast: https://docs.google.com/forms/d/e/1FAIpQLSfNjWlWyg8zROYmGX745a56AtagX_7cS16jyhjV2u_ebgc-tw/viewform?usp=sf_link


Hosted by Ausha. See ausha.co/privacy-policy for more information.

Share

Embed

You may also like

Description

As AI agents powered by large language models become more complex, developers often encounter erratic and unexpected behaviors during testing. From agents falling into infinite loops to models struggling with certain data formats, these issues can be tricky to diagnose and resolve. In this episode, Bradley Arsenault and Justin Macorin explore real-world examples of AI agents going off the rails. They discuss practical techniques like action governors, confusion matrix analysis, minimum task requirements, and targeted fine-tuning to create more robust and reliable agents. Tune in for valuable insights on taming unruly AI from two experienced practitioners at the forefront of prompt engineering and AI product development.


Continue listening to The Prompt Desk Podcast for everything LLM & GPT, Prompt Engineering, Generative AI, and LLM Security.
Check out PromptDesk.ai for an open-source prompt management tool.
Check out Brad’s AI Consultancy at bradleyarsenault.me
Add Justin Macorin and Bradley Arsenault on LinkedIn.
Please fill out our listener survey here to help us create a better podcast: https://docs.google.com/forms/d/e/1FAIpQLSfNjWlWyg8zROYmGX745a56AtagX_7cS16jyhjV2u_ebgc-tw/viewform?usp=sf_link


Hosted by Ausha. See ausha.co/privacy-policy for more information.

Description

As AI agents powered by large language models become more complex, developers often encounter erratic and unexpected behaviors during testing. From agents falling into infinite loops to models struggling with certain data formats, these issues can be tricky to diagnose and resolve. In this episode, Bradley Arsenault and Justin Macorin explore real-world examples of AI agents going off the rails. They discuss practical techniques like action governors, confusion matrix analysis, minimum task requirements, and targeted fine-tuning to create more robust and reliable agents. Tune in for valuable insights on taming unruly AI from two experienced practitioners at the forefront of prompt engineering and AI product development.


Continue listening to The Prompt Desk Podcast for everything LLM & GPT, Prompt Engineering, Generative AI, and LLM Security.
Check out PromptDesk.ai for an open-source prompt management tool.
Check out Brad’s AI Consultancy at bradleyarsenault.me
Add Justin Macorin and Bradley Arsenault on LinkedIn.
Please fill out our listener survey here to help us create a better podcast: https://docs.google.com/forms/d/e/1FAIpQLSfNjWlWyg8zROYmGX745a56AtagX_7cS16jyhjV2u_ebgc-tw/viewform?usp=sf_link


Hosted by Ausha. See ausha.co/privacy-policy for more information.

Share

Embed

You may also like