Discussing Stupid: A byte-sized podcast on stupid UX cover art

Discussing Stupid: A byte-sized podcast on stupid UX

Discussing Stupid: A byte-sized podcast on stupid UX

By: High Monkey
Listen for free

About this listen

Discussing Stupid returns to the airwaves to transform digital facepalms into teachable moments—all in the time it takes to enjoy your coffee break! Sponsored by High Monkey, this podcast dives into ‘stupid’ practices across websites and Microsoft collaboration tools, among other digital realms. Our "byte-sized" bi-weekly episodes are packed with expert insights and a healthy dose of humor. Discussions focus on five key areas: Business Process & Collaboration, UX/IA, Inclusive Design, Content & Search, and Performance & SEO. Join us and let’s start making the digital world a bit less stupid, one episode at a time. Visit our website at https://www.discussingstupid.com© 2025 Discussing Stupid: A byte-sized podcast on stupid UX Economics Marketing Marketing & Sales
Episodes
  • S3E4 - Intentional AI: The role of AI in content management
    Nov 11 2025

    In Episode 4 of the Intentional AI series, Cole and Virgil move further into the content lifecycle and this time they are focusing on content management.

    Once your content’s written, the real work begins. Editing, organizing, translating, tagging, all the behind-the-scenes steps that keep content consistent and usable. In this episode, the team looks at how AI can help streamline those tasks and where it still creates new challenges.

    Joined by returning guest Chad, they break down where AI fits, where it fails, and what happens when you trust it to translate complex content on its own.

    In this episode, they explore:

    • How AI supports the content management stage of the lifecycle
    • Common use cases like translation, auto-summary fields, and accessibility checks
    • Where automation makes sense and where it doesn’t
    • The biggest risks of AI content management, from oversimplification to data privacy
    • Why good input (clear, readable content) still determines good output
    • How readable, accessible writing improves both human and AI understanding

    This episode also continues the real-world experiment from previous episodes.

    Using the accessibility article originally created with Writesonic, the team tests how well three AI tools: Google Translate, DeepL, and ChatGPT, handle translating the piece into Spanish. The results reveal major differences in accuracy, tone, and overall usability across each model.

    A downloadable Episode Companion Guide is available below. It includes key takeaways, tool comparisons, and practical advice for using AI in the content management stage.

    DS-S3-E4-CompanionDoc.pdf

    🦃 Note: We’re taking a short Thanksgiving break, the next episode will drop on December 2, 2025.

    Upcoming episodes in the Intentional AI series:

    • Dec 2, 2025 — Accessibility
    • Dec 16, 2025 — SEO / AEO / GEO
    • Jan 6, 2026 — Content Personalization
    • Jan 20, 2026 — Front End Development & Wireframing
    • Feb 3, 2026 — Design & Media
    • Feb 17, 2026 — Back End Development
    • Mar 3, 2026 — Conversational Search (with special guest!)
    • Mar 17, 2026 — Chatbots & Agentic AI
    • Mar 31, 2026 — Series Finale & Tool Review

    Whether you’re managing websites, content workflows, or entire digital ecosystems, this conversation is about using AI intentionally, to work smarter without losing the human judgment that keeps content trustworthy.

    New episodes every other Tuesday.

    For more conversations about AI, digital strategy, and all the ways we get it wrong (and how to get it right), visit www.discussingstupid.com and subscribe on your favorite podcast platform.

    Chapters

    (0:00) - Intro

    (0:50) - Today's focus: Content management with AI

    (1:58) - Content management opportunities with AI

    (6:18) - Recurring series theme: Trust

    (8:34) - Refine your process one step at a time

    (9:53) - Better content = better everything

    (10:22) - We tested 3 AI translation tools

    (12:02) - Cole's "elephant in the room" test

    (14:28) - Poor content = poor translations

    (16:58) - True translation happens between people

    (18:45) - Closing takeaways

    Subscribe for email updates on our website:

    https://www.discussingstupid.com/

    Watch us on YouTube:

    https://www.youtube.com/@discussingstupid

    Listen on Apple Podcasts,...

    Show More Show Less
    21 mins
  • S3E3 - Intentional AI: Smarter content creation with AI
    Oct 28 2025

    In Episode 3 of the Intentional AI series, Cole and Virgil move into the next stage of the content lifecycle: content creation.

    AI can write faster than ever, but that doesn’t mean it writes well. From prompting and editing to maintaining voice and originality, AI-generated content still requires human effort and judgment. In this episode, the team explores where AI can help streamline production and where it can’t replace the creative process.

    In this episode, they explore:

    • How AI fits into the content creation stage of the lifecycle
    • Why AI-generated content often takes just as much time as writing from scratch
    • The key risks of AI content creation, including accuracy, effort, and authenticity
    • How to maintain your voice, tone, and originality when using AI tools
    • Why humans are still responsible for quality control and credibility
    • What happens when you test the same research prompt across three writing tools


    This episode also continues the real-world experiment from Episode 2. Using the research compiled with Perplexity, the team tests how three content-generation tools—Jenni AI, Perplexity Pro, and Writesonic—handle the same writing task. The results reveal just how differently each model performs when asked to create original, publishable content.


    A downloadable Episode Companion Guide is available below. It includes key takeaways, tool comparisons, and practical advice for using AI in the content creation stage.


    DS-S3-E3-CompanionDoc.pdf


    Upcoming episodes in the Intentional AI series:

    • Nov 11, 2025 — Content Management

    • Dec 2, 2025 — Accessibility

    • Dec 16, 2025 — SEO / AEO / GEO

    • Jan 6, 2026 — Content Personalization

    • Jan 20, 2026 — Front End Development & Wireframing

    • Feb 3, 2026 — Design & Media

    • Feb 17, 2026 — Back End Development

    • Mar 3, 2026 — Conversational Search (with special guest!)

    • Mar 17, 2026 — Chatbots & Agentic AI

    • Mar 31, 2026 — Series Finale & Tool Review

    Whether you’re a marketer, strategist, or developer, this conversation is about creating content intentionally and keeping your human voice at the center of it all.


    New episodes every other Tuesday.


    For more conversations about AI, digital strategy, and all the ways we get it wrong (and how to get it right), visit www.discussingstupid.com and subscribe on your favorite podcast platform.


    Chapters

    (0:00) - Intro

    (0:30) - Smarter content creation with AI

    (1:00) - Effort doesn't go away

    (3:20) - Tool / LLM differences

    (5:34) - Audience fit & voice

    (7:44) - We tested 3 tools for AI content creation

    (10:08) - Testing Jenni AI

    (13:23) - Testing Perplexity

    (14:55) - Testing Writesonic

    (16:55) - Key Takeaways


    Subscribe for email updates on our website:

    https://www.discussingstupid.com/

    Watch us on YouTube:

    https://www.youtube.com/@discussingstupid

    Listen on Apple Podcasts, Spotify, or Soundcloud:

    https://podcasts.apple.com/us/podcast/discussing-stupid-a-byte-sized-podcast-on-stupid-ux/id1428145024

    https://open.spotify.com/show/0c47grVFmXk1cco63QioHp?si=87dbb37a4ca441c0

    Show More Show Less
    20 mins
  • S3E2 - Intentional AI: Maximizing AI for research & analysis
    Oct 14 2025

    In Episode 2 of the Intentional AI series, Cole and Virgil dive into the first real stage of the content lifecycle: research and analysis.

    From brainstorming ideas to verifying data sources, AI is being used everywhere in the early stages of content creation. But how much of that information can you actually trust? In this episode, the team unpacks where AI helps, where it hurts, and why you still need to be the researcher of the research.

    In this episode, they explore:

    • How AI fits into the research and analysis stage of the content lifecycle
    • The major risks of using AI for research, including accuracy, bias, and misinformation
    • Why trust, verification, and validation are now part of your job
    • Security and legal concerns around AI scraping and data usage
    • How different tools handle citations, transparency, and usability
    • Why you can’t skip the human role in confirming, editing, and contextualizing AI outputs

    This episode also features the first step in a real experiment: researching a blog topic on digital accessibility using the tools Perplexity, ChatGPT, and Copilot. The results of that research will directly fuel the next episode on content creation.

    A downloadable Episode Companion Guide is available below. It includes key episode takeaways, tool comparisons, and practical guidance on how to use AI responsibly during the research stage.


    DS-S3-E2-CompanionDoc.pdf


    Upcoming episodes in the Intentional AI series:

    • Oct 28, 2025 — Content Creation
    • Nov 11, 2025 — Content Management
    • Dec 2, 2025 — Accessibility
    • Dec 16, 2025 — SEO / AEO / GEO
    • Jan 6, 2026 — Content Personalization
    • Jan 20, 2026 — Front End Development & Wireframing
    • Feb 3, 2026 — Design & Media
    • Feb 17, 2026 — Back End Development
    • Mar 3, 2026 — Conversational Search (with special guest!)
    • Mar 17, 2026 — Chatbots & Agentic AI
    • Mar 31, 2026 — Series Finale & Tool Review


    Whether you’re a marketer, strategist, or developer, this conversation is about making AI adoption intentional and keeping your critical thinking sharp.


    New episodes every other Tuesday.


    For more conversations about AI, digital strategy, and all the ways we get it wrong (and how to get it right), visit www.discussingstupid.com and subscribe on your favorite podcast platform.


    (0:00) - Intro

    (1:44) - Better research with AI

    (3:46) - Risk: Trust & reliability

    (5:29) - Risk: Security/legal concerns

    (7:04) - Risk: Hallucinations

    (9:17) - We tested 3 tools for AI research

    (11:03) - Testing Perplexity

    (14:38) - Testing ChatGPT

    (17:45) - Testing Copilot

    (19:54) - Comparing the tools and key takeaways

    (20:52) - Outro


    Subscribe for email updates on our website:

    https://www.discussingstupid.com/

    Watch us on YouTube:

    https://www.youtube.com/@discussingstupid

    Listen on Apple Podcasts, Spotify, or Soundcloud:

    https://podcasts.apple.com/us/podcast/discussing-stupid-a-byte-sized-podcast-on-stupid-ux/id1428145024

    https://open.spotify.com/show/0c47grVFmXk1cco63QioHp?si=87dbb37a4ca441c0

    Show More Show Less
    22 mins
No reviews yet
In the spirit of reconciliation, Audible acknowledges the Traditional Custodians of country throughout Australia and their connections to land, sea and community. We pay our respect to their elders past and present and extend that respect to all Aboriginal and Torres Strait Islander peoples today.