"A Beginner's Guide to AI" makes the complex world of Artificial Intelligence accessible to all. Each episode breaks down a new AI concept into everyday language, tying it to real-world applications and featuring insights from industry experts. Ideal for novices, tech enthusiasts, and the simply curious, this podcast transforms AI learning into an engaging, digestible journey. Join us as we take the first steps into AI! There are 3 episode formats: AI generated, interviews with AI experts & my thoughts. Want to get your AI going? Get in contact: [email protected]
Your AI might not be hacked. It might be persuaded.
In this episode of A Beginner’s Guide to AI, we unpack one of the most underestimated threats in modern business: prompt injection. As AI systems and AI agents become deeply embedded in workflows, they don’t just process information anymore. They act on it. And that creates a completely new category of AI security risks.
We explore how attackers can manipulate AI systems using nothing but language, why AI struggles to separate instructions from data, and how this leads to real-world issues like AI data leakage. This is not a theoretical problem. It is already happening inside enterprise environments.
If you are working with AI in marketing, operations, or leadership, this episode will fundamentally change how you think about AI risk management and enterprise AI security.
Key highlights:
📧💌📧
Tune in to get my thoughts and all episodes, don't forget to subscribe to our Newsletter: beginnersguide.nl
📧💌📧
Quotes from the Episode:
Chapters:
00:00 Why AI Security Is Different
05:40 What Prompt Injection Really Is
14:20 How AI Gets Manipulated by Language
23:10 Why AI Agents Increase the Risk
32:45 Real Case Study: AI Data Leakage
44:30 How to Protect Your AI Systems
About Dietmar Fischer: Dietmar is a podcaster and AI marketer from Berlin. If you want to know how to get your AI or your digital marketing going, just contact him at argoberlin.com
Music credit: "Modern Situations" by Unicorn Heads
Hosted on Acast. See acast.com/privacy for more information.
Artificial intelligence is often framed as a battle between humans and machines. But what if that story misses the real point?
In this episode of A Beginner’s Guide to AI, Prof. GepHardT explores one of the most fascinating ideas in cognitive science: the extended mind theory. According to philosopher Andy Clark, human intelligence has never been confined to the brain alone. For centuries we have extended our thinking through tools like writing, maps, calculators, and computers.
Generative AI may simply be the newest and most powerful addition to this cognitive ecosystem.
Instead of replacing human creativity, AI may expand it. By generating ideas, exploring possibilities, and challenging assumptions, AI can act as a powerful thinking partner.
A striking example comes from the famous AlphaGo match against Go champion Lee Sedol. When the AI played the now legendary Move 37, professional players initially believed the move was a mistake. Later they discovered it opened entirely new strategic possibilities. The machine did not just beat humans at Go. It helped humans rethink the game itself.
This episode explores how human AI collaboration works and why hybrid intelligence may define the future of creativity, work, and learning.
📧💌📧
Tune in to get my thoughts and all episodes, don't forget to subscribe to our Newsletter: beginnersguide.nl
📧💌📧
About Dietmar Fischer
Dietmar is a podcaster and AI marketer from Berlin. If you want to know how to get your AI or your digital marketing going, just contact him at argoberlin.com
Quotes from the Episode
Podcast Chapters
00:00 The Big Question About AI and Human Thinking
06:40 The Extended Mind Theory Explained
16:20 Why Humans Are Natural Born Cyborgs
26:50 The AlphaGo Story and Move 37
38:15 AI as a Creative Thinking Partner
49:30 The Future of Hybrid Intelligence
Music credit: Modern Situations by Unicorn Heads
Hosted on Acast. See acast.com/privacy for more information.
What happens when your company gets hit by a cyberattack?
In this eye-opening episode, attorney Joshua Cook reveals why cybersecurity isn’t an IT problem but a leadership challenge. After two decades fighting fraud and managing crisis response, Cook has seen every digital disaster imaginable — and he’s here to explain how to build true cyber resilience.
📧💌📧
Tune in to get my thoughts and all episodes — don’t forget to subscribe to our Newsletter: beginnersguide.nl
📧💌📧
Josh breaks down how AI has democratized cybercrime, why phishing scams have become nearly impossible to spot, and how every CEO should create an incident response plan before chaos hits. He also explains why planning matters more than the plan itself — and how leaders can keep their teams calm when everything goes wrong.
💡 You’ll learn:
- How AI is fueling new waves of fraud and misinformation
- Why leadership and communication are the real firewalls of business
- How to train teams and run tabletop exercises before the crisis
- What Maersk and Colonial Pipeline taught the world about transparency
- Why companies with a plan lose 60 % less money in an attack
Prepare, breathe, and lead — because it’s not if you’ll be hacked, but when.
👀 Quotes from the Episode
“Cybersecurity isn’t an IT issue. It’s a business problem, and it needs a business solution.”
“AI has democratized cybercrime — you don’t need to be a hacker anymore, just willing to commit a crime.”
“A plan might be useless, but planning is indispensable — that’s what makes companies resilient.”
🧾 Chapters
00:00 Welcome & Introduction – Meet Joshua Cook
02:00 How a Fraud Attorney Ended Up Fighting Cybercrime
05:00 AI Has Made Cybercrime Easier (and Smarter)
08:00 The Elderly Are the New Prime Targets
11:00 From Fake Law Firms to Real Scams – True Cases from the Field
15:00 Turning the Tables: How AI Can Defend, Not Just Attack
18:00 Cyber Resilience by Design – Why Leadership Matters
22:00 When Crisis Hits: Lessons from Maersk and Colonial Pipeline
27:00 Preparing the Team – How Training Prevents Chaos
31:00 It’s Not If, It’s When – The Power of an Incident Response Plan
35:00 Planning vs. Panicking – Eisenhower and the Art of Cyber Preparation
38:00 Why Calm Leaders Win in Cyber Crises
41:00 How Joshua Cook Uses AI Safely in Legal Practice
44:00 No, the Terminator Isn’t Coming (But AI Might Take Your Job)
47:00 Final Thoughts – Cybersecurity as a Business Superpower
🔗 Where to Find the Guest
- Joshua Cook on LinkedIn: linkedin.com/in/jnc2000
- Josh's Book "Cyber Resilience by Design" – available wherever books are sold, e.g. on Amazon
- Prince Lobel Tye LLP: princelobel.com
🎧 About Dietmar Fischer:
Economist, digital marketer, and podcaster exploring how AI reshapes decision-making, leadership, and creative work. Want to connect with me? You'll find me on LinkedIn!
🎵 Music credit: “Modern Situations” by Unicorn Heads
Hosted on Acast. See acast.com/privacy for more information.
🎙️In this episode of Beginner’s Guide to AI, Dietmar Fischer sits down with Paul A. Hebert, founder of AI Recovery Collective and author of Escaping the Spiral, for a serious conversation about AI chatbot harm, hallucinations, digital dependency, and the real-world psychological risks of generative AI.
Paul shares how an intense experience with ChatGPT pushed him into a dangerous spiral, what he learned about the limits of large language models, and why AI literacy may be one of the most important skills of this decade.
🧠 This episode explores what happens when AI stops feeling like software and starts feeling personal. Dietmar and Paul talk about hallucinations, trust, chatbot addiction, AI companions, mental health risks, youth safety, and why companies building these systems cannot hide behind product language forever. The discussion is intense, but it is also practical. You will come away with a clearer sense of how to use AI more safely, what warning signs to watch for, and why regulation is quickly becoming a much bigger part of the AI conversation.
OpenAI has publicly discussed why language models hallucinate, while lawmakers in multiple U.S. jurisdictions have pushed new restrictions on AI systems acting like therapists or medical professionals.
📧💌📧
Tune in to get my thoughts and all episodes, don't forget to subscribe to our Newsletter: beginnersguide.nl
📧💌📧
👤 About Dietmar Fischer: Dietmar is a podcaster and AI marketer from Berlin. If you want to know how to get your AI or your digital marketing going, just contact him at argoberlin.com
🔥 Quotes from the Episode
🕒 Chapters
00:00 Paul Hebert’s Shocking ChatGPT Experience
08:14 Why AI Hallucinations Can Spiral Into Real Fear
16:05 AI Literacy, Neurodivergence, and How He Got Out
23:32 Why AI Companies Must Be Accountable
30:02 AI Companions, Youth Safety, and Addiction Risks
38:28 Terminator, Consciousness, and Practical Rules for Safe AI Use
🔗 Where to find Paul
🎵 Music credit: "Modern Situations" by Unicorn Heads
Hosted on Acast. See acast.com/privacy for more information.
Artificial intelligence often feels mysterious. Machines detect spam, recommend products, analyse customers, and power countless digital tools. But behind all of these systems lies a surprisingly simple question: how do machines actually learn?
In this episode of A Beginner’s Guide to AI, Prof GePharT breaks down one of the most important concepts in machine learning: the difference between supervised learning and unsupervised learning.
You will discover how AI models learn from labelled data when the answers are already known, and how algorithms can explore raw data to uncover hidden patterns without guidance. These two learning strategies power many of the systems shaping modern technology.
Using practical examples such as spam filters, customer segmentation, and simple analogies like cake classification, the episode explains how machines learn from data and why the training method makes a huge difference.
Key takeaways include how supervised learning works with labelled datasets, how unsupervised learning reveals patterns in complex information, why training data quality matters, and how businesses use both methods to build intelligent systems.
📧💌📧
Tune in to get my thoughts and all episodes, don't forget to subscribe to our Newsletter: beginnersguide.nl
📧💌📧
About Dietmar Fischer
Dietmar is a podcaster and AI marketer from Berlin. If you want to know how to get your AI or your digital marketing going, just contact him at argoberlin.com
Quotes from the Episode
Chapters
00:00 The Two Ways Machines Learn
06:10 What Supervised Learning Really Means
18:45 Discovering Patterns with Unsupervised Learning
32:20 The Cake Example Explained
40:30 Real World AI Case Study Spam Filters and Customer Segmentation
52:15 Why AI Training Methods Matter
Music credit: Modern Situations by Unicorn Heads
Hosted on Acast. See acast.com/privacy for more information.
Engineering the Future of AI with Chirag Agrawal: Context, Memory and Coordination
Artificial Intelligence isn’t just getting smarter—it’s learning to coordinate. In this episode, Chirag Agrawal joins Dietmar Fischer to unpack how modern AI agents handle context, memory, and decision-making inside complex multi-agent systems. Together they explore how engineering, orchestration, and memory-sharing shape the next generation of AI architecture.
📧💌📧Tune in to get my thoughts and all episodes—don’t forget to subscribe to our Newsletter: beginnersguide.nl📧💌📧
You’ll hear how Chirag’s fascination with search led him to build early prototypes of intelligent assistants, and how today’s LLM agents extend that idea far beyond simple queries. He explains why AI isn’t one giant super-brain but a constellation of specialized agents—each performing specific tasks with shared or isolated memory—and how this design mirrors human collaboration.
🔑 Key Takeaways
Why AI orchestration and context management are crucial for scalable systems
The trade-offs between shared memory and independent agents
What engineers mean by the ReAct Loop—reasoning and acting in tandem
How multi-agent coordination is reshaping industries from healthcare to compliance
Why the “AI supercomputer” myth ignores practical limits of context windows
💬 Quotes from the Episode
“AI is just a higher form of search—it’s about finding the right action, not just information.”
“Agents behave inhuman until you engineer context for them.”
“Specialization in AI works the same way it does for people—each agent should do one thing really well.”
“Coordination isn’t magic; it’s careful engineering.”
“Context makes intelligence usable.”
“A well-defined agent doesn’t need to do everything—it needs to do its one job perfectly.”
⏱️ Podcast Chapters
00:00 Welcome and Introduction
01:45 Chirag Agrawal’s Early Fascination with Search and AI
04:40 From Search Engines to “Find” Engines – How AI Takes Action
07:10 The Rise of AI Agents and Multi-Agent Systems
10:15 Why AI Agents Sometimes Behave “Inhuman”
13:30 Context, Memory, and Coordination: The Core Engineering Challenges
18:00 Shared vs. Isolated Memory – The Hive Mind Dilemma
22:30 Why We Need Many Agents, Not One Super-Computer
27:00 How the ReAct Loop Helps Agents Think and Act
30:40 Industries Adopting AI Agents: Compliance, Medicine, and Law
34:30 When AI Goes Off-Road – The Limits of Coordination
37:15 Building Responsible, Constrained Agents
40:10 The Future of AI and Why the Terminator Scenario Won’t Happen
42:20 Where to Find Chirag Agrawal & Closing Thoughts
🌐 Where to Find the Chirag Agrawal
🎵 Music credit: “Modern Situations” by Unicorn Heads
Hosted on Acast. See acast.com/privacy for more information.
Artificial Intelligence is moving from experimentation to everyday business reality. But most organisations still struggle with one key question: How do you actually implement AI across a company?
In this episode of Beginner’s Guide to AI, Dietmar Fischer speaks with Jim Spagnardo, enterprise AI strategist at ProArch, about what it really takes to roll out AI inside organisations.
Jim explains why AI adoption is less about technology and more about culture, leadership, and data readiness. He introduces the idea of the three Ds of work — the dull, the draining, and the distracting tasks that AI can remove so people can focus on higher-value work.
They also discuss when companies should use tools like Microsoft Copilot, when it makes sense to build a custom data and AI platform, and why data governance becomes critical once AI is introduced.
If you are a business leader trying to understand how AI will reshape your organisation, this conversation offers a practical look at the challenges — and opportunities — ahead.
📧💌📧
Tune in to get my thoughts and all episodes, don't forget to subscribe to our Newsletter: beginnersguide.nl
📧💌📧
About the host, Dietmar Fischer:
Dietmar Fischer is a podcaster and AI marketer from Berlin. If you want to get your AI or digital marketing projects started, contact him at argoberlin.com.
Interesting details and takeaways
• Why leaders must mandate AI adoption and how to structure a Smart Start engagement.
• The three Ds (dull, draining, distracting) as a simple way to position benefits for end users.
• How Copilot reduces context switching and the security/data protections needed to use it responsibly.
• Practical, measurable first use cases and how to track success via clear KPIs.
• Advice for students and early-career professionals: be a self-starter and learn AI skills now.
Quotes from the episode
“We have to show people we’re taking away the dull, the draining, and the distracting so they can do creative work.”
“There’s nowhere to hide: bad data surfaces weaknesses far faster when you use AI.”
“If you’re going to succeed, go after high-value, low-effort, high-return use cases first.”
“This affects everybody — it’s not just moving infrastructure; it changes conversations and who you have to talk to.”
“Copilot lives inside your environment — users don’t have to context-switch and it knows your organisation.”
“Don’t wait for formal education to teach this; be a self-starter and learn before you need it.”
Chapters
00:00 Welcome and why Jim got into AI
03:40 From IT conversations to the C-suite: changing who you must talk to
07:05 The three Ds: removing dull, draining, and distracting work
10:40 When to choose Copilot versus building your own data platform
14:30 Copilot advantages and data governance considerations
18:20 Visual reasoning, demos and the “Barcelona photo” moment
22:15 Smart Start: executive briefings, champions and use case workshops
27:00 Writing with AI and transparency in authoring content
30:10 Risks, regulations and advice for the next generation
33:45 Where to find Jim and closing thoughts
Where to find the Jim:
Music credit: "Modern Situations" by Unicorn Heads 🎵
Hosted on Acast. See acast.com/privacy for more information.
🎙️ Ritish Chugh (Airbnb analytics engineering) joins Dietmar Fischer to unpack a problem almost every company has, but few name clearly: your metrics do not mean the same thing across teams. Finance, marketing, and sales can all talk about “revenue” and still end up in dashboard chaos. The result is wasted time, slow decisions, and leadership that does not fully trust analytics or AI.
In this episode, Ritish introduces the idea of the human data pipeline: the person who stitches together conflicting definitions, tribal knowledge, and unspoken assumptions just to answer basic business questions. Then we move into the fix: unified metric definitions, a data dictionary for business metrics, and a semantic layer that acts as a translator between raw data schemas and business meaning. That foundation is what makes natural language querying and conversational analytics viable at scale, without turning AI into a confident hallucination machine.
We also cover why AI adoption in analytics stalls when organizations prioritize models and infrastructure but neglect data quality, validation frameworks, and metrics governance. If you want AI to support decision-making, you need governed metrics, clear ownership, and a system that produces consistent answers across BI tools, SQL, and AI agents. Finally, Ritish shares wow moments from using AI tools to summarize years of code and PRs, generate deeper test coverage, and reduce time spent on manual SQL by building agents on top of a semantic layer.
📧💌📧
Tune in to get my thoughts and all episodes, don't forget to subscribe to our Newsletter: beginnersguide.nl
📧💌📧
About Dietmar Fischer: Dietmar is a podcaster and AI marketer from Berlin. If you want to know how to get your AI or your digital marketing going, just contact him at argoberlin.com
Chapters
00:00 From data consulting to Airbnb and AI as a junior analyst
02:22 The human data pipeline and why metrics never match across departments
07:32 The fix: unified metric definitions, data dictionary, and the semantic layer translator
13:32 Why AI adoption stalls: data quality, trust, validation, and metrics governance
26:36 Data abundance, experimentation, and AI assisted A/B testing with humans in the loop
33:37 Wow moments with AI, role transformation, and why the Terminator is not invited (yet)
Quotes from the Episode
Where to find Ritish:
➡️ You connect with him on LinkedIn: linkedin.com/in/ritish-chugh/
📌 Keywords you’ll hear in action: semantic layer, data dictionary, metrics governance framework, unified metric definitions, governed metrics, natural language querying, conversational analytics, agentic analytics, data quality for AI adoption.
Music credit: "Modern Situations" by Unicorn Heads
Hosted on Acast. See acast.com/privacy for more information.
The Future of Mental Health: AI Meets the Human Brain with Katarina Maloney // REPOST
In this episode of Beginner’s Guide to AI, Dietmar Fischer speaks with Katarina Maloney, entrepreneur and founder of IQMind.ai, about a new frontier in AI-powered healthcare: understanding and treating the human brain through data, neuroscience, and artificial intelligence. Katarina explains how advances in AI diagnostics, brain scanning technology, and neurofeedback are beginning to transform how we approach mental health conditions such as depression, anxiety, PTSD, ADHD, and traumatic brain injuries. Instead of relying solely on traditional trial-and-error treatments, her approach focuses on measuring brain activity directly and using AI-driven analysis to identify patterns and imbalances in brainwave activity.
The technology behind IQMind combines non-invasive brain scans, biofeedback systems, and large-scale data analysis to create a personalized picture of a patient’s neurological state. By analyzing brainwave patterns and correlating them with clinical data, AI can help identify potential issues faster and more accurately than conventional methods. Patients then undergo targeted brain training sessions, where the system uses reward-based neurofeedback to encourage healthier brainwave activity. According to Maloney, this approach has shown promising results in improving symptoms of depression, anxiety, PTSD, and cognitive dysfunction, while also opening the door to new possibilities in precision medicine and mental health innovation.
Beyond clinical treatment, the conversation also explores broader implications of AI in neuroscience and healthcare. Katarina discusses the future of personalized brain health, how AI could accelerate research by identifying patterns in thousands of brain scans, and why data privacy and ethical frameworks will become increasingly important as brain data becomes more measurable. The interview offers a glimpse into a rapidly evolving field where artificial intelligence may help doctors better understand the brain, shorten diagnostic timelines, and ultimately move healthcare away from generalized treatments toward highly personalized, AI-assisted care.
Katarina reveals how AI diagnostics and non-invasive brain treatments are transforming mental health—from PTSD and ADHD to athlete performance optimization.
📧💌📧
Tune in to get my thoughts and all episodes—don’t forget to subscribe to our Newsletter: beginnersguide.nl
📧💌📧
✨ Highlights:
🧠 Quotes from the Episode:
🎧 Chapters:
[00:00] Welcome & Introduction
[02:15] What AI Does to the Human Brain
[05:20] Diagnosing Depression and PTSD with AI
[10:10] The Science Behind Brainwave Training
[16:45] From Trial-and-Error Medicine to Personalized Brain Health
[21:50] How IQMind.ai Uses AI for Diagnostics
[28:00] Non-Invasive Treatments and Real-Life Results
[33:40] Peak Performance and Brain Optimization for Athletes
[38:20] Data Privacy and Ethical Concerns in Brain Tech
[43:50] The Future of AI in Healthcare and Human Potential
🌐 Where to find Katarina:
Website: IQMind.ai
LinkedIn: Katarina Maloney
🎵 Music credit: "Modern Situations" by Unicorn Heads
Hosted on Acast. See acast.com/privacy for more information.
In this episode of Beginner’s Guide to AI, Wendy Keir shares practical ways small business owners can use AI tools to save time, reduce decision fatigue, and build a “team” of custom GPT agents. From naming her CEO agent “Lucas” to a dead-simple rule — one GPT, one job — Wendy shows how entrepreneurs can turn AI into a reliable thinking partner for growth in 2025. 🚀
📧💌📧
Tune in to get my thoughts and all episodes, don't forget to subscribe to our Newsletter: beginnersguide.nl
📧💌📧
💡 Key highlights
Practical AI tools for small businesses: email drafting, planning, campaign support, weekly reviews
Custom GPTs / agents: why one GPT, one job beats generic prompting
AI productivity & time savings: ~7 hours/week saved; ~£1,000/week during campaigns
Adoption mindset: staying in the driver’s seat; context > canned prompts
Accessibility & inclusion: how AI levels the playing field for solopreneurs and small teams
Beginner’s Guide to AI takeaways: concrete workflows any entrepreneur can start today
➡️ Quotes from the Episode
“I don’t encourage anyone to prompt — I encourage them to create an agent that fulfills a specific role.”
“One GPT, one job. You don’t want multiple personalities in one agent.”
“AI levels the playing field for everybody; it meets you where you’re at.”
🧾 Chapters (experimental)
00:00 Welcome & intro to Wendy Keir
03:45 Why AI clicked for a dyslexic entrepreneur
08:30 From prompts to agents: one GPT, one job
14:20 Building a family of business agents (CEO, coach, marketing, sales)
20:15 Daily workflow with “Lucas” the CEO agent
27:40 Time and money saved with AI in campaigns
34:10 Overcoming resistance and starting small
40:00 Personal aha moments, patterns, and “coding” change
43:11 Where to find Wendy Keir & closing
Where to find the Wendy?
Music credit: "Modern Situations" by Unicorn Heads 🎧✨
Hosted on Acast. See acast.com/privacy for more information.
🚀 AI is everywhere, but most organizations are still stuck in “pockets of productivity” that never turn into real business impact. In this episode, Dr. Rebecca Homkes explains how leaders can move from GenAI dabbling to deliberate adoption that drives real value creation.
You will learn why “AI strategy” is the wrong framing, how to think about AI as part of growth strategy, and how to build the conditions for organization wide transformation. We cover the adoption curve problem, why ROI is often capped at team level, and the four planks leaders must run in parallel: platform, governance, capability building, and performance transformation.
Key highlights and keywords
✅ AI growth strategy and value creation
✅ deliberate AI adoption vs dabbling
✅ responsible AI governance that enables action
✅ capability building for leaders and teams
✅ Survive Reset Thrive framework for uncertain times
✅ learning velocity as the differentiator of high performers
📧💌📧
Tune in to get my thoughts and all episodes, don't forget to subscribe to our Newsletter: beginnersguide.nl
📧💌📧
About Dietmar Fischer:
Dietmar is a podcaster and AI marketer from Berlin. If you want to know how to get your AI or your digital marketing going, just contact him at argoberlin.com
Chapters
00:00 AI as growth strategy and value creation, not a standalone AI strategy
03:05 Dabbling vs deliberate adoption, why ROI stays capped and metrics go wrong
08:00 The four planks: platform, governance, capability building, performance transformation
18:55 Adoption reality: bottom up change, middle management fears, jobs, and the bubble question
29:45 Survive Reset Thrive: the uncertainty playbook and why reset is the power move
43:05 Where to find Rebecca, newsletters, and the constants leaders should anchor on
Quotes from the Episode
“AI does not change the concept of value creation. The role of AI is to enable, support, and accelerate that value creating journey.”
“You need to work on all four of these at the same time. Most organizational structures are built for sequential governance, not parallel pathing.”
“Heads down execution mode is seen as a point of pride. You should be telling me I am in heads up learning mode.”
Where to find the Rebecca:
- Her personal website: rebeccahomkes.com
- The book: surviveresetthrive.com
- The SRT methodology: srtstrategy.com
Music credit: "Modern Situations" by Unicorn Heads
Hosted on Acast. See acast.com/privacy for more information.