Description:
As generative AI tools like ChatGPT become more accessible, risk and compliance professionals are being called to both use them strategically and understand the risks they pose. This session explores how ChatGPT can support due diligence, policy drafting, monitoring, training, and more—while also unpacking the associated risks, such as data privacy, hallucinations, and regulatory scrutiny. You’ll learn practical ways to integrate ChatGPT into your compliance toolkit, including what tasks it can support and when human oversight is essential. We’ll also look at how regulators are beginning to address AI and what that means for governance, accountability, and risk management. Whether your organization is experimenting with ChatGPT or already using it, this session will help you lead conversations around safe, effective, and compliant use of AI in high-risk environments.
Who This Is Designed For: Risk and compliance professionals, internal auditors, legal teams, and financial crime investigators seeking practical, responsible ways to integrate ChatGPT and similar tools into regulated environments.
Agenda:
Explore how ChatGPT works and supports financial services workflows
Identify key risks: bias, hallucinations, data misuse, and leakage
Discover compliance use cases: audits, training, policy writing, monitoring
Review global regulatory updates impacting AI in financial sectors
Learn governance strategies to manage AI use responsibly and securely
Discuss real-world applications, challenges, and audience-submitted case examples
By the end of this webinar, you will know how to:
Apply ChatGPT and similar AI tools to support compliance, monitoring, and training tasks
Identify and mitigate risks including data leakage, bias, and hallucinations
Integrate AI responsibly into regulated workflows with appropriate human oversight
Understand emerging regulatory guidance and enforcement expectations around AI use
Lead organizational discussions on safe, effective, and compliant use of generative AI