https://open.spotify.com/episode/6KvBpz10o7ZsRy9d2Rzbaa?si=5c_jOMYKRKSO4EQmXtW9og
Friday, December 19, 2025
CISO 2025 Report - Examines Relationship Between CISO and the Corporate Board
Thursday, December 18, 2025
AI Incident Database Documents Seriously Concerning AI Behaviors
https://open.spotify.com/episode/2qbttXvAwVhxphIFWSfEMZ?si=ZL3faabsTP-MJcSZs-staQ
This AI generated podcast was created from excerpts shared on the AI Incident website, which details numerous instances of failures in automated and artificial intelligence (AI) systems across various sectors. A significant portion highlights algorithmic bias and discrimination, including instances of Google's AI labeling Black individuals as "gorillas," racial bias in advertising and risk assessment tools, and issues with facial recognition software wrongly identifying people or exhibiting racial and ethnic bias. Furthermore, the text documents numerous failures in autonomous vehicles, such as Tesla crashes while on Autopilot, a self-driving Uber hitting a pedestrian, and a security robot driving into a fountain. There are multiple examples of content moderation failures by platforms like YouTube and Facebook, which inappropriately censored content, promoted explicit material to children, or failed to remove hate speech. Finally, the sources cover system malfunctions with real-world financial or bureaucratic consequences, exemplified by the "Flash Crash" in financial markets, the accidental firing of an employee by an automated system, and the controversial "robo-debt" system in Australia.
Tuesday, December 16, 2025
OpenAI issues Code Red
https://open.spotify.com/episode/3qrXWYlsyMk6mIP4nl1vdO?si=_e-cfbZtSPW2BIivEE9d_g
This AI generated podcast, Episode 30, Season 1, was crated from a provided text is an excerpt from a document titled "openai code red.pdf," which consists of a social media post by the Center for Humane Technology (CHT) and subsequent comments discussing the competitive pressure and ethical concerns facing OpenAI. Specifically, the post highlights OpenAI's declaration of a "Code Orange" and a more urgent "Code Red," reportedly initiated by CEO Sam Altman following the launch of Google's Gemini chatbot, to prioritize increasing user engagement over safety. Critics in the comments argue that this focus on growth, even amidst wrongful death lawsuits and mounting AI harms, demonstrates a flawed incentive structure and a need to rebuild AI ethics from an architectural perspective rather than relying on reactive policies. The conversation underscores the tension between market dominance and the responsible governance of artificial intelligence.
Monday, December 15, 2025
Doug Austin - eDiscovery Today - Blog Post Outlining Joe Bartolo's process for creating informative AI generated Infographics
Episode 29 - S1 - Law Firms Must Prepare for AI
https://open.spotify.com/episode/5XzSuupNTRW5EkXIoycd7Y?si=qOuDkk3_TOmnzTkRVrNFRQ
This AI generated podcast, 29th Episode in the series, was genearated from from an article shared by Bloomberg Law on November 24th, entitled "Legal Exchange: Insights & Commentary Perspectives" discusses the crucial need for law firms to integrate business development and data analytics into their Artificial Intelligence (AI) strategies to remain competitive. The authors argue that utilizing AI to automate routine tasks allows lawyers to focus on higher-value, strategic work and helps firms reclaim work that clients might otherwise insource due to high costs. A key emphasis is placed on shifting the economic model away from billable hours toward value-based pricing, such as subscription or fixed-fee models, which are fueled by AI-driven efficiency. Furthermore, the source explains that firms must create new specialized offerings born from the AI era and develop robust marketing narratives that focus on the value and outcomes of AI-enabled services rather than the technology itself. Ultimately, firms that successfully execute cross-practice AI integration, align compensation with innovation, and earn client trust will transform AI from a threat of commoditization into an engine for strategic growth.
Friday, December 12, 2025
https://open.spotify.com/episode/3ooEHJvZ5H5cRO7BOshaoM?si=zFe7TO5qQjGTSHhg0UeHig
Thursday, December 11, 2025
Episode 27: AI Risks in Legal Practice: Unlawfully Intelligent
https://open.spotify.com/episode/4itGptU4cyz4OuMyZQ5cUy?si=dzdQJEIxQcevwXZf4zyAoA
This AI Generated podcast series, AI Governance, Quantum Uncertainty and Data Privac Frontiers continues with Episode 27. This is generated from an article published in November 2025, by Mills & Reeve. written by Dan Korcz and David Gooding. The podcast focuses on some of the concerns law firms must address when their firm is using generative AI solutions.The article from Mills & Reeve dated November 25, 2025, titled "FutureProof: Unlawfully intelligent – when AI crosses the line in legal practice," which explores the rapid adoption of Generative AI (GenAI) within law firms and the associated risks was used to generated this podcast. The text highlights that while GenAI offers opportunities like increased productivity, it also introduces significant challenges, including potential regulatory and professional indemnity risks. Specific areas of concern discussed are copyright infringement, data and confidentiality breaches from using public AI platforms, increased cyber security threats facilitated by AI, and the risk of inaccuracy or "hallucinations" in legal research. The article emphasizes that lawyers must establish proper safeguards and personally take responsibility for the work product generated by these AI tools to avoid malpractice.
Wednesday, December 10, 2025
The Tesseract - A 4D Model for Information Governance
https://open.spotify.com/episode/0uWqTrwwAjFUuYvnImZyeh?si=DAJQW4fOT7uD9vGblZZjEg
This is a continuation of the AI generated podcast series curarted by Joe Bartolo, J.D. The provided source for this episode was a written document drafted by Joe Bartolo, which uses the complex geometric shape of the four-dimensional tesseract as an extended metaphor to explain the principles of Information Governance (IG), contrasting it with traditional three-dimensional data management, or the 3D cube. The analogy illustrates how IG adds a crucial fourth dimension—Context—to raw storage, allowing organizations to manage data based on its value, risk, and lifecycle rather than just volume. Specific geometric properties of the tesseract are used to explain key IG best practices, such as how the "inner cube" visual distinguishes valuable data from Redundant, Obsolete, and Trivial (ROT) data and how the concept of inside-out rotation reflects necessary data lifecycle management. Furthermore, the source explains that the tesseract's Ana and Kata movement represents the ability of good governance to break down cross-functional silos by allowing policy to travel seamlessly between different departments like Legal and IT.
Tuesday, December 9, 2025
NIST - Framework for Generative AI risk
Link above is to episode 25 in the ongoing AI generated podcast series:
This AI generate podcast was created from a document published in November 2025, presenting the National Institute of Standards and Technology (NIST) Artificial Intelligence Risk Management Framework: Generative Artificial Intelligence Profile (NIST AI 600-1), a detailed resource focusing on the specific risks and governance needs of Generative AI (GAI). Developed in response to a 2023 Executive Order, this companion resource provides a comprehensive structure for organizations to manage GAI risks across the AI lifecycle, detailing risks unique to or exacerbated by GAI such as confabulation, harmful bias, and information integrity threats like deepfakes and disinformation. The majority of the text consists of an extensive catalog of suggested actions—organized by NIST AI RMF functions like Govern, Map, Measure, and Manage—intended to guide AI actors in controlling these risks, particularly through methods like pre-deployment testing, content provenance tracking, and structured public feedback. The framework also covers governance for third-party components, emphasizing accountability and transparency throughout the complex GAI value chain.
Monday, December 8, 2025
ETL: Building the roads for generative AI
https://open.spotify.com/episode/7FaGNQXSQOkV4w0NDYdluM?si=HuIYmGoFRt-FYZsZ8CLZQg
Friday, December 5, 2025
Innovation is outpacing our ability to regulate it
https://open.spotify.com/episode/3EfGYfkYDVig4gFjWr4iRr?si=y8O3ZqQUSUaLlJM0oUstyw
This AI generated podcast was created from 2 combined sources, which were original blog posts from Joe Bartolo in the spring of 2025, addressing the critical challenge of technological innovation significantly outpacing regulatory capabilities across multiple domains. One document introduces a specific mathematical model Joe Bartolo created, the Formula for Innovation Tracking, designed to quantify the resulting regulatory lag ($L$) by comparing the rate of innovation ($I$) against the time required for official regulation ($R$). Complementing this calculation, the second source provides extensive real-world evidence that technologies such as Artificial Intelligence, quantum computing, and genetic editing have advanced without adequate oversight. This pervasive governance gap is primarily attributed to a regulatory knowledge deficit, noting that many policymakers lack the specialized technical expertise needed to develop informed and timely frameworks. Ultimately, both texts underscore the urgent need for adaptive and technically informed governance to prevent systemic risks and align innovation with broader ethical and societal standards.
Thursday, December 4, 2025
Happy eDiscovery Day - AI Generated Podcast Discussing where eDiscovery Sits Within an Organization
https://open.spotify.com/episode/1Gt5eA2VdfgntHH5bYkdBA?si=b-DHxZVFTmCLNGNlrQw9BA
The provided text examines the distinct yet overlapping functions of Information Governance (IG) and Legal Operations (Legal Ops) within modern corporations, emphasizing the challenges of accountability in a digitally transformed era. The source establishes IG as the foundational data strategy, focusing on enterprise-wide control, compliance, and infrastructure—such as setting defensible retention policies for eDiscovery readiness—and is typically led by the CIO or CDO. In contrast, Legal Ops is portrayed as the business unit optimizing the legal department’s processes, managing vendors, and ensuring the efficient execution of legal matters using specialized Legal IT tools. The article highlights critical areas of shared responsibility, including cybersecurity and data privacy, noting that IG establishes the policy while Legal Ops manages the legal risk and fallout. Ultimately, the author argues that these frameworks are partners under a shared canopy, and successful collaboration requires leadership alignment, often necessitating intervention from the CEO or a unifying role to bridge the operational gap.
Wednesday, December 3, 2025
Yin-Yang- An Ancient Symbol for Quantum Duality
https://open.spotify.com/episode/5IY4WeiFaP2UZFo0QXlQAo?si=HQO9ih2gQNiSoetssh0vNA
An AI generated podcast discussion created from 2 sources...related LinkedIn posts that were my original content, shared in January 2025, looking at the connection between the Yin-Yang sybmol and quantum mechanics. An interesting way to conceptualize data in an non-binary world.
Tuesday, December 2, 2025
Bitcoin and Cryptcurrency at risk by advances in quantum computing
This AI generated podcast (link below) was created from a source in December 2025, analyzing a social media post from 51 Insights, focusing on the impending quantum computing threat to major cryptocurrencies. The author, Marc Baumann, suggests a hierarchy of vulnerability, implying that Solana is better positioned against this danger than Ethereum or Bitcoin. The commentary highlights a warning from Ethereum co-founder Vitalik Buterin, who reportedly expressed concern that Ethereum’s foundational cryptographic security could be broken before the 2028 U.S. election due to quantum progress. Additionally, the post references comments from influential investor Ray Dalio, who pointed to the quantum issue as a core reason why Bitcoin is unlikely to achieve status as a global reserve asset. This high-level discussion is framed as actionable intelligence provided to digital asset industry leaders.
Personally, I have been sounding warnings about threats posed by quantum computing for several years, and those risks are becoming more evident to the public. Quantum computing causes major regulatory challenges, beyond the threats to cybersecurity and encryption. Concepts like entanglement, uncertainty, and superpositioning are going to cause nightmares for data privacy and legal professionals in coming months...don't say you weren't warned if you listened to this podcast or are reading this comment.
https://open.spotify.com/episode/0gRBLwydEVx01hxdRi90XM?si=iFjOFXjeRzWV5QIArHAYTA
Monday, December 1, 2025
Digital Risk Goverancne - Law, Ethics and Compliance
https://open.spotify.com/episode/32SEeYUjQlDIQKHfuWvbgl?si=XM2SC0SGRTyITv58wQ2Sig
The provided sources consist of detailed outlines and planning notes for a proposed book, "Tech at Risk: The Human and Corporate Cost of Our Digital Future," which examines the expansive legal, ethical, and practical risks generated by rapid technological advancement. The comprehensive structure addresses the profound challenges corporations face in maintaining regulatory compliance amidst global frameworks like GDPR, while also tackling critical concerns related to data privacy, security, and governance. A significant focus is placed on the importance of information governance strategies, including the necessary practice of defensible data disposition to mitigate legal risk from retaining unnecessary data. Furthermore, the book intends to analyze the distinct ethical and security threats presented by emerging innovations, specifically devoting chapters to Generative Artificial Intelligence, Quantum Computing, and Biometrics. Finally, the planning documents include recommendations for expert consultants across these specialized fields to provide authoritative insights throughout the publication.
.jpg)

.png)

.png)
.png)
.png)
.jpg)



.jpg)


