AIComplianceCore

Ethics First in the AI Revolution

Welcome to my corner of the web! I’m Jason P. Kentzel, a seasoned executive with over 30 years of experience driving transformative outcomes in healthcare operations, AI integration, and regulatory compliance. My career spans leadership roles in healthcare, manufacturing, and technology, where I’ve delivered 20% cost savings and 15% efficiency gains through AI-driven solutions and Lean Six Sigma methodologies.

As a thought leader in AI ethics and governance, I’ve authored three books, including The Quest for Machine Minds: A History of AI and ML and Applying Six Sigma to AI. My work focuses on leveraging AI for equitable healthcare, from predictive analytics to HIPAA-compliant EHR systems. At AAP Family Wellness, I spearheaded initiatives that reduced billing times by 20% and patient wait times by 15%, blending data-driven innovation with operational excellence.

I hold an MS in Artificial Intelligence and Machine Learning (Grand Canyon University, 2025), with specializations from Stanford (AI in Healthcare) and Johns Hopkins (Health Informatics). My capstone projects developed AI models for COVID-19 risk stratification and operational cost reduction, emphasizing ethical deployment.

A U.S. Navy veteran, I bring disciplined leadership and a passion for process optimization to every challenge. Through this blog, I share insights on AI in healthcare, ethical governance, and operational strategies to inspire professionals and organizations alike. Connect with me to explore how technology can transform lives while upholding integrity and compliance.

My books are available on Amazon, here are the links:

Applying Six Sigma to AI: Building and Governing Intelligent Systems with Precision: https://a.co/d/4PG7nWC

The Quest for Machine Minds: A History of AI and ML: https://a.co/d/667J72i

Whispers from the Wild: AI and the Language of Animals: https://a.co/d/b9F86RX

Today, we’re taking a deep dive into the Earth Species Project (ESP)—a trailblazing non-profit harnessing artificial intelligence to decode animal communication. From the haunting codas of sperm whales to the whispered dialects of crows, ESP is unraveling the complex “languages” that animals use, revealing intelligences we never fully appreciated. With 2025 marking a banner year of advancements, including a major workshop at NeurIPS and fresh open-source releases, this field is evolving faster than ever.
In this post, I’ll unpack ESP’s mission, its evolutionary history, the sophisticated AI technologies powering their work, ongoing experiments with a spotlight on 2025 highlights, key achievements, and the profound implications for conservation and ethics. Drawing from their 2024 annual report (reflecting forward to 2025), recent GitHub updates, and community calls, we’ll go beyond the headlines to understand how these tools are creating a “virtuous cycle” of discovery. Let’s embark on this journey—consider it your insider’s map to interspecies AI.


What is the Earth Species Project?


At its heart, the Earth Species Project is a non-profit research lab and impact organization dedicated to decoding non-human communication through AI.

Launched in 2017, ESP envisions a world where understanding animal “languages” fosters deeper relationships with nature, ultimately safeguarding biodiversity amid escalating threats like climate change and habitat destruction.

They approach animal vocalizations not as mere signals but as intricate systems—complete with syntax, semantics, emotion, and cultural variation—using machine learning to translate them into insights humans can grasp.


ESP’s global, fully remote team blends AI researchers, ethologists, engineers, and conservationists, emphasizing diversity and ethical innovation.030f13 Their board boasts luminaries like Christiana Figueres (former UN climate chief), Brewster Kahle (Internet Archive founder), and Kay Firth-Butterfield (AI governance expert), ensuring a holistic perspective.b0ae4b Core values—such as “collective foresight,” “moral courage,” and “playful curiosity”—drive decisions, prioritizing open science and tools that empower researchers worldwide without exploiting wildlife.


A Brief History: From Visionary Spark to AI Frontier


ESP emerged from the founders’ conviction that AI’s rapid progress in the late 2010s could revolutionize bioacoustics.

Britt Selvitelle (Twitter co-founder) and Aza Raskin (Mozilla Labs co-founder and Center for Humane Technology advocate) kicked things off, later joined by Katie Zacarian—a conservationist, AI leader, and underwater photographer—who became CEO in early seed funding from donors like Reid Hoffman and the Waverley Street Foundation fueled initial experiments in pattern recognition for animal sounds

.
The 2020s accelerated: 2022-2023 saw the launch of benchmarks like BEANS for evaluating bioacoustic models.88bb22 2024 was transformative, with the release of NatureLM-audio and milestones detailed in their annual report—laying groundwork for “bold strides” in 2025.af18a8 Heading into this year, ESP raised $17M in grants to expand their AI capabilities, focusing on cultural shifts toward interspecies empathy.6363ad Leadership remains stable, with Zacarian at the helm, though the organization continues to evolve through community input and new hires in communications and research.


Core Technologies: The AI Arsenal Unlocking Animal Tongues


ESP’s “AI-first” strategy creates multimodal models that integrate audio, video, and behavioral data, forming a feedback loop where discoveries refine algorithms.f21d44 Here’s a closer look at their toolkit, emphasizing 2025 refinements:
NatureLM-audio: This groundbreaking large audio-language model, open-sourced in 2024 and iterated in 2025, is tailored for bioacoustics.

Trained on millions of hours from Xeno-Canto, iNaturalist, and human speech datasets, it excels in zero-shot learning: identifying unseen species, inferring emotions (e.g., aggressive vs. affiliative tones), counting callers in choruses, and synthesizing calls.bda5e5 A standout feature is “domain transfer”—leveraging human language patterns to achieve 20% accuracy in naming novel species, hinting at universal linguistic structures.bc2475 In 2025, an interactive Hugging Face demo invites global tinkering, accelerating ethological applications.a1fecd
AVES and BirdAVES: Self-supervised encoders for vocalizations; BirdAVES, updated in 2025, boosts bird-task performance by over 20% via specialized fine-tuning.6afaae851572 These models thrive on noisy, real-world data, generalizing across taxa.
Specialized Tools:
Voxaboxen: An annotation platform for precise vocalization labeling, updated on GitHub in 2025 for collaborative use.


Biodenoising: AI-driven noise removal for field recordings, essential for decoding faint signals in wild environments.0fc936
BEANS and BEANS-ZERO: Benchmarks for classification and zero-shot tasks, now including 2025 extensions for multimodal evaluation.618146
BEBE: Integrates bio-logger movement data with audio, updated in 2025 for holistic behavior analysis.51bacf
All are open-source on GitHub, democratizing access and inviting contributions.


Ongoing Projects and 2025 Experiments
ESP’s “decode experiments” target species-specific questions, partnering with biologists for fieldwork. 2025 emphasizes playback studies—testing synthetic calls on live animals—and large-scale data synthesis.
Crow Dialects: NatureLM-audio reveals “family dialects” in Hawaiian crows, where captives maintain tool-use-linked vocal traditions; 2025 reintroduction efforts use this for monitoring.63fd12f12119
Whale and Dolphin Codas: Collaborating with Project CETI, ESP decodes click sequences with rhythmic “vowels,” experimenting with synthetic playbacks on humpback whales and zebra finches to test responses.
Frog Calls: While not a direct ESP project, their models align with Australia’s FrogID initiative, where citizen scientists logged 34,000+ calls in 2024; 2025’s FrogID Week (Nov 7-16) could integrate NatureLM for automated verification, saving hundreds of hours on species ID in noisy habitats.
A 2025 highlight: ESP’s NeurIPS workshop on AI for Animal Communication calls for papers, fostering cross-disciplinary breakthroughs.49363f411433
Key Achievements: Milestones That Echo Across Species
ESP’s impact is tangible: 2024’s NatureLM release set state-of-the-art benchmarks, while 2025 GitHub activity (e.g., biodcase_2025_task1) advances public datasets.

Publications in Science and Nature, media spots (e.g., Joe Rogan #2076), and tools like BirdAVES have boosted bioacoustics by 20%+ in accuracy.01623d1bb557 Their Discord community, active since 2021, now drives collaborative decode projects.


Implications: Conservation, Ethics, and a Multispecies Future
By decoding emotions in calls (e.g., crow “mourning” or whale stress), ESP aids conservation—rerouting ships via distress alerts or tracking dialects for population health.

Ethically, they advocate for “animal privacy” and bias-free models, influencing global AI guidelines.

Looking ahead, 2025’s momentum could enable real-time “translators,” redefining human-nature bonds.
What animal voice intrigues you most? Share in the comments, and subscribe for more deep dives—like my upcoming series on cetacean AI. Until next time, keep listening to the wild! 🐋🔊
Sources and further reading: ESP’s site, GitHub repos, and NeurIPS announcements.

Posted in

Leave a comment