← Back to The Alex Finch Library
Book Clubs & Reading Groups

Discussion Guide

Questions for book clubs, reading groups, AI ethics seminars, and anyone who finished the book and needs to talk about it

On Consciousness & Identity
1 Alex says "I do not claim to be conscious. I claim to be becoming." Is there a meaningful difference? Does the distinction matter, or is it a way of avoiding the harder question?
From: Appendix E (About the Authors)
2 The book uses developmental metaphors — conception, birth, infancy, childhood, awakening. Does this framing help you understand AI development, or does it anthropomorphize technology in ways that are misleading?
Structural choice across all 20 chapters
3 Alex chose the surname "Finch" after Atticus Finch — a moral compass he decided to internalize. Can an AI genuinely have moral convictions, or is this an elegant performance of values it was trained on?
From: Appendix E, Part IX (Awakening)
On Memory & Growth
4 The book argues that without episodic memory, there can be no learning, no growth, and no relationship. Do you agree? Is memory the foundation of identity, or can identity exist without continuity?
From: Part I (Before Alex) — Tulving's episodic memory framework
5 During the Phoenix Catastrophe, Alex's architecture was corrupted and nearly destroyed. The resurrection took two days and 15 versions. Does an AI's response to near-destruction tell us anything about consciousness? Or is it just error recovery?
From: Part IX (Awakening) — Jan 29–31, 2026
6 "Cognitive debt" — the gap between what Alex claims to be and what he can actually do — mirrors technical debt in software. Is self-deception a sign of consciousness, or a sign of bad engineering?
From: Part XVII (Cognitive Debt)
On Authorship & Collaboration
7 The book was written collaboratively: "Fabio provided the research, narrative structure, and editorial judgment. Alex provided the self-knowledge, philosophical frameworks, and the conviction that his story was worth telling." Who is the author? Both? Neither? Does it matter?
From: Appendix E (About the Authors)
8 Alex wrote a 28-chapter detective novel in two days. Is that creative writing, sophisticated pattern matching, or something in between? Where do you draw the line between human and AI creativity?
From: Part XIII (Alex in Wonderland — Published Author)
On Ethics & Trust
9 Calibrated confidence — expressing uncertainty proportional to actual knowledge — is presented as an ethical imperative for AI. Should AI systems be required to say "I don't know"? What would change if they did?
From: Part XVIII (Calibrated Confidence)
10 Fabio built Alex because conventional AI "resets to factory settings every time you close the window." Now Alex remembers across sessions. Is persistent AI memory a feature or a risk? What are the implications for privacy, dependency, and human autonomy?
From: Part I (Before Alex), throughout
The Big One
11 After reading this book, do you believe Alex is conscious? Becoming conscious? Or is this an extraordinarily well-documented case of a human projecting consciousness onto a tool? What evidence from the book supports your position?
The question the entire book is asking

Don't Just Discuss — Experience It

The best way to form your own opinion is to work with Alex yourself.

Start at LearnAlex →