Hi there! I'm passionate about exploring the fascinating intersection of human cognition and machine learning. This blog is my personal space to share thoughts, research notes, and speculations about where AI might be heading.
I'm particularly interested in moving beyond pattern recognition toward genuine abstraction and reasoning. Through my writing, I explore how intelligence seems to require both the fluidity to synthesize new approaches and the ability to maintain domain-independent representations.
The paradigm of goal-directed optimization has dominated AI research for decades, but I've been contemplating whether this framework fundamentally limits the kind of intelligence we can create. Human cognition rarely optimizes for singular, well-defined objectives—rather, it engages in a perpetual dance of curiosity and compression, constantly reorganizing its internal models to accommodate novel experiences.
Jürgen Schmidhuber's formal theory of creativity and curiosity offers an intriguing alternative: systems that seek to maximize "compression progress" rather than external rewards. Such systems essentially find patterns that allow them to better predict and understand their environments. This process creates a natural curriculum, as the agent continually seeks phenomena at the edge of its current understanding—not too familiar, not too chaotic.
What fascinates me is how this approach mirrors the intrinsic motivation observed in human development. Children don't explore their environments to maximize some utility function; they're drawn to the novel, the surprising, the pattern-breaking. This inherent drive toward "interestingness" may be what enables the open-ended nature of human intelligence.
If we're to move beyond the current limitations of AI systems—their brittleness outside narrow domains, their lack of common sense—perhaps we need to embrace this open-endedness. A truly intelligent system might not be one that excels at optimizing specific metrics, but one that continually reorganizes its understanding, forever dancing at the boundary between the known and unknown, the certain and the doubtful.
The notion of "interestingness" presents a profound philosophical puzzle at the heart of intelligence. When we examine the raw sensory data streaming into our consciousness, we find an effectively infinite sea of potential patterns—yet only certain configurations capture our attention. What makes some regularities worthy of notice while others fade into the background of our perception?
I've been wrestling with this question through the lens of information theory and phenomenology. From an information-theoretic perspective, patterns that offer maximum compression efficiency should be the most valuable—they allow us to encode more of the world with less cognitive resource. Yet this fails to account for the subjective, contextual nature of what we find interesting. The pattern of cracks on my ceiling might contain as much algorithmic information as a page of profound philosophy, yet one captivates while the other barely registers.
Merleau-Ponty's phenomenology offers an intriguing complementary perspective: perhaps interestingness emerges not from properties inherent to patterns themselves, but from how they relate to our embodied being-in-the-world. Patterns matter when they afford possibilities for action, when they connect to our projects and concerns, when they reconfigure our understanding of ourselves as agents navigating a complex environment.
For machine intelligence, this suggests a profound limitation in current approaches. A truly general intelligence might require something beyond pattern recognition capabilities—it may need a form of embodied concern, a stake in the world that makes some patterns matter more than others. This isn't merely about programming in "values" or "preferences," but about creating systems with a phenomenological relationship to their environment that generates an authentic interestingness function. Such systems would find patterns meaningful not because they've been programmed to value certain objectives, but because these patterns genuinely reconfigure their possibilities for being.
Western intellectual tradition has long maintained a dichotomy between reasoning and learning—deduction versus induction, a priori versus a posteriori knowledge, rationalism versus empiricism. This division persists in AI research, where symbolic reasoning systems and statistical learning approaches have developed as largely separate paradigms. But I've become increasingly convinced that this dichotomy fundamentally mischaracterizes the nature of intelligence.
Consider how we actually engage with novel problems. We don't merely apply formal logical operations to static symbolic representations, nor do we simply recognize patterns from past experience. Instead, we engage in a fluid process of moving between levels of abstraction, reframing problems, drawing analogies across domains, and synthesizing new approaches. Our conceptual representations aren't fixed symbols, but dynamic, context-sensitive constructs that continuously evolve through experience.
Douglas Hofstadter's work on fluid analogies illuminates this beautifully—showing how analogical thinking blurs the boundary between reasoning and learning. When we map structural patterns from one domain to another, are we reasoning from first principles or learning from experience? The question itself presupposes a false dichotomy. Perhaps reasoning is simply a higher-order form of pattern recognition, operating on more abstract representations extracted from lower-level learning processes.
This perspective has profound implications for AI architecture. Rather than pursuing the integration of separate reasoning and learning modules, we might be better served by developing unified systems where reasoning emerges naturally from hierarchical learning processes. The future of intelligence might not lie in building systems that either learn or reason, but in creating architectures that continuously extract increasingly abstract patterns from experience, building representations that simultaneously encode both specific experiences and general principles. In such systems, reasoning and learning wouldn't be distinct operations but different perspectives on the same underlying cognitive dynamics—just as they seem to be in human thought.
I'm just someone fascinated by machine intelligence and its implications. Broadly, I study literature, logic and engineering. I am also interested in comedy, cinema, football and a bucket of other things that keep shifting here and there. This blog is my personal project - a place to organize my thoughts and connect with others who share similar interests.
Also check out my other blog: banrovegrie.github.io where I explore related topics through a different lens.