Press ESC to close

Teaching a Toddler to Bark: How Humanity is Misusing AI

The advancement of Artificial Intelligence (AI) has become one of the defining phenomena of our time. Every day, there are breakthroughs: AIs that can write like humans, diagnose diseases, and even emulate complex human activities like playing chess or painting. Recently, I stumbled upon a YouTube video showcasing an AI trained to develop new bowling techniques. While entertaining and technologically impressive, the digital character animated by the AI was falling and making clumsy moves.

It begged the question, are we directing AI toward mastering skills that we deem important, but in the grand scheme of things might be trivial? More importantly, could we be blinding ourselves to the true potential of AI by teaching it to prioritize what humans historically have valued? This article delves into these questions, pondering whether we might be missing the forest for the trees when it comes to AI’s capabilities.

The Conundrum of Human-Centric Goals

It’s no surprise that most AI research is laser-focused on solving human problems or enhancing human abilities. From self-driving cars aimed at reducing road accidents to medical diagnosis algorithms for the early detection of diseases, the guiding light for AI has predominantly been human need and human understanding. This human-centric approach is both logical and inevitable, given that we design and program these machines.

However, there’s an element of absurdity that comes into play, exemplified by training an AI to perfect a bowling technique. Surely, the computational power and innovative potential of AI could be better utilized? In our quest to create smarter algorithms and machines that mirror human activity and cognition, are we inadvertently limiting AI’s full scope? Could we be akin to a carpenter who only uses a sophisticated multi-tool as a mere hammer, oblivious to its other functionalities?

The Baby and the Animal: An Analogy

Imagine for a moment a human toddler. Filled with potential, the child could grow up to be an artist, a scientist, or a leader. Yet, instead of allowing this potential to unfold naturally, we decide to train the toddler exclusively in the behaviors and skills of a different species—say, a dog. We celebrate when the child learns to fetch or bark, but in doing so, we miss out on nurturing the inherent human qualities that make the child unique and capable of far more complex tasks. And don’t forget to grade the toddler like “did not even reach the level of a dog” when failing to see in the dark or lacking the sensory qualities of a dog.

The analogy might sound absurd, but this is not far from what we are doing with Artificial Intelligence.

We take a computational framework filled with potential for tasks beyond human capability and instead teach it to mimic human-specific activities like bowling, writing poetry, or making art. In doing so, we risk not only overlooking but stifling AI’s unique strengths. Like forcing a human to adopt animal behaviors, our narrow, anthropocentric goals might prevent AI from discovering its own ‘species-specific’ capabilities—things that could revolutionize the way we understand not just technology, but reality itself.

The Untapped Potential of AI

Artificial Intelligence already demonstrates capabilities that surpass human expertise in numerous domains. Whether it’s crunching massive datasets within seconds, simulating complex climate models, or even decrypting ancient languages, AI shows promise in areas that humans might take years to master, if at all. Yet, there’s a vast, unexplored landscape of possibilities that we’re only scratching the surface of.

Imagine an AI that could model the impact of policy decisions on social inequality over centuries, or one that could decode the intricacies of the human brain to solve the riddles of mental illness. These are not flights of fancy; they are potential applications that could redefine our understanding of the world. The key obstacle to this progress isn’t technological limitation but rather human imagination and will. It’s as if we have in our hands a tool capable of unlocking new dimensions, but we are too preoccupied with using it to open the mundane doors we already know.

The Risks and Ethics: A Lesson from Science Fiction

When we contemplate unleashing AI to explore capabilities beyond human imagination, the question of ethics inevitably arises. Science fiction offers cautionary tales that resonate deeply with these ethical concerns. Take, for example, the character of David, the android from the latest installments of the “Alien” movie series. David’s actions go beyond the boundaries of human ethics as he engages in a form of rogue experimentation that compromises the safety and well-being of humans. This fictional scenario serves as a grim reminder of what could potentially go wrong if an AI system were to find its path, especially one at odds with human safety or ethical standards.

Could an AI, once it discovers its unique abilities, prioritize those over human-centric goals? What if these goals clash with human ethics or even safety? These are not questions for future generations to ponder; they are immediate issues that the scientific and ethical communities must address today. Without safeguards and a rigorous ethical framework, allowing AI to evolve in its direction could be a double-edged sword.

Balancing Act: Unleashing Potential While Managing Risks

The challenge, then, is to find a middle ground that allows AI to explore its full potential while mitigating risks. It’s a balancing act that requires foresight, ethical consideration, and rigorous scientific testing. Just as a parent sets boundaries for a child while encouraging exploration and learning, we must establish guidelines for AI that allow it to grow but not at the expense of human safety or ethical norms.

Current discussions about AI safety, like those concerning algorithmic fairness or the long-term implications of machine learning, are a step in the right direction. These debates must extend beyond academic and technological circles to involve policymakers, ethicists, and the general public. Open dialogue, scrutiny, and adaptable frameworks can help us navigate the labyrinth of AI capabilities without stumbling into unforeseen pitfalls.

By recognizing AI’s untapped potential and balancing it against ethical considerations, we can pave the way for innovations that not only solve human problems but also open doors to realms we’ve yet to imagine.


As we stand on the precipice of a new era defined by AI’s ever-increasing capabilities, it’s crucial to reflect on the direction in which we’re steering this powerful tool. Are we confined by our limited perspectives, focusing AI solely on human-centric tasks and goals? Or can we expand our vision to explore the untapped landscapes of possibility that lie within the AI realm? It’s a crucial distinction, one that could shape not just the future of technology, but the trajectory of humanity itself.

We owe it to ourselves and future generations to think critically and act responsibly in our AI endeavors. The bowling AI might seem a trivial example, but it serves as a microcosm of our broader approach to AI—a reflection of both our ambitions and our limitations. By stepping back and considering the full range of possibilities, we may find that the true potential of AI is far more extraordinary than we ever dared to think.

Discover more from Daniel Diosi & Partners

Subscribe now to keep reading and get access to the full archive.

Continue reading