Abstract

The transition from predictive text models (Large Language Models) to Artificial General Intelligence (AGI) requires a fundamental shift from disembodied algorithms to physically interactive entities. The recent experiment by David Vivancos and Dr. José Sánchez—transplanting the Neuraxon v2.0 bio-inspired brain into a Sphero Mini robot—marks a critical milestone in #AliveAI. This article breaks down the theoretical framework, the hardware architecture, and the methodology for replicating this experiment.

1. The Theoretical Framework: Trinary Logic and Neural Growth

Traditional Artificial Neural Networks (ANNs) operate on binary logic, using continuous calculations that ultimately simulate "on/off" states. Neuraxon v2.0 completely departs from this by utilizing Trinary Logic, a system specifically designed to mimic the biological reality of human synapses.

In a biological brain, neurons do not merely excite one another; they also actively suppress signals to filter out noise and focus on important tasks. Neuraxon introduces this explicit third state: Inhibition. In this system, every artificial neuron can exist in one of three distinct conditions:

  • Excitation (+1): Actively passing the signal forward.

  • Rest (0): Remaining neutral and conserving energy.

  • Inhibition (-1): Actively blocking or suppressing the signal path.

How does a neuron make a decision? Instead of relying on rigid, pre-programmed rules, each Neuraxon neuron calculates a "weighted score" based on all incoming signals from its neighbors. If this combined signal is strong enough and surpasses a specific activation threshold, the neuron fires an Excitatory signal. If the incoming signals are overwhelmingly suppressive, it fires an Inhibitory signal. Otherwise, it stays at Rest.

Unlike static LLMs, Neuraxon employs a "Neural Growth Blueprint." This means the "weight" or importance of these connections physically alters its own network topology based on real-world feedback. When the Sphero Mini robot hits a wall, the negative physical feedback literally rewires the network's connections for the next attempt.

2. Hardware Architecture: Why the Sphero Mini?

To test physical cognition, the AI requires a "body" with sensory input and motor output. The Sphero Mini, despite its accessible ~$50 price point, serves as a perfect minimally viable organism.

It is equipped with an Inertial Measurement Unit (IMU), which is crucial for the AI to understand physics (gravity, momentum, and spatial orientation).

  • Sensory Input (Afferent Pathways): The 3-axis gyroscope and 3-axis accelerometer feed real-time spatial data back to the Neuraxon brain.

  • Motor Output (Efferent Pathways): The AI calculates the required trinary signals to drive the internal dual-motor system, dictating speed and heading.

3. Experimental Methodology: Replicating the Setup

For researchers looking to experiment with open-science #AliveAI, the protocol is straightforward:

Step 1: Hardware Preparation

Acquire a Sphero Mini robot. Ensure it is fully charged and Bluetooth is enabled on your host machine.

Step 2: Access the Neuraxon Brain Interfaces

Navigate to the open-source Hugging Face spaces provided by David Vivancos:

  • For Locomotion (Neuraxon2MiniControl): This interface acts as the motor cortex, allowing you to observe how the neural network calculates basic navigation paths based on spatial input.

  • For Fine Motor Skills (Neuraxon2MiniWrite): This requires higher-level cognitive processing. The AI must calculate the exact physical trajectories, accounting for physical friction and momentum, to draw specific letters or words on a surface.

Step 3: The Feedback Loop

Connect the Sphero to the interface via the Web Bluetooth API. Do not simply execute commands; observe the neural growth. When the Sphero attempts to write a letter, monitor how the Neuraxon code (available on GitHub) processes the physical drift and attempts to correct its trajectory in subsequent movements.

4. Analytical Implications

This experiment proves that intelligence cannot be fully realized in a vacuum. By forcing the AI to interact with physical laws, Qubic and the Vivancos team are building the foundational nervous system for future robotics. Today, it drives a sphere; tomorrow, this exact trinary, bio-inspired architecture could regulate the complex kinematics of a humanoid robot.

Key Takeaways: The Future of #AliveAi

  • From "Dead" to "Alive" AI: Moving beyond static Large Language Models (LLMs), Neuraxon v2.0 introduces embodied cognition, allowing AI to learn and adapt through real-world physical interaction and failure.

  • Trinary Logic Superiority: By utilizing a -1 (Inhibit), 0 (Rest), and 1 (Excite) framework, Neuraxon mimics true biological brain efficiency, drastically reducing the computational waste seen in traditional binary systems.

  • Accessible Open Science: The integration with a $50 Sphero Mini robot democratizes AI testing. It proves that developing physical AI doesn't require multi-million-dollar robotics labs.

  • The Blueprint for AGI: Powered by the decentralized Qubic network, this "brain transplant" experiment lays the foundational nervous system for the complex kinematics of future humanoid robotics.

#Qubic #AGI

Neuraxon2MiniControl 👉https://huggingface.co/spaces/DavidVivancos/Neuraxon2MiniControl