In the demo below, our prototype uses trial & error to learn a robotics task from only 3 examples.
Guesses theories about what's there and how it works.
Tests its theories by interacting with its environment.
Learns from its mistakes by revising its theories.
Based on principles from philosophy. No deep learning or statistics.
Sight, sound, etc.
One recording is the "training" episode; the other is the "testing" episode.
The agent's core principles were developed independent of the task, and are the foundation for solving more complex tasks.
The sensory input at t0
is the first "data" ever seen by the agent.
First, the agent looks for any new problems in its theories and attempts to reconcile them.
In the first timestep, the agent has no theories yet, so there are no problems to reconcile.
We choose the goal for now, and we can modify the goal at any time.
It chooses a strategy for reaching its goal. If it doesn't have any strategies, it needs to invent one.
This time, the agent's sensory input reveals a problem with its theories. It reconciles the problem by rejecting one or more theories.
This time, it can rule out more theories because it has more sensory experience.
t3
, it reaches the goal for the first time.But it doesn't yet understand how it reached the goal...
It discards the problematic theories despite reaching the goal.
Let's skip ahead to when it reaches the goal for the third time, at t8
.
t8
, the agent reaches the goal for the third time...Importantly, it didn't experience any problems on the way to the goal.
It knows how to reach its goal without encountering any problems.
A problem can come from an error in its theories or from not-knowing how to reach its goal.
If the agent doesn't discover any new problems, then it has no "fuel" to continue learning.
When the agent's theories are good-enough to reliably reach the goal without encountering problems, then learning is complete.
The agent retains all of its knowledge from training.
It needs to conjecture theories about what's there rather than how it works.
The agent got lucky and guessed its situation on the first try.
This indicates a threshold level of understanding of this situation, simulation, and task.
Tasks that require less abstraction and reasoning, e.g. assembly, welding, factory tasks.
Tasks that require more abstraction and reasoning, e.g. language, math, decision-making, multi-step processes.
Tasks that require creating and testing new designs or new theories.
We've been able to make meaningful progress in a relatively short time, and we want to reach this threshold as quickly as possible.
There are many interesting problems ahead of us.
Weโll systematically work to address increasingly complex tasks and environments.
Contact Collin Kindrom for more information, or with any other questions or ideas.
collinkindrom@gmail.com