AVA - Advanced Verbal Assistant Robot


This is Ava.  This is my 3rd bot and is the product of 3 years of work.  Ava combines advanced software, hardware, 3D sensors, machine learning models, and 22-degree of freedom movement to deliver what I believe is my smartest, most capable, and hopefully best looking bot yet. 

This bot uses visual SLAM to map and locate itself.  I plan on using this bot to demonstrate what is possible with state of the art natural language and vision processing using a combination of a custom neural net based architecture and advanced "transformer" based machine leaning models.


A Few Pictures…


Advanced Neural Net Architecture

Conceptually, Ava's brain is a neural net of neural nets.  Each node in the high-level brain can accept inputs, have custom activation rules, and send outputs to other nodes.  Every node has a name and one or more aliases.  Any node can also be autonomous.  Internally each node can have its own internal neural net, fuzzy logic rules, verbal "English Paragraph" script, or procedural code.


Natural Language Orchestration

People can talk to the bot using language.  Because every node in the brain has a name, a person can talk to any node and get, set, increase, decrease, etc. its value.  In the same way, nodes can talk to each other...using natural language.  This allows complex operations (like coordinated movements) to be performed using simple sentences and paragraphs.



Ava is self-learning.  It builds knowledge graphs from linked open data sources on the web (such as dbpedia, wikipedia, concept-net etc), databases, APIs, etc.  Ava integrates all the data for any given topic into a single graph.  Ava can build a graph around people, places, users, or any given word or topic.


Natural Language Question Answering & Comprehension

Based on the knowledge graphs, Ava can then answer natural language questions.  Ava can also answer questions based on text sources using a "transformer" BERT-based model for question answering.  Ava can also compare and contrast topics..."Compare hydrogen and helium."


Machine Learning Models

Currently, this bot uses several off-the shelf machine learning models (from the Intel Model Zoo or HuggingFace) to do object recognition (like Yolov3) and other visual functions as well as many natural language-based functions.  Over time, more models will be added.  Some of these models run on the neural compute stick and other run on the LattePanda.

1 Latte Panda Alpha 864
1 Intel Movidius Neural Compute Stick 2
1 Arduino Mega ADK
1 Arduino Mega Sensor Board
1 SSC-32U Servo Controller
1 Sabertooth 2x5 Motor Controller
1 Orbec Astra S 3D Depth Camera (for color and depth)
1 Intel T265 Tracking Camera (SLAM)
1 IMU (on the T265)
1 Tilt Compensated Compass
2 Adafruit AMG8833 Thermal Cameras
12 Devantech SRF-01 Sonars
2 Sharp IR Distance Sensors
3 Microphones
1 Bluetooth Keyboard, Mouse, Game Controller, Earpiece
1 Android-based voice remote
2 12V Planetary Gear Motors
8 Actuonix Linear Actuators
4 Servo-City Gearboxes and Servos
4 Hobby Servos
1 24-LED NeoPixel Ring in Chest
4 7-LED NeoPixel Jewels
1 5in HDMI screen
2 Lasers
All Rights