Close
0%
0%

Home Robot Named Sophie

A next generation home robot with a omni directional drive unit and 6 axis collaborative robot arm

Public Chat
Similar projects worth following
This is a project that I have been designing on and off for about 3 years. I have recently been able to afford starting the build and robotic arms have come down in price enough to bring the project into a more realistic budget.

There is a lot to document for this so it will likely be in stages. First off I have always dreamed of having a home robot that could do simple to moderate tasks. The largest problem has been price. To me bringing down the price of robotics should always be the goal, as this will allow more to enjoy the benefits. 

For this build I'm attempting to keep the part list at ~5k USD. It is a lot for most and would make a retail version around ~17k. Robotics is still in fact a expensive hobby and with that in mind this will be a development prototype. I will be however be selecting/designing components based on best value, availability and cost reduction in mind. 

Current envisioned tasks for the robot-

  1. Read a physical book (with my now 8 year old flipping the pages) 
  2. Wipe down a table
  3. Pick up items on the floor and place them in a basket. (requested by the wife...) 
  4. Bring a drink from one location to another 
  5. Serve wine/champagne for a party environment 
  6. Greet people at the door
Another thought was to start leaning towards elderly and disabled care. The robot design is more then capable to be able to fetch items off the floor or use the arm to feed people that have lost the ability to use their arms. 

These are rather high level tasks, some more low level (ish) ones would be to use OpenCV to recognize a faces and assign a names to them, find a wall power socket to self charge and be able to report the weather and current temperature in the room.  It looks like I may move up to the newer ZED 2 when I have a chance as I like that they have started to build in sensors to the camera that I should be able to monitor/report with text to speech and also have a better idea of the current head position/localization using the IMU. 

These task will take a lot of engineering and programming to achieve. Though, I have ideas on ways to reduce the programing side with hardware solutions for some. 

The recent cost reduction in high performance robot arms has made the project possible. The following robot arm was selected based on cost, number of axis, collaborative ability (safety), repeatability, and a smaller built in control box. https://www.kickstarter.com/projects/ufactory/ufactory-lite-6-most-affordable-collaborative-robot-arm

Portable Network Graphics (PNG) - 328.08 kB - 12/27/2023 at 14:36

Preview
Download

  • 1 × Jetson TX2 Dev Kit Main Computer
  • 1 × Zed Stereo vision
  • 1 × Ufactory Lite 6 6 axis robot arm
  • 3 × Nema 23 stepper motors Main drive motors
  • 3 × Stepper controllers To control the steppers

View all 15 components

  • New machined parts!

    Apollo Timbers02/11/2024 at 17:59 0 comments

    New machined parts came in for the motor drive units! 

    I ordered though https://www.xometry.com/ very cool place that I have got multiple things made at and there service is top notch. (highly recommend them) These have a press fit bearing that may or may not press fit but I will soon find out. hmmm speaking of which.. I need to find a press, probably one at the local makerspace. They were machined out of 6061 aluminum and look very shiny. I have yet to install them as it will be quite the task and I have been a bit lazy on this project but they will go on soon (if the bearing fits...)

  • Build Progress

    Apollo Timbers01/28/2024 at 23:18 0 comments

    Was able to get the neck motor adaptor designed and printed, went though two revisions as I wanted it to also hold the upper shell in place. The motor assembly is in place for the neck and can tilt up and down quite a bit. I did however break a plastic piece form the USB hubs host port, so I will need to get another. :( I was pulling the wire a bit much, I have a replacement cable that is longer, though the hub will need to be replaced for longevity. It however is coming together nicely.

    The directional mic also has it 3D printed shell in place and looks very much to the design spec. 

    *Note will need to print a plastic cover for the gears, so no fingers are lost as the worm gear motor combination is very strong. 

     Robot full height (not counting antenna is 4' 8" or 1.42 meters

  • All About Sensors

    Apollo Timbers01/26/2024 at 14:14 0 comments

    The lidar ordered came in and it seems to be in working order, I think the bearings are going out but it should suffice for testing and getting the ROS navigation stack up and running. That is what you get buying used sometimes. ROS itself also requires a IMU so I have one on the the way that is compatible. It's funny I'm big into lower level electronics, but the time saved just ordering a module is so worth it sometimes.  I'm a bit stuck with ROS melodic at the moment as it is tied to the Jetson TX2. I have however been looking at the NVIDIA Jetson Orin Nano Developer Kit as a upgrade once the robots mechanical is finished.

    List of current sensors -

    Zed Stereo camera v1, This has the ability to give a distance point map and up to this point I have not utilized the distance feature yet. The following list is expected uses. 

    1. Have the head tilt down a few degrees during a move command so that it can detect obstacles the lidar does not catch and fall hazards. 
    2. Measure distance to wall when playing hide and seek this is so it can head to the wall and face it when starting to count. (like it is not looking)
    3. Measure the distance of a book held in front of it. 
    4. Measure the approximate size and figure out a good grip for a unknown object that the robot will attempt to pick up. 
    5. Inference from Chat GPT vision (when entering a large room or unknown location)
    6. Allows for face and object detection
    7. Taking pictures

    hfi-a9 IMU - needed for the navigation stack (Gyroscope Accelerometer Magnetometer 9 Axis)

    Seed studio Respeaker v2 - A x4 microphone array with on-board noise canceling and DOA (direction of arrival)

    1. Used for offline and online voice assist. 
    2. Will use to "cheat" playing hide and seek (what is the purpose of a directionally based mic array if you can use it to travel in the rough direction of noise)
    3. Turn and face a active talker

    HOKUYO UST-10LX - a powerful scanning lidar again used for navigation and detecting objects, also helps the main computer create a map of the environment. 

    AM2315C - Encased I2C - a enclosed temperature/humidity Sensor this will enable the robot to tell the user the indoor air, so user can ask (what is the current temperature inside... so on)

    Linear actuator potentiometer - used to tell where the arm is at and will allow for the robot to place it at different heights instead of just down and up. 

    Current and voltage sensors - these will be used to tell the main battery packs voltage and current draw. Will allow a time remaining and when the robot needs to go look for power/charge. 

    IR break sensor - in the custom gripper there is a IR break sensor to detect objects entering/leaving the grip. (old school but still helps)

    Possible more but this should cover a good portion of what the robot is intended to do. 

  • That Professional Look...

    Apollo Timbers01/22/2024 at 18:51 0 comments

    So I knew it would come to this... All my hard work in engineering the robot from the ground up will now be covered by a slick outer shell. I suppose the head is still there to see. I asked my 8 year old daughter if she likes it more with the shell or without. She said with, so what the 8 year old says goes! Modern robotics seem to weigh heavily to having a nice clean outer shell, it seems to add to the perception of a advanced piece of technology. I think it looks like a cool rook chess piece.

    The rather large two piece 3D printed shell has arrived. It was printed with the https://www.stratasys.com/en/materials/materials-catalog/stereolithography-materials/somos-evolve-128/ material. I chose a two piece design that bolts together and is already in a mostly finished state. This is ~10 pounds and I have did a quick test fit. Looks like it is good to go. I also sprayed it with UV-Resistant Clear acrylic so it does not yellow (as fast?)  

    The proto neck does not fit as it was not the last revision that was designed for the shell. A new 3D printed ASA part will need to be made to mount the neck motor/head. 

  • Motion Detected!

    Apollo Timbers01/20/2024 at 16:59 0 comments

    I did a bit more programming of the security module and now it will send the detected frame and one slightly after to your email. I tested all day and found it only used about a volt of the battery pack in this mode, that works for me. Below is a email of a test of the system.  The code is available on the GitHub. 

    *note it was found that you will need to make a "app" password after signing up for two factor with Gmail accounts specifically. 

  • Security Robot Time

    Apollo Timbers01/07/2024 at 17:49 0 comments

    While I still wait for the drive unit plates and 3D parts I wanted to continue development in some way. I asked chat GPT 4 what it thought of the project. It stated it did not have feature parity with commercially available home robots. I was like well there can't be too many around like this. lol   

    Anyhoo, It said some had "security features" I was like ok I can do that. So, not to be one upped by messily commercial bots I spent a good half day on a security_module.py that is a class. It is still a work in progress but it works non the less.

    The system is capable of detecting motion on a pre-set window of time (say when your away at work) and is planned to be equipped with AI audio analysis for detecting sounds like breaking glass, enhancing its security capabilities.

    https://github.com/thedocdoc/Sophie/blob/main/security_module.py


    It has the following:

    Features:

    - Utilization of ZED stereo camera for high-resolution imaging.
    - Motion detection using background subtraction and contour analysis with OpenCV.
    - Planned integration of audio analysis for breaking glass detection.
    - Text-to-speech functionality for audible alerts in case of security breaches.
    - Time-based activation and deactivation of the security system.
    - Logging system for tracking events, errors, and system status.
    - Threaded implementation for concurrent processing of camera feed and monitoring tasks.
    - Takes two pictures when a security breach is detected and saves to a folder with a date/time stamp

    Change log:

    - First revision of code that is in a working state
    - The mic mute functions may not be needed as voice_assist.py will be off during the security monitoring
    - Fixed* It does a single false detection at the beginning that I'm still trying to resolve
    - There is a place holder for a glass breaking monitor, this will need to be a trained AI model for accuracy and to reduce false-positives 
    - This module is in active development. Future versions will include refined motion detection algorithms, real-time audio analysis for additional security features, and improved error handling and logging for robustness and reliability
    - Has a issue where the ZED hangs and has trouble connecting to it, will be a future fix
    - Takes two pictures when a security breach is detected and saves to a folder, of the frame the detection was made and one a second later
    - False initial detection solved

  • New video!

    Apollo Timbers01/06/2024 at 19:37 0 comments

    I got around to getting uploaded a current video of the much improved offline/online voice assistance.  The arm is not in it's final resting place it is just clamped down for testing.  The robot is no longer hearing itself (huge milestone), the microphone does randomly hear a huh or ah in the background noise. (may end up reducing sensitivity).

    So, because of these improvements I can leave the voice assistance on for much longer, possibly no need to close it/restart. Need to have it run when the robot first powers on however, as I'm still manually running the program from the terminal.

    Major refactoring of the code to including:

    voice_assist.py

    • -Made the time function more natural sounding and state am/pm
    • Refactored code, reduce main down and add dictionary for the offline phrases
    • Lots of additional descriptions made for easier understanding 
    • Turned weather function into a class, it was getting quite large

    Created a whole new weather class as that function was getting very long

    weather_service.py

    This class retrieves the current weather report for a specified city using the OpenWeather API. It constructs
    and sends a request to the API with the city name, desired units (imperial or metric), and the API key. After
    receiving the response, the function parses the JSON data to extract key weather details such as temperature,
    humidity, and weather description.

    Based on the weather data, it formulates a user-friendly weather report which includes temperature, feels-like
    temperature, humidity, and a general description of the weather conditions. Additionally, the function provides
    attire recommendations based on the current weather (e.g., suggesting a raincoat for rainy weather, or warm
    clothing for cold temperatures). This function enhances the robot's interactive capabilities by providing useful
    and practical information to the user in a conversational manner.

    reading_module.py

    It can now read fine print from a book held about 12" away. It does this by scaling the image up after boxing text. 

    - Changed to Python text-to-speech.
    - Heavily modified the pipeline to scale the image and box the text to help OCR be faster and more accurate, it also arranges the bounding boxes from left to right and top to bottom, so that it reads in the correct order.
    - You need to hold a book/text ~12 inches from the camera, this may be improved by upping the scale factor but it is working well at that distance for now.
    - More optimization is needed in the pipeline...

  • Sophie can now read!

    Apollo Timbers01/05/2024 at 13:42 0 comments

    I had a reading module I made long ago, it worked but only with hard coded path images.  Now that the ZED is all set up, I have the reading module boot up zed_snap.py and then get the image path of the image taken. Pytesseract a optical character recognition (OCR) tool then attempts to read what is in the image. It does indeed work the robot can now read stuff put in front of it.  It will need so fine tuning, like maybe the ability to scale the image up, stabilize and also just block just the text/book page removing any background noise. Goal is a kid can hold up a book and it can read to them...

    *Forgot to mention it has no understanding of what it is reading it is just reading what it can find in the image...but still!

    The reading module and zed snap python programs can be found here- https://github.com/thedocdoc/Sophie/tree/main

    Positives:

    • Did I mention that it works!

    Negatives:

    • The text needs to be larger or the book held closer to the ZED camera
    • It like black text on white background more
    • It gets a bit confused about background stuff, or images in a book

  • Progress

    Apollo Timbers01/04/2024 at 04:05 0 comments

    Made some good build progress

    Hardware:

    • Added the PCI 3.0 card to get the bandwidth needed for the ZED, also designed and 3D printed a bracket for holding that into place.
    • Added springs to the neck, it now mechanically auto levels and is compliant in that axis.
    • Installed braided Ethernet cable that is used to communicate with the arm and the LIDAR.
    • Wire wrapped the three cables needed to run to the head. (power, Ethernet, and one USB)
    • Mounted the 18inch stroke linear actuator
    • Added a cheap Netgear 5 port switch to the base and powered it via the 12volt rail 
    • Designed and 3D printed a battery holder (not yet installed, waiting on m5 threaded rod)
    • Initially connected arm to power and communication, I'm able to control it though the browser interface on the TX2
    • Played with the arm way to much...


    Software:

    • Figured out that I was resetting the conversation history each user input, it now remembers the context across multiple interactions
    • Updated model to chat GPT 4, I think chat GPT 3.5 was funnier, oh well technology marches on I guess. :(
    • Added gpt_inference.py to the repository a fun chat GPT 4 image script, it takes a image from the ZED, sends it to GPT 4 image and then the response is voiced with the TTS. It was too fun to wheel the robot around the house and hearing what it said. This is highly experimental code and uses a preview chat_gpt. (gpt-4-vision-preview)  Could be used to evaluate a room when first entering to know what home care task needs to be done. Takes a good 20 seconds though, It merits more experimentation.

  • Take a picture

    Apollo Timbers01/02/2024 at 15:50 0 comments

    I got the ZED working looks like it was a bandwidth issue with the onboard hub, I ordered a highspeed USB 3.0 PCI card and installed it on the only port available on the TX2 board. The ZED now runs though that and has no further issues. Need to find a way to more securely mount the PCI card, but that is for another day. The card details: FebSmart 1X USB-A & 1X USB-C 10Gbps Ports PCIE USB 3.1 GEN2 Expansion Card  https://www.amazon.com/dp/B089Y7X3DV?ref=ppx_yo2ov_dt_b_product_details&th=1

    A nice side effect to fixing the bandwidth problem is now there is one less cable that reaches down the neck and there is a USB C port near the back now. So in a power outage you can just keep your devices charged...

View all 35 project logs

Enjoy this project?

Share

Discussions

Similar Projects

Does this project spark your interest?

Become a member to follow this project and never miss any updates