Skip to content

ryanstout/micropilot

Repository files navigation

Self Driving (Toy) Car

MicroPilot is an autopilot for a DIY self driving (Toy) car, originally built during the Stem Club at St. John Paul the Great Catholic Highschool.

Put down masking tape and you can teach it to stay in the lines using only the frames from a Webcam.

driving1.mov
driving2.mov

The car does a great job of stays in the lanes. This was a fun project that the students seemed really excited about.

Due to limited time, we had to gloss over quite a few details, but I think this did a good job of getting them excited about computer science, engineering, and AI. There were lots of "where do I learn more" questions.

How It Works

MicroPilot takes in frames from an off the shelf webcam and predicts the steering angle.

We created a course by putting down masking tape for lanes. The goal was to get the car to follow the track and stay in the lanes using only the webcam footage as input.

It's probably possible to run these models on the Pi itself (especially the Pi5), but since there's no cooling in our setup, the Pi gets thermal throttled quite a bit. I found it easier to just connect the Pi to the Wifi, then record to and run inference on my machine. (Sending frames over a TCP connection)

Screenshot 2024-05-27 at 5 42 15 PM

To train the car, you attach a USB Joystick and drive the course. When the left upper button is held down, it records frames from the webcam and the steering angle of the joystick. The goal is to get a few types of training data:

  • Footage driving the center of the track
  • Footage of the car correcting when its pointing the wrong way. (So make the car face more towards the tape, then point the wheels hard in the correcting direction and start recording you correcting it back to the center) Also have the car slightly go over the lane and record data of it correcting back into the lane.

All recorded data should be with the wheels pointing in the direction you would use to correct the car back to the center. So let off the record button when you position the car closer to the edges.

The trick is also to make sure the wheel direction changes smoothly. It can be tempting to drive by toggling between strong adjustments and straight. For good training data, you'll need to make your steering adjustments smooth. As you get closer to the edge, make your adjustments strong back towards the center of the lane. The training data should convey that when you're in the middle you need little adjustments and as you get closer to the tape you need stronger adjustments.

Once we collected training data, we built a model that uses the features from a pretrained Vision Transformer (Vit). The features from the Vit are fed into a small model that uses a series of Linear layers to predict the steering angle. Finally a Tanh layer constraints the output to between -1 and 1.

The Hardware

Originally I had hoped we could build a car from scratch and 3d print a chassis, but as the semester went by, I realized time constraints meant we needed something that would work so we could focus more on the software/AI side. This project should work on any hardware, but v1 was built with a RaspberryPi 5, an off the shelf webcam and a kit I found on amazon for the car, motors, etc..

  • Raspberry Pi - We used a Pi5, originally thinking we would run the model on the car. Any Pi should work with the way we ended up doing it. Also, I went with the Pi because I figured the drivers and packages would be easy to install, but I actually feel like RaspberryPiOS makes things harder not easier. I ended up turning off their managed pip package thing. (This might be a debian thing now?)

  • SunFounder Smart Kit - A car kit with most everything you need. While this makes it easy to get started with, the car itself is very slow and the steering has a lot of play in it and top speed is ridiculously slow. Were I to start over I would explore other options or building something DIY.

  • Any Random Webcam - Any webcam that linux supports should be fine.

  • Game Controller - For driving the car while training it.

The Build

The SunFounder Smart Kit (PiCar) is a good start. I didn't end up using the "head" part of the car. Instead my son built a small stand out of legos for us to tape the webcam to. You want the webcam towards the front of the car, fixed in place, and a bit higher up so it can see the angle on the road lines. We faced the camera down just a bit.

How to Train

To simplify things I used tailscale to map the Pi into my network. config.py has an ip variable you can change.

Setup

I use poetry for the client and the server, but PiCarX needs to be installed manually since it's not on pip:

Follow the directions at https://github.com/sunfounder/picar-x

Then poetry install on the client (your machine) and the server (the Pi)

Collecting Training Data

Plug the USB Joystick into the car. Then make sure you've set ip in config.py, then run python train_server.py on the Pi and python train_client.py on your machine. (I used deploy.sh to rsync the project to the Pi)

Hold down the top left button to start recording. Make sure your steering angles are correct when recording. See the "How it Works" section for details on what good training data looks like.

Training the Model

Each time you run the train_client.py, it creates a new folder in data/. The folders will contain the webcam frame, and the filename will include the recorded steering angle.

SteeringDataModule in model.py takes in a list of directories with training. Update the paths on line 238 to the training folders you want to use. Then start training with:

python model.py

This will take some time :-) About 30 minutes on my M1 Mac before loss stops improving. We used PyTorch Lightning for the model and the excellent timm library for the pretrained Vit. There is also a version based on mobilenetv3 (which is a much smaller model, but it's not quite as good) Change the vit = True in config.py if you want to try the mobilenet version. You could definitely run the mobilenet version on the car itself. (And maybe the Vit also)

As the model trains, it will save the top epochs in the checkpoints folder. I also have it setup to log to wandb, but this is optional. Once you have a top performing checkpoint, you can update the model checkpoint path in config.py

Run Autopilot

If you're not on a Apple Silicon Mac, be sure to change your accelerator in config.py. To run the autopilot, put the car on the track and run python drive_server.py on the Pi, and python drive_client.py on your machine.

About

an autopilot for a self driving (Toy) car

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published