Sparkfun Edge, MacOS X, FTDI

More for my reference than anything else. I’ve been trying to get the toolchain set up to use a Sparkfun Edge. I had the Edge, the Beefy3 FTDI breakout, and a working USB cable.

Blurry pic of cats taken using Sparkfun Edge and HIMAX camera

This worked great for the speech example, for me (although the actual tensorflow part never understands my “yes” “no” etc, but anyway, I was able to successfully upload it)

$ git clone --depth 1
$ cd tensorflow
$ gmake -f tensorflow/lite/micro/tools/make/Makefile TARGET=sparkfun_edge micro_speech_bin
$ cp tensorflow/lite/micro/tools/make/downloads/AmbiqSuite-Rel2.2.0/tools/apollo3_scripts/ tensorflow/lite/micro/tools/make/downloads/AmbiqSuite-Rel2.2.0/tools/apollo3_scripts/
$ python3 tensorflow/lite/micro/tools/make/downloads/AmbiqSuite-Rel2.2.0/tools/apollo3_scripts/ --bin tensorflow/lite/micro/tools/make/gen/sparkfun_edge_cortex-m4_micro/bin/micro_speech.bin --load-address 0xC000 --magic-num 0xCB -o main_nonsecure_ota --version 0x0
$ python3 tensorflow/lite/micro/tools/make/downloads/AmbiqSuite-Rel2.2.0/tools/apollo3_scripts/ --load-address 0x20000 --bin main_nonsecure_ota.bin -i 6 -o main_nonsecure_wire --options 0x1
$ export BAUD_RATE=921600
$ export DEVICENAME=/dev/cu.usbserial-DN06A1HD
$ python3 tensorflow/lite/micro/tools/make/downloads/AmbiqSuite-Rel2.2.0/tools/apollo3_scripts/ -b ${BAUD_RATE} ${DEVICENAME} -r 1 -f main_nonsecure_wire.bin -i 6

But then I couldn’t figure out how to generalise it to use other examples – I wanted to use the camera because ages ago I bought a load of tiny cameras to use with the Edge.

So I tried this guide, but couldn’t figure out where it the installer had put the compiler. Seems basic but….??

So in the end I used the first instructions to download the tools, and then the second to actually do the compilation and installation on the board.

$ find . | grep lis2dh12_accelerometer_uart
# you might need this - 
# mv tools/apollo3_scripts/ tools/apollo3_scripts/ 
$ cd ./tensorflow/lite/micro/tools/make/downloads/AmbiqSuite-Rel2.2.0/boards_sfe/edge/examples/lis2dh12_accelerometer_uart/gcc/
$ export PATH="/Users/libbym/personal/mayke2021/tensorflow/tensorflow/lite/micro/tools/make/downloads/gcc_embedded/bin/:$PATH"
$ make clean
$ make COM_PORT=/dev/cu.usbserial-DN06A1HD bootload_asb ASB_UPLOAD_BAUD=921600

etc. Your COM port will be different, find it using

ls /dev/cu*

If like me the FTDI serial port KEEPS VANISHING ARGH – this may help (I’d installed 3rd party FTDI drivers ages ago and they were conflicting with the Apple’s ones. Maybe. Or the reboot fixed it. No idea).

Then you have to use a serial programme to get the image. I used the arduino serial since it was there and then copy and pasted the output into a textfile, at which point you can use


to convert it to a png. Palavers.

ESP32 M5StickC, https, websockets, and Slack

I got one of these lovely M5StickCs for a present, and had a play with it as part of Makevember. I wanted to make a “push puppet” (one of those toys that you push upwards and they collapse) that reacted to Slack commands. Not for any reason really, though I like the idea of tiny colleagues that stand up when addressed on slack. Makevember doesn’t need a reason. Or at any rate, it doesn’t need a good reason.

Here are some notes about https and websockets on the ESP32 pico which is the underlying board for the M5StickC.

I made a “slack wobbler” a couple of years ago, also in makevember – an ESP8266 that connected to slack, then wobbled when someone was mentioned, using a servo. Since then I ran into some https problems, obviously also encountered by Jeremy21212121 who fixed it using a modified version of a websockets server. This works for the ESP8266 – turns out you can also get the same result using httpsClient.setInsecure() using BearSSL. I’ve put an example of that here.

For ESP32 it seems a bit different. As far as I can tell you need the certificate not the fingerprint in this case. You can get it using openssl s_client -connect

For ESP32 you also need to use the correct libraries for wifi and wifimulti. The websocket client library is this one.

And a final note – the M5StickC is very cool but doesn’t enable you to use many of its GPIO ports. The only one I can find that allows you to use a servo directly is on the Grove connector, which I bodged some female jumper wires into, though you can get a grove to servo converter (there are various M5Stick hats you can use for multiple servos). Here’s some code. And a video.

Sock-puppet – an improved, simpler presence robot

Makevember and lockdown have encouraged me to make an improved version of libbybot, which is a physical version of a person for remote participation. I’m trying to think of a better name – she’s not all about representing me, obviously, but anyone who can’t be somewhere but wants to participate. [update Jan 15: she’s now called “sock-puppet”].

This one is much, much simpler to make, thanks to the addition of a pan-tilt hat and a simpler body. It’s also more expressive thanks to these lovely little 5*5 led matrixes.

Her main feature is that – using a laptop or phone – you can see, hear and speak to people in a different physical place to you. I used to use a version of this at work to be in meetings when I was the only remote participant. That’s not much use now of course. But perhaps in the future it might make sense for some people to be remote and some present.

New recent features:

  • easy to make*
  • wears clothes**
  • googly eyes
  • expressive mouth (moves when the remote participant is speaking, can be happy, sad, etc, whatever can be expressed in 25 pixels)
  • can be “told” wifi details using QR codes
  • can move her head a bit (up / down / left / right)

* ish
**a sock

I’m still writing docs, but the repo is here.

Libbybot-lite – portrait by Damian

NetworkManager, dhcpcd, Raspberry Pi Zero Gadget and Balena AP

This is probably pretty niche.

tl;dr: you can keep gadget mode working on the Pi Zero with Balena’s soft-AP mode if you disable dhcpcd only for wifi (see how below).

Jasmine, Henry and I are doing a workshop at work (we’re making Radiodan-neues). We’ve got 20 people making their own, using Raspberry Pi Zeros. Some of them have no experience with Pis, command line interfaces, linux etc, so we wanted to make getting the device on the wifi network easy.

This is something we’ve been going on about in Radiodan for years – it’s the difficult part of making a device from a Pi that’s end-user-ready. If you’re happy with the command line and text files you can create a wpa_supplicant.conf file, put it in /boot while the SD card is in your laptop and all is well (it gets copied to the correct place on the Pi on boot). But we thought in our case it would be tough to lead 20 people through that process remotely.

Wifi-enabled Arduino-like things like ESP 8266 and ESP 32 can be made to create their own access point, offer up a webpage that you can enter your wifi credentials into, then join the specified wifi network using those creds (“soft-AP mode“). Andrew Nicolaou modified Balena’s similar technique to work with the Pi without Docker or their infrastructure.

The way Balena do this trick is by using NetworkManager (and some other things like a web server) wrapped in some Rust code. It only offers the access points that NetworkManager finds on a scan, and you can’t add your own wifi network if it’s not on the list. For some reason, only on a Pi Zero, my wifi network doesn’t show up. Lots of other networks show up, it’s just mine. :head-desk-emoji:.

So I’ve been setting up Balena’s technique using a Zero in the amazing gadget mode (which means you can access the device via USB). Because it can’t find my wifi network, when I test it by unplugging the USB, I’ve essentially bricked that card (I can still access it over its own AP but I can’t do anything that requires going on a network). Gadget networking stops working, I’m guessing because it depends on dhcpcd which is uninstalled by the Balena code, because it’s not compatible with NetworkManager, at least on the Pi.

So, we were looking for way to keep the Balena AP stuff but also have a back up method, e.g. the wpa_supplicant file technique, or the gadget network.

You can keep gadget mode working if you disable dhcpcd only for wifi. Which is dead easy – edit /etc/dhcpcd.conf and add denyinterfaces wlan0 to the bottom, and restart dhcpcd: sudo systemctl restart dhcpcd.service

It took me so long to figure that out. I do not enjoy linux networking stuff.

Digging into it a bit more:

You can’t use the wpa_supplicant.conf trick with Balena. This is because NetworkManager (and so Balena) can’t access the wifi when dhcpcd has control over it, and this is how wpa_supplicant works on the Pi, via dhcpcd. With NetworkManager dhcpcd both enabled, you get wlan0 unavailable:

$ nmcli dev status
 wlan0   wifi      unavailable  --
 usb0    ethernet  unmanaged    --
 lo      loopback  unmanaged    --

But we can use gadget mode and that’s good enough for us I think: we can rescue any people having trouble with wifi and lead them individually through gadget mode, but we don’t have to do that for everyone.

The final step is to include a bash scipt like this and run it on startup e.g. with systemd:

 sleep 30s
 iwgetid -r
 if [ $? -eq 0 ]; then
     printf 'Got wifi - skipping WiFi Connect\n'
     printf 'No wifi, trying ethernet'
     output=$(curl -Is | head -n 1)
     if [[ $output =~ $pat ]]; then  # we have ethernet
       printf 'Got ethernet - skipping WiFi Connect\n'
       printf 'No wifi or ethernet found - starting WiFi Connect\n'
       sudo wifi-connect -o 7777
       if [ $? -eq 0 ]; then
          printf 'ok\n'

An additional note: I discovered that if the Zero is in gadget mode then you can access its Samba-enabled directories, so you could for example, allow people to use Samba to drag a wpa_supplicant.conf file to the /boot directory. There’s an example of how to enable Samba on a Pi which Dan Nuttall set up for radiodan-neue here. This is no good for us, but pretty cool nonetheless.

Libbybot – a posable remote presence bot made from a Raspberry Pi 3 – updates

A couple of people have asked me about my presence-robot-in-a-lamp, libbybot – unsurprising at the moment maybe – so I’ve updated the code in github to use the most recent RTCMultiConnection (webRTC) library and done a general tidy up.

I gave a presentation at EMFCamp about it a couple of years ago – here are the slides: