João Victor "Teddy" Martins

Hello everyone, I'm João (but everyone calls me Teddy), and I'm a Field Application Engineer at Toradex! I’m an Computer Engineering undergrad (finishing this semester!) with a deep passion for electrical systems and embedded computing. What really excites me is the intersection between hardware and software.
My journey into embedded systems started with a classic: with Raspberry Pis and emulators. That curiosity gradually evolved into a professional path focused on System-on-Modules, BSP customization, Yocto builds, device tree debugging, and performance optimization for real-world products.
Outside of work I've always been a DIY person with a love for the outdoors. From camping and gardening to track days with an old Ford Escort 96, for which I even am developing a telemetry platform based on the AM62 from TI!
I love understanding how things work and turning ideas into projects into reality! I'll always be chasing the next project!


Session

05-27
16:15
40min
From Track to Edge: Shipping Real-Time AI on Embedded Linux
João Victor "Teddy" Martins

Modern motorsport telemetry systems generate massive amounts of data including GPS traces, IMU measurements, CAN signals, and vehicle dynamics. In most cases, analysis happens after the session, often in the cloud. By the time insights are available, the opportunity to correct driving behavior in real time is already gone. For deterministic feedback during a session, cloud-dependent approaches are too slow, too fragile, and sometimes simply unavailable.

In this talk, we walk through the engineering journey of building a real-time telemetry analysis system that runs entirely at the edge on embedded Linux. The objective was straightforward: detect driving patterns and performance anomalies during a session without relying on connectivity. Achieving that goal required solving a set of practical system-level challenges that extend far beyond data acquisition and model training.

We begin with the development pipeline: training a model offline, exporting to ONNX or TFLite, quantizing for constrained hardware, and deploying to embedded System-on-Modules. We compare CPU-only execution against NPU acceleration, highlighting latency, memory footprint, and sustained-load behavior. Real benchmark results demonstrate where hardware acceleration delivers measurable gains and where it introduces additional constraints.

Running inference once is not the hard part. Shipping a complete embedded systems product is.

The talk then focuses on the integration and production aspects of edge AI systems. We examine kernel driver and user-space runtime alignment, accelerator operator support limitations, memory pressure under sustained workloads, and thermal behavior during continuous inference. We discuss containerized deployment on embedded Linux, using Torizon OS as a reference implementation, including hardware access from containers, separation of sensor ingestion and inference pipelines, reproducible builds, and safe over-the-air model updates without reflashing the device.

By the end of the session, attendees will have a practical blueprint for taking an AI model from experimentation to a production-ready embedded deployment. More importantly, they will gain an honest understanding of what breaks, what scales, and what must be designed early when building real-time intelligence on embedded Linux systems.

This is not a showcase of AI capabilities, but a systems engineering story about building, benchmarking, integrating, and maintaining edge AI under real-world constraints.

Auditorium