The Problem: Fences Hold Factories Back

In most factories today, robots work behind physical barriers — fences, cages, light curtains. The moment a person steps too close, the robot stops. It’s safe, but it’s also rigid. These barriers waste valuable floor space, slow down workflows, and make it impossible for humans and robots to truly work side by side.

The Cynergy4MIE project is working to change that. The goal: remove the fences and replace them with something smarter — a system that always knows where every person is.

Our Approach: Seeing Through Many Eyes

Instead of physical barriers, we use multiple cameras positioned around the workspace, each covering the area from a different angle. An AI model (a neural network called YOLO) analyzes every camera feed in real time, detecting each person visible in the frame.

But detecting people in camera images is only the first step. The system then projects each detection onto a floor plan — converting pixel positions into real-world coordinates in meters. Since multiple cameras often see the same person from different viewpoints, a fusion algorithm combines all observations into a single, consistent map.

The result: a live bird’s-eye view of the workspace showing exactly where every person is, updated several times per second. Each person receives a persistent identity — the system doesn’t just see “a person,” it knows it’s the same person moving through the space.

To make tracking even more robust, the system can also fuse camera data with workers authorization, based for example on Ultra-Wideband (UWB) or RFID. Small wireless tags worn by workers authorize or even localize them, which is combined with the camera-based tracking. If realized by UWB, this adds a layer of redundancy — if a person is temporarily hidden from all cameras, the UWB signal can keep their position updated until they become visible again.

Testing in a Virtual Factory

Before installing real cameras, we needed a way to develop and test the tracking system quickly and safely. So we built a virtual factory — a 3D simulation of an industrial workspace, complete with walls, obstacles, and animated people walking around.

The virtual cameras in this simulation stream video to our tracking system in exactly the same way real cameras would. This means we can experiment freely: try different camera placements, add more people, rearrange obstacles — all without touching any hardware.

The video below shows the system in action. On the top, you can see the four camera views with detected people highlighted. On the bottom, the room map displays each person’s tracked position in real time as colored dots with motion trails.

The multi-camera tracking system running in a simulated factory. Top: four camera views with AI-detected people (green boxes). Bottom: real-time room map showing each person’s tracked position.

This simulation-first approach lets us push the tracking algorithms to their limits and verify reliability before moving to real-world deployment.

What’s Next

With the tracking system proven in simulation, the next step is deployment on real hardware — industrial Basler cameras in a physical testbed. The long-term vision is a workspace where robots can move and operate freely alongside people, guided by a reliable, always-on awareness of their human coworkers.

This work is part of Cynergy4MIE, a European collaborative project bringing together partners from across industry and academia to enable smarter, safer manufacturing through advanced sensing and AI.

Blog signed by: BUT team