Figure 2 | Scientific Reports

Figure 2

From: Pedestrian orientation dynamics from high-fidelity measurements

Figure 2

(a, c) Pedestrian trajectories (purple) superimposed to depth snapshots (gray). Orientation estimates and local velocities (directions of motion) are reported, respectively, in red and yellow. We estimate shoulder orientation on a snapshot-by-snapshot basis, considering depth “imagelets” centered on a pedestrian. The sub-panel (b) reports an example of such an imagelet with the \(x-y\) coordinate system considered. We employ instantaneous direction of motion \(\theta _v\) extracted from preexisting trajectory data as training labels for a neural network. This yields a reliable estimator for the orientation \(\theta\), accurate even in cases challenging for humans, like in (c). Due to clothing, arms and body posture, presence of bag-packs or errors in depth reconstruction, the overhead pedestrian shape might appear substantially different from an ellipse elongated in the direction of the shoulders.

Back to article page