## Abstract

We investigate in real-life conditions and with very high accuracy the dynamics of body rotation, or yawing, of walking pedestrians—a highly complex task due to the wide variety in shapes, postures and walking gestures. We propose a novel measurement method based on a deep neural architecture that we train on the basis of generic physical properties of the motion of pedestrians. Specifically, we leverage on the strong statistical correlation between individual velocity and body orientation: the velocity direction is typically orthogonal with respect to the shoulder line. We make the reasonable assumption that this approximation, although instantaneously slightly imperfect, is correct on average. This enables us to use velocity data as training labels for a highly-accurate point-estimator of individual orientation, that we can train with no dedicated annotation labor. We discuss the measurement accuracy and show the error scaling, both on synthetic and real-life data: we show that our method is capable of estimating orientation with an error as low as \(7.5^\circ\). This tool opens up new possibilities in the studies of human crowd dynamics where orientation is key. By analyzing the dynamics of body rotation in real-life conditions, we show that the instantaneous velocity direction can be described by the combination of orientation and a random delay, where randomness is provided by an Ornstein–Uhlenbeck process centered on an average delay of \(100\,\hbox {ms}\). Quantifying these dynamics could have only been possible thanks to a tool as precise as that proposed.

## Introduction

The orientation of our body and shoulder-line changes continuously as we walk. When our gait is regular, these changes are nearly periodic and follow the swinging trend of our trajectories as we balance our weight between our feet^{1}. At times, motion direction and body orientation remain temporarily decoupled. This happens, for instance, when we sidestep or in proximity of turns and distractions.

Shoulder-line yawing is not just a mechanical reflection of the walking action, it rather becomes an essential dynamic ingredient as our motion gets geometrically constrained, e.g. by a dense crowd or by a narrow environment. In both cases, as we need to make our way to our destination, we, consciously or unconsciously, rotate our bodies sideways to minimize collisions or maintain comfort distances with other pedestrians or the environment. The capability of measuring and understanding the orientation dynamics of our body and shoulders comes both with societal and fundamental relevance. As a proxy for sight direction, shoulder orientation can be used to assess individual visual attention^{2} or even to increase our capacity to identify anomalous behavior. Moreover, augmenting the traditional position-centered modeling of pedestrians with the orientation degree of freedom, strengthens the connection between human dynamics and other active matter systems, where shape and nematic ordering are key elements to individual and collective behaviors, particularly at high densities (e.g.,^{3,4}).

The dynamics of shoulder-line rotation has been scarcely investigated from a quantitative viewpoint. The data currently available is extremely limited and has been acquired via few laboratory experiments (e.g. Refs.^{9,10,11}). Such scarceness of accurate data hinders the capability of statistic characterizations of the rotation dynamics beyond the estimation of average properties, to include, e.g., fluctuations and rare events. We believe that this is connected with the inherent technical complexity of measuring body yawing accurately and in real-life conditions. Real-life measurements campaigns, in fact, need to rely only on non-intrusive imaging data (or alike) of pedestrians, and cannot be supported by ad hoc wearable sensors, such as accelerometers^{10}. Indeed, even the accurate estimation of the position of an individual in real-life, a more “macroscopic” or “coarser-scale” degree of freedom than orientation, is a recognized technical challenge^{12}. Since few years, overhead depth-sensing^{7,8,13,14}, as used in this work, has been successfully employed to perform accurate pedestrian localization and prolonged tracking campaigns (see example in Fig. 1 and Ref.^{6}). Overhead depth data, not only allows privacy respectful data acquisition, but enables also accurate position measurements even in high-density conditions (for a highly-accurate algorithm leveraging on machine learning-based analyses see, e.g. Ref.^{15}).

In this paper we propose a novel method to measure—in real-life conditions and with very high accuracy—the shoulder rotation of walking pedestrians. Our measurement method is based on a deep Convolutional Neural Network^{16} (CNN) point-estimator which operates on overhead depth images centered on individual pedestrians—from now on referred to as “imagelets”. Intuition suggests that pedestrians seen from an overhead perspective have a well-defined elongated, elliptic-like, shape. In our measurements this is true only in a small fraction of cases in which pedestrians walk carrying their arms alongside the body. Conversely, we found a majority of exceptions, impossible to address by hand-made algorithms (cf. Fig. 2). This marks an ideal use-case for supervised deep learning^{16}.

It is well known that the high performance of Deep Neural Network methods come also at the price of, often prohibitively, labor-intensive manual annotations of training data (frequently in the order of millions of individual images). Depending on the context, the reliability of human annotations can furthermore be arguable, this is the case whenever different experts are in frequent mutual disagreement about the annotation value. Shoulder orientation in depth imagelets falls in such a case. Here by relying on the strong statistical correlation between individual velocity and body orientation, we manage to produce potentially limitless annotations. While walking on straight paths, our velocity direction is (on average) in very good approximation orthogonal to our shoulder line. On this basis, we can employ the velocity direction as a singularly slightly imperfect, but correct on average, annotation for the orientation. Notably, the zero-average residual error between the velocity direction and the actual orientation gets averaged out as we train our CNN point-estimator with gradient descent. This (self) amends for annotation errors.

We investigate the orientation measurement accuracy of our method and consider its error scaling vs. the size of the training set using both real-life and synthetic depth imagelets. Combining extensive training with the enforcement of *O*(2) symmetry of the estimator, we show that we can deliver an orientation estimator with an error as low as \(7.5^\circ\). Our tool enables us to characterize the stochastic process that connects the instantaneous velocity direction to the shoulder orientation. We show that the velocity orientation can be well described by delaying the orientation dynamics through a stochastic process centered on, about, \(100\,\hbox {ms}\) and with Ornstein–Uhlenbeck (OU) statistics.

Conceptually speaking, although our tool has been devised for depth imagelets, it can be easily extended to other computer vision-based pedestrian tracking approaches and, more in general, can be used for any system in which there is a statistical connection between (average) individual “particle” velocity and (average) shape.

## Orientation measurements: problem definition

Let \(\mathcal {I}\) be a overhead imagelet centered on a pedestrian, see examples in Fig. 2 (for convenience we opt for imagelets of square shape, yet this is not a constraint).

We define the shoulder-line orientation angle, \(\theta\), as the angle between the direction normal to the shoulder-line and a fixed reference, here the *y* axis (direction \(\mathbf {e}_y\), cf. Fig. 2a,b). According to this definition, a body rotation of \(180^\circ\) leaves \(\theta\) unchanged. Thus, we aim at a function *f* such that

where \(f(\mathcal {I}) = \theta _o\) approximates the actual orientation \(\theta\) (with \(-\pi /2\equiv \pi /2\), i.e. \(\theta\) is an element of the real projective line \(\mathbf{P} ^1(\mathbb {R})\), see e.g. Ref.^{17}). We report a list of the symbols employed in Table 1.

We model the mapping *f* via a deep neural network that we train in a supervised, end-to-end, fashion (see structure in the Supporting Information, SI). The network returns the estimate of \(\theta\) as a discrete probability distribution, \(h_{\mathrm{pred}}\), on \(\mathbf{P} ^1(\mathbb {R})\) (quantized in \(B=45\) uniform bins, \(4^\circ\) wide, via soft-max activation function in the final layer). We retain the \(\mathbf{P} ^1(\mathbb {R})\)-average (“circular average”) of the distribution \(h_{\mathrm{pred}}\), as final output. It formulas, the output \(\theta _o\) reads

we leave the details to the Supplementary information.

We train with orientation data with a “two-hot” encoding: each orientation \(\theta\) is unambiguously represented in terms of a probability distribution non-vanishing on (up to) two adjacent angular bins (we chose “two-hot” in opposition to the typical one-hot training data for classification problems, in which the annotations are Dirac probability distributions on the ground-truth class). We will refer to this encoding, that avoids quantization errors, as \(h_2(\theta )\) (we observed no strong sensitivity on the number of bins when these were more or equal than 10). As usual, we use a cross-entropy loss, \(\mathcal {H}(\cdot ,\cdot )\).

We employ pedestrian velocity information to tackle the need for huge amounts of accurately annotated data to train the free parameters of the deep neural network (usually in the millions, \(\approx 1.3\;{\text {M}}\) in our case).

Let \(\theta _v(t) \in \mathbf{P} ^1(\mathbb {R})\) be the angle between the walking velocity and a reference at time \(t>0\), i.e.

where \(\mathbf {v}(t)\) is the instantaneous velocity, and \(\angle (\cdot ,\cdot )\) denotes the angle comprised the directions in its argument (with \(\pi\)-periodicity). When we walk, either for the periodic sway or when we make turns, our shoulder line is most-frequently, and in very good approximation, orthogonal with respect to the walking velocity, i.e.

Therefore, velocities provide a meaningful “proxy” annotation for orientation. We used the “approximately equal” sign in Eq. (4) because we can have frequent, yet small, disagreements between velocity and orientation. These can be due to small loss of alignment between the two (e.g. because something attracted our attention) or they can be due to inaccuracies, e.g., in the velocity measurements. It is also possible, yet less likely, that velocity and orientation remain misaligned for longer time intervals. This holds, e.g., for people walking sideways. We retain these as rare occasions, which we expect to occur symmetrically for both left and right sides, with no relevant weight in our training dataset. This hypothesis reasonably holds on unidirectional pedestrian flows happening on rectilinear corridors, but might be invalid in case, e.g., of curved paths. Formally, for a walking person, we model the relation in Eq. (4) as

with \(\epsilon\) being a small, symmetric, and zero-centered residual.

We train our neural network using the labels \(h_2(\theta _v)\) as a proxy for \(h_2(\theta )\). The training process aims at the minimization of the (average) loss \(\mathbb {E}_{\theta _v}[\mathcal {H}(h_{pred},h_2(\theta _v))]\). As such, the output \(h_{pred}\) converges to the distribution of annotations of similar imagelets, whose average is the correct point-estimation of the orientation:

We refer to the Supplementary information for a formal proof with simplifying assumptions and a simulation-based proof in the general case.

Finally, once a pedestrian with shoulder orientation \(\theta\) rigidly rotates around the vertical axis by an angle \(\alpha\), their orientation becomes \(\theta + \alpha\). Similarly, “mirroring” a pedestrian around the \(\mathbf {e}_y\) direction, their orientation changes sign.

The map *f* must satisfy such symmetry with respect to imagelets rotations and mirroring. In other words, *f* must be co-variant^{18} with respect to the group, *O*(2), of the orthogonal transformations of the plane. In formulas, this reads

for all transformations \(\phi =\Phi R_\alpha \in O(2)\), that concatenate a rotation of \(\alpha\), \(R_\alpha\), and, possibly, a reflection (i.e. \(\Phi \in \{\text {Id}, J\}\), respectively the identity, \(\text {Id}\), and the reflection, *J*, from which the sign change given by the determinant of the transformation: \(\det (\phi )=\det (\Phi )=\pm 1\)).

Symmetries in neural networks are often injected at training time, by augmenting the training set by all the symmetry group orbits. Similarly, we include multiple copies of the same imagelets with multiple random rotations with and without flipping. This also ensures that the training set spans \(\mathbf{P} ^1(\mathbb {R})\) uniformly. Yet, this does not yield a strictly *O*(2)-symmetric estimator Eq. (7). We further enforce this symmetry by constructing a new map, \(\tilde{f}\), as the *O*(2)-group average of *f*, which is thus strictly respecting Eq. (7). In formulas it holds

we leave the proof of this identity, the *O*(2)-symmetry of \(\tilde{f}\) and further details on \(\mathbf{P} ^1(\mathbb {R})\)-averages to the SI. In the following, we consider approximations of the integral in Eq. (9) by equi-spaced and random sampling of *O*(2).

## CNN: training and testing

We consider two types of training/testing imagelets: algorithmically generated, “synthetic”, imagelets, of which the orientation angle \(\theta\) is known, and real-life imagelets. In the first case we mimic a velocity-based training by adding a centered noise to labels known exactly [following Eq. (5)]. In the second case, as we have no manually annotated ground truth, of which the accuracy would nevertheless be debatable, we propose a validation based on the convergence towards low-pass filtered orientation signals. In both cases, we show that the average prediction error [ARMSE, Eq. (11)] is about \(7.5^\circ\) degrees or, possibly, lower, should the training set size *N* be large enough. Specifically, the datasets are as follows:

### Synthetic dataset

We generate synthetic imagelets mimicking the overhead shape of people in terms of a superposition of two ellipses: one for the body/shoulders, \(E_b\), and another one, \(E_h\), at lower depth values (i.e. higher on the ground), for the head. We generate synthetic imagelets mimicking the overhead shape of people in terms of a superposition of two ellipses: one for the body/shoulders, \(E_b\), and another one, \(E_h\) at lower depth values (i.e. objects higher above the ground are closer to the overhead mounted depth sensors and are thus associated with smaller depth values), for the head.

We report examples of such imagelets in Fig. 3, while the details of the generation algorithm are left to the SI.

By construction, the rotation angle \(\alpha _{b}\) of \(E_b\) represents the pedestrian orientation, i.e. it is the ground truth for the training. We train the network with such synthetic imagelets and a small centered Gaussian noise \(\epsilon \sim \mathcal {N}\left( 0^\circ , 20^\circ \right)\) superimposed to \(\alpha _{b} = \theta _{gt}\) to imitate velocity-based training. Hence, we train using labels \(\alpha _{gt} + \epsilon\) while we validate with \(\alpha _{gt}\) (cf. Eq. 5).

### Real-life dataset

We consider depth images and velocity data from a real-life measurement campaign conducted during a city-wide festival (GLOW) in Eindhoven, The Netherlands, in Nov. 2017. The measurements involve a uni-directional crowd flow passing through a corridor-shaped exhibit (tracking area: \(12\;{\text {m}} \times 6\;{\text {m}}\)), for further details see^{6}. The dataset leverages on high-resolution individual localization and tracking based on overhead depth images (as in Fig. 1) and with \(30\,\hbox {Hz}\) time sampling. The localization and tracking algorithms employed are analogous to what employed in previous works^{7, 8}. To ensure that our velocity data provides a well-defined proxy for orientation, we restrict to pedestrians having average velocity above \(0.65\,\hbox {m/s}\). Moreover, for each trajectory we extract imagelets and velocity data with a time sampling of \(\Delta T \approx 0.5\,\hbox {s}\), which increases the independence between training data.

Additionally, we apply random rotations and random horizontal flips to all imagelets (and, correspondingly, to labels). This aims at training with a dataset uniformly distributed on \(\mathbf{P} (\mathbb {R}^1)\).

In absence of ground truth, we build our test set as follows: we rely on our neural network trained with \(1\,\hbox {M}\) different imagelets (i.e. twice as much the largest training dataset considered in Fig. 4d,e, on which we perform random augmentation and final *O*(2)-averaging of the operator), hence the most accurate, to make orientation predictions over complete pedestrian trajectories. As an orientation signal \(\theta (t)\) needs to be continuous in time, we smoothen the predicted \(\theta (t)\) in time (low-pass Butterworth filter^{19} of order \(n=1\), cutoff frequency \(c_f=2.0\,\hbox {Hz}\) and window length \(l=52\)) to eliminate random noise. The final dataset contains values \(\theta (t)\) from different trajectories and sampled at different, independent, time instants. Therefore, we build the dataset on the basis of two independent elements: a heavily trained network and a physics-based time-regularity hypothesis on orientation signals.

We assess the prediction performance as the training set size, *N*, increases. To compute exhaustive performance statistics, we train the network on *M* independent datasets for every *N* (in a cross-validation-like setting). We can distinguish two kinds of errors, systematic and random^{20}. The first is an error that always, and in the same manner, interferes with the outcome of the measurements (e.g. a constant rotation offset for all predictions); the second, also referred to as variance, is caused by unexplained variability of the model with respect to the observed imagelets (i.e. the prediction accuracy may vary between different imagelets). To quantify the network performance in relation to these two sources of error, we employ the two following measures. Given a reference orientation (e.g. ground truth), \(\theta _r(\mathcal {I})\), for an imagelet \(\mathcal {I}\) from dataset \(D_k\) (\(k=1,2,\dots ,M\)), we quantify the systematic error as the average prediction bias, \(\hat{b}\), evaluated as the root-mean-square of the individual network biases, \(\hat{b}_k\):

Additionally, we consider the average root-mean-square error (ARMSE), which quantifies the total error, as superposition of systematic and random components. In formulas, this reads

where \(V_k\) is the variance of the prediction of the *k*-th network.

## Results

In Fig. 4a–c, we report the orientation signals as estimated by the networks in three different real-life contexts. The network is capable of accurate predictions that, as expected, are independent of the actual instantaneous velocity. Hence, it remains accurate in case of a pedestrian walking sideways (Fig. 4b), in which the orientation signal loses temporarily coupling with the velocity orientation and in case of a pedestrian temporarily stopping and standing (Fig. 4c), in which the velocity orientation is even undefined (note that these cases were excluded from the training).

We include in Fig. 4d–f the values of average prediction bias and ARMSE as the training set size increases, in case of synthetic and real-life imagelets [respectively, in (**d**) and (**e**)]. In both cases the network performance increases with *N*, with slightly slower convergence rate for the ARMSE for the real-life dataset, which is likely more challenging to learn than the synthetic one. In both cases the predictions are free of bias (cf. sub-panels). With the largest number of training imagelets considered (\(N\approx 5 \times 10^5\)), we measured an ARMSE of about \(5^\circ\) for the synthetic data and \(11^\circ\) for the real-life data. We managed to further reduce this error to, respectively, \(4.5^\circ\) and \(7.5^\circ\) by enforcing *O*(2) symmetry. Note that we could trivially apply Eq. (8) as we are in a bias-free context, else a systematic correction for the bias would have been necessary. In Fig. 4f, we report the network performance as we approximate better and better the *O*(2) group average. We stress that in case of real-life data, the network predictions, on which no time-smoothing has been applied, converge to test data that underwent time-smoothing. Thus, as *N* grows, the network predictions show increasing robustness and consistency approaching jitter-free signals.

## Real-life orientation dynamics

We are now capable of investigating with high-resolution, and in real-life conditions, the connection between shoulder orientation and velocity direction—which, in the previous sections, we reduced to the error term \(\epsilon\). In particular, we can characterize a stochastic delay signal, *d*(*t*), which allows us to model the relation between velocity and orientation as

where *A* is a positive constant.

First, thanks to the high-accuracy of our tool, we measure a velocity-dependent delay between velocity orientation and shoulder orientation whose probability distribution function is in Fig. 5 (see Supplementary information for details on the delay measurement algorithm). The velocity orientation follows in time the shoulder yawing, with a delay that decreases (on average) between \(160\,\hbox {ms}\) and \(100\,\hbox {ms}\) as the average walking velocity, \(\hat{v}\), increases from \(0.6\,\hbox {m/s}\) to \(1.4\,\hbox {m/s}\) (respectively walking speed values in leisure and normal walking regimes, see, e.g. Ref.^{21}).

The structure of *d*(*t*) appears well-modeled by a OU process:

where \(\hat{d}>0\) is the average delay (\(\hat{d} \approx \langle d(t) \rangle\)), \(\tau >0\) is the OU time-scale and \(\xi >0\) is the intensity of the \(\delta\)-correlated white noise \(\dot{W}\). In particular, in Fig. 6 we compare statistical observables of measurements and simulations considering the case of normal walking speed (average velocity \(\hat{v} \approx 1.3\,\hbox {m/s}\)), of which we retain the measured orientation signals, \(\theta (t)\), as a basis for Eq. (12) (simulation parameters: \(A=1.85\), \(\hat{d} =0.08\,\hbox {s}\), \(\tau =1.2\,\hbox {s}\) and \(\xi =1.85\)). In Fig. 6a, we report the pdf of the difference between orientation and velocity orientation when one is shifted in time by, \(\hat{d}\), to compensate for the average delay. Measurements and simulations, in excellent mutual agreement, follow a Gaussian statistics. Thanks to a stochastic delay, we achieve a very good quantitative agreement in the delay distributions (Fig. 6b). In Fig. 6c, we report the Power Spectral Density (psd) of \(\theta _v(t)\) and \(\theta (t)\) computed by averaging all the psds obtained from individual velocity direction and orientation signals. We observe that the stochastic delay does not substantially modify the psd of orientations, especially at low frequencies. As an effect, the peak around \(f=1\,\hbox {Hz}\), connected with the periodic swinging in walking (see Fig. 4a), is reproduced (yet it is slightly underestimated). Moreover, the psd shows another peak around \(0.2\,\hbox {Hz}\), which is connected to large-scale motion (a pedestrian might curve along their path) and/or have a non-periodic orientation signal (which yields low-frequency spectral artifacts). Also this peak, not modeled by Eq. (13), is reproduced by the model but with a slight overestimation.

## Discussion

In this paper we presented an extremely accurate estimator for the pedestrian shoulder-line orientation based on deep convolutional neural networks. We leveraged on statistic aspects of pedestrian dynamics to overcome two outstanding issues related to deep networks training: the labor-intensive annotation of training data in sufficient amounts (generally millions of images) and the accuracy of annotations in non-trivial contexts.

Thanks to the strong statistical correlation of shoulder-line and velocity direction, which are typically orthogonal, we can employ the velocity direction as a training label. Although often slightly incorrect, it remains correct on the average, to which our point-estimator converges. Notably, the relation between velocity and orientation holds regardless of the quality of the raw imaging data employed. In case of overhead depth maps, as used here, often we had disagreement between human annotators, which would unavoidably yield low quality labels. By using velocity we can circumvent this issue and produce training data in arbitrarily large amounts. We stress that this correlation assumption is crucial only for training the estimator. As evidenced in the paper, once trained, the estimator can be used to successfully measure pedestrian orientation when shoulder-line orientation and velocity direction are systematically not orthogonal (like it happens for people walking sideways), or even for vanishing walking velocity (where the velocity orientation is not defined). We mention additionally that our approach can be conceptually extended to other imaging formats, such as color images, provided accurate and sufficiently prolonged tracking data are available.

Our tool unlocked the possibility to accurately investigate the relation between velocity direction and shoulder orientation. We could measure a velocity-dependent delay of about \(100\,\hbox {ms}\) between velocity and orientation, that we are able to quantitatively reproduce in terms of a simple Ornstein–Uhlenbeck process. In particular, on the basis of measured orientation signals, we could generate velocity directions featuring amplitude with respect to the orientation signal, velocity-orientation delay distribution and power spectral density in very good agreement with the measurements.

We currently employed our velocity-trained network to investigate dynamics at relatively low density. We expect the network to be capable to operate and deliver accurate orientation estimates in different scenarios from those considered. As such, natural next steps are the investigation of static and dynamic high-density crowds, clogged bottlenecks conditions, or other scenarios in which the “nematic” ordering of the crowd is expected to play a key role in the dynamics. Additionally, the tool developed can be applied to do (real-time) analyses of orientation, e.g. to gather a first estimate of sight/attention direction and/or possibly extend anomaly detection capabilities for crowd motion.

## References

Pontzer, H., Holloway, J. H., Raichlen, D. A. & Lieberman, D. E. Control and function of arm swing in human walking and running.

*J. Exp. Biol.***212**, 523–534. https://doi.org/10.1242/jeb.024927 (2009).Gallup, A.

*et al.*Visual attention and the acquisition of information in human crowds.*Proc. Natl. Acad. Sci.***109**, 7245–7250 (2012).Romanczuk, P., Bär, M., Ebeling, W., Lindner, B. & Schimansky-Geier, L. Active Brownian particles.

*Eur. Phys. J. Spec. Top.***202**, 1–162 (2012).Marchetti, M. C.

*et al.*Hydrodynamics of soft active matter.*Rev. Mod. Phys.***85**, 1143–1189. https://doi.org/10.1103/RevModPhys.85.1143 (2013).Microsoft Corp. Kinect for Xbox 360. Redmond, WA, USA (2012).

Corbetta, A.

*et al.*A large-scale real-life crowd steering experiment via arrow-like stimuli.*Collect. Dyn.***5**, 61–68. https://doi.org/10.17815/CD.2020.34 (2020).Corbetta, A., Lee, C., Benzi, R., Muntean, A. & Toschi, F. Fluctuations around mean walking behaviours in diluted pedestrian flows.

*Phys. Rev. E***95**, 032316. https://doi.org/10.1103/PhysRevE.95.032316 (2017).Corbetta, A., Meeusen, J., Lee, C., Benzi, R. & Toschi, F. Physics-based modeling and data representation of pairwise interactions among pedestrians.

*Phys. Rev. E***98**, 062310. https://doi.org/10.1103/PhysRevE.98.062310 (2018).Yamamoto, H., Yanagisawa, D., Feliciani, C. & Nishinari, K. Body-rotation behavior of pedestrians for collision avoidance in passing and cross flow.

*Transp. Res. Part B Methodol.***122**, 486–510. https://doi.org/10.1016/j.trb.2019.03.008 (2019).Feliciani, C. & Nishinari, K. Pedestrians rotation measurement in bidirectional streams. In

*Pedestrian and Evacuation Dynamics 2016*121–129 (University of Science and Technology of China Press, Hefei, 2016).Feliciani, C. & Nishinari, K. Estimation of pedestrian crowds’ properties using commercial tablets and smartphones.

*Transp. B Transp. Dyn.***7**, 865–896. https://doi.org/10.1080/21680566.2018.1517061 (2019).Boltes, M. & Seyfried, A. Collecting pedestrian trajectories.

*Neurocomputing***100**, 127–133. https://doi.org/10.1016/j.neucom.2012.01.036 (2013).Seer, S., Brändle, N. & Ratti, C. Kinects and human kinetics: A new approach for studying pedestrian behavior.

*Transp. Res. C Emerg.***48**, 212–228. https://doi.org/10.1016/j.trc.2014.08.012 (2014).Brščić, D., Kanda, T., Ikeda, T. & Miyashita, T. Person tracking in large public spaces using 3-d range sensors.

*IEEE Trans. Hum. Mach. Syst.***43**, 522–534. https://doi.org/10.1109/THMS.2013.2283945 (2013).Kroneman, W, T., Corbetta A. Accurate pedestrian localization in overhead depth images via height-augmented HOG.

*Collect. Dyn.***5**(2020).LeCun, Y., Bengio, Y. & Hinton, G. Deep learning.

*Nature***521**, 436. https://doi.org/10.1038/nature14539 (2015).Gowers, T., Barrow-Green, J. & Leader, I.

*The Princeton Companion to Mathematics*(Princeton University Press, Princeton, 2008).Grove, L. C.

*Classical Groups and Geometric Algebra*Vol. 39 (American Mathematical Society, Washington, DC, 2002).Selesnick, I. & Burrus, C. Generalized digital butterworth filter design.

*IEEE Trans. Signal Process.***46**, 1688–1694. https://doi.org/10.1109/78.678493 (1998).Goodfellow, I.J., Bengio, Y. & Courville, A.

*Deep Learning*(MIT Press, Cambridge, 2016). http://www.deeplearningbook.org.Fruin, J. J.

*Pedestrian Planning and Design*(Elevator World Inc., Ansterdam, 1987).

## Acknowledgements

A.C. acknowledges the Talent Scheme (Veni) research programme (project N. 16771) financed by the Netherlands Organization for Scientific Research.

## Author information

### Authors and Affiliations

### Contributions

A.C. and F.T. designed the study. A.C., V.M., F.T. co-supervised the research. J.W. developed the software and performed the analyses. A.C. designed the software infrastructure and wrote the paper. All the authors approved the manuscript.

### Corresponding author

## Ethics declarations

### Competing interests

The authors declare no competing interests.

## Additional information

### Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

## Supplementary information

## Rights and permissions

**Open Access** This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.

## About this article

### Cite this article

Willems, J., Corbetta, A., Menkovski, V. *et al.* Pedestrian orientation dynamics from high-fidelity measurements.
*Sci Rep* **10**, 11653 (2020). https://doi.org/10.1038/s41598-020-68287-6

Received:

Accepted:

Published:

DOI: https://doi.org/10.1038/s41598-020-68287-6

## Comments

By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.