scholarly journals Multimodal Joint Head Orientation Estimation in Interacting Groups via Proxemics and Interaction Dynamics

Author(s):  
Stephanie Tan ◽  
David M. J. Tax ◽  
Hayley Hung

Human head orientation estimation has been of interest because head orientation serves as a cue to directed social attention. Most existing approaches rely on visual and high-fidelity sensor inputs and deep learning strategies that do not consider the social context of unstructured and crowded mingling scenarios. We show that alternative inputs, like speaking status, body location, orientation, and acceleration contribute towards head orientation estimation. These are especially useful in crowded and in-the-wild settings where visual features are either uninformative due to occlusions or prohibitive to acquire due to physical space limitations and concerns of ecological validity. We argue that head orientation estimation in such social settings needs to account for the physically evolving interaction space formed by all the individuals in the group. To this end, we propose an LSTM-based head orientation estimation method that combines the hidden representations of the group members. Our framework jointly predicts head orientations of all group members and is applicable to groups of different sizes. We explain the contribution of different modalities to model performance in head orientation estimation. The proposed model outperforms baseline methods that do not explicitly consider the group context, and generalizes to an unseen dataset from a different social event.

2016 ◽  
Vol 48 (6) ◽  
pp. 1757-1772 ◽  
Author(s):  
Dua K. S. Y. Klaas ◽  
Monzur Alam Imteaz ◽  
Arul Arulrajah

Abstract To assess the effect of three grid cell properties (size, mean slope of the surface and distance between centre of grid and observation well) on groundwater models' performances, a tropical karst catchment characterized by monsoonal season in Rote Island, Indonesia was selected. Here, MODFLOW was used to develop models with five different spatial discretization schemes: 10 × 10 m, 20 × 20 m, 30 × 30 m, 40 × 40 m and 50 × 50 m. Using parameter estimation method, hydraulic conductivity and specific yield values over a selection of pilot points were estimated. The trends of the performances were calculated at each observation well in order to recommend the most appropriate location for observation well placement in terms of topographical characteristic. It is confirmed that the deterioration of model performance is mainly controlled by the increase of distance between well and centre of the cell, and the mean slope of the surface. Results reveal that model performance increases substantially for areas of low slope (<3%) and medium slope (3–10%) for a smaller grid cell size. Therefore, to improve model performance, it is recommended that the observations wells are placed in areas of low and medium slopes.


Author(s):  
Qiang Yang ◽  
Yuanqing Zheng

Voice interaction is friendly and convenient for users. Smart devices such as Amazon Echo allow users to interact with them by voice commands and become increasingly popular in our daily life. In recent years, research works focus on using the microphone array built in smart devices to localize the user's position, which adds additional context information to voice commands. In contrast, few works explore the user's head orientation, which also contains useful context information. For example, when a user says, "turn on the light", the head orientation could infer which light the user is referring to. Existing model-based works require a large number of microphone arrays to form an array network, while machine learning-based approaches need laborious data collection and training workload. The high deployment/usage cost of these methods is unfriendly to users. In this paper, we propose HOE, a model-based system that enables Head Orientation Estimation for smart devices with only two microphone arrays, which requires a lower training overhead than previous approaches. HOE first estimates the user's head orientation candidates by measuring the voice energy radiation pattern. Then, the voice frequency radiation pattern is leveraged to obtain the final result. Real-world experiments are conducted, and the results show that HOE can achieve a median estimation error of 23 degrees. To the best of our knowledge, HOE is the first model-based attempt to estimate the head orientation by only two microphone arrays without the arduous data training overhead.


Electronics ◽  
2019 ◽  
Vol 8 (2) ◽  
pp. 220 ◽  
Author(s):  
Ruibin Guo ◽  
Keju Peng ◽  
Dongxiang Zhou ◽  
Yunhui Liu

Orientation estimation is a crucial part of robotics tasks such as motion control, autonomous navigation, and 3D mapping. In this paper, we propose a robust visual-based method to estimate robots’ drift-free orientation with RGB-D cameras. First, we detect and track hybrid features (i.e., plane, line, and point) from color and depth images, which provides reliable constraints even in uncharacteristic environments with low texture or no consistent lines. Then, we construct a cost function based on these features and, by minimizing this function, we obtain the accurate rotation matrix of each captured frame with respect to its reference keyframe. Furthermore, we present a vanishing direction-estimation method to extract the Manhattan World (MW) axes; by aligning the current MW axes with the global MW axes, we refine the aforementioned rotation matrix of each keyframe and achieve drift-free orientation. Experiments on public RGB-D datasets demonstrate the robustness and accuracy of the proposed algorithm for orientation estimation. In addition, we have applied our proposed visual compass to pose estimation, and the evaluation on public sequences shows improved accuracy.


Sign in / Sign up

Export Citation Format

Share Document