103 Results
October 28, 2019
DenseRaC: Joint 3D Pose and Shape Estimation by Dense Render-and-Compare
International Conference on Computer Vision (ICCV)
We present DenseRaC, a novel end-to-end framework for jointly estimating 3D human pose and body shape from a monocular RGB image. Our two-step framework takes the body pixel-to-surface correspondence map (i.e., IUV map) as proxy representation and then performs estimation of parameterized human pose and shape.
By: Yuanlu Xu, Song-Chun Zhu, Tony TungOctober 28, 2019
Ray tracing 3D spectral scenes through human optics models
Journal of Vision
Scientists and engineers have created computations and made measurements that characterize the first steps of seeing. ISETBio software integrates such computations and data into an open-source software package. The initial ISETBio implementations modeled image formation (physiological optics) for planar or distant scenes. The ISET3d software described here extends that implementation, simulating image formation for three-dimensional scenes.
By: Trisha Lian, Kevin J. MacKenzie, David H. Brainard, Nicolas P. Cottaris, Brian A. WandellOctober 27, 2019
Habitat: A Platform for Embodied AI Research
International Conference on Computer Vision (ICCV)
We present Habitat, a platform for research in embodied artificial intelligence (AI). Habitat enables training embodied agents (virtual robots) in highly efficient photorealistic 3D simulation.
By: Manolis Savva, Abhishek Kadian, Oleksandr Maksymets, Yili Zhao, Erik Wijmans, Bhavana Jain, Julian Straub, Jia Liu, Vladlen Koltun, Jitendra Malik, Devi Parikh, Dhruv BatraSeptember 19, 2019
Virtual Grasping Feedback and the Virtual Hand Ownership
Symposium on Applied Perception (SAP)
In this study, we analyze the performance, user preference, and sense of ownership for eight virtual grasping visualizations. Six are classified as either a tracked hand visualization or an outer hand visualization. The tracked hand visualizations are those that allow the virtual hand to enter the object being grasped, whereas the outer hand visualizations do not, thereby simulating a realistic interaction.
By: Ryan Canales, Aline Normoyle, Yu Sun, Yuting Ye, Massimiliano Di Luca, Sophie JörgSeptember 9, 2019
Flexible binaural resynthesis of room impulse responses for augmented reality research
EAA Spatial Audio Signal Processing Symposium (SASP)
A basic building block of audio for Augmented Reality (AR) is the use of virtual sound sources layered on top of real sources present in an environment. In order to perceive these virtual sources as belonging to the natural scene it is important to carefully replicate the room acoustics of the listening space. However, it is unclear to what extent the real and virtual room impulse responses (RIR) need to be matched in order to generate plausible scenes in which virtual sound sources blend seamlessly with real sound sources. This contribution presents an auralization framework that allows binaural rendering, manipulation and reproduction of room acoustics in augmented reality scenarios, in order to get a better understanding of the perceptual relevance of individual room acoustic parameters.
By: Sebastià V. Amengual Garí, W. Owen Brimijoin, Henrik G. Hassager, Philip W. RobinsonSeptember 6, 2019
Perceptual comparison of ambisonics-based reverberation methods in binaural listening
EAA Spatial Audio Signal Processing Symposium (SASP)
Reverberation plays a fundamental role in the auralisation of enclosed spaces as it contributes to the realism and immersiveness of virtual 3D sound scenes. However, rigorous simulation of interactive room acoustics is computationally expensive, and it is common practice to use simplified models at the cost of accuracy. In the present study, two subjective listening tests were carried out to explore trade-offs between algorithmic complexity (and approach) and perceived spatialisation quality in a binaural spatialisation context.
By: Isaac Engel, Craig Henry, Sebastià V. Amengual Garí, Philip Robinson, David Poirier-Quinot, Lorenzo PicinaliSeptember 5, 2019
C3DPO: Canonical 3D Pose Networks for Non-Rigid Structure From Motion
International Conference on Computer Vision (ICCV)
We propose C3DPO, a method for extracting 3D models of deformable objects from 2D keypoint annotations in unconstrained images. We do so by learning a deep network that reconstructs a 3D object from a single view at a time, accounting for partial occlusions, and explicitly factoring the effects of viewpoint changes and object deformations.
By: David Novotny, Nikhila Ravi, Benjamin Graham, Natalia Neverova, Andrea VedaldiAugust 12, 2019
Efficient Segmentation: Learning Downsampling Near Semantic Boundaries
International Conference on Computer Vision (ICCV)
Many automated processes such as auto-piloting rely on a good semantic segmentation as a critical component. To speed up performance, it is common to downsample the input frame. However, this comes at the cost of missed small objects and reduced accuracy at semantic boundaries. To address this problem, we propose a new content-adaptive downsampling technique that learns to favor sampling locations near semantic boundaries of target classes.
By: Dmitrii Marin, Zijian He, Peter Vajda, Priyam Chatterjee, Sam Tsai, Fei Yang, Yuri BoykovJuly 31, 2019
Neural Volumes: Learning Dynamic Renderable Volumes from Images
SIGGRAPH
To overcome memory limitations of voxel-based representations, we learn a dynamic irregular grid structure implemented with a warp field during ray-marching. This structure greatly improves the apparent resolution and reduces grid-like artifacts and jagged motion. Finally, we demonstrate how to incorporate surface-based representations into our volumetric-learning framework for applications where the highest resolution is required, using facial performance capture as a case in point.
By: Stephen Lombardi, Tomas Simon, Jason Saragih, Gabriel Schwartz, Andreas Lehrmann, Yaser SheikhJuly 12, 2019
The contributions of skin stretch and kinesthetic information to static weight perception
World Haptics
In this study, we examined the contributions of kinesthetic and skin stretch cues, in isolation and together, to the static perception of weight. In two psychophysical experiments, we asked participants either to detect on which hand a weight was presented or to compare between two weight cues. Two closed-loop controlled haptic devices were used to present weights with a precision of 0.05g to an end-effector held in a pinch grasp.
By: Femke E. van Beek, Raymond J. King, Casey Brown, Massimiliano Di Luca