AudioEar: Single-View Ear Reconstruction for Personalized Spatial Audio
Creators
- 1. Shanghai Jiao Tong University
- 2. None
Description
Abstract
We introduce AudioEar3D, a high-quality 3D ear dataset consisting of 112 point cloud ear scans with RGB images, to benchmark the ear reconstruction task. We further collect a 2D ear dataset composed of 2,000 images, each one with manual annotation of occlusion and 55 landmarks, named AudioEar2D. To our knowledge, both datasets have the largest scale and best quality of their kinds for public use.
The code is publicly available at https://github.com/seanywang0408/AudioEar.
Citation
If you find this project useful, currently please cite the paper as:
Xiaoyang Huang, Yanjun Wang, Yang Liu, Bingbing Ni, Wenjun Zhang, Jinxian Liu, Teng Li. "AudioEar: Single-View Ear Reconstruction for Personalized Spatial Audio". arXiv preprint arXiv:2301.12613, 2023.
or using bibtex:
@article{huang2023audioear, title={AudioEar: Single-View Ear Reconstruction for Personalized Spatial Audio}, author={Huang, Xiaoyang and Wang, Yanjun and Liu, Yang and Ni, Bingbing and Zhang Wenjun and Liu Jinxian and Li, Teng}, journal={arXiv preprint arXiv:2301.12613}, year={2023} }
License
The dataset is licensed under Creative Commons Attribution 4.0 International (CC BY 4.0).
The code is under Apache-2.0 License.
Mirror Link
We recommend users to download the data from Zenodo official link. However, if you find any downloading problem, you can also use this mirror link from Google Drive.
Changelog
v1.1: Add README.md.
v1.0: Initial repository of AudioEar3D and AudioEar2D.
Files
AudioEar2D.zip
Additional details
Related works
- Is described by
- Preprint: https://arxiv.org/abs/2301.12613 (URL)