Page Not Found
Page not found. Your pixels are in another canvas.
A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.
Page not found. Your pixels are in another canvas.
About me
This is a page not in th emain menu
Published:
This post will show up by default. To disable scheduling of future posts, edit config.yml
and set future: false
.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Short description of portfolio item number 1
Short description of portfolio item number 2
Published in Journal of the Korea Institute of Information and Communication Engineering, 2017
Recommended citation: Dong-Hyun Hwang and Kyung-Sik Jang. "Hand-Gesture Recognition Using Concentric-Circle Expanding and Tracing Algorithm." Journal of the Korea Institute of Information and Communication Engineering 21.3 (2017): 636-642. https://www.koreascience.or.kr/article/JAKO201713647760782.page
Published in Journal of the Korea Institute of Information and Communication Engineering, 2017
Recommended citation: Dong-Hyun Hwang and Kyung-Sik Jang. "Fast Hand-Gesture Recognition Algorithm For Embedded System." Journal of the Korea Institute of Information and Communication Engineering 21.7 (2017): 1349-1354. https://www.koreascience.or.kr/article/JAKO201723839836707.page
Published in The 2018 CHI Conference on Human Factors in Computing Systems (ACM CHI 2018, Poster), 2018
Recommended citation: Dong-Hyun Hwang, and Hideki Koike. "Parapara: Synthesizing pseudo-2.5 d content from monocular videos for mixed reality." Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems. 2018. https://dl.acm.org/doi/abs/10.1145/3170427.3188596
Published in International Conference on Artificial Reality and Telexistence, Eurographics Symposium on Virtual Environments 2018 (ICAT-EGVE 2018, Short paper), 2018
Recommended citation: Ikeda et al. "AR based Self-sports Learning System using Decayed Dynamic Time Warping Algorithm." ICAT-EGVE 2018. 2018. https://diglib.eg.org/handle/10.2312/egve20181330
Published in The 2018 ACM International Conference on Interactive Surfaces and Spaces (ISS 2018, Full paper), 2018
Recommended citation: Sato et al. "MlioLight: Projector-camera Based Multi-layered Image Overlay System for Multiple Flashlights Interaction." Proceedings of the 2018 ACM International Conference on Interactive Surfaces and Spaces. 2018. https://dl.acm.org/doi/abs/10.1145/3279778.3279797
Published in The 2018 ACM International Conference on Interactive Surfaces and Spaces (ISS 2018, Full paper), 2018
Recommended citation: Ikeda, Atsuki, Dong-Hyun Hwang, and Hideki Koike. "Real-time Visual Feedback for Golf Training Using Virtual Shadow." Proceedings of the 2018 ACM International Conference on Interactive Surfaces and Spaces. 2018. https://dl.acm.org/doi/abs/10.1145/3279778.3279927
Published in The 26th IEEE Conference on Virtual Reality and 3D User Interfaces (IEEE VR 2019, Poster), 2019
Recommended citation: Hwang et al. "MonoEye: Monocular Fisheye Camera-based 3D Human Pose Estimation." 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR). IEEE, 2019. https://ieeexplore.ieee.org/document/8798267
Published in ACM SIGGRAPH 2019 (Emerging Technologies), 2019
Recommended citation: Ikeda et al. "Golf training system using sonification and virtual shadow." ACM SIGGRAPH 2019 Emerging Technologies. 2019. https://dl.acm.org/doi/abs/10.1145/3379337.3415856
Published in IEEE/CVF Winter Conference on Applications of Computer Vision 2020 (IEEE/CVF WACV 2020, Full paper), 2020
Recommended citation: Hwang et al. "Lightweight 3D human pose estimation network training using teacher-student learning." Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 2020. https://openaccess.thecvf.com/content_WACV_2020/html/Hwang_Lightweight_3D_Human_Pose_Estimation_Network_Training_Using_Teacher-Student_Learning_WACV_2020_paper.html
Published in The 33rd Annual ACM Symposium on User Interface Software and Technology (ACM UIST 2020, Full paper), 2020
Recommended citation: Hwang et al. "MonoEye: Multimodal Human Motion Capture System Using A Single Ultra-Wide Fisheye Camera." Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology. 2020. https://dl.acm.org/doi/abs/10.1145/3379337.3415856
Published in Augmented Humans Conference 2021 (AHs 2021, Short Paper), 2021
Recommended citation: Aso et al. "Portable 3D Human Pose Estimation for Human-Human Interaction using a Chest-Mounted Fisheye Camera." Augmented Humans Conference 2021. 2021. https://dl.acm.org/doi/10.1145/3458709.3458986
Published in Journal of Applied Sciences 2021, 11(17), 7946 (SCIE, IF: 2.679, Full paper), 2021
Recommended citation: Dong-Hyun Hwang and Hideki Koike. 2021. "MonoMR: Synthesizing Pseudo-2.5D Mixed Reality Content from Monocular Videos" Applied Sciences 11, no. 17: 7946. https://doi.org/10.3390/app11177946 https://www.mdpi.com/2076-3417/11/17/7946
Published in ACM SIGGRAPH Asia 2021, Tokyo, Japan (Poster), 2021
Recommended citation: https://arxiv.org/abs/2105.10153
Published:
Free-viewpoint video (FVV) is a kind of advanced media that provides a more immersive user experience than traditional media. It allows users to interact with content because users can view media at the desired viewpoint and is becoming a next-generation media. In creating FVV content, existing systems require complex and specialized capturing equipment and has low end-user usability because it needs a lot of expertise to use the system. This becomes an inconvenience for individuals or small organizations who want to create content and limits the end user’s ability to create FVV-based user-generated content (UGC) and inhibits the creation and sharing of various created content. To tackle these problems, ParaPara is proposed in this work. ParaPara is an end-to-end system that uses a simple yet effective method to generate pseudo-2.5D FVV content from monocular videos, unlike the previously proposed systems. First, the system detects persons from the monocular video through a deep neural network, calculates the real-world homography matrix based on the minimal user interaction, and estimates the pseudo-3D positions of the detected persons. Then, person textures are extracted using general image processing algorithms and placed at the estimated real-world positions. Finally, the pseudo-2.5D content is synthesized from these elements. The content, which is synthesized by the proposed system, is implemented on Microsoft HoloLens; the user can freely place the generated content on the real world and watch it on a free viewpoint.
Published:
We present MoVNect, a lightweight deep neural network to capture 3D human pose using a single RGB camera. To improve the overall performance of the model, we apply the teacher-student learning method based knowledge distillation to 3D human pose estimation. Real-time post-processing makes the CNN output yield temporally stable 3D skeletal information, which can be used in applications directly. We implement a 3D avatar application running on mobile in real-time to demonstrate that our network achieves both high accuracy and fast inference time. Extensive evaluations show the advantages of our lightweight model with the proposed training method over previous 3D pose estimation methods on the Human3.6M dataset and mobile devices.