Inpyo Song
Hello I'm Inpyo Song, a M.S. student at Sungkyunkwan University. I'm interested in computer vision and deep learning. Specifically, I'm working on human pose estimation, anomaly detection and video question answering.
Publications
Motion-Aware Heatmap Regression for Human Pose Estimation in Videos
Inpyo Song, Jongmin Lee, Moonwook Ryu, and Jangwon Lee
In IJCAI2024 (International Joint Conferences on Artificial Intelligence), Aug. 2024, (Poster, Lower than 20.0% acceptance rate)
We present an approach to solving 2D human pose estimation in videos. The problem of human pose estimation in videos differs from estimating human poses in static images since videos contain a lot of motion related information. Thus, we investigate how to utilize by the information of the human body movements across in a sequence of video frames for estimating human poses in videos. To do this, we introduce a novel heatmap regression method what we call motion-aware heatmap regression. Our approach computes motion vectors in joint keypoints from adjacent frames. We then design a new style of heatmap that we call Motion-Aware Heatmaps to reflect the motion uncertainty of each joint point. Unlike traditional heatmaps, our motion-aware heatmaps not only consider the current joint locations but also account how joints move over time. Furthermore, we introduce a simple yet effective framework designed to incorporate motion information into heatmap regression.


Video Question Answering for People with Visual Impairments Using an Egocentric 360-Degree Camera
Inpyo Song, Minjun Ju, Joonhyung Kwon, Jangwon Lee
In CVPR (IEEE Conference on Computer Vision and Pattern Recognition) EgoVis Workshop, Jun. 2024
This paper addresses the daily challenges encountered by visually impaired individuals, such as limited access to information, navigation difficulties, and barriers to social interaction. To alleviate these challenges, we introduce a novel visual question answering dataset. Our dataset of fers two significant advancements over previous datasets: Firstly, it features videos captured using a 360-degree egocentric wearable camera, enabling observation of the entire surroundings, departing from the static image-centric nature of prior datasets. Secondly, unlike datasets centered on singular challenges, ours addresses multiple real-life obstacles simultaneously through an innovative visual-question answering framework. We validate our dataset using various state-of-the-art VideoQA methods and diverse metrics. Results indicate that while progress has been made, satisfactory performance levels for AI-powered assistive services remain elusive for visually impaired individuals. Additionally, our evaluation highlights the distinctive features of the proposed dataset, featuring ego-motion in videos captured via 360-degree cameras across varied scenarios.


Scale and Motion Adaptive Multi-Object Tracking Algorithm for Unmanned Aerial Vehicles
Inpyo Song, Jangwon Lee
In RO-MAN (IEEE International Conference on Robot & Human Interactive Communication) Late Breaking Report, Aug. 2023
We introduce a novel method to overcome UAV challenges (unpredictable motion, small size of target objects). Specifically, we present a new tracking strategy that involves initiating the tracking of target objects from low-confidence detections, which are frequently encountered in various UAV application scenarios. Additionally, we propose revisiting traditional appearance matching algorithms to improve the association of low-confidence detections.
Action-Conditioned Contrastive Learning for 3D Human Pose and Shape Estimation in Videos
Inpyo Song, Moonwook Ryu, Jangwon Lee
In CKAIA (Conference of Korean Artificial Intelligence Association), Jul. 2023
We propose a novel approach called the action-conditioned mesh recovery (ACMR) model, which improves accuracy without compromising temporal consistency by leveraging human action information. Our ACMR model outperforms existing methods that prioritize temporal consistency in terms of accuracy, while also achieving comparable temporal consistency with other state-of-the-art methods.
Smooth and Consistent Pose Estimation with Intersections Module
Jongmin Lee*, Inpyo Song* (Equal Contribution), Moonwook Ryu, Jangwon Lee
In KIBME (The Korean Institute of Broadcast and Media Engineers Summer Conference) Jun. 2024
This paper addresses the challenge of achieving stable 2D human pose estimation in videos. While existing methods primarily focus on accurately localizing keypoints, they often neglect the temporal stability of these keypoints, resulting in jittering noise. This instability can result in unrealistic outputs, limiting practical applications. To tackle this issue, we propose using evaluation metrics that measure jittering noise and introduce a simple yet effective heatmap aggregation module, termed Intersections. Our approach demonstrates improved stability in keypoint localization over time, enhancing the reliability of pose estimation in video.
Multitask Learning in Facial Kinship Verification
Inpyo Song, Jangwon Lee
In KIBME (The Korean Institute of Broadcast and Media Engineers Summer Conference) Jun. 2023
This research leverages a novel multi-task learning approach to enhance kinship verification. By training a model on interconnected tasks—family classification, kin-relationship prediction, and age regression— we improve the system's performance and overcome challenges associated with traditional feature extraction techniques.
Awards and Projects
Contact Me
Let's start a conversation
Address
Sungkyunkwan University Central Library 507
LinkedIn
https://www.linkedin.com/in/ipsong/
Email Address
inpyosong.deep
@gmail.com
I want to stay in touch with you!
Feel free to contact me.