프로그램
상세 프로그램:
DAY 1 (8/8 Monday) |
DAY 2 (8/9 Tuesday) |
DAY 3 (8/10 Wednesday3 |
DAY 4 (8/11 Thursday) |
||||||
08:50 | - | 09:00 | Workshop (08:50 - 12:30) |
||||||
09:00 | - | 09:10 | Registration (09:00 - 18:00) / Exhibition (10:00 - 18:00) |
||||||
09:10 | - | 09:20 | |||||||
09:20 | - | 09:30 | |||||||
09:30 | - | 09:40 | |||||||
09:40 | - | 09:50 | Opening (09:40 - 10:00) |
||||||
09:50 | - | 10:00 | |||||||
10:00 | - | 10:10 | Oral 1 (10:00 - 11:00) |
Registration / Exhibition (10:00 - 18:00) |
Invited Talk 2 (10:00 - 11:00) Jitendra Malik (UC Berkeley) |
Registration / Exhibition (10:00 - 18:00) |
Oral 5 (10:00 - 11:00) |
||
10:10 | - | 10:20 | |||||||
10:20 | - | 10:30 | |||||||
10:30 | - | 10:40 | |||||||
10:40 | - | 10:50 | |||||||
10:50 | - | 11:00 | |||||||
11:00 | - | 11:10 | Doctoral Consortium (11:00 - 12:00) |
KCVS General Meeting (11:00 - 12:00) |
Invited Talk 3 (11:00 - 12:00) Chelsea Finn (Stanford Univ.) |
||||
11:10 | - | 11:20 | |||||||
11:20 | - | 11:30 | |||||||
11:30 | - | 11:40 | |||||||
11:40 | - | 11:50 | |||||||
11:50 | - | 12:00 | |||||||
12:00 | - | 12:10 | Lunch (12:00 - 13:20) |
Lunch (12:00 - 13:20) |
Lunch (12:00 - 13:20) |
||||
12:10 | - | 12:20 | |||||||
12:20 | - | 12:30 | |||||||
12:30 | - | 12:40 | Lunch (12:30 - 13:30) |
||||||
12:40 | - | 12:50 | |||||||
12:50 | - | 13:00 | |||||||
13:00 | - | 13:10 | |||||||
13:10 | - | 13:20 | |||||||
13:20 | - | 13:30 | Oral 2 (13:20 - 14:20) |
Oral 3 (13:20 - 14:20) |
Oral 6 (13:20 - 14:20) |
||||
13:30 | - | 13:40 | Tutorial (13:30 - 18:00) |
||||||
13:40 | - | 13:50 | |||||||
13:50 | - | 14:00 | |||||||
14:00 | - | 14:10 | |||||||
14:10 | - | 14:20 | |||||||
14:20 | - | 14:30 | Industry 1 (14:20 - 15:20) |
Industry 2 (14:20 - 15:00) |
Industry 3 (14:20 - 15:00) |
||||
14:30 | - | 14:40 | |||||||
14:40 | - | 14:50 | |||||||
14:50 | - | 15:00 | |||||||
15:00 | - | 15:10 | Poster / Demo 2 (15:00 - 16:40) |
Poster / Demo 3 (15:00 - 16:40) |
|||||
15:10 | - | 15:20 | |||||||
15:20 | - | 15:30 | Poster / Demo 1 (15:20 - 17:00) |
||||||
15:30 | - | 15:40 | |||||||
15:40 | - | 15:50 | |||||||
15:50 | - | 16:00 | |||||||
16:00 | - | 16:10 | |||||||
16:10 | - | 16:20 | |||||||
16:20 | - | 16:30 | |||||||
16:30 | - | 16:40 | |||||||
16:40 | - | 16:50 | Oral 4 (16:40 - 18:00) |
Oral 7 (16:40 - 18:00) |
|||||
16:50 | - | 17:00 | |||||||
17:00 | - | 17:10 | Invited Talk 1 (17:00 - 18:00) Antonio Torralba (MIT) |
||||||
17:10 | - | 17:20 | |||||||
17:20 | - | 17:30 | |||||||
17:30 | - | 17:40 | |||||||
17:40 | - | 17:50 | |||||||
17:50 | - | 18:00 |
Monday, August 8 | |||||||
09:00 | - | 18:00 | Registration | ||||
09:40 | - | 10:00 | Opening | ||||
10:00 | - | 18:00 | Exhibition | ||||
10:00 | - | 11:00 | Oral 1 좌장: 김현우 교수(고려대) |
Authors | Title | ||
10:00 | - | 10:20 | MON-O-01 | Juil Koo, Ian Huang, Panos Achlioptas, Leonidas J. Guibas, Minhyuk Sung | PartGlot: Learning Shape Part Segmentation from Language Reference Games | ||
10:20 | - | 10:40 | MON-O-02 | Sohyun Lee, Taeyoung Son, Suha Kwak | FIFO: Learning Fog-invariant Features for Foggy Scene Segmentation | ||
10:40 | - | 11:00 | MON-O-03 | Hanul Kim, Su-Min Choi, Chang-Su Kim, Yeong Jun Koh | Representative Color Transform for Image Enhancement | ||
11:00 | - | 12:00 | Doctoral Consortium 좌장: 권준석 교수(중앙대) |
Presenter | Title |
||
11:00 | - | 11:15 | DC-01 | Walid Abdullah (한국외대) | Reinforcement Learning Agents for Anatomical Landmark Localization | ||
11:15 | - | 11:30 | DC-02 | 박송 (네이버 AI Research) | Few-shot font style transfer with localized style representations | ||
11:30 | - | 11:45 | DC-03 | 조나단 사무엘 (서울대) | Pseudo-data Exploration and Dynamic Adaptive Learning: Leveraging the Both Worlds to Tackle Various Neural-based Vision Tasks | ||
11:45 | - | 12:00 | DC-04 | 김범수 (LG AI Research) | Towards Real-Time Human-Object Interaction Detection | ||
13:20 | - | 14:20 | Oral 2 좌장: 김원준 교수(건국대) |
Authors | Title | ||
13:20 | - | 13:40 | MON-O-04 | Jinwoo Jeon, Jaechang Kim, Kangwook Lee, Sewoong Oh, Jungseul Ok | Gradient Inversion with Generative Image Prior | ||
13:40 | - | 14:00 | MON-O-05 | Gwanghyun Kim, Taesung Kwon, Jong Chul Ye | DiffusionCLIP: Text-Guided Diffusion Models for Robust Image Manipulation | ||
14:00 | - | 14:20 | MON-O-06 | Kyoungkook Kang, Seongtae Kim, Sunghyun Cho | GAN Inversion for Out-of-Range Images With Geometric Transformations | ||
14:20 | - | 15:20 | Industry 1 좌장: 곽수하 교수(POSTECH) |
Presenter | Title |
||
14:20 | - | 14:40 | MON-S-01 | 루닛(Lunit) - Sergio Pereira(VP of Research) |
AI in Oncology and Histopathology: Leveraging Large Image Data in Practice |
||
14:40 | - | 15:00 | MON-S-02 | 스트라드비전(StradVision) - 김준환(CEO), 김기재(Head of P&C) | AI Assisted Driving for Everyone |
||
15:00 | - | 15:20 | MON-S-03 | 포티투닷(42dot) - 정성균(이사) | An Introduction to Autonomous Driving for Real-road Environment |
||
15:20 | - | 17:00 | Poster 1 | Authors | Title | ||
MON-P-01 | Juil Koo, Ian Huang, Panos Achlioptas, Leonidas J. Guibas, Minhyuk Sung | PartGlot: Learning Shape Part Segmentation from Language Reference Games | |||||
MON-P-02 | Sohyun Lee, Taeyoung Son, Suha Kwak | FIFO: Learning Fog-invariant Features for Foggy Scene Segmentation | |||||
MON-P-03 | Hanul Kim, Su-Min Choi, Chang-Su Kim, Yeong Jun Koh | Representative Color Transform for Image Enhancement | |||||
MON-P-04 | Jinwoo Jeon, Jaechang Kim, Kangwook Lee, Sewoong Oh, Jungseul Ok | Gradient Inversion with Generative Image Prior | |||||
MON-P-05 | Gwanghyun Kim, Taesung Kwon, Jong Chul Ye | DiffusionCLIP: Text-Guided Diffusion Models for Robust Image Manipulation | |||||
MON-P-06 | Kyoungkook Kang, Seongtae Kim, Sunghyun Cho | GAN Inversion for Out-of-Range Images With Geometric Transformations | |||||
MON-P-07 | Juwon Kang, Sohyun Lee, Namyup Kim, Suha Kwak | Style Neophile: Constantly Seeking Novel Styles for Domain Generalization | |||||
MON-P-08 | Geon Yeong Park, Sang Wan Lee | Information-Theoretic Regularization for Multi-Source Domain Adaptation | |||||
MON-P-09 | Dayoung Gong, Joonseok Lee, Manjin Kim, Seong Jong Ha, Minsu Cho | Future Transformer for Long-term Action Anticipation | |||||
MON-P-10 | Jaeyoung Yoo, Hojun Lee, Inseop Chung, Geonseok Seo, Nojun Kwak | Training Multi-Object Detector by Estimating Bounding Box Distribution for Input Image | |||||
MON-P-11 | Soohyun Kim, Jongbeom Baek, Jihye Park, Gyeongnyeon Kim, Seungryong Kim | InstaFormer: Instance-Aware Image-to-Image Translation with Transformer | |||||
MON-P-12 | Sunghwan Hong, Seungryong Kim | Deep Matching Prior: Test-Time Optimization for Dense Correspondence | |||||
MON-P-13 | Kyungjune Baek, Hyunjung Shim | Commonality in Natural Images Rescues GANs: Pretraining GANs with Generic and Privacy-free Synthetic Data | |||||
MON-P-14 | Chanyong Jung, Gihyun Kwon, Jong Chul Ye | Exploring Patch-wise Semantic Relation for Contrastive Learning in Image-to-Image Translation Tasks | |||||
MON-P-15 | Gihyun Kwon, Jong Chul Ye | CLIPstyler: Image Style Transfer with a Single Text Condition | |||||
MON-P-16 | Seungwook Kim, Juhong Min, Minsu Cho | TransforMatcher: Match-to-Match Attention for Semantic Correspondence | |||||
MON-P-17 | Namyup Kim, Dongwon Kim, Cuiling Lan, Wenjun Zeng, Suha Kwak | ReSTR: Convolution-free Referring Image Segmentation Using Transformers | |||||
MON-P-18 | Sungmin Cha, beomyoung kim, YoungJoon Yoo, Taesup Moon | SSUL: Semantic Segmentation with Unknown Label for Exemplar-based Class-Incremental Learning | |||||
MON-P-19 | Jin Kim, Jiyoung Lee, Jungin Park, Dongbo Min, Kwanghoon Sohn | Pin the Memory: Learning to Generalize Semantic Segmentation | |||||
MON-P-20 | Jaehui Hwang, Jun-Hyuk Kim, Jun-Ho Choi, Jong-Seok Lee | Just One Moment: Structural Vulnerability of Deep Action Recognition Against One Frame Attack | |||||
MON-P-21 | Sangwon Jung, Sanghyuk Chun, Taesup Moon | Learning Fair Classifiers with Partially Annotated Group Labels | |||||
MON-P-22 | Hyomin Kim, Jungeon Kim, Jaewon Kam, Jaesik Park, Seungyong Lee | Deep Virtual Markers for Articulated 3D Shapes | |||||
MON-P-23 | Seokju Lee, Francois Rameau, Fei Pan, In So Kweon | Attentive and Contrastive Learning for Joint Depth and Motion Field Estimation | |||||
MON-P-24 | Donghun Kang, Hyeonjoong Jang, Jungeon Lee, Chong-Min Kyung, Min H. Kim | Uniform Subdivision of Omnidirectional Camera Space for Efficient Spherical Stereo Matching | |||||
MON-P-25 | Sukjun Hwang, Miran Heo, Seoung Wug Oh, Seon Joo Kim | Video Instance Segmentation using Inter-Frame Communication Transformers | |||||
MON-P-26 | Minsu Kim, Joanna Hong, Se Jin Park, Yong Man Ro | Multi-Modality Associative Bridging Through Memory: Speech Sound Recollected From Face Video | |||||
MON-P-27 | Hyesong Choi, Hunsang Lee, Sunkyung Kim, Sunok Kim, Seungryong Kim, Kwanghoon Sohn, Dongbo Min | Adaptive Confidence Thresholding for Monocular Depth Estimation | |||||
MON-P-28 | Jungin Park, Jiyoung Lee, Ig-Jae Kim, Kwanghoon Sohn | Probabilistic Representations for Video Contrastive Learning | |||||
MON-P-29 | Junyong Lee, Myeonghee Lee, Sunghyun Cho, Seungyong Lee | Reference-based Video Super-Resolution Using Multi-Camera Video Triplets | |||||
MON-P-30 | Jinwoo Nam, Daechul Ahn, Dongyeop Kang, Seong Jong Ha, Jonghyun Choi | Zero-Shot Natural Language Video Localization | |||||
MON-P-31 | Yeongwoo Nam, Mohammad Mostafavi, Kuk-Jin Yoon, Jonghyun Choi | Stereo Depth from Events Cameras: Concentrate and Focus on the Future | |||||
MON-P-32 | Chaoning Zhang, Kang Zhang, Trung X. Pham, Axi Niu, Zhinan Qiao, Chang D. Yoo, In So Kweon | Towards Understanding and Simplifying MoCo: Dual Temperature Helps Contrastive Learning without Many Negative Samples | |||||
MON-P-33 | GyuTae Park, SungJoon Son, JaeYoung Yoo, SeHo Kim, Nojun Kwak | MatteFormer: Transformer-Based Image Matting via Prior-Tokens | |||||
15:20 | - | 17:00 | Demo 1 | Presenter | Title | ||
MON-D | 임종우 교수(한양대 컴퓨터비전연구실) | 360-degree Omnidirectional Depth and Motion Sensing | |||||
17:00 | - | 18:00 | Invited Talk 1 좌장: 이경무 교수(서울대) |
Presenter | Title | ||
17:00 | - | 18:00 | MON-I | Antonio Torralba (MIT) | Learning to See by Looking at Noise Abstract: The importance of data in modern computer vision is hard to overstate. The ImageNet dataset, with its millions of labelled images, is widely thought to have spurred the era of deep learning, and since then the scale of vision datasets has been increasing at a rapid pace. These datasets come with costs: curation is expensive, and they inherit human biases. To counter these costs, interest has surged in learning with unlabeled images as it avoids the curation efforts, or using simulated environments, but content creation is also labor intensive. In this talk I will describe our work in trying to reduce the need for data. We will start by trying to get rid of the annotation effort an explore several self-supervised tasks using multimodal data such as auditory and tactile information. Finally, we will go a step further and ask if we can do away with real image datasets entirely, instead learning from noise processes. Noise processes produce images that are reminiscent of abstract art, where images contain textures and shapes, but there are no recognizable objects. Our findings show that good performance on real images can be achieved even with training images that are far from realistic. |
||
Tuesday, August 9 | |||||||
10:00 | - | 18:00 | Registration / Exhibition | ||||
10:00 | - | 11:00 | Invited Talk 2 좌장: 권인소 교수(KAIST) |
Presenter | Title | ||
10:00 | - | 11:00 | TUE-I | Jitendra Malik (UC Berkeley) | Perception and Action Abstract: I believe that the primary challenge for creating AI is to first master the link between sensory perception and motor control which, over the course of biological evolution, has provided the substrate for the development of capabilities such as language and abstract thought. Recently, we have seen major progress in core computer vision problems such as recognition and reconstruction, and the time is ripe to bootstrap these to advance core robotics problems such as locomotion, navigation, and manipulation. In my talk I will present various results along these directions. |
||
11:00 | - | 12:00 | KCVS General Meeting | ||||
13:20 | - | 14:20 | Oral 3 좌장: 김광인 교수(UNIST) |
Authors | Title | ||
13:20 | - | 13:40 | TUE-O-01 | Nayoung Kim, Seong Jong Ha, Je-Won Kang | Video Question Answering Using Language-Guided Deep Compressed-Domain Video Feature | ||
13:40 | - | 14:00 | TUE-O-02 | Jaeho Lee, Jihoon Tack, Namhoon Lee, Jinwoo Shin | Meta-Learning Sparse Implicit Neural Representations | ||
14:00 | - | 14:20 | TUE-O-03 | Yoonwoo Jeong, Seokjun Ahn, Christopher Choy, Anima Anandkumar, Minsu Cho, Jaesik Park | Self-Calibrating Neural Radiance Fields | ||
14:20 | - | 15:00 | Industry 2 좌장: 최종현 교수(연세대) |
Presenter | Title |
||
14:20 | - | 14:40 | TUE-S-01 | 현대자동차(Hyundai Motors) - 이재호(팀장) | 모빌리티의 미래 '로보틱스' |
||
14:40 | - | 15:00 | TUE-S-02 | 퓨리오사AI(FuriosaAI) - 백준호(CEO) | FuriosaAI WARBOY : AI Inference Chip for the Most Advanced Vision Applications |
||
15:00 | - | 16:40 | Poster 2 | Authors | Title | ||
TUE-P-01 | Nayoung Kim, Seong Jong Ha, Je-Won Kang | Video Question Answering Using Language-Guided Deep Compressed-Domain Video Feature | |||||
TUE-P-02 | Jaeho Lee, Jihoon Tack, Namhoon Lee, Jinwoo Shin | Meta-Learning Sparse Implicit Neural Representations | |||||
TUE-P-03 | Yoonwoo Jeong, Seokjun Ahn, Christopher Choy, Anima Anandkumar, Minsu Cho, Jaesik Park | Self-Calibrating Neural Radiance Fields | |||||
TUE-P-04 | HyeongJoo Hwang, Geon-Hyeong Kim, Seunghoon Hong, Kee-Eung Kim | Multi-View Representation Learning via Total Correlation Objective | |||||
TUE-P-05 | Bumsoo Kim, Jonghwan Mun, Kyoung-Woon On, Minchul Shin, Junhyun Lee, Eun-Sol Kim | MSTR: Mutli-Scale Transformer for End-to-End Human-Object Interaction Detection | |||||
TUE-P-06 | Inhwan Bae, Jin-Hwi Park, Hae-Gon Jeon | Non-Probability Sampling Network for Stochastic Human Trajectory Prediction | |||||
TUE-P-07 | Jihwan Park, SeungJun Lee, Hwan Heo, Hyeong Kyu Choi, Hyunwoo J. Kim | Consistency Learning via Decoding Path Augmentation for Transformers in Human Object Interaction Detection | |||||
TUE-P-08 | Hochang Rhee, Yeong Il Jang, Seyun Kim, Nam Ik Cho | LC-FDNet: Learned Lossless Image Compression with Frequency Decomposition Network | |||||
TUE-P-09 | Youmin Kim, Jinbae Park, YounHo Jang, Muhammad Ali, Tae-Hyun Oh, Sung-Ho Bae | Distilling Global and Local Logits With Densely Connected Relations | |||||
TUE-P-10 | Hyolim Kang, Jinwoo Kim, Taehyun Kim, Seon Joo Kim | UBoCo: Unsupervised Boundary Contrastive Learning for Generic Event Boundary Detection | |||||
TUE-P-11 | Junho Kim, Jaehyeok Bae, Gangin Park, Dongsu Zhang, Young Min Kim | N-ImageNet: Towards Robust, Fine-Grained Object Recognition With Event Cameras | |||||
TUE-P-12 | Suhyeon Lee, Hongje Seong, Seongwon Lee, Euntai Kim | WildNet: Learning Domain Generalized Semantic Segmentation from the Wild | |||||
TUE-P-13 | Jin-Man Park, Ue-Hwan Kim, Seon-Hoon Lee, Jong-Hwan Kim | Dual Task Learning by Leveraging Both Dense Correspondence and Mis-Correspondence for Robust Change Detection With Imperfect Matches | |||||
TUE-P-14 | Beomyoung Kim, YoungJoon Yoo, Chae Eun Rhee, Junmo Kim | Beyond Semantic to Instance Segmentation: Weakly-Supervised Instance Segmentation via Semantic Knowledge Transfer and Self-Refinement | |||||
TUE-P-15 | Minhyun Lee, Dongseob Kim, Hyunjung Shim | Threshold Matters in WSSS: Manipulating the Activation for the Robust and Accurate Segmentation Model Against Thresholds | |||||
TUE-P-16 | Ahyun Seo, Byungjin Kim, Suha Kwak, Minsu Cho | Reflection and Rotation Symmetry Detection via Equivariant Learning | |||||
TUE-P-17 | Sanghun Jung, Jungsoo Lee, Daehoon Gwak, Sungha Choi, Jaegul Choo | Standardized Max Logits: A Simple yet Effective Approach for Identifying Unexpected Road Obstacles in Urban-Scene Segmentation | |||||
TUE-P-18 | Kyungsu Lee, Haeyun Lee, Jae Youn Hwang | Self-Mutating Network for Domain Adaptive Segmentation in Aerial Images | |||||
TUE-P-19 | Byung-Kwan Lee, Junho Kim, Yong Man Ro | Masking Adversarial Damage: Finding Adversarial Saliency for Robust and Sparse Network | |||||
TUE-P-20 | Philip Chikontwe, Soopil Kim, Sang Hyun Park | CAD: Co-Adapting Discriminative Features for Improved Few-Shot Classification | |||||
TUE-P-21 | Sukmin Yun, Hankook Lee, Jaehyung Kim, Jinwoo Shin | Patch-level Representation Learning for Self-supervised Vision Transformers | |||||
TUE-P-22 | Junho Kim, Yunjey Choi, Youngjung Uh | Feature Statistics Mixing Regularization for Generative Adversarial Networks | |||||
TUE-P-23 | Obin Kwon, Nuri Kim, Yunho Choi, Hwiyeon Yoo, Jeongho Park, Songhwai Oh | Visual Graph Memory With Unsupervised Representation for Visual Navigation | |||||
TUE-P-24 | Youngho Yoon, Inchul Chung, Lin Wang, Kuk-Jin Yoon | SphereSR: 360° Image Super-Resolution with Arbitrary Projection via Continuous Spherical Image Representation | |||||
TUE-P-25 | Hyeokjun Kweon, Sung-Hoon Yoon, Hyeonseong Kim, Daehee Park, Kuk-Jin Yoon | Unlocking the Potential of Ordinary Classifier: Class-Specific Adversarial Erasing Framework for Weakly Supervised Semantic Segmentation | |||||
TUE-P-26 | Jungsoo Lee, Eungyeup Kim, Juyoung Lee, Jihyeon Lee, Jaegul Choo | Learning Debiased Representation via Disentangled Feature Augmentation | |||||
TUE-P-27 | Jaewon Lee, Kyong Hwan Jin | Local Texture Estimator for Implicit Representation Function | |||||
TUE-P-28 | Junho Kim, Changwoon Choi, Hojun Jang, Young Min Kim | PICCOLO: Point Cloud-Centric Omnidirectional Localization | |||||
TUE-P-29 | Sihyeon Kim, Sanghyeok Lee, Dasol Hwang, Jaewon Lee, Seong Jae Hwang, Hyunwoo J. Kim | Point Cloud Augmentation With Weighted Local Transformations | |||||
TUE-P-30 | Chunghyun Park, Yoonwoo Jeong, Minsu Cho, Jaesik Park | Fast Point Transformer | |||||
TUE-P-31 | Pilhyeon Lee, Hyeran Byun | Learning Action Completeness From Points for Weakly-Supervised Temporal Action Localization | |||||
TUE-P-32 | Inkyu Shin, Yi-Hsuan Tsai, Bingbing Zhuang, Samuel Schulter, Buyu Liu, Sparsh Garg, In So Kweon, Kuk-Jin Yoon | MM-TTA: Multi-Modal Test-Time Adaptation for 3D Semantic Segmentation | |||||
TUE-P-33 | Hyunmin Lee, Jaesik Park | Instance-wise Occlusion and Depth Orders in Natural Scenes | |||||
15:00 | - | 16:40 | Demo 2 | Presenter | Title | ||
15:00 | - | 16:40 | TUE-D | 이수찬 교수(국민대), 박상준 교수(분당서울대병원) | Seoul Retinal Vessel Analysis Library | ||
16:40 | - | 18:00 | Oral 4 좌장: 조민수 교수(POSTECH) |
Authors | Title | ||
16:40 | - | 17:00 | TUE-O-04 | HyeongJoo Hwang, Geon-Hyeong Kim, Seunghoon Hong, Kee-Eung Kim | Multi-View Representation Learning via Total Correlation Objective | ||
17:00 | - | 17:20 | TUE-O-05 | Bumsoo Kim, Jonghwan Mun, Kyoung-Woon On, Minchul Shin, Junhyun Lee, Eun-Sol Kim | MSTR: Mutli-Scale Transformer for End-to-End Human-Object Interaction Detection | ||
17:20 | - | 17:40 | TUE-O-06 | Inhwan Bae, Jin-Hwi Park, Hae-Gon Jeon | Non-Probability Sampling Network for Stochastic Human Trajectory Prediction | ||
17:40 | - | 18:00 | TUE-O-07 | Jihwan Park, SeungJun Lee, Hwan Heo, Hyeong Kyu Choi, Hyunwoo J. Kim | Consistency Learning via Decoding Path Augmentation for Transformers in Human Object Interaction Detection | ||
Wednesday, August 10 | |||||||
10:00 | - | 18:00 | Registration / Exhibition | ||||
10:00 | - | 11:00 | Oral 5 좌장: 황원준 교수(아주대) |
Authors | Title | ||
10:00 | - | 10:20 | WED-O-01 | Daehee Kim, Youngjun Yoo, Seunghyun Park, Jinkyu Kim, Jaekoo Lee | SelfReg: Self-Supervised Contrastive Regularization for Domain Generalization | ||
10:20 | - | 10:40 | WED-O-02 | Byeong-Ju Han, Kuhyeun Ko, Jae-Young Sim | End-to-End Trainable Trident Person Search Network Using Adaptive Gradient Propagation | ||
10:40 | - | 11:00 | WED-O-03 | Seunghun Lee, Wonhyeok Choi, Changjae Kim, Minwoo Choi, Sunghoon Im | ADAS: A Direct Adaptation Strategy for Multi-Target Domain Adaptive Semantic Segmentation | ||
11:00 | - | 12:00 | Invited Talk 3 좌장: 손광훈 교수(연세대) |
Presenter | Title | ||
11:00 | - | 12:00 | WED-I | Chelsea Finn (Stanford Univ.) | Robust Deep Networks through Invariance and Adaptation Abstract: While we have seen immense progress in machine learning, a critical shortcoming of current methods lies in handling distribution shift between training and deployment. Distribution shift is pervasive in real-world problems ranging from natural variation in the distribution over locations or domains, to shift in the distribution arising from different decision making policies, to shifts over time as the world changes. In this talk, I'll start by discussing our efforts in benchmarking machine learning methods under multiple natural occurrences of distribution shift, including domain shift, subpopulation shift, and gradual shift over time. Then, I will talk about two new algorithms that can mitigate certain forms of covariate shift. The first leverages domain information to learn domain invariant functions, without using any explicit regularizers. This leads to significant and consistent gains on a variety of natural distribution shift problems. The second instead leverages unlabeled target distribution data to learn a diverse set of functions. In doing so, it is able to address major limitations of prior robustness works: it doesn’t require labeled data from the test distribution to tune hyperparameters, and it can handle an extreme version of spurious correlations where there is a perfect correlation between the spurious attribute and label. I'll conclude by discussing important open questions for future work.
|
||
13:20 | - | 14:20 | Oral 6 좌장: 임성훈 교수(DGIST) |
Authors | Title | ||
13:20 | - | 13:40 | WED-O-04 | Sang-Heon Shim, Sangeek Hyun, DaeHyun Bae, Jae-Pil Heo | Local Attention Pyramid for Scene Image Generation | ||
13:40 | - | 14:00 | WED-O-05 | Hyeonjun Sim, Jihyong Oh, Munchurl Kim | XVFI: eXtreme Video Frame Interpolation | ||
14:00 | - | 14:20 | WED-O-06 | Jeany Son | Contrastive Learning for Space-Time Correspondence via Self-cycle Consistency | ||
14:20 | - | 15:00 | Industry 3 좌장: 김태현 교수(한양대) |
Presenter | Title |
||
14:20 | - | 14:40 | WED-S-01 | 네이버랩스(NAVER LAbS) - 김수정(리더), 연수용(리더) | Computer Vision for In/Outdoor Mobility |
||
14:40 | - | 15:00 | WED-S-02 | 퀄컴(Qaulcomm) - 김덕훈(상무) |
Qualcomm Autonomous Driving |
||
15:00 | - | 16:40 | Poster 3 | Authors | Title | ||
WED-P-01 | Daehee Kim, Youngjun Yoo, Seunghyun Park, Jinkyu Kim, Jaekoo Lee | SelfReg: Self-Supervised Contrastive Regularization for Domain Generalization | |||||
WED-P-02 | Byeong-Ju Han, Kuhyeun Ko, Jae-Young Sim | End-to-End Trainable Trident Person Search Network Using Adaptive Gradient Propagation | |||||
WED-P-03 | Seunghun Lee, Wonhyeok Choi, Changjae Kim, Minwoo Choi, Sunghoon Im | ADAS: A Direct Adaptation Strategy for Multi-Target Domain Adaptive Semantic Segmentation | |||||
WED-P-04 | Sang-Heon Shim, Sangeek Hyun, DaeHyun Bae, Jae-Pil Heo | Local Attention Pyramid for Scene Image Generation | |||||
WED-P-05 | Hyeonjun Sim, Jihyong Oh, Munchurl Kim | XVFI: eXtreme Video Frame Interpolation | |||||
WED-P-06 | Jeany Son | Contrastive Learning for Space-Time Correspondence via Self-cycle Consistency | |||||
WED-P-07 | Dong-Hwan Jang, Sanghyeok Chu, Joonhyuk Kim, Bohyung Han | Pooling Revisited: Your Receptive Field is Sub-optimal | |||||
WED-P-08 | Dongkwon Jin, Wonhui Park, Seong-Gyun Jeong, Heeyeon Kwon, Chang-Su Kim | Eigenlanes: Data-Driven Lane Descriptors for Structurally Diverse Lanes | |||||
WED-P-09 | Farkhod Makhmudkhujaev, Sungeun Hong, In Kyu Park | Re-Aging GAN: Toward Personalized Face Age Transformation | |||||
WED-P-10 | Jihwan Bang, Hyunseo Koh, Seulki Park, Hwanjun Song, Jung-Woo Ha, Jonghyun Choi | Online Continual Learning on a Contaminated Data Stream with Blurry Task Boundaries | |||||
WED-P-11 | Wencan Cheng, Jae Hyun Park, Jong Hwan Ko | HandFoldingNet: A 3D Hand Pose Estimation Network Using Multiscale-Feature Guided Folding of a 2D Hand Skeleton | |||||
WED-P-12 | Nyeong-Ho Shin, Seon-Ho Lee, Chang-Su Kim | Moving Window Regression: A Novel Approach to Ordinal Regression | |||||
WED-P-13 | Jung Hyun Lee, Jihun Yun, Sung Ju Hwang, Eunho Yang | Cluster-Promoting Quantization With Bit-Drop for Minimizing Network Quantization Loss | |||||
WED-P-14 | Mijeong Kim, Seonguk Seo, Bohyung Han | InfoNeRF: Ray Entropy Minimization for Few-Shot Neural Volume Rendering | |||||
WED-P-15 | Guoyuan An, Yuchi Huo, Sung-eui Yoon | Hypergraph Propagation and Community Selection for Objects Retrieval | |||||
WED-P-16 | Geonwoon Jang, Wooseok Lee, Sanghyun Son, Kyoung Mu Lee | C2N: Practical Generative Noise Modeling for Real-World Denoising | |||||
WED-P-17 | Junghun Oh, Heewon Kim, Seungjun Nah, Cheeun Hong, Jonghyun Choi, Kyoung Mu Lee | Attentive Fine-Grained Structured Sparsity for Image Restoration | |||||
WED-P-18 | Dohyung Kim, Junghyup Lee, Bumsub Ham | Distance-Aware Quantization | |||||
WED-P-19 | Haechan Noh, Taeho Kim, Jae-Pil Heo | Product Quantizer Aware Inverted Index for Scalable Nearest Neighbor Search | |||||
WED-P-20 | Hongsuk Choi, Gyeongsik Moon, JoonKyu Park, Kyoung Mu Lee | Learning to Estimate Robust 3D Human Mesh from In-the-Wild Crowded Scenes | |||||
WED-P-21 | Young Kyun Jang, Nam Ik Cho | Self-Supervised Product Quantization for Deep Unsupervised Image Retrieval | |||||
WED-P-22 | Jongin Lim, Sangdoo Yun, Seulki Park, Jin Young Choi | Hypergraph-Induced Semantic Tuplet Loss for Deep Metric Learning | |||||
WED-P-23 | Seulki Park, Youngkyu Hong, Byeongho Heo, Sangdoo Yun, Jin Young Choi | The Majority Can Help the Minority: Context-rich Minority Oversampling for Long-tailed Classification | |||||
WED-P-24 | Wonyong Jeong, Hayeon Lee, Geon Park, Eunyoung Hyung, Jinheon Baek, Sung Ju Hwang | Task-Adaptive Neural Network Search with Meta-Contrastive Learning | |||||
WED-P-25 | Youngwan Lee, Jonghee Kim, Jeffrey Willette, Sung Ju Hwang | MPViT : Multi-Path Vision Transformer for Dense Prediction | |||||
WED-P-26 | Jaesung Choe, Sunghoon Im, Francois Rameau, Minjun Kang, In So Kweon | VolumeFusion: Deep Depth Fusion for 3D Scene Reconstruction | |||||
WED-P-27 | Junyoung Byun, Seungju Cho, Myung-Joon Kwon, Hee-Seon Kim, Changick Kim | Improving the Transferability of Targeted Adversarial Examples through Object-Based Diverse Input | |||||
WED-P-28 | Donghyeon Baek, Youngmin Oh, Bumsub Ham | Exploiting a Joint Embedding Space for Generalized Zero-Shot Semantic Segmentation | |||||
WED-P-29 | Wonchul Son, Jaemin Na, Junyong Choi, Wonjun Hwang | Densely Guided Knowledge Distillation Using Multiple Teacher Assistants | |||||
WED-P-30 | Kunliang Liu, Ouk Choi, Jianming Wang, Wonjun Hwang | CDGNet: Class Distribution Guided Network for Human Parsing | |||||
WED-P-31 | Taekyung Kim, Jaehoon Choi, Seokeon Choi, Dongki Jung, Changick Kim | Just a Few Points Are All You Need for Multi-View Stereo: A Novel Semi-Supervised Learning Method for Multi-View Stereo | |||||
WED-P-32 | Minsoo Kang, Jaeyoo Park, Bohyung Han | Class-Incremental Learning by Knowledge Distillation with Adaptive Feature Consolidation | |||||
15:00 | - | 16:40 | Demo 3 | Presenter | Title | ||
15:00 | - | 16:40 | WED-D | 이승용 교수(POSTECH 컴퓨터그래픽스연구실) | Deep Computational Photography Library | ||
16:40 | - | 18:00 | Oral 7 좌장: 심현정 교수(KAIST) |
Authors | Title | ||
16:40 | - | 17:00 | WED-O-07 | Dong-Hwan Jang, Sanghyeok Chu, Joonhyuk Kim, Bohyung Han | Pooling Revisited: Your Receptive Field is Sub-optimal | ||
17:00 | - | 17:20 | WED-O-08 | Dongkwon Jin, Wonhui Park, Seong-Gyun Jeong, Heeyeon Kwon, Chang-Su Kim | Eigenlanes: Data-Driven Lane Descriptors for Structurally Diverse Lanes | ||
17:20 | - | 17:40 | WED-O-09 | Farkhod Makhmudkhujaev, Sungeun Hong, In Kyu Park | Re-Aging GAN: Toward Personalized Face Age Transformation | ||
17:40 | - | 18:00 | WED-O-10 | Jihwan Bang, Hyunseo Koh, Seulki Park, Hwanjun Song, Jung-Woo Ha, Jonghyun Choi | Online Continual Learning on a Contaminated Data Stream with Blurry Task Boundaries | ||
Thursday, August 11 (ZOOM) | |||||||
08:50 | - | 12:30 | Workshop | Organizer | Title | ||
08:50 | - | 12:30 | THU-W | 42dot |
42dot Autonomous Tech Day |
||
13:30 | - | 18:00 | Programming Practice | Presenter | Title | ||
13:30 | - | 15:00 | THU-PP-01 | 김선주 교수(연세대) | Video Instance Segmentation using Inter-Frame Communication Transformers | ||
15:00 | - | 16:30 | THU-PP-02 | 심현정 교수(KAIST) | Threshold Matters in WSSS: Manipulating the Activation for the Robust and Accurate Segmentation Model Against Thresholds, | ||
16:30 | - | 18:00 | THU-PP-03 | 김현우 교수(고려대) | Consistency Learning via Decoding Path Augmentation for Transformers in Human Object Interaction Detection |