[DGIF연사미리보기 #4] Ming-Hsuan Yang 교수님과 그의 연구

Home[I] 정보통신정보통신・컴퓨터[I]

curated by DGIST LIBRARY
 

DGIF 2017 정보통신융합전공은 "Visual Intelligence"라는 핫한 주제로 10명의 연사들의 강연이 마련되어 있습니다. 캘리포니아 메르세드대학의 Ming-Hsuan Yang 교수님이 "Recent Results on Image Editing and Learning Filters"를 주제로 발표합니다. FWCI 10이상으로 논문 영향력이 어마어마 하신 분의 연구는 어떨까요? 관련 논문과 책 등을 미리 읽어보시고, Computer Vision을 연구하시는 DGIST 곽수하, 조성현 교수님의 강연도 경청해주세요!

DGIF 2017: Visual Intelligence

Program

[Meeting Abstract] Recent Results on Image Editing and Learning Filters

▶ Date: Nov. 30 16:40~17:25
▶ Place:  L29, E7, DGIST
▶ Speaker: Ming-Hsuan Yang, UC Merced
▶ Title: Recent Results on Image Editing and Learning Filters
▶ Abstract
In the first part of this talk, I will present recent results on semantic-aware image editing. Skies are common backgrounds in photos but are often less interesting due to the time of photographing. Professional photographers correct this by using sophisticated tools with painstaking efforts that are beyond the command of ordinary users. In this work, we propose an automatic background replacement algorithm that can generate realistic, artifact-free images with diverse styles of skies. The key idea of our algorithm is to utilize visual semantics to guide the entire process including sky segmentation, search and replacement. First, we train a deep convolutional neural network for semantic scene parsing, which is used as visual prior to segment sky regions in a coarse-to-fine manner. Second, in order to find proper skies for replacement, we propose a data-driven sky search scheme based on semantic layout of the input image. Finally, to re-compose the stylized sky with the original foreground naturally, an appearance transfer method is developed to match statistics locally and semantically. We show that the proposed algorithm can automatically generate a set of visually pleasing results. In addition, we demonstrate the effectiveness of the proposed algorithm with extensive user studies.

In the second part, ....

연사의 이력(CV)

[Vision and Learning Lab] University of California, MercedVLLab - UC Merced

Ming-Hsuan Yang 교수님은 University of Illinois at Urbana-Champaign에서 박사학위를 받고, 2008년부터 현재까지 캘리포니아 메르세드대학(University of California, Merced)에서 교수로 재직중입니다. 2019년 IEEE 컴퓨터비전 컨퍼런스 공동의장으로 활동하고 있어 관련 전공 연구자들은 알아두시면 좋을 것 같네요! 

Ming-Hsuan Yang is a professor in Electrical Engineering and Computer Science at University of California, Merced. He received the PhD degree in Computer Science from the University of Illinois at Urbana-Champaign in 2000. He serves as an area chair for several conferences including IEEE Conference on Computer Vision and Pattern Recognition, IEEE International Conference on Computer Vision, European Conference on Computer Vision, Asian Conference on Computer, AAAI National Conference on Artificial Intelligence, and IEEE International Conference on Automatic Face and Gesture Recognition. He serves as a program co-chair for IEEE International Conference on Computer Vision in 2019 as well as Asian Conference on Computer Vision in 2014, and general co-chair for Asian Conference on Computer Vision in 2016. He serves as an associate editor of the IEEE Transactions on Pattern Analysis and Machine Intelligence (2007 to 2011), International Journal of Computer Vision, Computer Vision and Image Understanding, Image and Vision Computing, and Journal of Artificial Intelligence Research. Yang received the Google faculty award in 2009, and the Distinguished Early Career Research award from the UC Merced senate in 2011, the Faculty Early Career Development (CAREER) award from the National Science Foundation in 2012, and the Distinguished Research Award from UC Merced Senate in 2015.

Ming-Hsuan Yang's Reseasrch

Keyword

[Infographic] Research keywords

138건의 논문 FWCI가 10.59, 국제협력이 100건, 
Citation 수가 7,437건 

알고리즘, 컴퓨터 비전, 트랙킹, 패턴인식, 모델, 이미지 프로세싱, 픽셀 등 키워드가 많고 증가하고 있음 

Book

[DGIST Library Catalog] Face detection and gesture recognition for human-computer interaction소장자료

▶Author: Yang Ming-Hsuan, Ahuja Narendra
▶Title: Face detection and gesture recognition for human-computer interaction
▶Publisher: Kluwer Academic,2001.
▶Page: 182 p.
▶ISBN: 9780792374091, 0792374096 
▶Subject: Human-computer interaction, Image processingDigital techniques.

Book chapter

[SpringerLink] Toward Robust Online Visual Tracking

Yang MH., Ho J. (2011) Toward Robust Online Visual Tracking. In: Bhanu B., Ravishankar C., Roy-Chowdhury A., Aghajan H., Terzopoulos D. (eds) Distributed Video Sensor Networks. Springer, London

Article

[Article]  발표 주제와 관련한 주요 최신 논문들

1. Chi, Z.; Li, H.; Lu, H.; Yang, M. H., Dual Deep Network for Visual Tracking. ITIP 2017, 26 (4), 2005-2015.

2. Song, Y.; Bao, L.; He, S.; Yang, Q.; Yang, M. H., Stylizing face images via multiple exemplars. CVIU 2017, 162, 135-145.

3. Zhong, G.; Tsai, Y. H.; Yang, M. H., Weakly-supervised video scene co-parsing. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2017; Vol. 10111 LNCS, pp 20-36.

4. Tsai, Y. H.; Shen, X.; Lin, Z.; Sunkavalli, K.; Yang, M. H., Sky is not the limit: Semantic-aware sky replacement. ACM Transactions on Graphics 2016, 35 (4).

5. Zhang, L.; Yang, C.; Lu, H.; Ruan, X.; Yang, M. H., Ranking Saliency. ITPAM 2017, 39 (9), 1892-1904.

6. Ma, C.; Yang, C.-Y.; Yang, X.; Yang, M.-H., Learning a no-reference quality metric for single-image super-resolution. CVIU 2017, 158 (Supplement C), 1-16.

DGIST Speakers

Suha Kwak

[Meeting Abstract] Learning with Minimum Supervision for Semantic Segmentation

Suha Kwak
Daegu Gyeongbuk Intstitute of Science and Technology (DGIST)
Dec. 1 09:20~09:50

Semantic segmentation is a visual recognition task aiming to estimate pixel-level class labels in images. This problem has been recently handled by Deep Convolutional Neural Networks (DCNNs), and the state of the art based on DCNNs achieve impressive records on public benchmarks. However, learning DCNNs demands a large number of annotated training data while segmentation annotations in existing datasets are significantly limited in terms of both quantity and diversity due to the heavy annotation cost. Weakly supervised approaches tackle this issue by leveraging weak annotations such as bounding boxes and scribbles, which are either readily available in existing large-scale datasets or easily obtained thanks to their low annotation costs. In this talk, I will introduce our recent approaches to weakly supervised semantic segmentation based on image-level class label, which is the form of minimum supervision indicating only presence or absence of a semantic entity in an image. Learning semantic segmentation with image-level class label is a significantly ill-posed problem since neither object location nor shape is informed by the label. We tackled this challenging problem by employing (1) unsupervised techniques revealing low-level image structures, (2) web-crawled videos as additional data sources, and (3) DCNN architectures appropriate for learning segmentation with incomplete pixel-level annotations. I will conclude this talk with a few suggestions for future research directions worth to investigate for further improvement.

Sunghyun Cho

[Meeting Abstract] Recent Results in Image Deblurring

Sunghyun Cho
Daegu Gyeongbuk Intstitute of Science and Technology (DGIST)
Dec. 1 09:50~10:20

One goal in computational photography is to overcome limitations of traditional cameras such as blur. Blur, which is one of the most annoying artifacts in photographs, is caused by long exposure time or lens aberration. Deblurring is a problem to remove blur from a blurry image so that valuable hidden information can be revealed. Deblurring can benefit consumer cameras, medical imaging, aerial and satellite imaging, and many others. In this talk, I will briefly introduce some of our past and recent results in image deblurring including convergence analysis of deblurring, and deblurring in low-light environments. Besides deblurring, I will share my ideas on possible future directions in computational photography research.

* 불법, 상업/광고, 욕설 등의 댓글은 삭제될 수 있습니다.