{"id":882,"date":"2017-05-10T14:41:46","date_gmt":"2017-05-10T14:41:46","guid":{"rendered":"https:\/\/labsites.rochester.edu\/gsharma\/?page_id=882"},"modified":"2020-11-24T06:36:40","modified_gmt":"2020-11-24T06:36:40","slug":"computer-vision","status":"publish","type":"page","link":"https:\/\/labsites.rochester.edu\/gsharma\/research\/computer-vision\/","title":{"rendered":"Computer Vision\/Image Processing"},"content":{"rendered":"<p><img loading=\"lazy\" decoding=\"async\" class=\"alignleft wp-image-8182\" src=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/11\/cover_figure_PRIME-FP20.png\" alt=\"\" width=\"236\" height=\"115\" srcset=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/11\/cover_figure_PRIME-FP20.png 1251w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/11\/cover_figure_PRIME-FP20-300x146.png 300w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/11\/cover_figure_PRIME-FP20-1024x498.png 1024w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/11\/cover_figure_PRIME-FP20-768x374.png 768w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/11\/cover_figure_PRIME-FP20-624x304.png 624w\" sizes=\"auto, (max-width: 236px) 100vw, 236px\" \/><span style=\"font-size: 14pt;\"><strong><span style=\"font-size: 12pt;\">Deep Retinal Vessel Segmentation For Ultra-Widefield Fundus Photography<\/span>\u00a0<span id=\"sample-permalink\" style=\"font-size: 10pt;\">[<a href=\"https:\/\/labsites.rochester.edu\/gsharma\/research\/computer-vision\/deep-retinal-vessel-segmentation-for-ultra-widefield-fundus-photography\/\">project page<\/a>]<br \/>\n<\/span><\/strong><\/span>We propose an annotation-efficient method for vessel segmentation in ultra-widefield (UWF) fundus photography (FP) that does not require de novo labeled ground truth. Our method utilizes concurrently captured UWF fluorescein angiography (FA) images and iterates between a multi-modal registration and a weakly-supervised learning step. We construct a new dataset to facilitate further work on this problem.<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-size: 14pt;\"><strong><span style=\"font-size: 12pt;\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-6382 alignleft\" src=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/04\/cover_figure.jpg\" alt=\"\" width=\"236\" height=\"90\" srcset=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/04\/cover_figure.jpg 1961w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/04\/cover_figure-300x114.jpg 300w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/04\/cover_figure-1024x391.jpg 1024w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/04\/cover_figure-768x293.jpg 768w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/04\/cover_figure-1536x586.jpg 1536w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2020\/04\/cover_figure-624x238.jpg 624w\" sizes=\"auto, (max-width: 236px) 100vw, 236px\" \/>Deep Retinal Vessel Segmentation For Fluorescein Angiography<\/span>\u00a0<span id=\"sample-permalink\" style=\"font-size: 10pt;\">[<a href=\"https:\/\/labsites.rochester.edu\/gsharma\/research\/computer-vision\/deep-retinal-vessel-segmentation-for-fluorescein-angiography-fa-retinal-images\/\">project page<\/a>]<br \/>\n<\/span><\/strong><\/span>We propose a novel deep learning pipeline to detect retinal vessels in fluorescein angiography, a modality that has received limited attention in prior works, that reduces the effort required for generating labeled ground truth data. We release a new dataset to facilitate further research on this problem.<\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignleft wp-image-3932\" src=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/07\/WAMIRegProbFormul-300x208.png\" alt=\"\" width=\"236\" height=\"164\" srcset=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/07\/WAMIRegProbFormul-300x208.png 300w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/07\/WAMIRegProbFormul-624x432.png 624w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/07\/WAMIRegProbFormul.png 684w\" sizes=\"auto, (max-width: 236px) 100vw, 236px\" \/><\/p>\n<p><span style=\"font-size: 14pt;\"><strong><span style=\"font-size: 12pt;\">Large Scale Visual Data Analytics for Geospatial Applications<\/span>\u00a0<span id=\"sample-permalink\" style=\"font-size: 10pt;\">[<a href=\"https:\/\/labsites.rochester.edu\/gsharma\/research\/computer-vision\/geospatialanalytics\/\">project page<\/a>]<br \/>\n<\/span><\/strong><\/span><span style=\"font-size: 1rem;\">The widespread availability of high-resolution aerial imagery covering wide geographical areas is spurring a revolution in large scale visual data analytics. This project focuses on how analytics for wide-area motion imagery can be enhanced by incorporating information from geographical information systems enabled by pixel-accurate registration of vector roadmaps to the WAMI frames and exploitation of this information in vehicle tracking and 3D georegistration. The work highlights how computer vision applications are a fertile ground for incorporating machine learning and data science methodologies.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-4522 alignleft\" src=\"https:\/\/www.hajim.rochester.edu\/ece\/lding6\/wp-content\/uploads\/2017\/04\/sampleLidar_Cam-300x222.jpg\" alt=\"\" width=\"235\" height=\"174\" \/><\/p>\n<p><strong><br \/>\nFusing SfM and Lidar for Dense Accurate Depth Map Estimation [<a href=\"https:\/\/labsites.rochester.edu\/gsharma\/research\/computer-vision\/fusing-structure-from-motion-and-lidar-for-dense-accurate-depth-map-estimation\/\">project page<\/a>]<\/strong><br \/>\nWe present a novel framework for precisely estimating dense depth maps by combining 3D lidar scans with a set of uncalibrated camera RGB color images for the same scene. The approach is based on fusing structure from motion and lidar to precisely recover the transformation from 3D lidar space to 2D image plane. The 3D to 2D map is then utilized to estimate a dense depth map for each image. The framework does not require the relative position of lidar and camera to be fixed and the sensor can conveniently be deployed independently for data acquisition.<\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-1852 alignleft\" src=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/05\/ComputerVision-300x186.png\" alt=\"\" width=\"236\" height=\"146\" srcset=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/05\/ComputerVision-300x186.png 300w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/05\/ComputerVision.png 400w\" sizes=\"auto, (max-width: 236px) 100vw, 236px\" \/><\/p>\n<p><strong><span style=\"font-size: 14pt;\"><span style=\"font-size: 12pt;\"><br \/>\nPoint Cloud Analytics and Applications<\/span> <span id=\"sample-permalink\" style=\"font-size: 10pt;\">[<a href=\"https:\/\/labsites.rochester.edu\/gsharma\/research\/computer-vision\/point-cloud-analytics-and-application-architectural-biometrics\/\">project page<\/a>]<br \/>\n<\/span><\/span><\/strong><span id=\"sample-permalink\"><\/span><span id=\"edit-slug-buttons\"><\/span>Recent advances in the sensors used for lidar 3D scanners have substantially reduced their cost spurring an increase in the number of applications for which they are deployed. The work concentrates on point cloud analytics and its applications on different areas, e.g., humanities, sensor fusion, etc.<\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-5272 alignleft\" src=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2019\/06\/Fig2_JEI190015.png\" alt=\"\" width=\"236\" height=\"139\" srcset=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2019\/06\/Fig2_JEI190015.png 1850w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2019\/06\/Fig2_JEI190015-300x176.png 300w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2019\/06\/Fig2_JEI190015-768x452.png 768w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2019\/06\/Fig2_JEI190015-1024x602.png 1024w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2019\/06\/Fig2_JEI190015-624x367.png 624w\" sizes=\"auto, (max-width: 236px) 100vw, 236px\" \/><\/p>\n<p><strong><span style=\"font-size: 12pt;\"><br \/>\nLocal-Linear-Fitting-Based Matting for Joint Hole Filling and Depth Upsampling of RGB-D Images <\/span><span id=\"sample-permalink\" style=\"font-size: 10pt;\">[<a href=\"https:\/\/labsites.rochester.edu\/gsharma\/research\/computer-vision\/joint-hole-filling-and-depth-upsampling-for-rgb-d-images\/\" target=\"wp-preview-952\" rel=\"noopener noreferrer\">project page<\/a>]<br \/>\n<\/span><\/strong>We propose an approach for jointly filling holes and upsampling depth information for RGB-D images, where RGB color information is available at all pixel locations whereas depth information is only available at lower resolution and entirely missing in small regions referred to as \u201choles\u201d.<\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignleft wp-image-3732\" style=\"margin-top: 0.857143rem; margin-right: 1.71429rem; margin-bottom: 0.857143rem;\" src=\"https:\/\/www.hajim.rochester.edu\/ece\/lding6\/wp-content\/uploads\/2017\/03\/physical_model-300x173.jpg\" alt=\"\" width=\"235\" height=\"136\" \/><\/p>\n<p><span style=\"font-family: arial,helvetica,sans-serif;\"><strong><span style=\"font-size: 14pt;\"><br \/>\nHazerd<span style=\"font-size: 12pt;\">: An Outdoor Scene Dataset and Benchmark for Single Image Dehazing<\/span> <span style=\"font-size: 10pt;\"><span id=\"sample-permalink\">[<a href=\"https:\/\/labsites.rochester.edu\/gsharma\/research\/hazerd\" target=\"wp-preview-952\" rel=\"noopener noreferrer\">project page<\/a>]<br \/>\n<\/span><span id=\"edit-slug-buttons\"><\/span><\/span><\/span><\/strong><\/span>We provide a new dataset, HazeRD (Haze Realistic Dataset) for benchmarking dehazing algorithms under realistic haze conditions. HazeRD contains ten real outdoor scenes, for each of which five different weather conditions are simulated. All images are of high resolution, typically six to eight megapixels.<\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\" wp-image-3142 alignleft\" src=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/05\/multipitch-300x188.png\" alt=\"\" width=\"233\" height=\"146\" srcset=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/05\/multipitch-300x188.png 300w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/05\/multipitch.png 459w\" sizes=\"auto, (max-width: 233px) 100vw, 233px\" \/><\/p>\n<p><span style=\"font-size: 14pt;\"><strong><span style=\"font-size: 12pt;\">Visually Informed Multi-Pitch Analysis of String Ensembles<\/span> <span style=\"font-size: 10pt;\"><a href=\"http:\/\/www.ece.rochester.edu\/~gsharma\/papers\/DineshEnhMultiPitchEstimWVidICASSP2017.pdf\">[paper]<\/a><br \/>\n<\/span><\/strong><\/span>Multi-pitch analysis of polyphonic music requires estimating concurrent pitches (estimation) and organizing them into temporal streams according to their sound sources (streaming). This is challenging for approaches based on audio alone due to the polyphonic nature of the audio signals. Video of the performance, when available, can be useful to alleviate some of the difficulties.<\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\" wp-image-3132 alignleft\" src=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/05\/association-300x181.png\" alt=\"\" width=\"236\" height=\"142\" srcset=\"https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/05\/association-300x181.png 300w, https:\/\/labsites.rochester.edu\/gsharma\/wp-content\/uploads\/2017\/05\/association.png 465w\" sizes=\"auto, (max-width: 236px) 100vw, 236px\" \/><span style=\"font-size: 14pt;\"><strong><span style=\"font-size: 12pt;\">See and Listen: Score-Informed Association of Sound Tracks to Players in Chamber Music Performance Videos<\/span> <span style=\"font-size: 10pt;\"><a href=\"http:\/\/www.ece.rochester.edu\/~gsharma\/papers\/LiAudioVideoAssocICASSP2017.pdf\">[paper]<\/a><br \/>\n<\/span><\/strong><\/span>Both audio and visual aspects of a musical performance, especially their association, are important for expressing players\u2019 ideas and for engaging the audience. We present a framework for combining audio and video analyses of multi-instrument chamber music performances to associate players in the video to the individual separated instrument sources from the audio, in a score-informed fashion.<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-size: 14pt;\"><strong>Selected Publications<\/strong><\/span><\/p>\n<div class=\"teachpress_pub_list\"><form name=\"tppublistform\" method=\"get\"><a name=\"tppubs\" id=\"tppubs\"><\/a><\/form><div class=\"teachpress_message_error\"><p>Sorry, no publications matched your criteria.<\/p><\/div><\/div>\n","protected":false},"excerpt":{"rendered":"<p>Deep Retinal Vessel Segmentation For Ultra-Widefield Fundus Photography\u00a0[project page] We propose an annotation-efficient method for vessel segmentation in ultra-widefield (UWF) fundus photography (FP) that does not require de novo labeled ground truth. Our method utilizes concurrently captured UWF fluorescein angiography (FA) images and iterates between a multi-modal registration and a weakly-supervised learning step. We construct [&hellip;]<\/p>\n","protected":false},"author":32,"featured_media":0,"parent":1592,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"page-templates\/full-width.php","meta":{"jetpack_post_was_ever_published":false,"footnotes":""},"class_list":["post-882","page","type-page","status-publish","hentry"],"jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/Paivks-ee","_links":{"self":[{"href":"https:\/\/labsites.rochester.edu\/gsharma\/wp-json\/wp\/v2\/pages\/882","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/labsites.rochester.edu\/gsharma\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/labsites.rochester.edu\/gsharma\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/labsites.rochester.edu\/gsharma\/wp-json\/wp\/v2\/users\/32"}],"replies":[{"embeddable":true,"href":"https:\/\/labsites.rochester.edu\/gsharma\/wp-json\/wp\/v2\/comments?post=882"}],"version-history":[{"count":31,"href":"https:\/\/labsites.rochester.edu\/gsharma\/wp-json\/wp\/v2\/pages\/882\/revisions"}],"predecessor-version":[{"id":8682,"href":"https:\/\/labsites.rochester.edu\/gsharma\/wp-json\/wp\/v2\/pages\/882\/revisions\/8682"}],"up":[{"embeddable":true,"href":"https:\/\/labsites.rochester.edu\/gsharma\/wp-json\/wp\/v2\/pages\/1592"}],"wp:attachment":[{"href":"https:\/\/labsites.rochester.edu\/gsharma\/wp-json\/wp\/v2\/media?parent=882"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}