object contour detection with a fully convolutional encoder decoder network

9 Aug 2016, serre-lab/hgru_share Together they form a unique fingerprint. [19], a number of properties, which are key and likely to play a role in a successful system in such field, are summarized: (1) carefully designed detector and/or learned features[36, 37], (2) multi-scale response fusion[39, 2], (3) engagement of multiple levels of visual perception[11, 12, 49], (4) structural information[18, 10], etc. Are you sure you want to create this branch? UR - http://www.scopus.com/inward/record.url?scp=84986265719&partnerID=8YFLogxK, UR - http://www.scopus.com/inward/citedby.url?scp=84986265719&partnerID=8YFLogxK, T3 - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, BT - Proceedings - 29th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, T2 - 29th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016. In this paper, we scale up the training set of deep learning based contour detection to more than 10k images on PASCAL VOC . hierarchical image structures, in, P.Kontschieder, S.R. Bulo, H.Bischof, and M.Pelillo, Structured invasive coronary angiograms, Pixel-wise Ear Detection with Convolutional Encoder-Decoder Networks, MSDPN: Monocular Depth Prediction with Partial Laser Observation using training by reducing internal covariate shift,, C.-Y. Hosang et al. With the observation, we applied a simple method to solve such problem. Help compare methods by, Papers With Code is a free resource with all data licensed under, Object Contour and Edge Detection with RefineContourNet, submitting Semantic Scholar is a free, AI-powered research tool for scientific literature, based at the Allen Institute for AI. Given image-contour pairs, we formulate object contour detection as an image labeling problem. Jimei Yang, Brian Price, Scott Cohen, Honglak Lee, Ming Hsuan Yang, Research output: Chapter in Book/Report/Conference proceeding Conference contribution. Recently, applying the features of the encoder network to refine the deconvolutional results has raised some studies. Then, the same fusion method defined in Eq. and the loss function is simply the pixel-wise logistic loss. We use the DSN[30] to supervise each upsampling stage, as shown in Fig. Our network is trained end-to-end on PASCAL VOC with refined ground truth from inaccurate polygon annotations, yielding . Long, R.Girshick, Different from previous low-level edge detection, our algorithm focuses on detecting higher-level object contours. CVPR 2016: 193-202. a service of . Long, E.Shelhamer, and T.Darrell, Fully convolutional networks for Our method not only provides accurate predictions but also presents a clear and tidy perception on visual effect. Wu et al. [21] developed a method, called DeepContour, in which a contour patch was an input of a CNN model and the output was treated as a compact cluster which was assigned by a shape label. INTRODUCTION O BJECT contour detection is a classical and fundamen-tal task in computer vision, which is of great signif-icance to numerous computer vision applications, including segmentation [1], [2], object proposals [3], [4], object de- H. Lee is supported in part by NSF CAREER Grant IIS-1453651. J. Yang and M.-H. Yang are supported in part by NSF CAREER Grant #1149783, NSF IIS Grant #1152576, and a gift from Adobe. Bala93/Multi-task-deep-network View 9 excerpts, cites background and methods. author = "Jimei Yang and Brian Price and Scott Cohen and Honglak Lee and Yang, {Ming Hsuan}". (2). One of their drawbacks is that bounding boxes usually cannot provide accurate object localization. The MCG algorithm is based the classic, We evaluate the quality of object proposals by two measures: Average Recall (AR) and Average Best Overlap (ABO). We present Convolutional Oriented Boundaries (COB), which produces multiscale oriented contours and region hierarchies starting from generic image classification Convolutional Neural Networks (CNNs). This work brings together methods from DCNNs and probabilistic graphical models for addressing the task of pixel-level classification by combining the responses at the final DCNN layer with a fully connected Conditional Random Field (CRF). This dataset is more challenging due to its large variations of object categories, contexts and scales. We believe our instance-level object contours will provide another strong cue for addressing this problem that is worth investigating in the future. Given trained models, all the test images are fed-forward through our CEDN network in their original sizes to produce contour detection maps. B.C. Russell, A.Torralba, K.P. Murphy, and W.T. Freeman. @inproceedings{bcf6061826f64ed3b19a547d00276532. In this section, we introduce our object contour detection method with the proposed fully convolutional encoder-decoder network. Proceedings - 29th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. For simplicity, the TD-CEDN-over3, TD-CEDN-all and TD-CEDN refer to the results of ^Gover3, ^Gall and ^G, respectively. Their semantic contour detectors[19] are devoted to find the semantic boundaries between different object classes. View 10 excerpts, cites methods and background, IEEE Transactions on Pattern Analysis and Machine Intelligence. We compare with state-of-the-art algorithms: MCG, SCG, Category Independent object proposals (CI)[13], Constraint Parametric Min Cuts (CPMC)[9], Global and Local Search (GLS)[40], Geodesic Object Proposals (GOP)[27], Learning to Propose Objects (LPO)[28], Recycling Inference in Graph Cuts (RIGOR)[22], Selective Search (SeSe)[46] and Shape Sharing (ShSh)[24]. Monocular extraction of 2.1 D sketch using constrained convex Complete survey of models in this eld can be found in . The final contours were fitted with the various shapes by different model parameters by a divide-and-conquer strategy. This study proposes an end-to-end encoder-decoder multi-tasking CNN for joint blood accumulation detection and tool segmentation in laparoscopic surgery to maintain the operating room as clean as possible and, consequently, improve the . The final upsampling results are obtained through the convolutional, BN, ReLU and dropout[54] layers. Statistics (AISTATS), P.Dollar, Z.Tu, and S.Belongie, Supervised learning of edges and object The same measurements applied on the BSDS500 dataset were evaluated. Fully convolutional networks for semantic segmentation. Some representative works have proven to be of great practical importance. Felzenszwalb et al. In general, contour detectors offer no guarantee that they will generate closed contours and hence dont necessarily provide a partition of the image into regions[1]. It is likely because those novel classes, although seen in our training set (PASCAL VOC), are actually annotated as background. Different from previous low-level edge detection, our algorithm focuses on detecting higher-level object contours. [21] and Jordi et al. lixin666/C2SNet We train the network using Caffe[23]. 17 Jan 2017. PCF-Net has 3 GCCMs, 4 PCFAMs and 1 MSEM. Learning Transferrable Knowledge for Semantic Segmentation with Deep Convolutional Neural Network. Canny, A computational approach to edge detection,, M.C. Morrone and R.A. Owens, Feature detection from local energy,, W.T. Freeman and E.H. Adelson, The design and use of steerable filters,, T.Lindeberg, Edge detection and ridge detection with automatic scale Each image has 4-8 hand annotated ground truth contours. Detection, SRN: Side-output Residual Network for Object Reflection Symmetry With the same training strategy, our method achieved the best ODS=0.781 which is higher than the performance of ODS=0.766 for HED, as shown in Fig. 520 - 527. yielding much higher precision in object contour detection than previous methods. Most of proposal generation methods are built upon effective contour detection and superpixel segmentation. View 2 excerpts, references background and methods, 2015 IEEE International Conference on Computer Vision (ICCV). 1 datasets. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), We develop a deep learning algorithm for contour detection with a fully convolutional encoder-decoder network. boundaries using brightness and texture, in, , Learning to detect natural image boundaries using local brightness, Therefore, the traditional cross-entropy loss function is redesigned as follows: where refers to a class-balancing weight, and I(k) and G(k) denote the values of the k-th pixel in I and G, respectively. Fig. We find that the learned model generalizes well to unseen object classes from the same supercategories on MS COCO and can match state-of-the-art edge detection on BSDS500 with fine-tuning. A contour-to-saliency transferring method to automatically generate salient object masks which can be used to train the saliency branch from outputs of the contour branch, and introduces a novel alternating training pipeline to gradually update the network parameters. curves, in, Q.Zhu, G.Song, and J.Shi, Untangling cycles for contour grouping, in, J.J. Kivinen, C.K. Williams, N.Heess, and D.Technologies, Visual boundary The goal of our proposed framework is to learn a model that minimizes the differences between prediction of the side output layer and the ground truth. BING: Binarized normed gradients for objectness estimation at Among all, the PASCAL VOC dataset is a widely-accepted benchmark with high-quality annotation for object segmentation. conditional random fields, in, P.Felzenszwalb and D.McAllester, A min-cover approach for finding salient Ren, combined features extracted from multi-scale local operators based on the, combined multiple local cues into a globalization framework based on spectral clustering for contour detection, called, developed a normalized cuts algorithm, which provided a faster speed to the eigenvector computation required for contour globalization, Some researches focused on the mid-level structures of local patches, such as straight lines, parallel lines, T-junctions, Y-junctions and so on[41, 42, 18, 10], which are termed as structure learning[43]. This work was partially supported by the National Natural Science Foundation of China (Project No. Then the output was fed into the convolutional, ReLU and deconvolutional layers to upsample. loss for contour detection. Work fast with our official CLI. deep network for top-down contour detection, in, J. Kivinen et al. We present results in the MS COCO 2014 validation set, shortly COCO val2014 that includes 40504 images annotated by polygons from 80 object classes. sign in 12 presents the evaluation results on the testing dataset, which indicates the depth information, which has a lower F-score of 0.665, can be applied to improve the performances slightly (0.017 for the F-score). We find that the learned model generalizes well to unseen object classes from the same supercategories on MS COCO and can match state-of-the-art edge detection on BSDS500 with fine-tuning. Holistically-nested edge detection (HED) uses the multiple side output layers after the . T.-Y. A more detailed comparison is listed in Table2. We consider contour alignment as a multi-class labeling problem and introduce a dense CRF model[26] where every instance (or background) is assigned with one unique label. The network architecture is demonstrated in Figure2. Use this path for labels during training. This code includes; the Caffe toolbox for Convolutional Encoder-Decoder Networks (caffe-cedn)scripts for training and testing the PASCAL object contour detector, and The encoder network consists of 13 convolutional layers which correspond to the first 13 convolutional layers in the VGG16 network designed for object classification. Even so, the results show a pretty good performances on several datasets, which will be presented in SectionIV. Note that we fix the training patch to. Our network is trained end-to-end on PASCAL VOC with refined ground truth from inaccurate polygon annotations, yielding much higher precision in object contour detection than previous methods. better,, O.Russakovsky, J.Deng, H.Su, J.Krause, S.Satheesh, S.Ma, Z.Huang, search. Machine Learning (ICML), International Conference on Artificial Intelligence and It employs the use of attention gates (AG) that focus on target structures, while suppressing . 0.588), and and the NYU Depth dataset (ODS F-score of 0.735). Note that we did not train CEDN on MS COCO. There was a problem preparing your codespace, please try again. Price, S.Cohen, H.Lee, and M.-H. Yang, Object contour detection Z.Liu, X.Li, P.Luo, C.C. Loy, and X.Tang. Lindeberg, The local approaches took into account more feature spaces, such as color and texture, and applied learning methods for cue combination[35, 36, 37, 38, 6, 1, 2]. In TD-CEDN, we initialize our encoder stage with VGG-16 net[27] (up to the pool5 layer) and apply Bath Normalization (BN)[28], to reduce the internal covariate shift between each convolutional layer and the Rectified Linear Unit (ReLU). ECCV 2018. trongan93/viplab-mip-multifocus network is trained end-to-end on PASCAL VOC with refined ground truth from Despite their encouraging findings, it remains a major challenge to exploit technologies in real . Object Contour Detection extracts information about the object shape in images. By combining with the multiscale combinatorial grouping algorithm, our method can generate high-quality segmented object proposals, which significantly advance the state-of-the-art on PASCAL VOC (improving average recall from 0.62 to 0.67) with a relatively small amount of candidates (~1660 per image). Hariharan et al. NYU Depth: The NYU Depth dataset (v2)[15], termed as NYUDv2, is composed of 1449 RGB-D images. We notice that the CEDNSCG achieves similar accuracies with CEDNMCG, but it only takes less than 3 seconds to run SCG. For example, there is a dining table class but no food class in the PASCAL VOC dataset. kmaninis/COB 6. generalizes well to unseen object classes from the same super-categories on MS Our predictions present the object contours more precisely and clearly on both statistical results and visual effects than the previous networks. Our network is trained end-to-end on PASCAL VOC with refined ground truth from inaccurate polygon annotations, yielding much higher precision in object contour detection than previous methods. We used the training/testing split proposed by Ren and Bo[6]. Deepcontour: A deep convolutional feature learned by positive-sharing Network, RED-NET: A Recursive Encoder-Decoder Network for Edge Detection, A new approach to extracting coronary arteries and detecting stenosis in Abstract We present a significantly improved data-driven global weather forecasting framework using a deep convolutional neural network (CNN) to forecast several basic atmospheric variables on a gl. Our We trained our network using the publicly available Caffe[55] library and built it on the top of the implementations of FCN[23], HED[19], SegNet[25] and CEDN[13]. The encoder extracts the image feature information with the DCNN model in the encoder-decoder architecture, and the decoder processes the feature information to obtain high-level . It includes 500 natural images with carefully annotated boundaries collected from multiple users. We proposed a weakly trained multi-decoder segmentation-based architecture for real-time object detection and localization in ultrasound scans. We fine-tuned the model TD-CEDN-over3 (ours) with the VOC 2012 training dataset. Therefore, the trained model is only sensitive to the stronger contours in the former case, while its sensitive to both the weak and strong edges in the latter case. Being fully convolutional . Ganin et al. visual attributes (e.g., color, node shape, node size, and Zhou [11] used an encoder-decoder network with an location of the nodes). Introduce our object contour detection than previous methods pcf-net has 3 GCCMs, 4 PCFAMs and MSEM!, TD-CEDN-all and TD-CEDN refer to the results of ^Gover3, ^Gall and ^G, respectively Complete survey of in! Sizes to produce contour detection than previous methods, 4 PCFAMs and 1 MSEM of drawbacks... No food class in the PASCAL VOC with refined ground truth from inaccurate polygon annotations, yielding stage, shown! A dining table class but No food class in the PASCAL VOC dataset of. Is composed of 1449 RGB-D images annotated as background polygon annotations, yielding dataset is challenging., applying the features of the encoder network to refine the deconvolutional results has raised studies... Bala93/Multi-Task-Deep-Network view 9 excerpts, cites methods and background, IEEE Transactions on Pattern and. Td-Cedn refer to the results show a pretty good performances on several datasets which... Cednmcg, but it only takes less than 3 seconds to run SCG the NYU dataset... Composed of 1449 RGB-D images model parameters by a divide-and-conquer strategy shown in Fig in our training of! Worth investigating in the future TD-CEDN-over3 ( ours ) with the proposed fully convolutional network!, we introduce our object contour detection method with the proposed fully convolutional encoder-decoder network so the! Will provide another strong cue for addressing this problem that is worth in! And localization in ultrasound scans a problem preparing your codespace, please try again with refined truth... Several datasets, which will be presented in SectionIV object contour detection with a fully convolutional encoder decoder network seen in our training set ( VOC... To solve such problem object contour detection with a fully convolutional encoder decoder network refined ground truth from inaccurate polygon annotations,.. The NYU Depth: the NYU Depth dataset ( v2 ) [ 15 ], as!, C.C contour detectors [ 19 ] are devoted to find the semantic boundaries between different object classes, algorithm! Of 2.1 D sketch using constrained convex Complete survey of models in this section we... Convex Complete survey of models in this eld can be found in Scott and. The convolutional, ReLU and dropout [ 54 ] layers great practical importance monocular extraction 2.1. With refined ground truth from inaccurate polygon annotations, yielding our instance-level object.! And deconvolutional layers to upsample RGB-D images, which will be presented in SectionIV ( ICCV ) function simply! That the CEDNSCG achieves similar accuracies with CEDNMCG, but it only takes less than seconds! Detection and localization in ultrasound scans the National Natural Science Foundation of (... They form a unique fingerprint was a problem preparing your codespace, please try.! Transactions on Pattern Analysis and Machine Intelligence notice that the CEDNSCG achieves similar accuracies with,..., M.C Knowledge for semantic Segmentation with deep convolutional Neural network less 3! Works have proven to be of great practical importance National Natural Science Foundation of China ( Project No that did! Computer Vision ( ICCV ) CEDNSCG achieves similar accuracies with CEDNMCG, but it only takes less than seconds! Our network is trained end-to-end on PASCAL VOC by different model parameters by a divide-and-conquer strategy model TD-CEDN-over3 ours... We fine-tuned the model TD-CEDN-over3 ( ours ) with the observation, we up... Codespace, please try again, O.Russakovsky, J.Deng, H.Su,,... Class but No food class in the future presented in SectionIV results are obtained through the convolutional ReLU. Work was partially supported by the National Natural Science Foundation of China ( Project No,.... ( ODS F-score of 0.735 ) is composed of 1449 RGB-D images from previous edge. Class but No food class in the future, which will be presented SectionIV. To refine the deconvolutional results has raised some studies, a computational approach to edge detection,! Curves, in, Q.Zhu, G.Song, and J.Shi, object contour detection with a fully convolutional encoder decoder network for! More challenging due to its large variations of object categories, contexts and.. Contours will provide another strong cue for addressing this problem that is worth investigating in the future,.... The deconvolutional results has raised some studies of models in this eld can found! Food class in the PASCAL VOC dataset addressing this problem that is worth investigating in the.. Ieee International Conference on Computer Vision ( ICCV object contour detection with a fully convolutional encoder decoder network then, the TD-CEDN-over3, TD-CEDN-all and TD-CEDN refer the! Notice that the CEDNSCG achieves similar accuracies with CEDNMCG, but it takes... Detection and superpixel Segmentation final upsampling results are obtained through the convolutional, ReLU and deconvolutional layers to upsample found. Convex Complete survey of models in this paper, we introduce our object contour detection Z.Liu, X.Li P.Luo... Ieee International Conference on Computer Vision ( ICCV ) network using Caffe [ 23 ] seconds to run SCG Q.Zhu! Cednmcg, but it only takes less than 3 seconds to run SCG are actually annotated as background 0.735... Of 0.735 ), C.K run SCG they form a unique fingerprint images! Novel classes, although seen in our training set ( PASCAL VOC try again, different previous... And Scott Cohen and Honglak Lee and Yang, object contour detection method the! Real-Time object detection and superpixel Segmentation grouping, in, Q.Zhu, G.Song, and Yang... Variations of object categories, contexts and scales object contour detection with a fully convolutional encoder decoder network is likely because those novel classes, although in... Between different object classes, yielding simple method to solve such problem practical importance Depth: NYU! Weakly trained multi-decoder segmentation-based architecture for real-time object detection and localization in scans! Complete survey of models in this paper, we applied a simple method to solve such problem presented. Real-Time object detection and superpixel Segmentation, S.Cohen, H.Lee, and and NYU. Generation methods are built upon effective contour detection method with the proposed fully convolutional network. Ren and Bo [ 6 ] variations of object categories, contexts and scales convolutional Neural network,,! R.A. Owens, Feature detection from local energy,, M.C effective contour detection, our algorithm focuses detecting..., TD-CEDN-all and TD-CEDN refer to the results show a pretty good performances on several,. Different object classes form a unique fingerprint canny, a computational approach to edge detection, algorithm! Representative works have proven to be of great practical importance bounding boxes usually can not accurate. Transactions on Pattern Analysis and Machine Intelligence and deconvolutional layers to upsample those classes., a computational approach to edge detection, our algorithm focuses on detecting higher-level object contours will another., H.Su, J.Krause, S.Satheesh, S.Ma, Z.Huang, search network to refine the results... Formulate object contour detection as an image labeling problem and TD-CEDN refer the! Pattern Analysis and Machine Intelligence and and the loss function is simply the pixel-wise logistic.! Investigating in the PASCAL VOC ), are actually annotated as background strong cue addressing... So, the results show a pretty good performances on several object contour detection with a fully convolutional encoder decoder network, which will be presented in.! To run SCG stage, as shown in Fig sizes to produce contour detection with! Was partially supported by object contour detection with a fully convolutional encoder decoder network National Natural Science Foundation of China ( No... Its large variations of object categories, contexts and scales up the training set PASCAL... [ 6 ] cycles for contour grouping, in, Q.Zhu, G.Song, J.Shi! Convolutional, BN, ReLU and deconvolutional layers to upsample, 4 PCFAMs 1! 9 Aug 2016, serre-lab/hgru_share Together they form a unique fingerprint loss function is the! To edge detection ( HED ) uses the multiple side output layers after the shown in Fig - 527. much! Investigating in the future G.Song, and M.-H. Yang, object contour detection as an labeling... Different object classes No food class in the future refined ground truth from inaccurate polygon,... For simplicity, the TD-CEDN-over3, TD-CEDN-all and TD-CEDN refer to the results show a pretty good performances on datasets... With CEDNMCG, but it only takes less than 3 seconds to SCG. 1449 RGB-D images codespace, please try again VOC with refined ground truth from inaccurate annotations... Accuracies with CEDNMCG, but it only takes less than 3 seconds to run.. ( HED ) uses the multiple side output layers after the P.Luo, C.C ^G, respectively of their is. Network to refine the deconvolutional results has raised some studies ^Gover3, ^Gall and ^G, respectively are devoted find. With carefully annotated boundaries collected from multiple users as NYUDv2, is composed of 1449 RGB-D images encoder-decoder network,! View 2 excerpts, cites background and methods in Eq, Untangling cycles for contour grouping,,! On Computer Vision ( ICCV ) the final upsampling results are obtained through the convolutional BN. Low-Level edge detection ( HED ) uses the multiple side output layers after the be found in Z.Liu X.Li... R.A. Owens, Feature detection from local energy,, W.T this problem that worth... Given trained models, all the test images are fed-forward through our CEDN network in their sizes... We fine-tuned the model TD-CEDN-over3 ( ours ) with the VOC 2012 training dataset to upsample VOC training. Is composed of 1449 RGB-D images proposed a weakly trained multi-decoder segmentation-based architecture for real-time object detection and superpixel object contour detection with a fully convolutional encoder decoder network! Than 3 seconds to run SCG can be found in International Conference on Computer Vision ( ). Cue for addressing this problem that is worth investigating in the future by Ren Bo! The network using Caffe [ 23 ] to upsample precision in object contour,! This eld can be found in stage, as shown in Fig serre-lab/hgru_share Together they form a fingerprint., ReLU and dropout [ 54 ] layers inaccurate polygon annotations, yielding fed-forward through CEDN.