Mimicking Very Efficient Network for Object Detection

Total Page:16

File Type:pdf, Size:1020Kb

Mimicking Very Efficient Network for Object Detection Mimicking Very Efficient Network for Object Detection Quanquan Li1, Shengying Jin2, Junjie Yan1 1SenseTime 2Beihang University [email protected], [email protected], [email protected] Abstract Method MR−2 Parameters test time (ms) Inception R-FCN 7.15 2.5M 53.5 Current CNN based object detectors need initialization 1/2-Inception 7.31 625K 22.8 from pre-trained ImageNet classification models, which are Mimic R-FCN usually time-consuming. In this paper, we present a fully 1/2-Inception finetuned 8.88 625K 22.8 convolutional feature mimic framework to train very effi- from ImageNet cient CNN based detectors, which do not need ImageNet pre-training and achieve competitive performance as the Table 1: The parameters and test time of large and small mod- els. Tested on TITANX with 1000×1500 input. The 1/2-Inception large and slow models. We add supervision from high-level model trained by mimicking outperforms that fine-tuned from Im- features of the large networks in training to help the small ageNet pre-trained model. Moreover, it obtains similar perfor- network better learn object representation. More specifi- mance as the large Inception model with only 1/4 parameters and cally, we conduct a mimic method for the features sampled achieves a 2.5× speed-up. from the entire feature map and use a transform layer to map features from the small network onto the same dimen- sion of the large network. In training the small network, we AlexNet gets 56.8% AP. Due to this phenomenon, nearly optimize the similarity between features sampled from the all the modern detection methods can only train networks same region on the feature maps of both networks. Exten- which have been trained on ImageNet before and cannot sive experiments are conducted on pedestrian and common train a network from scratch to achieve comparable results. object detection tasks using VGG, Inception and ResNet. The result is that we can only use networks designed for On both Caltech and Pascal VOC, we show that the modi- classification task such as AlexNet [23], ZFNet [35], VG- × fied 2.5 accelerated Inception network achieves competi- GNet [30] and ResNet [17], which are not necessarily op- tive performance as the full Inception Network. Our faster timal for detection. Due to the limitation, if we want to × model runs at 80 FPS for a 1000 1500 large input with sweep different network configurations and find a more ef- only a minor degradation of performance on Caltech. ficient network, we will need to pre-train these models on ImageNet classification task and then fine-tune them on de- tection task. The process is very expensive considering 1. Introduction that training a ImageNet classification model needs several Object detection is a fundamental problem in image un- weeks even on multiple GPUs. Moreover, in experiments derstanding. It aims to determine where in the image the ob- we find that smaller networks always perform poor on Im- jects are and which category each object belongs to. Many ageNet classification so that fine-tuning them on detection popular deep convolutional neural network based object de- also leads to poor detection performance. tection methods have been proposed, such as Faster R-CNN In this paper, we want to train more efficient detection [28], R-FCN [6] and SSD [25]. Compared with traditional networks without ImageNet pre-training. More importantly, methods such as DPM [12], these CNN based frameworks we still need to achieve competitive performance as the achieve good performance on challenging dataset. large ImageNet pre-trained models. The basic idea is that if Since the pioneering work R-CNN [14], CNN based we already have a network that achieves satisfying perfor- object detectors need a pre-trained ImageNet classifica- mance for detection, the network can be used to supervise tion model for initialization to get the desired performance. other network training for detection. The question then be- According to the experiments in [22], the Fast R-CNN comes how to use a detection network to supervise a more [13] with AlexNet trained from scratch gets the 40.4% efficient network and keeps its accuracy for detection. AP on Pascal VOC 2007, while with ImageNet pre-trained Similar ideas have been used in standard classification 6356 task, such as [18, 2]. However, we find that they do not performance on Caltech detection tasks. work well for this more complex detection task. The main problems are how and where to add the supervision from 2. Related Work detection ground-truth and the one from a different network. The related work includes recent CNN based object de- Our solution for mimicking in object detection comes tections, network mimicking and network training, as well from observation of modern CNN based detectors, includ- as network acceleration. ing Faster R-CNN [28], R-FCN [6], SSD [25] and YOLO A seminal CNN based object detection method is R- [27]. They all calculate a feature map and then use different CNN [14], which uses the fine-tuned CNN to extract fea- methods to decode detection results from the feature map. tures from object proposals and SVM to classify them. The In this way, detector can actually be divided into the jointly spatial pyramid pooling [16] and Fast R-CNN [13] extract trained feature extractor and the feature decoder. The dif- features on top of a shared feature map to speed up the ferences between the large network and the more efficient R-CNN. Faster R-CNN [28] further improves by predict- network lie in the feature extractor. To this end, our phi- ing region proposals and classifying proposals in the shared losophy is that the mimicking supervision should be added feature map. A very recent work R-FCN [6] proposes in the feature map generated by the feature extractor; the the position-sensitive score map to share more computa- ground-truth supervision should be added on the final fea- tion. The R-CNN series takes object detection as a two-shot ture decoder. For the mimicking supervision, we define a problem, including region proposal generation and region transformation on top of the feature map generated by the classification. Recently, one-shot methods have been pro- small network to a new feature. We want to minimize the posed, such as YOLO and SSD. All these methods need to Euclidean distance between this new feature and the feature calculate the feature map which takes most of the computa- generated by the large network. tion. The mimicking technique we proposed is validated on For the ground-truth supervision, it is the same as the Faster R-CNN and R-FCN, but it can be naturally extended origin detector, such as the joint classification and localiza- to SSD, YOLO and other CNN feature map based methods. tion loss in Fast R-CNN. In training, we first extract the Network mimicking or distilling are recently introduced feature map of each training image generated by the large model acceleration and compression approaches by [18, 2] network, and then use the feature maps and detection anno- aiming to train a more compact model that can learn tations to jointly train the detector with the small network from the output of a large model. [29] further improves initialized from scratch. One problem is that the feature this method by way of the implementation of deeper stu- map is of very high dimension, and we find that directly dent models and hints from the intermediate representation mimicking the feature map does not converge as expected. learned by the teacher network. However, all these mimick- Since the feature extractor is region or proposal based, we ing works, to our best knowledge, have only been validated sample features from regions to optimize, which leads to on easy classification tasks [18, 2, 29, 33]. In this paper, satisfying results. we show how to extend the mimicking techniques for more The feature map mimicking technique proposed in the challenging object detection tasks, and how to use it to train paper can naturally be extended. The first extension is that more efficient object detector. we can mimic across scales. In CNN based detection, we Some works have been proposed to give better initial- only need 1/4 computation if we can reduce the width and ization or replace the ImageNet pre-train. [22] sets the height of the input image by half. However, it usually leads weights of a network such that all units in the network train to significantly performance drop. We show that we can de- at roughly the same rate to avoid vanishing or exploding fine a simple transformation to up-sample the feature map gradients. [1] and [8] learn an unsupervised representation to a large scale and then mimic the transformed feature from videos, and [8] uses spatial context as the source to map. Another extension is that we can extend the mimick- provide supervision signal for training. These methods per- ing technique to a two-stage procedure that further improves form much better than being trained randomly from scratch, the performance. but they still have a large performance gap between the pre- We conduct experiments on Caltech pedestrian detec- trained method from ImageNet. The recent work [19] ana- tion and Pascal VOC object detection using R-FCN and lyzes the ImageNet features in detail. Faster R-CNN. On both Caltech and Pascal VOC, we Our work is also related to works of network accelera- show that the mimicked models demonstrate superior per- tion. [7, 20, 24] accelerate single layer of CNN through formance than the models fine-tuned from ImageNet pre- linear decomposition , while [38] considers the nonlinear trained model.
Recommended publications
  • Detection-Based Object Tracking Applied to Remote Ship Inspection
    sensors Article Detection-Based Object Tracking Applied to Remote Ship Inspection Jing Xie 1,* , Erik Stensrud 1 and Torbjørn Skramstad 2 1 Group Technology and Research, DNV GL, Veritasveien 1, 1363 Høvik, Norway; [email protected] 2 Department of Computer Science, Norwegian University of Science and Technology, NO-7491 Trondheim, Norway; [email protected] * Correspondence: [email protected] Abstract: We propose a detection-based tracking system for automatically processing maritime ship inspection videos and predicting suspicious areas where cracks may exist. This system consists of two stages. Stage one uses a state-of-the-art object detection model, i.e., RetinaNet, which is customized with certain modifications and the optimal anchor setting for detecting cracks in the ship inspection images/videos. Stage two is an enhanced tracking system including two key components. The first component is a state-of-the-art tracker, namely, Channel and Spatial Reliability Tracker (CSRT), with improvements to handle model drift in a simple manner. The second component is a tailored data association algorithm which creates tracking trajectories for the cracks being tracked. This algorithm is based on not only the intersection over union (IoU) of the detections and tracking updates but also their respective areas when associating detections to the existing trackers. Consequently, the tracking results compensate for the detection jitters which could lead to both tracking jitter and creation of redundant trackers. Our study shows that the proposed detection-based tracking system has achieved a reasonable performance on automatically analyzing ship inspection videos. It has proven the feasibility of applying deep neural network based computer vision technologies to automating remote ship inspection.
    [Show full text]
  • Machine Learning for Blob Detection in High-Resolution 3D Microscopy Images
    DEGREE PROJECT IN COMPUTER SCIENCE AND ENGINEERING, SECOND CYCLE, 30 CREDITS STOCKHOLM, SWEDEN 2018 Machine learning for blob detection in high-resolution 3D microscopy images MARTIN TER HAAK KTH ROYAL INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE Machine learning for blob detection in high-resolution 3D microscopy images MARTIN TER HAAK EIT Digital Data Science Date: June 6, 2018 Supervisor: Vladimir Vlassov Examiner: Anne Håkansson Electrical Engineering and Computer Science (EECS) iii Abstract The aim of blob detection is to find regions in a digital image that dif- fer from their surroundings with respect to properties like intensity or shape. Bio-image analysis is a common application where blobs can denote regions of interest that have been stained with a fluorescent dye. In image-based in situ sequencing for ribonucleic acid (RNA) for exam- ple, the blobs are local intensity maxima (i.e. bright spots) correspond- ing to the locations of specific RNA nucleobases in cells. Traditional methods of blob detection rely on simple image processing steps that must be guided by the user. The problem is that the user must seek the optimal parameters for each step which are often specific to that image and cannot be generalised to other images. Moreover, some of the existing tools are not suitable for the scale of the microscopy images that are often in very high resolution and 3D. Machine learning (ML) is a collection of techniques that give computers the ability to ”learn” from data. To eliminate the dependence on user parameters, the idea is applying ML to learn the definition of a blob from labelled images.
    [Show full text]
  • Object Detection in 20 Years: a Survey
    1 Object Detection in 20 Years: A Survey Zhengxia Zou, Zhenwei Shi, Member, IEEE, Yuhong Guo, and Jieping Ye, Senior Member, IEEE Abstract—Object detection, as of one the most fundamental and challenging problems in computer vision, has received great attention in recent years. Its development in the past two decades can be regarded as an epitome of computer vision history. If we think of today’s object detection as a technical aesthetics under the power of deep learning, then turning back the clock 20 years we would witness the wisdom of cold weapon era. This paper extensively reviews 400+ papers of object detection in the light of its technical evolution, spanning over a quarter-century’s time (from the 1990s to 2019). A number of topics have been covered in this paper, including the milestone detectors in history, detection datasets, metrics, fundamental building blocks of the detection system, speed up techniques, and the recent state of the art detection methods. This paper also reviews some important detection applications, such as pedestrian detection, face detection, text detection, etc, and makes an in-deep analysis of their challenges as well as technical improvements in recent years. Index Terms—Object detection, Computer vision, Deep learning, Convolutional neural networks, Technical evolution. F 1 INTRODUCTION BJECT detection is an important computer vision task O that deals with detecting instances of visual objects of a certain class (such as humans, animals, or cars) in digital images. The objective of object detection is to develop computational models and techniques that provide one of the most basic pieces of information needed by computer vision applications: What objects are where? As one of the fundamental problems of computer vision, object detection forms the basis of many other computer vision tasks, such as instance segmentation [1–4], image captioning [5–7], object tracking [8], etc.
    [Show full text]
  • Cyclesegnet: Object Co-Segmentation with Cycle Refinement and Region Correspondence
    1 CycleSegNet: Object Co-Segmentation with Cycle Refinement and Region Correspondence Chi Zhang∗, Guankai Li∗, Guosheng Lin, Qingyao Wu, Rui Yao Abstract—Image co-segmentation is an active computer vision task that aims to segment the common objects from a set of images. Recently, researchers design various learning-based algorithms to undertake the co-segmentation task. The main difficulty in this task is how to effectively transfer information between images to make conditional predictions. In this paper, we present CycleSegNet, a novel framework for the co-segmentation task. Our network design has two key components: a region correspondence module which is the basic operation for exchanging information between local image regions, and a cycle refinement module, which utilizes ConvLSTMs to progressively update image representations and exchange information in a cycle and iterative manner. Extensive experiments demonstrate that our proposed method significantly outperforms the state-of-the-art methods on four popular benchmark datasets — PASCAL VOC dataset, MSRC dataset, Internet dataset, and iCoseg dataset, by 2.6%, 7.7%, 2.2%, and 2.9%, respectively. Index Terms—deep learning, co-segmentation, cycle refinement, attention F 1 INTRODUCTION Image co-segmentation is an active computer vision topic with a long research history, which aims to segment the common objects jointly from a set of images. Image co- segmentation algorithms have shown their usages in var- ious computer vision tasks, such as image retrieval [47], 3D reconstruction [37], photo collections [41], image match- ing [6], [58], [59], and video object tracking [32], [33], [41]. Recently, data-driven deep neural networks based meth- ods attract wide interest in the literature.
    [Show full text]
  • Towards Efficient Video Detection Object Super-Resolution with Deep Fusion Network for Public Safety
    Hindawi Security and Communication Networks Volume 2021, Article ID 9999398, 14 pages https://doi.org/10.1155/2021/9999398 Research Article Towards Efficient Video Detection Object Super-Resolution with Deep Fusion Network for Public Safety Sheng Ren , Jianqi Li , Tianyi Tu , Yibo Peng , and Jian Jiang School of Computer and Electrical Engineering, Hunan University of Arts and Science, Changde 415000, China Correspondence should be addressed to Jianqi Li; [email protected] Received 22 March 2021; Revised 14 April 2021; Accepted 14 May 2021; Published 24 May 2021 Academic Editor: David Meg´ıas Copyright © 2021 Sheng Ren et al. *is is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Video surveillance plays an increasingly important role in public security and is a technical foundation for constructing safe and smart cities. *e traditional video surveillance systems can only provide real-time monitoring or manually analyze cases by reviewing the surveillance video. So, it is difficult to use the data sampled from the surveillance video effectively. In this paper, we proposed an efficient video detection object super-resolution with a deep fusion network for public security. Firstly, we designed a super-resolution framework for video detection objects. By fusing object detection algorithms, video keyframe selection al- gorithms, and super-resolution reconstruction algorithms, we proposed a deep learning-based intelligent video detection object super-resolution (SR) method. Secondly, we designed a regression-based object detection algorithm and a key video frame selection algorithm.
    [Show full text]
  • Deepco3: Deep Instance Co-Segmentation by Co-Peak Search and Co-Saliency Detection
    DeepCO3: Deep Instance Co-segmentation by Co-peak Search and Co-saliency Detection Kuang-Jui Hsu1,2 Yen-Yu Lin1 Yung-Yu Chuang1,2 1Academia Sinica, Taiwan 2National Taiwan University, Taiwan Abstract In this paper, we address a new task called instance co- segmentation. Given a set of images jointly covering object instances of a specific category, instance co-segmentation aims to identify all of these instances and segment each of them, i.e. generating one mask for each instance. This task is important since instance-level segmentation is preferable for humans and many vision applications. It is also chal- lenging because no pixel-wise annotated training data are available and the number of instances in each image is un- known. We solve this task by dividing it into two sub-tasks, co-peak search and instance mask segmentation. In the for- mer sub-task, we develop a CNN-based network to detect Figure 1. Two examples of instance co-segmentation on categories bird and sheep, respectively. An instance here refers to an object the co-peaks as well as co-saliency maps for a pair of im- appearing in an image. In each example, the top row gives the ages. A co-peak has two endpoints, one in each image, that input images while the bottom row shows the instances segmented are local maxima in the response maps and similar to each by our method. The instance-specific coloring indicates that our other. Thereby, the two endpoints are potentially covered by method produces a segmentation mask for each instance.
    [Show full text]
  • A Survey of Deep Learning-Based Object Detection
    1 A Survey of Deep Learning-based Object Detection Licheng Jiao, Fellow, IEEE, Fan Zhang, Fang Liu, Senior Member, IEEE, Shuyuan Yang, Senior Member, IEEE, Lingling Li, Member, IEEE, Zhixi Feng, Member, IEEE, and Rong Qu, Senior Member, IEEE Abstract—Object detection is one of the most important and understanding, object detection has been widely used in many challenging branches of computer vision, which has been widely fields of modern life, such as security field, military field, applied in peoples life, such as monitoring security, autonomous transportation field, medical field and life field. Furthermore, driving and so on, with the purpose of locating instances of semantic objects of a certain class. With the rapid development many benchmarks have played an important role in object of deep learning networks for detection tasks, the performance detection field so far, such as Caltech [1], KITTI [2], ImageNet of object detectors has been greatly improved. In order to [3], PASCAL VOC [4], MS COCO [5], and Open Images V5 understand the main development status of object detection [6]. In ECCV VisDrone 2018 contest, organizers have released pipeline, thoroughly and deeply, in this survey, we first analyze a novel drone platform-based dataset [7] which contains a large the methods of existing typical detection models and describe the benchmark datasets. Afterwards and primarily, we provide a amount of images and videos. comprehensive overview of a variety of object detection methods • Two kinds of object detectors in a systematic manner, covering the one-stage and two-stage Pre-existing domain-specific image object detectors usually detectors.
    [Show full text]
  • Yolomask, an Instance Segmentation Algorithm Based on Complementary Fusion Network
    mathematics Article YOLOMask, an Instance Segmentation Algorithm Based on Complementary Fusion Network Jiang Hua 1, Tonglin Hao 2, Liangcai Zeng 1 and Gui Yu 1,3,* 1 Key Laboratory of Metallurgical Equipment and Control Technology, Ministry of Education, Wuhan University of Science and Technology, Wuhan 430081, China; [email protected] (J.H.); [email protected] (L.Z.) 2 School of Automation, Wuhan University of Science and Technology, Wuhan 430081, China; [email protected] 3 School of Mechanical and Electrical Engineering, Huanggang Normal University, Huanggang 438000, China * Correspondence: [email protected] Abstract: Object detection and segmentation can improve the accuracy of image recognition, but traditional methods can only extract the shallow information of the target, so the performance of algorithms is subject to many limitations. With the development of neural network technology, semantic segmentation algorithms based on deep learning can obtain the category information of each pixel. However, the algorithm cannot effectively distinguish each object of the same category, so YOLOMask, an instance segmentation algorithm based on complementary fusion network, is proposed in this paper. Experimental results on public data sets COCO2017 show that the proposed fusion network can accurately obtain the category and location information of each instance and has good real-time performance. Keywords: image segmentation; deep learning; instance segmentation; fusion network Citation: Hua, J.; Hao, T.; Zeng, L.; Yu, G. YOLOMask, an Instance Segmentation Algorithm Based on Complementary Fusion Network. 1. Introduction Mathematics 2021, 9, 1766. https:// Object detection and segmentation based on RGB images are the basis of 6D pose doi.org/10.3390/math9151766 estimation, and it is also the premise of successful robot grasping.
    [Show full text]
  • Deep Object Co-Segmentation Via Spatial-Semantic Network Modulation
    The Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI-20) Deep Object Co-Segmentation via Spatial-Semantic Network Modulation Kaihua Zhang,1 Jin Chen,1 Bo Liu,2∗ Qingshan Liu1 1B-DAT and CICAEET, Nanjing University of Information Science and Technology, Nanjing, China 2JD Finance America Corporation {zhkhua, kfliubo}@gmail.com Abstract Object co-segmentation is to segment the shared objects in (a) multiple relevant images, which has numerous applications in computer vision. This paper presents a spatial and se- mantic modulated deep network framework for object co- segmentation. A backbone network is adopted to extract multi-resolution image features. With the multi-resolution (b) features of the relevant images as input, we design a spa- tial modulator to learn a mask for each image. The spatial modulator captures the correlations of image feature descrip- tors via unsupervised learning. The learned mask can roughly localize the shared foreground object while suppressing the Figure 1: Object co-segmentation examples by our ap- background. For the semantic modulator, we model it as a proach. (a) Horse group; (b) Horse group co-segmentation supervised image classification task. We propose a hierar- maps. chical second-order pooling module to transform the image features for classification use. The outputs of the two mod- ulators manipulate the multi-resolution features by a shift- features cannot well handle the challenging cases in co- and-scale operation so that the features focus on segmenting segmentation such as background clutter and large-scale co-object regions. The proposed model is trained end-to-end appearance variations of the co-objects in images.
    [Show full text]
  • Towards Real-Time Object Detection on Edge with Deep Neural Networks
    TOWARDS REAL-TIME OBJECT DETECTION ON EDGE WITH DEEP NEURAL NETWORKS A Dissertation presented to the Faculty of the Graduate School at the University of Missouri In Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy by ZHI ZHANG Dr. Zhihai He, Dissertation Supervisor December 2018 The undersigned, appointed by the Dean of the Graduate School, have examined the dissertation entitled: TOWARDS REAL-TIME OBJECT DETECTION ON EDGE WITH DEEP NEURAL NETWORKS presented by Zhi Zhang, a candidate for the degree of Doctor of Philosophy and hereby certify that, in their opinion, it is worthy of acceptance. Dr. Zhihai He Dr. Guilherme DeSouza Dr. Dominic Ho Dr. Jianlin Cheng ACKNOWLEDGMENTS This is the perfect time to replay my memories during my doctorate life, and I suddenly recall so many people to thank. First and foremost, I would like to sincerely thank my advisor Dr. Zhihai He, who guided me into the research community. I am fortunate enough to have Dr. He’s professionalism backing my research and study. Dr. He is an awesome friend and tutor as well, which is warm and nice considering I am studying ten thousand miles away from home. I would also like to express my deep gratitude to Dr Guilherme DeSouza, Dr. Dominic Ho, Dr Jianlin Chen, for being so supportive committee members. Besides, it is a good time to sincerely thank professors and department faculties Dr. Tony Han, Dr. Michela Becchi, Dr. James Keller and so many more, who teach me knowledge and help me walk through the doctoral degree. And of course, how can I forget my dear colleagues and friends in Mizzou and Columbia: Xiaobo Ren, Yifeng Zeng, Chen Huang, Guanghan Ning, Zhiqun Zhao, Yang Li, Hao Sun and Hayder Yousif.
    [Show full text]
  • Content-Based Image Copy Detection Using Convolutional Neural Network
    electronics Article Content-Based Image Copy Detection Using Convolutional Neural Network Xiaolong Liu 1,2,* , Jinchao Liang 1, Zi-Yi Wang 3, Yi-Te Tsai 4, Chia-Chen Lin 3,5,* and Chih-Cheng Chen 6,* 1 College of Computer and Information Science, Fujian Agriculture and Forestry University, Fuzhou 350002, China; [email protected] 2 Digital Fujian Institute of Big Data for Agriculture and Forestry, Fujian Agriculture and Forestry University, Fuzhou 350002, China 3 Department of Computer Science and Information Engineering, National Chin-Yi University of Technology, Taichung 41170, Taiwan; [email protected] 4 Department of Computer Science and Communication Engineering, Providence University, Taichung 43301, Taiwan; [email protected] 5 Department of Computer Science and Information Management, Providence University, Taichung 43301, Taiwan 6 Department of Aeronautical Engineering, Chaoyang University of Technology, Taichung 413310, Taiwan * Correspondence: [email protected] (X.L.); [email protected] (C.-C.L.); [email protected] (C.-C.C.) Received: 27 October 2020; Accepted: 23 November 2020; Published: 1 December 2020 Abstract: With the rapid development of network technology, concerns pertaining to the enhancement of security and protection against violations of digital images have become critical over the past decade. In this paper, an image copy detection scheme based on the Inception convolutional neural network (CNN) model in deep learning is proposed. The image dataset is transferred by a number of image processing manipulations and the feature values in images are automatically extracted for learning and detecting the suspected unauthorized digital images. The experimental results show that the proposed scheme takes on an extraordinary role in the process of detecting duplicated images with rotation, scaling, and other content manipulations.
    [Show full text]
  • Multiple Object Recognition Using Opencv
    Multiple Object Recognition Using OpenCV D. Kavitha1; B.P. Rishi Kiran2; B. Niteesh3; S. Praveen4 1Assistant Professor, Department of Computer Science and Engineering, SRM Institute of Science and Technology, Ramapuram Campus, Chennai, India. [email protected] 2Department of Computer Science and Engineering, SRM Institute of Science and Technology, Ramapuram Campus, Chennai, India. [email protected] 3Department of Computer Science and Engineering, SRM Institute of Science and Technology, Ramapuram Campus, Chennai, India. [email protected] 4Department of Computer Science and Engineering, SRM Institute of Science and Technology, Ramapuram Campus, Chennai, India. [email protected] Abstract For automatic vision systems used in agriculture, the project presents object characteristics analysis using image processing techniques. In agriculture science, automatic object characteristics identification is important for monitoring vast areas of crops, and it detects signs of object characteristics as soon as it occurs on plant leaves. Image content characterization and supervised classifier type neural network are used in the proposed deciding method. Pre-processing, image segmentation, and detection are some of the image processing methods used in this form of decision making. An image data will be rearranged and, if necessary, a region of interest will be selected during preparation. For network training and classification, colour and texture features are extracted from an input. Colour characteristics such as mean and variance in the HSV colour space, as well as texture characteristics such as energy, contrast, homogeneity, and correlation. The device will be trained to automatically identify test images in order to assess object characteristics. With some training samples of that type, an automated classifier NN could be used for classification supported learning in this method.
    [Show full text]