{"id":1995,"date":"2018-10-22T17:51:43","date_gmt":"2018-10-22T20:51:43","guid":{"rendered":"https:\/\/impresee.com\/?p=1995"},"modified":"2020-09-04T16:13:16","modified_gmt":"2020-09-04T20:13:16","slug":"understanding-object-detection-methods-2","status":"publish","type":"post","link":"https:\/\/impresee.com\/es\/understanding-object-detection-methods-2\/","title":{"rendered":"Entiendo m\u00e9todos de detecci\u00f3n de objetos"},"content":{"rendered":"<p class=\"graf graf--p\">Object detection is a common problem when it comes to computer vision. Knowledge and understanding in this area, however, is maturing rapidly, mainly because of advances in deep-learning models. The goal of object detection is to pinpoint and classify objects of interest appearing in an image. This task is naturally more complex than straightforward image classification because there may be a range of objects we are focused on, as well as the additional requirement of needing to predict bounding boxes that will encompass the detected object. A bounding box is defined by four parameters [x, y, w, h], where the first two parameters (x,y) indicate a reference spatial position in the box, commonly the center of the box or the upper-left corner, and the last two are set for the width and height of the box, respectively. An example of this task is showed in Figure 1.<\/p>\n<figure class=\"graf graf--figure\">\n<p><div style=\"width: 462px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"graf-image\" src=\"https:\/\/cdn-images-1.medium.com\/max\/800\/1*eDfWOUre6cqZTNKEieGTdQ.png\" alt=\"\" width=\"452\" height=\"272\"><p class=\"wp-caption-text\">Figure 1. Object detection in&nbsp;action.<\/p><\/div><\/figure>\n<p class=\"graf graf--p\">Deep-learning based models, through convolutional neural networks, have had a positive impact on advances in this area, and this has already led to the technology being applied to industry models. In this post, we will discuss two of the main strategies for addressing object detection. The first one is a two-stage based approach mainly represented by the Faster-RCNN [1] architecture, and the second one is a one-stage approach represented by YOLO [3].<\/p>\n<p class=\"graf graf--p\">Before the emergence of deep-learning, the task of object detection was addressed by means of a costly strategy known as \u201c<em class=\"markup--em markup--p-em\">the sliding window<\/em>\u201d, where a rectangle with different sizes would move over the whole image trying to find relevant objects. For this purpose, a classifier for each of the classes we were interested in needed to be applied for each window. It was highly costly! Nowadays, state-of-the-art methods for object detection rely on a <strong class=\"markup--strong markup--p-strong\">convolutional neural network<\/strong> which also implements a sliding window approach but in a more efficient way. Convolution layers are key!<\/p>\n<figure class=\"graf graf--figure\">\n<div style=\"width: 678px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"graf-image\" src=\"https:\/\/cdn-images-1.medium.com\/max\/800\/1*PGSpDxt6I495aiFyt5S33Q.png\" alt=\"\" width=\"668\" height=\"372\"><p class=\"wp-caption-text\">Figure 2. Scheme of a convolutional neural network [copyright Serengil].<\/p><\/div><figcaption class=\"imageCaption\"><\/figcaption><\/figure>\n<p class=\"graf graf--p\"><strong class=\"markup--strong markup--p-strong\">1. Convolutional Neural Networks<\/strong><\/p>\n<p class=\"graf graf--p\">A convolutional neural network (CNN) is a special kind of neural network (see Figure 2) mainly focused on processing image data but it also includes extensions to other kinds of signals which can be thought as a 2-D grid of pixels. The term <em class=\"markup--em markup--p-em\">convolution <\/em>indicates that the network is composed of layers resembling the <em class=\"markup--em markup--p-em\">convolution<\/em> operation used frequently in image processing. Convolutional networks are the most successful examples within the <em class=\"markup--em markup--p-em\">connectivist<\/em> branch of the machine learning field. This can be seen as a simplified, cartoon view of brain function, sharing some properties with the primary visual cortex V1.<\/p>\n<ul class=\"postList\">\n<li class=\"graf graf--li\">V1 and CNN are arranged in a spatial map where each neuron is affected by a local receptive field. Each neuron of a convolution layer is affected by a local region in the input image, this region is known as the receptive field.<\/li>\n<li class=\"graf graf--li\">V1 and CNN contain simple and complex cells which allow for the learning of a hierarchical representation from images. Indeed, the shallow layers in a CNN learn low-level features, similar to the well known Gabor filters, while the deeper layers bring more semantic information.<\/li>\n<\/ul>\n<p class=\"graf graf--p\">As a convolutional neural network is capable of learning highly-discriminative features, we can exploit the learned features in other problems that may be different from the classification one. In fact, we have seen successful results on similarity search, object detection, face recognition, image segmentation, image captioning, among others.<\/p>\n<p class=\"graf graf--p\">Focusing on the object detection problem, there have been significant advances in the last few years, particularly since the publication of Faster-RCNN, which brought two specific contributions: i) the use of a regions proposal network, and ii) the use of anchors to deal with the variable size of objects. A performance evolution on the object detection problem is showed in Figure 3.<\/p>\n<figure class=\"graf graf--figure\">\n<div style=\"width: 810px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"graf-image\" src=\"https:\/\/cdn-images-1.medium.com\/max\/800\/1*7Mrku99K2v7SZP4rw31RLA.png\" alt=\"\" width=\"800\" height=\"399\"><p class=\"wp-caption-text\">Figure 3. Performance Evolution on Object Detection [copyright Girshick]<\/p><\/div><figcaption class=\"imageCaption\"><\/figcaption><\/figure>\n<p class=\"graf graf--p\"><strong class=\"markup--strong markup--p-strong\">2. An end-to-end Convolutional Neural Network for Object Detection<\/strong><\/p>\n<p class=\"graf graf--p\">Faster-RCNN proposed by Ren et al. [1] was one of the first proposals aiming to utilize an end-to-end trained convolutional model for the object detection problem. This proposal came up as an improvement of Fast-RCNN, a previous work which required a list of candidate objects generated by a separated module known as the <em class=\"markup--em markup--p-em\">objectness module<\/em>. To this end, the <em class=\"markup--em markup--p-em\">selective search<\/em> approach [5] was commonly used. Having an extra module for acquiring object candidates would mean additional costs. Therefore, the main objective of Faster-RCNN was to reduce this overhead by a convolutional architecture trained end-to-end for the object detection task.<\/p>\n<p class=\"graf graf--p\">Faster-RCNN is composed of two blocks sharing a backbone. This in turn aims to produce discriminative features that will be used to estimate candidate objects as well as to predict the class of those objects. Of course, the backbone is also a convolutional neural network! To this end, a ResNet architecture has recently been used. The first block, termed Region Proposal Network (RPN), is devoted to find regions where an object is highly likely to appear, while the second block is just the Fast-RCNN focusing on predicting the class of proposed objects.<\/p>\n<p class=\"graf graf--p\"><strong class=\"markup--strong markup--p-strong\">2.1. Region Proposal Network<\/strong><\/p>\n<p class=\"graf graf--p\">The Region Proposal Network (RPN) is a convolutional network devoted to detect regions in the image where objects may be found. This works as a class-agnostic stage.<\/p>\n<p class=\"graf graf--p\">The backbone of the RPN is a convolutional neural network. To this end, a ResNet arquitecture has been recently used. The deeper convolutional layer is used as the feature map and this map is arranged into HxW nodes. Each node is related to a receptive field with a certain size in the input image. The exact size will depend on the feature map depth.<\/p>\n<p class=\"graf graf--p\">The RPN places a convolutional layer immediately after the feature map, aiming to learn 256 features for each node. This is carried out by simply connecting a 256-channel convolutional using a 3&#215;3 kernel. The result of this layer is the new feature map that will be used to predict the occurrence of an object.<\/p>\n<p class=\"graf graf--p\">The last component of the RPN is the predictor. Actually, we have two kinds of predictors, one in charge of predicting whether a region contains an object or not, and the other in charge of predicting the bounding box covering the object, if it occurred. The first one is a kind of classifier, while the second is a regressor.<\/p>\n<p class=\"graf graf--p\">To accelerate the training, RPN uses a set of reference regions per node, <em class=\"markup--em markup--p-em\">the anchors, <\/em>which actually represents a rectangular region in the input image which is positioned in the center of the receptive field with respect to each node in the feature map. An anchor is defined by its scale (e.g. 64&#215;64) and its aspect ratio (e.g. 1:1, 1:2, 2:1). The choice of the anchors (scale and aspect ratios) strongly depends on the application being used. In the Figure 4, a set of anchors is illustrated.<\/p>\n<figure class=\"graf graf--figure\">\n<p><div style=\"width: 810px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"graf-image\" src=\"https:\/\/cdn-images-1.medium.com\/max\/800\/1*SE_ZBRktaWReju7hTCPdQg.png\" alt=\"\" width=\"800\" height=\"495\"><p class=\"wp-caption-text\">Figure 4. Examples of anchor for a&nbsp;node.<\/p><\/div><figcaption class=\"imageCaption\"><\/figcaption><\/figure>\n<figure class=\"graf graf--figure\">\n<p><div style=\"width: 810px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"graf-image\" src=\"https:\/\/cdn-images-1.medium.com\/max\/800\/1*orNyHi561QiZPS1MjNV2sA.png\" alt=\"\" width=\"800\" height=\"317\"><p class=\"wp-caption-text\">Figure 5. RPN Arquitecture<\/p><\/div><figcaption class=\"imageCaption\"><\/figcaption><\/figure>\n<p class=\"graf graf--p\">Supposing we have defined K anchors per node, the classification block will need to learn 2K classifiers. That is, it will learn an object-classifier as well as a non-object classifier per anchor. On the other hand, the regressor will learn four parameters per anchor. These parameters represent certain transformations that an anchor must undergo to tightly cover the objects with respect to the four parameters which define it. Therefore, the number of regressors that RPN needs to learn is 4K. In Figures 5 and 6, the RPN architecture is shown. For training, the classifier set uses a classical cross-<em class=\"markup--em markup--p-em\">entropy<\/em> loss, while the regressor set uses the <em class=\"markup--em markup--p-em\">smooth-L1<\/em> loss function.<\/p>\n<figure class=\"graf graf--figure\">\n<p><div style=\"width: 810px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"graf-image\" src=\"https:\/\/cdn-images-1.medium.com\/max\/800\/1*JWWv0GyNl3UGFY0bFERjjA.png\" alt=\"\" width=\"800\" height=\"334\"><p class=\"wp-caption-text\">Figure 6. 2K classifiers and 4K regressors in the&nbsp;RPN<\/p><\/div><figcaption class=\"imageCaption\"><\/figcaption><\/figure>\n<p class=\"graf graf--p\"><strong class=\"markup--strong markup--p-strong\">2.2. Detection by Fast-RCNN<\/strong><\/p>\n<p class=\"graf graf--p\">This is the second block of the Faster-RCNN, and actually this corresponds to the Fast-RCNN model. However, in this case, the set of candidate objects will be provided by the RPN block. The Fast-RCNN is focused on predicting the most probable class (given a set of predefined classes) plus an extra class representing the background. For this purpose, the model again depends on a convolutional neural network to provide discriminative features for each region of interest (ROI). This is the backbone of the model and it is the same as that used by the RPN block.<\/p>\n<p class=\"graf graf--p\">Having a feature map computed from the input image through the convolutional backbone, the Fast-RCNN extracts a subregion according to the candidate ROIs proposed by the RPN. We must be aware that the diversity of sizes within the area of interest is really a matter of defining the architecture of a convolutional model. To address this problem, Fast-RCNN proposes a special layer termed \u201cROI Pooling Layer\u201d which, through an average pooling operation, transforms the feature map with respect to a given ROI into a feature map of HxW nodes. The result is then passed through a couple of fully connected layers to finally pass to a set of classifiers to predict the most probable class, as well as to a set of regressors in charge of adjusting the input ROI using semantic information. In Figure 7, a scheme of Fast-RCNN is shown.<\/p>\n<p><div style=\"width: 863px\" class=\"wp-caption alignnone\"><img loading=\"lazy\" decoding=\"async\" class=\"progressiveMedia-zoom js-progressiveMedia-inner\" style=\"font-weight: bold; color: #666666; font-size: 0.8125rem; font-style: italic;\" src=\"https:\/\/cdn-images-1.medium.com\/max\/1000\/1*jEj8s0qkX8rRCtXtNZ-B1Q.png\" alt=\"\" width=\"853\" height=\"425\"><p class=\"wp-caption-text\">Figure 7. Fast-RCNN (image from original paper[2])<\/p><\/div><span style=\"color: #333333; font-size: 1rem;\">The training process is carried out in an alternate way between RPN and Fast-RCNN. This model has shown high performance results, but it is still low for real-time applications.<\/span><\/p>\n<p class=\"graf graf--p\" style=\"text-align: left;\"><strong class=\"markup--strong markup--p-strong\">3. Toward Real Time Detection with YOLO<\/strong><\/p>\n<p class=\"graf graf--p\">The two-stage based approach proposed by Faster-RCNN is the main drawback of the model. It would be more efficient if a one-stage approach could do the same work, right? Well, YOLO [3] has been proposed to address this drawback. In fact, YOLO is an acronym of the expression \u201cYou Only Look Once\u201d.<\/p>\n<p class=\"graf graf--p\">The first proposal via YOLO is somewhat minimalist, which also makes it attractive. Like Faster-RCNN, the backbone of the proposal is a convolutional network which produces a feature map from an input image. This feature map contains SxS nodes, each one related with a receptive field in the input image. Each node also predicts B bounding boxes as well as a score for each box. This score represents how confident a model is about finding an object in that box. A score of zero will mean that there is no object in the box. Consequently, five predictors are required for each bounding box. These predictors estimate the (x,y) coordinates representing the center of the box, the width (w) and height (h) of the box as well as the confidence of the prediction. In addition, each node also predicts C conditional class probabilities P(C_i | Object). The proposal predicts one set of class probabilities per each node, regardless of the number of boxes. Therefore, at the end, the model needs to estimate (SxSx(Bx5 + C)) regressors as showed in Figure 8.<\/p>\n<figure class=\"graf graf--figure\">\n<p><div style=\"width: 810px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"graf-image\" src=\"https:\/\/cdn-images-1.medium.com\/max\/800\/1*XIcahjCcj8F3cqXZPgTvzA.png\" alt=\"\" width=\"800\" height=\"290\"><p class=\"wp-caption-text\">Figure 8. YOLO in action. The colored volume represents the set of learned regressors.<\/p><\/div><figcaption class=\"imageCaption\"><\/figcaption><\/figure>\n<p class=\"graf graf--p\">Although, the prediction speed of the model is very high, the performance is not yet comparable with Faster-RCNN, undergoing a poor localization. However, in the follow-up version of YOLO[3] (YOLO9000 [4]), the authors incorporated the use of anchors which then allowed them to improve the localization.<\/p>\n<p class=\"graf graf--p\">In the last year, there have been other improvements focused on powering up the performance of one-stage models leveraging the multi-scale information embedded into the convolutional neural nets and addressing the class imbalance present within the object detection problem. <strong class=\"markup--strong markup--p-strong\">However, these new improvements will be discussed in the next posts<\/strong>.<\/p>\n<p class=\"graf graf--p\"><strong class=\"markup--strong markup--p-strong\">References<\/strong><\/p>\n<p class=\"graf graf--p\">[1] <a class=\"markup--anchor markup--p-anchor\" href=\"https:\/\/arxiv.org\/search\/cs?searchtype=author&amp;query=Ren%2C+S\" target=\"_blank\" rel=\"noopener noreferrer\">Shaoqing Ren<\/a>, <a class=\"markup--anchor markup--p-anchor\" href=\"https:\/\/arxiv.org\/search\/cs?searchtype=author&amp;query=He%2C+K\" target=\"_blank\" rel=\"noopener noreferrer\">Kaiming He<\/a>, <a class=\"markup--anchor markup--p-anchor\" href=\"https:\/\/arxiv.org\/search\/cs?searchtype=author&amp;query=Girshick%2C+R\" target=\"_blank\" rel=\"noopener noreferrer\">Ross Girshick<\/a>, <a class=\"markup--anchor markup--p-anchor\" href=\"https:\/\/arxiv.org\/search\/cs?searchtype=author&amp;query=Sun%2C+J\" target=\"_blank\" rel=\"noopener noreferrer\">Jian Su. n<\/a>Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. <a class=\"markup--anchor markup--p-anchor\" href=\"https:\/\/arxiv.org\/abs\/1506.01497\" target=\"_blank\" rel=\"nofollow noopener noreferrer\">https:\/\/arxiv.org\/abs\/1506.01497<\/a><\/p>\n<p class=\"graf graf--p\">[2] <a class=\"markup--anchor markup--p-anchor\" href=\"https:\/\/arxiv.org\/search\/cs?searchtype=author&amp;query=Girshick%2C+R\" target=\"_blank\" rel=\"noopener noreferrer\">Ross Girshick<\/a>, Fast R-CNN. <a class=\"markup--anchor markup--p-anchor\" href=\"https:\/\/arxiv.org\/abs\/1504.08083\" target=\"_blank\" rel=\"nofollow noopener noreferrer\">https:\/\/arxiv.org\/abs\/1504.08083<\/a><\/p>\n<p class=\"graf graf--p\">[3] <em class=\"markup--em markup--p-em\">Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. <\/em>You Only Look Once: Unified, Real-Time Object Detection. <a class=\"markup--anchor markup--p-anchor\" href=\"https:\/\/pjreddie.com\/media\/files\/papers\/yolo_1.pdf\" target=\"_blank\" rel=\"nofollow noopener noreferrer\">https:\/\/pjreddie.com\/media\/files\/papers\/yolo_1.pdf<\/a><\/p>\n<p class=\"graf graf--p\">[4] <em class=\"markup--em markup--p-em\">Joseph Redmon, Ali Farhadi. <\/em>YOLO9000: Better, Faster, Stronger. <a class=\"markup--anchor markup--p-anchor\" href=\"https:\/\/pjreddie.com\/media\/files\/papers\/YOLO9000.pdf\" target=\"_blank\" rel=\"nofollow noopener noreferrer\">https:\/\/pjreddie.com\/media\/files\/papers\/YOLO9000.pdf<\/a><\/p>\n<p class=\"graf graf--p\">[5] J. R. R. Uijlings, K. E. A. van de Sande. T. Gevers. A.W.M.Smeulders. Selective Search for Object Recognition<\/p>","protected":false},"excerpt":{"rendered":"<p>Object detection is a common problem when it comes to computer vision. Knowledge and understanding in this area, however, is maturing rapidly, mainly because of advances in deep-learning models. The goal of object detection is to pinpoint and classify objects of interest appearing in an image. This task is naturally more complex than straightforward image [&hellip;]<\/p>","protected":false},"author":1,"featured_media":1236,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_et_pb_use_builder":"","_et_pb_old_content":"","_et_gb_content_width":"","footnotes":""},"categories":[23],"tags":[10,59,58],"class_list":["post-1995","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-impresee-research-center","tag-computer-vision","tag-object-detection","tag-research"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v24.1 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Understanding Object Detection Methods convolutional neural network<\/title>\n<meta name=\"description\" content=\"Object detection is a common problem when it comes to computer vision. Knowledge and understanding in this area, advances in deep-learning models.\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/impresee.com\/es\/understanding-object-detection-methods-2\/\" \/>\n<meta property=\"og:locale\" content=\"es_ES\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Understanding Object Detection Methods convolutional neural network\" \/>\n<meta property=\"og:description\" content=\"Object detection is a common problem when it comes to computer vision. Knowledge and understanding in this area, advances in deep-learning models.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/impresee.com\/es\/understanding-object-detection-methods-2\/\" \/>\n<meta property=\"og:site_name\" content=\"Impresee: Search bar and navigation solutions for eCommerce\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/impresee\/\" \/>\n<meta property=\"article:published_time\" content=\"2018-10-22T20:51:43+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2020-09-04T20:13:16+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/impresee.com\/wp-content\/uploads\/2018\/08\/1_HJ8AqiN5wxZ_uU_iJ-x5Ew.png\" \/>\n\t<meta property=\"og:image:width\" content=\"714\" \/>\n\t<meta property=\"og:image:height\" content=\"405\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"Impresee\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@impresee\" \/>\n<meta name=\"twitter:site\" content=\"@impresee\" \/>\n<meta name=\"twitter:label1\" content=\"Escrito por\" \/>\n\t<meta name=\"twitter:data1\" content=\"Impresee\" \/>\n\t<meta name=\"twitter:label2\" content=\"Tiempo estimado de lectura\" \/>\n\t<meta name=\"twitter:data2\" content=\"10 minutos\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/\"},\"author\":{\"name\":\"Impresee\",\"@id\":\"https:\/\/impresee.com\/#\/schema\/person\/f4b36cfb4f415642ee02f2d6eda34b62\"},\"headline\":\"Understanding Object Detection Methods\",\"datePublished\":\"2018-10-22T20:51:43+00:00\",\"dateModified\":\"2020-09-04T20:13:16+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/\"},\"wordCount\":2013,\"publisher\":{\"@id\":\"https:\/\/impresee.com\/#organization\"},\"image\":{\"@id\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/impresee.com\/wp-content\/uploads\/2018\/08\/1_HJ8AqiN5wxZ_uU_iJ-x5Ew.png\",\"keywords\":[\"computer vision\",\"object detection\",\"research\"],\"articleSection\":[\"research center impresee\"],\"inLanguage\":\"es-CL\"},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/\",\"url\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/\",\"name\":\"Understanding Object Detection Methods convolutional neural network\",\"isPartOf\":{\"@id\":\"https:\/\/impresee.com\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/impresee.com\/wp-content\/uploads\/2018\/08\/1_HJ8AqiN5wxZ_uU_iJ-x5Ew.png\",\"datePublished\":\"2018-10-22T20:51:43+00:00\",\"dateModified\":\"2020-09-04T20:13:16+00:00\",\"description\":\"Object detection is a common problem when it comes to computer vision. Knowledge and understanding in this area, advances in deep-learning models.\",\"breadcrumb\":{\"@id\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#breadcrumb\"},\"inLanguage\":\"es-CL\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"es-CL\",\"@id\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#primaryimage\",\"url\":\"https:\/\/impresee.com\/wp-content\/uploads\/2018\/08\/1_HJ8AqiN5wxZ_uU_iJ-x5Ew.png\",\"contentUrl\":\"https:\/\/impresee.com\/wp-content\/uploads\/2018\/08\/1_HJ8AqiN5wxZ_uU_iJ-x5Ew.png\",\"width\":714,\"height\":405,\"caption\":\"Figure 1. Resolution vs Semantic\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/impresee.com\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Understanding Object Detection Methods\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/impresee.com\/#website\",\"url\":\"https:\/\/impresee.com\/\",\"name\":\"Impresee\",\"description\":\"AI search bar, filters and discount games for eCommerce to get more visitors, increase conversion and get repeat sales. Available for Shopify, Woocomerce, Magento and others\",\"publisher\":{\"@id\":\"https:\/\/impresee.com\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/impresee.com\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"es-CL\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/impresee.com\/#organization\",\"name\":\"Impresee\",\"url\":\"https:\/\/impresee.com\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"es-CL\",\"@id\":\"https:\/\/impresee.com\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/impresee.com\/wp-content\/uploads\/2019\/01\/logo_color_negro_sf_web-e1547827174322.png\",\"contentUrl\":\"https:\/\/impresee.com\/wp-content\/uploads\/2019\/01\/logo_color_negro_sf_web-e1547827174322.png\",\"width\":400,\"height\":152,\"caption\":\"Impresee\"},\"image\":{\"@id\":\"https:\/\/impresee.com\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/www.facebook.com\/impresee\/\",\"https:\/\/x.com\/impresee\",\"https:\/\/www.instagram.com\/impresee.ai\/\",\"https:\/\/www.linkedin.com\/company\/impresee\",\"https:\/\/www.youtube.com\/channel\/UC5ZniKfJGvG8t4GrOOgYczA\"]},{\"@type\":\"Person\",\"@id\":\"https:\/\/impresee.com\/#\/schema\/person\/f4b36cfb4f415642ee02f2d6eda34b62\",\"name\":\"Impresee\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"es-CL\",\"@id\":\"https:\/\/impresee.com\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/ec921996d4b354556f36179b07308eff2ad5cbbb42fce1cd037b95bfcd0d6eb5?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/ec921996d4b354556f36179b07308eff2ad5cbbb42fce1cd037b95bfcd0d6eb5?s=96&d=mm&r=g\",\"caption\":\"Impresee\"},\"url\":\"https:\/\/impresee.com\/es\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Understanding Object Detection Methods convolutional neural network","description":"Object detection is a common problem when it comes to computer vision. Knowledge and understanding in this area, advances in deep-learning models.","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/impresee.com\/es\/understanding-object-detection-methods-2\/","og_locale":"es_ES","og_type":"article","og_title":"Understanding Object Detection Methods convolutional neural network","og_description":"Object detection is a common problem when it comes to computer vision. Knowledge and understanding in this area, advances in deep-learning models.","og_url":"https:\/\/impresee.com\/es\/understanding-object-detection-methods-2\/","og_site_name":"Impresee: Search bar and navigation solutions for eCommerce","article_publisher":"https:\/\/www.facebook.com\/impresee\/","article_published_time":"2018-10-22T20:51:43+00:00","article_modified_time":"2020-09-04T20:13:16+00:00","og_image":[{"width":714,"height":405,"url":"https:\/\/impresee.com\/wp-content\/uploads\/2018\/08\/1_HJ8AqiN5wxZ_uU_iJ-x5Ew.png","type":"image\/png"}],"author":"Impresee","twitter_card":"summary_large_image","twitter_creator":"@impresee","twitter_site":"@impresee","twitter_misc":{"Escrito por":"Impresee","Tiempo estimado de lectura":"10 minutos"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#article","isPartOf":{"@id":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/"},"author":{"name":"Impresee","@id":"https:\/\/impresee.com\/#\/schema\/person\/f4b36cfb4f415642ee02f2d6eda34b62"},"headline":"Understanding Object Detection Methods","datePublished":"2018-10-22T20:51:43+00:00","dateModified":"2020-09-04T20:13:16+00:00","mainEntityOfPage":{"@id":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/"},"wordCount":2013,"publisher":{"@id":"https:\/\/impresee.com\/#organization"},"image":{"@id":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#primaryimage"},"thumbnailUrl":"https:\/\/impresee.com\/wp-content\/uploads\/2018\/08\/1_HJ8AqiN5wxZ_uU_iJ-x5Ew.png","keywords":["computer vision","object detection","research"],"articleSection":["research center impresee"],"inLanguage":"es-CL"},{"@type":"WebPage","@id":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/","url":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/","name":"Understanding Object Detection Methods convolutional neural network","isPartOf":{"@id":"https:\/\/impresee.com\/#website"},"primaryImageOfPage":{"@id":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#primaryimage"},"image":{"@id":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#primaryimage"},"thumbnailUrl":"https:\/\/impresee.com\/wp-content\/uploads\/2018\/08\/1_HJ8AqiN5wxZ_uU_iJ-x5Ew.png","datePublished":"2018-10-22T20:51:43+00:00","dateModified":"2020-09-04T20:13:16+00:00","description":"Object detection is a common problem when it comes to computer vision. Knowledge and understanding in this area, advances in deep-learning models.","breadcrumb":{"@id":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#breadcrumb"},"inLanguage":"es-CL","potentialAction":[{"@type":"ReadAction","target":["https:\/\/impresee.com\/understanding-object-detection-methods-2\/"]}]},{"@type":"ImageObject","inLanguage":"es-CL","@id":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#primaryimage","url":"https:\/\/impresee.com\/wp-content\/uploads\/2018\/08\/1_HJ8AqiN5wxZ_uU_iJ-x5Ew.png","contentUrl":"https:\/\/impresee.com\/wp-content\/uploads\/2018\/08\/1_HJ8AqiN5wxZ_uU_iJ-x5Ew.png","width":714,"height":405,"caption":"Figure 1. Resolution vs Semantic"},{"@type":"BreadcrumbList","@id":"https:\/\/impresee.com\/understanding-object-detection-methods-2\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/impresee.com\/"},{"@type":"ListItem","position":2,"name":"Understanding Object Detection Methods"}]},{"@type":"WebSite","@id":"https:\/\/impresee.com\/#website","url":"https:\/\/impresee.com\/","name":"Impresee","description":"AI search bar, filters and discount games for eCommerce to get more visitors, increase conversion and get repeat sales. Available for Shopify, Woocomerce, Magento and others","publisher":{"@id":"https:\/\/impresee.com\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/impresee.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"es-CL"},{"@type":"Organization","@id":"https:\/\/impresee.com\/#organization","name":"Impresee","url":"https:\/\/impresee.com\/","logo":{"@type":"ImageObject","inLanguage":"es-CL","@id":"https:\/\/impresee.com\/#\/schema\/logo\/image\/","url":"https:\/\/impresee.com\/wp-content\/uploads\/2019\/01\/logo_color_negro_sf_web-e1547827174322.png","contentUrl":"https:\/\/impresee.com\/wp-content\/uploads\/2019\/01\/logo_color_negro_sf_web-e1547827174322.png","width":400,"height":152,"caption":"Impresee"},"image":{"@id":"https:\/\/impresee.com\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/impresee\/","https:\/\/x.com\/impresee","https:\/\/www.instagram.com\/impresee.ai\/","https:\/\/www.linkedin.com\/company\/impresee","https:\/\/www.youtube.com\/channel\/UC5ZniKfJGvG8t4GrOOgYczA"]},{"@type":"Person","@id":"https:\/\/impresee.com\/#\/schema\/person\/f4b36cfb4f415642ee02f2d6eda34b62","name":"Impresee","image":{"@type":"ImageObject","inLanguage":"es-CL","@id":"https:\/\/impresee.com\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/ec921996d4b354556f36179b07308eff2ad5cbbb42fce1cd037b95bfcd0d6eb5?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/ec921996d4b354556f36179b07308eff2ad5cbbb42fce1cd037b95bfcd0d6eb5?s=96&d=mm&r=g","caption":"Impresee"},"url":"https:\/\/impresee.com\/es\/"}]}},"_links":{"self":[{"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/posts\/1995","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/comments?post=1995"}],"version-history":[{"count":4,"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/posts\/1995\/revisions"}],"predecessor-version":[{"id":7255,"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/posts\/1995\/revisions\/7255"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/media\/1236"}],"wp:attachment":[{"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/media?parent=1995"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/categories?post=1995"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/impresee.com\/es\/wp-json\/wp\/v2\/tags?post=1995"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}