{"id":1613,"date":"2021-07-13T10:00:00","date_gmt":"2021-07-13T01:00:00","guid":{"rendered":"https:\/\/blog.testworks.co.kr\/en\/?p=1613"},"modified":"2022-12-28T15:39:04","modified_gmt":"2022-12-28T06:39:04","slug":"automation-in-ai-data-processing","status":"publish","type":"post","link":"https:\/\/blog.aiworkx.ai\/en\/automation-in-ai-data-processing\/","title":{"rendered":"Automation in AI Data Processing"},"content":{"rendered":"\n<div style=\"height:100px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"439\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/10\/\ub370\uc774\ud130-\uac00\uacf5-\uc790\ub3d9\ud654-\ubaa8\ub378\uacfc-\ud65c\uc6a9_cover-image-1024x439.png\" alt=\"\" class=\"wp-image-1667\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/10\/\ub370\uc774\ud130-\uac00\uacf5-\uc790\ub3d9\ud654-\ubaa8\ub378\uacfc-\ud65c\uc6a9_cover-image-1024x439.png 1024w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/10\/\ub370\uc774\ud130-\uac00\uacf5-\uc790\ub3d9\ud654-\ubaa8\ub378\uacfc-\ud65c\uc6a9_cover-image-300x129.png 300w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/10\/\ub370\uc774\ud130-\uac00\uacf5-\uc790\ub3d9\ud654-\ubaa8\ub378\uacfc-\ud65c\uc6a9_cover-image-768x330.png 768w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/10\/\ub370\uc774\ud130-\uac00\uacf5-\uc790\ub3d9\ud654-\ubaa8\ub378\uacfc-\ud65c\uc6a9_cover-image.png 1235w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure><\/div>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-1 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\"><\/div>\n<\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<p style=\"font-size:18px\">Most of the deep learning models are open source, so if you are a developer with knowledge in the field of artificial intelligence, you can make use of the models without difficulty by reading related papers and finding and running the implemented source on GitHub.&nbsp;What determines the performance of a deep learning model is the set of weight values of the model learned through training.<\/p>\n\n\n\n<p style=\"font-size:18px\">Training a deep learning model requires hundreds, thousands, and even millions of data point. Manually processing all the data is wasteful, labor-intensive, repetitive, and inefficient.<\/p>\n\n\n\n<p style=\"font-size:18px\">For efficient data processing, an automation model that has been pre-trained is used to label the data first, and then reviewed and fixed by humans-in-the-loop as the second step.<\/p>\n\n\n\n<p style=\"font-size:18px\">If you use the auto-labeling model, you can focus on improving the quality of data by reducing the resources required for the initial annotation and paying more attention on reviewing and fixing the annotation errors in the second phase.<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"249\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-1-1-1024x249.png\" alt=\"\" class=\"wp-image-1638\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-1-1-1024x249.png 1024w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-1-1-300x73.png 300w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-1-1-768x187.png 768w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-1-1.png 1235w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><figcaption><strong>[Figure 1] Data processing workflow using the automation model<\/strong><\/figcaption><\/figure><\/div>\n\n\n\n<div style=\"height:100px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>Automation Model<\/strong><\/h2>\n\n\n\n<p style=\"font-size:18px\">If you are not in the field of artificial intelligence data processing, the term \u201cdata processing automation model\u201d may seem unfamiliar and ambiguous in its meaning.&nbsp;This is because data is processed to train a deep learning model and using a deep learning model to process this data sounds like the chicken-and-egg problem.&nbsp;In fact, the chicken-and-egg problem is a good way to think about the automation model that auto-labels the data and the model trained with the preprocessed data.&nbsp;The difference is that the automation model that does the preliminary auto-labeling is trained with relatively limited data to achieve \u2018good enough\u2019 performance in a short time. On the other hand, the model trained for production use needs to have the best possible performance so more training data is used, together with various hyper-parameters optimization, and numerous tests over a long period of time.<\/p>\n\n\n\n<p style=\"font-size:18px\">Currently, about 80% of the processing projects carried out by Testworks are related to image-based data, and automation models for image data are developed and deployed to improve productivity and quality.&nbsp;In the case of text or voice data, customized voice transcription tools are used.<\/p>\n\n\n\n<div style=\"height:100px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>The Status of Automation Models<\/strong><\/h2>\n\n\n\n<p style=\"font-size:18px\">TestWorks has developed an automation model that can recognize approximately 260 objects in 20 categories in 7 fields such as autonomous driving, pedestrian walking, medical imaging, and security.<\/p>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>1. Object Detection<\/strong><\/h4>\n\n\n\n<p style=\"font-size:18px\">The object detection model is a model that finds the type of object and its location based on a rectangle grid and draws a bounding box around the target object as shown in our blackolive screen below.<\/p>\n\n\n\n<p style=\"font-size:18px\">The Object Detection model consists of two modules: a CNN-based classification model and a localization model that detects the location of an object. The automation model can be a 1-Stage Detector or a 2-Stage Detector depending on how the two modules are configured.<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"500\" height=\"312\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-2.png\" alt=\"\" class=\"wp-image-1621\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-2.png 500w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-2-300x187.png 300w\" sizes=\"auto, (max-width: 500px) 100vw, 500px\" \/><figcaption><strong>[Figure 2] Data processing automation and management solution, blackolive \u2013 Example of bounding boxes for object recognition<\/strong><\/figcaption><\/figure><\/div>\n\n\n\n<p style=\"font-size:18px\">The 2-Stage Detector is slow but has high accuracy, and the 1-Stage Detector is fast but has a relatively low-accuracy model.<\/p>\n\n\n\n<p style=\"font-size:18px\">Representative object recognition models<br> \u2013 2-Stage Detector: Fast RCNN, Faster RCNN, FPN (Feature Pyramid Network)<br> \u2013 1-Stage Detector: YOLO, SSD (Single Shot MultiBox Detector), RetinaNet, CenterNet<\/p>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>2. Semantic Segmentation<\/strong><\/h4>\n\n\n\n<p style=\"font-size:18px\">Semantic segmentation refers to segmenting objects of an image into meaningful units. The prediction is done at the pixel level.&nbsp;The image on the left shows the areas where the disabled can move. The result is obtained by applying an automation model to the images collected through the \u201cSidewalk Image AI Training Data\u201d project. Testworks was the leading company of the project which is sponsored by the National Intelligence Service (NIA).&nbsp;This type of prediction cannot be done by an object detection model which classifies objects in the form of bounding boxes.<\/p>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-3 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:100%\">\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-2 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-3-1.jpg\" alt=\"\" class=\"wp-image-1623\" width=\"471\" height=\"320\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-3-1.jpg 736w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-3-1-300x204.jpg 300w\" sizes=\"auto, (max-width: 471px) 100vw, 471px\" \/><figcaption><strong>[Figure 3] Example of Segmentation of a Sidewalk Pedestrian Area<\/strong><\/figcaption><\/figure><\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-4.png\" alt=\"\" class=\"wp-image-1624\" width=\"471\" height=\"320\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-4.png 471w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-4-300x204.png 300w\" sizes=\"auto, (max-width: 471px) 100vw, 471px\" \/><figcaption>[Figure 4] <strong>Example <\/strong>of highway vehicle segmentation<\/figcaption><\/figure><\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n\n\n\n<p style=\"font-size:18px\">Representative Semantic Segmentation Models<br> \u2013 Fully Convolutional Network (FCN), U-Net, DeepLab, Mask RCNN<\/p>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>3. Key-Point<\/strong><\/h4>\n\n\n\n<p style=\"font-size:18px\">During the 2020 NIA \u201cSign Language Image AI Data\u201d construction project, data for sign language recognition was labeled for human face, body, and fingers.&nbsp;For sign language recognition, it is necessary to process continuous frames (movies) for the duration of 10 seconds, rather than one or two still images.&nbsp;For example, for a video at 30 fps per second, you need to work with 300 images, so automation is absolutely necessary.<\/p>\n\n\n\n<p style=\"font-size:18px\">Although there are not many models that can extract key-points from face, body, and fingers, for sign language image data construction project, the preliminary auto-labeling is done using the Openpose model that is developed and managed by CMU (Carnegie Mellon University).<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large is-resized\"><img decoding=\"async\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-5.png\" alt=\"\" class=\"wp-image-1625\" width=\"400\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-5.png 233w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-5-218x300.png 218w\" sizes=\"(max-width: 233px) 100vw, 233px\" \/><figcaption><strong>[Figure 5] Keypoint extraction using Openpose \u2013 Sign language image data example<\/strong><\/figcaption><\/figure><\/div>\n\n\n\n<p style=\"font-size:18px\">However, there are times when finger key-point extraction could not be done properly when the left and right hands overlap. In such cases, we used different algorithms like Kalman filter to correct the problems.<\/p>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>4. Semi-Auto Polygon Labeling<\/strong><\/h4>\n\n\n\n<p style=\"font-size:18px\">When annotating polygons, if the object type has a similar domain to that of the previously developed model, the preliminary labeling can be performed using the automated model. If not, annotators must manually draw the object boundaries to conduct polygon annotation.&nbsp;This can be a daunting task that takes more than five minutes for a single object, even for experienced annotators.<\/p>\n\n\n\n<p style=\"font-size:18px\">Semi-auto polygon labeling can reduce the processing time. When the user selects the target object in a rectangular shape, it automatically separates the background and annotates the foreground object into polygons.<\/p>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-4 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large is-resized\"><img decoding=\"async\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-6.png\" alt=\"\" class=\"wp-image-1626\" width=\"400\"\/><figcaption><strong>[Figure 6] Select the target in a rectangular shape<\/strong><\/figcaption><\/figure><\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large is-resized\"><img decoding=\"async\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-7.png\" alt=\"\" class=\"wp-image-1627\" width=\"400\"\/><figcaption><strong>[Figure 7] Automatically processed result<\/strong><\/figcaption><\/figure><\/div>\n<\/div>\n<\/div>\n\n\n\n<p style=\"font-size:18px\">Using the Salient Object Detection (SOD) model, it is possible to find the most important and attention-grabbing object in an image and segment that whole object by following its contour line.<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"473\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-8-1024x473.png\" alt=\"\" class=\"wp-image-1628\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-8-1024x473.png 1024w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-8-300x138.png 300w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-8-768x354.png 768w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-8.png 1235w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><figcaption><strong>[Figure 8] Extraction of important objects using SOD model <span class=\"has-inline-color has-vivid-cyan-blue-color\">[i]<\/span><\/strong><\/figcaption><\/figure><\/div>\n\n\n\n<p style=\"font-size:18px\">Representative SOD models<br> \u2013 BASNet, U<sup>2<\/sup>Net<\/p>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>5. De-Identification<\/strong><\/h4>\n\n\n\n<p style=\"font-size:18px\">When constructing a dataset for artificial intelligence, one of the concerns is whether personal information is included in the data.&nbsp;Autonomous driving-related datasets that include a lot of people and vehicles should pay particular attention to de-identification of personal information.&nbsp;This is because, if an issue related to personal information arises, not only the data but also the model trained using the data may become liable.<\/p>\n\n\n\n<p style=\"font-size:18px\">Testworks uses an automation model to detect a person&#8217;s face and license plate, and de-identifies personal information by blurring the identified objects.<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"408\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-9-1024x408.png\" alt=\"\" class=\"wp-image-1629\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-9-1024x408.png 1024w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-9-300x120.png 300w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-9-768x306.png 768w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-9.png 1235w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><figcaption><strong>[Figure 9] Face and license plate de-identification processing<\/strong><\/figcaption><\/figure><\/div>\n\n\n\n<p style=\"font-size:18px\">The current de-identification method is either blurring or pixelating the target which changes the pixels so that humans cannot recognize them. However, these methods have disadvantages in some cases when the anonymized objects are needed for training. Thus, if these objects are altered, the data cannot be used at all.<\/p>\n\n\n\n<p style=\"font-size:18px\">To compensate for these shortcomings, Testworks is researching a de-identification solution that can be used for these models without damaging the image, using GAN with the goal of making it available in the first half of next year (2022).&nbsp;(&nbsp;Refer to the&nbsp;blog post&nbsp;<a rel=\"noreferrer noopener\" href=\"https:\/\/blog.testworks.co.kr\/portrait-of-the-data-age_how-much-is-your-face\/\" target=\"_blank\">\u201cPortrait of the Data Age, Advances in De-identification Technology<\/a>&nbsp;\u201d&nbsp;<span class=\"has-inline-color has-vivid-cyan-blue-color\"><strong>[ii]<\/strong><\/span>&nbsp;)<\/p>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>Model performance comparison example<\/strong><\/h4>\n\n\n\n<p style=\"font-size:18px\">Regarding model performance, we train our deep learning models using data collected and processed from several projects after they are bootstrapped using existing weights. This process is known as &#8216;transfer learning&#8217;. Due to this, our model shows better performance than other public models.&nbsp;The table below shows performance comparison \u200b\u200bfor major classes of Faster RCNN, a benchmark model for object detection.<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"249\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-10-1-1024x249.png\" alt=\"\" class=\"wp-image-1639\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-10-1-1024x249.png 1024w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-10-1-300x73.png 300w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-10-1-768x187.png 768w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-10-1.png 1235w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><figcaption><strong>[Figure 10] Example of model performance comparison<\/strong><\/figcaption><\/figure><\/div>\n\n\n\n<div style=\"height:50px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>Effect of applying automation model<\/strong><\/h4>\n\n\n\n<p style=\"font-size:18px\">By applying the automation model, it is possible to save about 53% of resources compared to full-scale manual labeling.<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"249\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-11-1-1024x249.png\" alt=\"\" class=\"wp-image-1640\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-11-1-1024x249.png 1024w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-11-1-300x73.png 300w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-11-1-768x187.png 768w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Automation-in-AI-Data-Processing_Figure-11-1.png 1235w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><figcaption><strong>[Figure 11] Effect of using the automation model on NIA &#8216;Traffic Safety AI Data&#8217; annotation project.<\/strong><\/figcaption><\/figure><\/div>\n\n\n\n<div style=\"height:100px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h2 class=\"wp-block-heading\"><strong>Conclusion<\/strong><\/h2>\n\n\n\n<p style=\"font-size:18px\">There is a saying in statistics, &#8220;If you put garbage in, garbage comes out&#8221;.&nbsp;This means that if you use bad data, the model will not perform well either.&nbsp;Although AI models are improved to some extent as they develop, a good dataset is still an essential element for creating high-performance deep learning models.&nbsp;To create a large amount of high-quality data, an auto labeling automation model is essential, and with the development of deep learning models, auto-labeling will become easier and easier.<\/p>\n\n\n\n<p style=\"font-size:18px\">Recently, research is ongoing not only on automating annotation but also on the review process.&nbsp;For instance, you can verify annotation results by cross validating the predictions made by multiple models on the same dataset. Another active area of research is active learning where the confidence value of the object recognition model during prediction is used to notify the user if it is below the threshold value so that only the needed data are reviewed and re-labeled. All such development shows that automation in AI data processing is far from done and we need to continue to keep abreast of the latest research development.<\/p>\n\n\n\n<div style=\"height:100px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<hr class=\"wp-block-separator is-style-wide\"\/>\n\n\n\n<p style=\"font-size:18px\">[i]&nbsp;Image excerpt from article U2-Net: Going Deeper with Nested U-Structure for Salient Object Detection<\/p>\n\n\n\n<p style=\"font-size:18px\">[ii]&nbsp;<a rel=\"noreferrer noopener\" href=\"https:\/\/blog.testworks.co.kr\/portrait-of-the-data-age_how-much-is-your-face\/\" target=\"_blank\">Portrait of the data age, development of de-identification technology, Testworks Hyungbok Kim<\/a><\/p>\n\n\n\n<hr class=\"wp-block-separator is-style-wide\"\/>\n\n\n\n<div style=\"height:100px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-media-text alignwide is-stacked-on-mobile\" style=\"grid-template-columns:15% auto\"><figure class=\"wp-block-media-text__media\"><img loading=\"lazy\" decoding=\"async\" width=\"648\" height=\"649\" src=\"https:\/\/blog.testworks.co.kr\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Im-Jeong-hyeon-1.png\" alt=\"\" class=\"wp-image-1633 size-full\" srcset=\"https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Im-Jeong-hyeon-1.png 648w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Im-Jeong-hyeon-1-300x300.png 300w, https:\/\/blog.aiworkx.ai\/en\/wp-content\/uploads\/sites\/3\/2021\/09\/Im-Jeong-hyeon-1-150x150.png 150w\" sizes=\"auto, (max-width: 648px) 100vw, 648px\" \/><\/figure><div class=\"wp-block-media-text__content\">\n<p><strong>Jeonghyun Im<\/strong><\/p>\n\n\n\n<p><strong>Senior Researcher, AI Model Development Team<\/strong><\/p>\n\n\n\n<p>Chosun University, B.S. in Computer Engineering,<\/p>\n\n\n\n<p>(Former) Mobile Reader Senior Researcher, Participated in Samsung Keis development<\/p>\n\n\n\n<p>After developing Windows applications for more than 15 years, he changed his field of interest to artificial intelligence and started developing a geological analysis system using a deep learning model.<\/p>\n<\/div><\/div>\n\n\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Introducing a machining automation model to improve the quality of data and the performance of deep learning models<\/p>\n","protected":false},"author":1,"featured_media":1636,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[4],"tags":[23,428,440,68,69,433,93,377,445,442,444,441,138,381],"class_list":["post-1613","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-insight","tag-ai-data-2","tag-ai-model-development","tag-ai-technology","tag-automation","tag-automation-model-2","tag-data-processing","tag-de-identification","tag-deep-learning","tag-jeonghyeon-im","tag-object-recognition","tag-save-resource","tag-semi-auto-polygon","tag-testworks","tag-vision-keypoint"],"_links":{"self":[{"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/posts\/1613","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/comments?post=1613"}],"version-history":[{"count":20,"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/posts\/1613\/revisions"}],"predecessor-version":[{"id":1831,"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/posts\/1613\/revisions\/1831"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/media\/1636"}],"wp:attachment":[{"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/media?parent=1613"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/categories?post=1613"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blog.aiworkx.ai\/en\/wp-json\/wp\/v2\/tags?post=1613"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}