[go: up one dir, main page]

Next Article in Journal
Complex Field Network Coding for Multi-Source Multi-Relay Single-Destination UAV Cooperative Surveillance Networks
Next Article in Special Issue
Real-Time Monitoring System for Shelf Life Estimation of Fruit and Vegetables
Previous Article in Journal
Refining Network Lifetime of Wireless Sensor Network Using Energy-Efficient Clustering and DRL-Based Sleep Scheduling
Previous Article in Special Issue
Data-Driven Calibration of Soil Moisture Sensor Considering Impacts of Temperature: A Case Study on FDR Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Analysis of the Functionality of the Feed Chain in Olive Pitting, Slicing and Stuffing Machines by IoT, Computer Vision and Neural Network Diagnosis

by
Alberto Lucas Pascual
1,
Antonio Madueño Luna
2,
Manuel de Jódar Lázaro
1,*,
José Miguel Molina Martínez
1,
Antonio Ruiz Canales
3,
José Miguel Madueño Luna
4 and
Meritxell Justicia Segovia
3
1
Food Engineering Department, Technical University of Cartagena, 30203 Cartagena, Spain
2
Aerospace Engineering and Fluid Mechanical Department, University of Seville, 41013 Seville, Spain
3
Engineering Department, Miguel Hernández University of Elche, 03312 Orihuela, Spain
4
Graphics Engineering Department, University of Seville, 41013 Seville, Spain
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(5), 1541; https://doi.org/10.3390/s20051541
Submission received: 15 January 2020 / Revised: 2 March 2020 / Accepted: 6 March 2020 / Published: 10 March 2020
(This article belongs to the Special Issue IoT Technologies and the Agricultural Value Chain)
Figure 1
<p>Example of the use of region of interest (ROI) in an olive in a normal position. The red point corresponds to the origin of coordinates for pixels. The green rectangle comprises the pocket within its walls and depth and limited by the previous and the next pocket of the feed chain. The red rectangle is the ROI.</p> ">
Figure 2
<p>Example of the use of ROI in an olive in a “boat” position between [80°, 90°] and [−80°, −90°].</p> ">
Figure 3
<p>Selected ROI.</p> ">
Figure 4
<p>Punch needles in an olive pitting machine.</p> ">
Figure 5
<p>Magnetic sensor used to detect the passage of the pockets in the chain.</p> ">
Figure 6
<p>Electronic circuit of the external trigger and LED.</p> ">
Figure 7
<p>Electronic circuit of the PC with CM1K chip.</p> ">
Figure 8
<p>Qt Creator application for IoT management of olive pitting, slicing and stuffing (DRR) machines.</p> ">
Figure 9
<p>The system implemented on a “DRR machine”.</p> ">
Figure 10
<p>PC-CM1K communication system.</p> ">
Figure 11
<p>IoT control system.</p> ">
Figure 12
<p>Confusion matrix using 10 × 10 resolution images.</p> ">
Figure 13
<p>Matrix of the obtained results using MATLAB (11 × 11 pixels resolution).</p> ">
Figure 14
<p>Matrix of the obtained results using MATLAB (16 × 16 pixels resolution).</p> ">
Figure 15
<p>Graphical User Interface (GUI) created in Matlab to analyze data from the “DRR machines”.</p> ">
Figure 16
<p>Settings dialogue box of the GUI from Matlab.</p> ">
Figure 17
<p>GUI of “boat” olives and doubles for the “DRR machine” diagnosis.</p> ">
Figure 18
<p>Multiple accumulated values (boats, empty, doubles and small parts).</p> ">
Figure 19
<p>Instantaneous values for boats, doubles, small parts and empty.</p> ">
Figure 20
<p>Instantaneous values of speed and production.</p> ">
Versions Notes

Abstract

:
Olive pitting, slicing and stuffing machines (DRR in Spanish) are characterized by the fact that their optimal functioning is based on appropriate adjustments. Traditional systems are not completely reliable because their minimum error rate is 1–2%, which can result in fruit loss, since the pitting process is not infallible, and food safety issues can arise. Such minimum errors are impossible to remove through mechanical adjustments. In order to achieve this objective, an innovative solution must be provided in order to remove errors at operating speed rates over 2500 olives/min. This work analyzes the appropriate placement of olives in the pockets of the feed chain by using the following items: (1) An IoT System to control the DRR machine and the data analysis. (2) A computer vision system with an external shot camera and a LED lighting system, which takes a picture of every pocket passing in front of the camera. (3) A chip with a neural network for classification that, once trained, classifies between four possible pocket cases: empty, normal, incorrectly de-stoned olives at any angles (also known as a “boat”), and an anomalous case (foreign elements such as leafs, small branches or stones, two olives or small parts of olives in the same pocket). The main objective of this paper is to illustrate how with the use of a system based on IoT and a physical chip (NeuroMem CM1K, General Vision Inc.) with neural networks for sorting purposes, it is possible to optimize the functionality of this type of machine by remotely analyzing the data obtained. The use of classifying hardware allows it to work at the nominal operating speed for these machines. This would be limited if other classifying techniques based on software were used.

1. Introduction

Olives were pitted and stuffed by hand until the 1970s. As labor costs gradually increased, the need to mechanize both processes arose. The industrialization of the olive industry progressed at a slow pace, and the first continuous olive pitting machines appeared in 1975 [1]. Current models have improved and are able to pit olives at a rate up to 2500 olives/min. However, there has been a lack of electronic improvements because the ongoing mechanical improvements in olive pitting machines have limited the incorporation of new technologies and fine adjustments. It is worth mentioning however some of the few advances in this area, including performance optimization and remote error detection [2]. The olive pitting minimum error rate of traditional systems is 1–2% (if the machine is correctly adjusted) and cannot be improved through mechanical adjustments although this error can increase due to the poor calibration of an olive pitting machine. The consequences of these error increases include a greater fruit loss and food safety issues, because the pitting process is not completely ensured, it may result in a low-quality product. The growing competitiveness in producer countries, due to a globalized market, and the increase of food safety measures have led to new technological solutions. These solutions are aimed at improving the reliability of olive pitting machines during the pitting process by removing the stone completely and increasing their productivity for better competition with other producers.
As a biological product characterized by a heterogeneous morphology, olives challenge detection technologies. Therefore, neural networks are the tools chosen to identify every possible case. Cases using deep learning and computer vision have been confirmed to deliver great results and are promising solutions for solving the aforementioned problem. Yunchao et al. [3] presented an article of interest on the use of real-time detection in recycled aggregate concrete technology. However, for this particular case, it does not prioritize the processing speed. Therefore, this technology does not have the same applicability in the present article. Another study that could be considered for determining an applicable solution is the analysis developed by Mingyou et al. [4], who used a multi-camera system, which is interesting for the analysis of several pitting machines or increase the perception range of vision. However, the use of several cameras is not the subject of this study. Therefore, it is necessary to explore other solutions that provide sufficient processing speed to analyze images at the speed prescribed by the pitting machines. Although these studies use interesting technology for the detection and classification of images, their results are not applicable here because processing speed is not relevant.
There are applications which use computer visions with rapid detection technologies for industrial purposes like product classification or image analysis. These systems use morphological operations to extract the sought characteristics in the image through filters and specific analysis algorithms like Canny Edge, Otsu, local binary pattern algorithm, K-means, etc. In order to develop these algorithms, specific software is used, such as the integrated development setting of Microsoft Visual Studio and specific computer vision libraries, as in the case with Opencv. Maoyon, et al. [5] proposes a study using these types of techniques in the classification of apples on an industrial level. Although the described procedures are of great interest, these types of techniques require high-performance computing hardware and elevated costs for the type of application evaluated in the current analysis, where more economic solutions and reduced hardware size are sought. Lucas et al. [6] propose a specific example done with olives which uses these fast detection techniques based on C++ programming language and OpenCV artificial vision libraries.
Specific studies in the agri-food sector and the related behavior of neural networks have been considered. It is worth mentioning the studies conducted by Guichao et al. [7,8,9] on detection systems for the artificial visualization of fruits through RGB space color analysis and robotic systems. In the case of the classification of apples, Yang et al. [10] provide positive results. Yang’s investigation used three layers of 9-6-3 neurons, with 96.6% accuracy. Nagata et al. [11] developed a grading system for fruit and vegetables using neural network technologies, obtaining a high percentage of accuracy for strawberries and green peppers (94% to 98% and 89%, respectively). Likewise, Behrooz et al. [12] applied machine vision and artificial neural network (ANN) for modelling and controlling the grape drying process, offering a new method for predictive modelling of the grape drying process for the on-line monitoring and controlling of the process.
There have been some techniques used to classify table olives using computer vision and neural networks. Gatika et al. [13] proposed an interesting paper about olive fruit recognition using neural networks. Olive fruit recognition is performed by analyzing RGB images taken from olive trees. Another study used a neural network based on the backpropagation method. Mancuso et al. [14] identified olive (Olea europaea) cultivars using artificial neural networks. Backpropagation neural networks (BPNNs) were used to distinguish 10 olive (Olea europaea L.) cultivars that originate throughout the Mediterranean basin. The book Computer Vision Technology for Food Quality Evaluation, 2nd Edition [15], includes a specific chapter based on the different technologies use to analyze the quality of food. Chapters 11 to 13 describe the quality evaluations of apples, citrus fruits, and strawberries, respectively. Of particular interest to the problem at hand, Chapter 14 [16] explains the classification and evaluation of table olives and describes how to classify them by color, shape, or external defects made by insects. The usual way to classify olives is by using computer vision. That paper analyzes the images captured by a camera connected to a PC; these images allow the analysis of 66 olives per matrix. In that paper, the author used a Bayesian math model for pre-classification to perform this process. Neural network software was used, with 15 sorting parameters and a hidden layer. The result was successful. The network was able to classify more than four types of olives. The results, however, could be improved by using high-resolution images.
Previous examples employed software-based neural networks, where the processing speed is not relevant. Thus, although the detection and classification technology in these studies is applicable, it is necessary to look for technology that achieves high processing speeds. In addition, it is preferable to use systems with limited physical implementation that can be installed with low intrusion at lower costs, instead of traditional hardware, such as industrial PC’s. Therefore, neural network technology based on a physical chipset ensures its successful implementation and low cost.
The possible applications of this type of chip are very broad. The following article is interesting because CM1K neural chip of General Vision Inc. (Petaluma, CA 94952 USA) was implemented in a common industrial process used to fill bottles in factories. This process is quite simple; however, it may be necessary to use an intelligent device to inspect the process and ensure food security [17]. In another example, parallel neural network chips were used offshore for fish inspection before filleting [18]. Each network chip system uses four neural network chips (accounting for 312 neurons) based on a natively parallel, hard-wired architecture that performs real-time learning and nonlinear classification (RBF). The use of CM1K in real time verifies its use for the purposes of the present study.
The combination of computer vision and neural networks offers a way to perform tasks that could be more complex. Liu et al. [19] proposed a neural network chip for license plate recognition. This chip combines a video image-processing module with a neural network module by using equalized image processing algorithms and network classification algorithms. Santu Sardar et al. [20] published a paper based on automated facial recognition, which is a technique employed in a wide range of practical applications, including personnel access control and identification systems. Image recognition is simpler than image processing methods for facial recognition, mainly due to the lack of a fixed pattern for comparison purposes; these applications reveal the tremendous possibilities of CM1K.
Later CM1K chipsets have delivered satisfactory results. An interesting aspect to consider in future research would be to comparison our chosen chip with these other technologies described in the following studies [21,22,23,24,25,26,27,28,29,30].
The use of the IoT is widely extended nowadays and there are numerous cases, for example [31,32,33,34,35] in precision agriculture, irrigation, temperature control, monitoring of the agricultural production process or automated-olive-chain. In the case presented in this paper, we are going to be able to monitor and analyze the operation of DRR machines through the internet.

2. Materials and Methods

In this section, the internal structure of the CM1K chip (Section 2.1 and Section 2.2) and the implementation of Matlab or neural networks with characteristics similar to those of the CM1K chip are analyzed (Section 2.3).
A chip such as the CM1K has a limitation in that the entry vector it can support is reduced in this case (256 bytes), meaning that if it is going to process an image, that image cannot be larger in pixel size than this value (256 bytes) and that it could only be B/W with 256 levels of gray. In order to assure that a neural network with these characteristics (equipped with high speed as it is pure hardware, but of limited vector entry) can be acceptable for the purpose of classification in olive pitting machines, a neural network is simulated in Matlab which works with a delay and allows us to assure that correct classification will be carried out in various categories with images of just 16 × 16 pixels in B/W.
In Section 2.4 and Section 2.5 the physical implementation of the classification system is shown with a real olive pitting machine and in Section 2.6 the commercial version of the CM1K chip that has been used in the trials is analyzed. For these sections, the description supplied in [36] will be used by the authors of this paper.
In Section 2.7 the communication hardware of the CM1K-PC for classification in real time is described and lastly, in Section 2.8, the IoT system to control the olive pitting, slicing and stuffing machine (DRR in Spanish) and data analysis is analyzed.

2.1. Neural Network

In this work a supervised neural network with backpropagation has been chosen because there are no relevant differences with respect to the calculation processes used by other networks. Moreover, this system is one of the most widespread systems used for image identification purposes, and it is widely used in the MATLAB Neural Network Toolbox employed in this paper. Backpropagation is the most efficient way to set this value [37]. First, errors are calculated in output units by considering the difference between the desired and predetermined values. Next, they propagate through the network using the weights and obtain the minimum value in the most optimal way.
The software emulation of a neural network is computationally demanding, which indicates that its operations will take a long period of time to be carried out. The use of a chip that physically implements a neural network (as in this paper with the CM1K chip or Intel Curie chip) will speed up the training and response processes of the neural network. A comparison with other existing hardware neural networks like Google Coral Edge TPU [38], Intel® Movidius™ Neural Computer Stick 2 [39] or Nvidia-Jetson-Nano [40] and optimized results will be provided in future papers.

2.2. Operation of the CM1K Chip

The CM1K is a pattern recognition accelerator chip that is trainable in real-time by learning examples [41]. It is a fully parallel silicon neural network for either learning or recognition [42], that can store and process information simultaneously. It is composed of four modules [43], has two possible classifiers: K-nearest neighbor (KNN) or radial basis function (RBF) [44] and uses the “Winner-Takes-All” strategy [45]. It has 1024 neurons of 256 bytes. Intel Curie chip is a diminished version with only 128 neurons of 128 bytes.

2.3. The MATLAB Neural Network Toolbox

In this work we are using the MATLAB Neural Network [46] by using a specific network called ‘autoencoder’ for classification purposes [47,48], it will allow us to analyze the effect of image resolution on the sorting capacity of the network.

2.3.1. Preliminary Tests: Maximum Resolution Available

The maximum resolution implemented under camera board is 144 × 176 meeting the size requirements of the pocket according to used lenses. After the first test using the above-mentioned resolution, a 25,344 pixels at 1-byte color depth grayscale image is obtained, which implies a high processing rate (nearly two minutes using a 2.5 GHz CPU, RAM 8 GB, Microsoft Windows 10, Redmond, Washington, USA, and MathWorks MATLAB 2019, Massachusetts, USA).
Preliminary tests suggest that processing optimization measures need to be carried out. These measures will focus on two main aspects:
  • Establishing a region of interest (ROI) on the image.
  • Testing different resolutions that allow the identification of the image at the minimum processing rate.
In addition to these aspects, it is important to consider the processing rate of the physical chips: Intel Curie (128 neurons with a 128-byte input vector) and NeuroMem CM1K (1024 neurons with a 256-byte input vector).
Considering the above-mentioned restrictions, the maximum resolution of the image must be between 128 and 256 pixels for a square image, as finally used: 16x16 pixels for NeuroMem CM1K 256-byte chip and 11 × 11 pixels (121 bytes < 128 bytes) for Intel Curie chip.

2.3.2. Preliminary Tests: Minimum Resolution

Several tests to identify the minimum resolution accepted by physical chips are carried out with the purpose of setting the lowest processing rate. To do so, real images are to be processed in MATLAB and it should be estimated if the system is able to identify them.
The first step is to set the ROI, which must be a part of the image with enough information in order to identify the position of the olive in the pocket. Therefore, the olive must be completely displayed on the image.
The following Figure 1 and Figure 2 show some examples of 176 × 144 pixel-images with the (X,Y) reference system. As shown in the figures, the olive is restricted in the X-axis (the same pocket confines the olive to its space) and, therefore, the selection of the parameters of ROI in that axis is constant. There is a possibility for the olive to move in Y-axis, consequently leading to a deviation of the dimensions of ROI. However, it is empirically proven that, due to the inclination of the chain and the upward movement of the olive, gravity, inertia and some elements of the machine (brushes, for example) forces the olive to stay in a similar position. Variations may be identified in orientation, which is the subject of this paper.
After previous considerations, the following conditions are set out (measured in pixels and referred to the origin of coordinates, Figure 3):
Once the ROI parameters in X and Y are set, several tests are carried out in order to establish if the system is still able to identify images using the selected resolutions. A set of 10,000 images of 176 × 144 pixels in B/W of the olives in normal position and boat and empty buckets are available, there are no anomalous cases (i.e., leaves, small branches, stones, double olives, small parts) in this set. This set is available to other researchers [49]. With an application in Matlab, each image is binarized and the orientation of each olive is calculated. For the tests described in this work, they are randomly selected in groups of 300 olives. The orientation of each olive (normal, intermediate position or boat) is defined by the position relative to the punches (Figure 4).
The classification by angles is as follows (see Table 1):
  • [0°, 10°] and [0°, −10°] (Normal)
  • [10°, 20°] and [−10°, −20°] (Intermediate)
  • [20°, −30°] and [−20°, −30°] (Intermediate)
  • [30°, −40°] and [−30°, −40°] (Intermediate)
  • [40°, −50°] and [−40°, −50°] (Intermediate)
  • [50°, −60°] and [−50°, −60°] (Intermediate)
  • [60°, −70°] and [−60°, −70°] (Intermediate)
  • [70°, −80°] and [−70°, −80°] (Intermediate)
  • [80°, −90°] and [−80°, −90°] (Boat)

2.4. Hardware Used in Image Capture

The system developed has two functions:
  • Obtain images for deferred analysis with Matlab, the Intel Curie and CM1K neural chips to evaluate the operation of the latter.
  • Characterize the real-time operation through IoT of the pitting machines of an olive factory that reach speeds of up to 2500 olives/min.
The system consists of two parts:
  • The hardware that lets the camera [50] control and lighting based on a magnetic sensor that allows to obtain images of each bucket facing the camera (Figure 5 and Figure 6).
  • A personal computer (PC) which allows, for the first case, the storage of images for deferred analysis and in the second case together with a CM1K chip for real-time analysis of the same over the internet with an application developed in Qt Creator [51] (Figure 7).

2.5. Application for IoT Management of DRR Machines

As we said, for the remote control of the DRR machines, an application in Qt Creator (Figure 8) has been developed which allows the most important operating parameters to be seen every 1000 images. This data is sent via Dropbox [52] to a folder which can later be consulted to analyze the operation of the machine (see Section 3.3).
The application previously determines the size in pixels of the olive to generate two cases not contemplated by the neuronal network: small parts and doubles, these two cases are not sent to the neural network. Figure 9 shows the system (Industrial PC, “DRR machine”, camera and led) in a table olive factory.

2.6. Neural Chips Used

As mentioned before in Section 2.2., two neuromorphic chips have been used in this work. The first one is an Intel Curie from an Arduino-Genuino 101, which includes an ANN module (128 neurons with a 128-byte input vector). This module is a scale case of the second chip CogniMem CM1K with 1024 neurons and an input vector of 256 bytes, and it has been used in a BrainCard [53,54]. Both chips use RBF or k-NN techniques for sorting purposes.
121 pixels images (11 × 11) will be processed with the Intel Curie Arduino 101 and with the CM1K chip. 256 pixels images (16 × 16) only with the CM1K chip. We have used three classification options:
  • First, there is a complex and thorough process on both chips by which olives are classified according to existing angles, see previous Section 2.3.2.
  • Later, an intermediate position is set to simplify the rank of angles and the following positions are only taken into account: normal (interval [0°, −10°] and [0°, 10°]), “boat” ([−80°, −90°] and [80°, 90°]), intermediate (from 10° to 80° and −10° to −80°) and empty pocket.
  • Finally, this classification is simplified even more using only three out of four pocket cases: normal, “boat” and empty.
For training purposes, a maximum of 300 olives per case has been taken into account as an empirical value to prevent overtraining, which may lead the system to detect even the most trifling detail and, therefore, an unnecessary overuse of neurons. Every case has been tested using 300 random olives replicated 10 times, which accounts for 3000 per every case.

2.7. CM1K-PC Hardware Communication for Real Time Classification

One of the main problems found in the implementation of this neural chip for use in olive pitting machines, which reach speeds of up to 2500 olives/min, is the speed of communication between the chip and the external PC supplying the image to be classified. The problem arises because this chip uses I2C [55] communication and requires an intermediate interface (a UART-USB microcontroller with a FT232RL [56] or a CH340G [57] converter) which transforms the I2C communication into serial. Normally, this communication reaches a maximum speed of 115,200 bauds with the PC, Figure 10. Conversely, if the information is transmitted in 1 byte packets (default case), the communication is slowed because the PC-microcontroller combination has to restart the process of serial communication every time 1 byte is sent.
To eliminate this problem, a microcontroller has been used which directly includes direct conversion to USB generating a “virtual port”. Specifically, these two options have been analyzed:
(1)
The INTEL CURIE [58] (which internally incorporates a limited version of the CM1K chip with 128 neurons with a 128 byte vector) but which also can be used just for a part of the communication with the PC At the present time, it is no longer produced.
(2)
A TEENSY 4.0 [59] which includes a ARM Cortex-M7 (NXP iMXRT1062) operating at a speed of 600 MHz.
Conversely, to avoid restarting the communication series for each byte, complete 256 byte packets have been sent with each image (16 × 16 bytes) which allow them to arrive to the CM1K in just one delivery. Each delivery, is made up of 258 bytes: 1 byte as the header (to indicate the type of instruction, for example training the neural network or challenging), 256 bytes of data (with a 16 × 16 image) and a last byte, in case the of a challenge, includes an answer from the neural network (olive in a normal position, boat position, empty, anomaly, etc.).
In Table 2, the differences in the times expressed in ms during the communication with a UART-USB, an INTEL CURIE and a TEENSY 4.0 standard converter are shown. The average value is of 1000 packages of 258 × 1 byte in the first row and in the second a single package of 258 bytes.
If we consider that the reception process for the information of the byte classified has an unappreciable duration, we find that the delivery of 1 × 258 byte in the case of CH340G, CURIE and TEENSY show processing rates of 2.72, 3.87 and 6.56 images per second respectively or 163.2, 232.2 and 393.6 olives/min, clearly insufficient for the standard speed of the olive pitting machines which can reach 2500 olives/min.
In the case of the delivery of 258 bytes with a CH304 chip, it would reach the level of 44.6 images per second (2677.3 olives/min), but we consider that any load process in the PC could generate a loss of processed images, and for that reason we will only use the cases of delivery of a single data package of 258 bytes with CURIE or TEENSY.

2.8. IoT System to Control the DRR Machine and Data Analysis

An IoT system has been implemented for the analysis and remote control consistent with a client connected to an olive pitting machine which includes: a PC with Windows 10 and specific hardware which has an air conditioning system, the magnetic sensor, camera, LED lighting and chip CM1K (Figure 11).
The system is connected to the internet and every 1000 images generates information via an application (see Section 2.5) which is stored in a file and is sent by Dropbox, so that it can be consulted at any time from any other PC Another specific application has been developed to allow the analysis of this data and with it, the evolution of the operation of the olive pitting machine (see Section 3.3). In order to control the system remotely, a remote desktop from Google [60] is used.

3. Results

3.1. Results Obtained Using a MATLAB Neural Network

The structure of the autoencoder of the neural network has been trained with 45 iterations, fixed by the MATLAB neural network library. Three settings have been tested: The first one uses 1-byte depth greyscale images, 10 × 10 pixels ROI-scaled, the second one uses 1-byte depth greyscale images, 11 × 11 pixels ROI-scaled and the last one uses 1-byte depth greyscale images, 16 × 16 pixels ROI-scaled.
For training purposes, a set of 9 images of empty pockets, 11 of “boat” olives and 10 of normal olives has been used. Unusual case images were not included since the purpose of the test was to prove the feasibility of the classification of these low-resolution images using a neural network. The results using a resolution of 10x10 pixels show an unsatisfactory outcome, as shown in Figure 12 with 26.7% error rate.
As shown in Figure 13, the neural network has carried out an appropriate classification (<5% error rate) with 11 × 11 pixels images processed. Tests have been repeated with 16 × 16 pixels images (Figure 14), and the results were better (<4% error rate), which means that it is also possible to classify the cases with the images (11 × 11 pixels).

3.2. Results Obtained Using Neuromorphic Chips

The following results are given (Table 3 and Table 4) according to the chip that has been used, the type of classification (degrees, intermediate position and simple) and number of neurons. The total error rate is the average obtained of the ten repetitions.
The optimal option is Braincard 256 chip (0.63% error rate < 1%) and only needs four neurons to classify between normal, empty and boat olives. If we use an Intel curie chip error rate is 4.80% < 5% and only needs four neurons too. This means that the suggested system is capable of detecting up to a 95–99% of the deficiencies (boats and empty) presented by the machine.

3.3. Analysis of the Results of the Operation of the DRR Machine

An application (“LoggerDRR”) has been developed to analyze the operation of the olive pitting machine with GUI by Matlab (Figure 15). The most relevant parameters of this application will be indicated below:
The representative data is indicated in the following box (Figure 16:
  • Speed: Indicated in pockets per minute
  • Production: Real value in olives per minute (without empty pocket)
  • “Boat” olives, normal, double olives, empty, small pieces, anomalies: Percentage of the total production.
  • Accumulated values: Data added in the selected period.
In Figure 17, a practical example of use is shown. There are two accumulated graphs, the olive green one corresponding to “boat” olives detected and the ochre color corresponding to double olives entering in the pocket. Each data update point corresponds to a count of 1000 olives.
The spaces correspond to periods when the machine was stopped (1), this can be due to changes in the feed chain (change of container where the olives come from) or stops to adjust the machine or personnel breaks.
In (2), the machine is observed to have a misalignment (due to the misalignment of the brush which straightens the olives in the feeding chain), followed by a stop for recalibration (3). Immediately after that, we can observe how the slope decreases (4)→(5) which means the number of “boat” olives has decreased. Additionally, the number of double olives (various in the same pocket) increases (6), meaning that the recalibration has not been optimal and the feed plate has been poorly adjusted.
Figure 18 shows an example of various combined parameters (boat accumulates, empty accumulated, doubles accumulated and small parts accumulated).
This type of graph allows us to observe if the calibration process entails a comprehensive improvement in performance or, conversely, whether an improvement in a parameter after an adjustment entails a reduction in goodness of fit in another. In (1), (2) and (3) it can be observed how the feeding plate goes out of adjustment a little more each time, which has an impact on the number of doubles and small parts. During all of this time, the growth in the number of “boat” olives is monotonous which means that there are no misalignments in the straightening brush. For its part, the number of empty (4) increases abruptly each time a lack of product in the feeder when the DRR machine is working. At times, the data analysis without accumulating data can be interesting in order to see the effect of specific actions. In Figure 19, the peaks in the “boat” olives/doubles (1), of only doubles (2), empty (3) are observed due to the action of isolated adjustments or operation of the machine. In all of the cases shown, these actions are unfavorable (either calibration or operation).
Finally, Figure 20 shows instances in which the performance decreases when working with a machine with a poorly adjusted feeding plate and straightening brush (1), only with the feed plate misaligned (2) or due to the lack of product in the feeder (3).

4. Conclusions

With the current paper, an IoT control system for a DRR machine has been developed which allows its operation to be improved:
An IoT system has been developed which allows for the reception of useful operational information for the DRR machine remotely in order to do an operational diagnosis based on this information, and, in this case, carry out calibration operations. This IoT system allows for configuration of the equipment via a remote desktop.
In the second place, the classification system has been made to work in an industrial setting at a nominal speed for the DRR machine (up to 2500 olives per minute) through the use of a PC-CM1K connection which uses a microcontroller with an Intel Curie or Teensy “virtual port” as an interface.
In the third place, the training and testing of a neural network based on a physical chip to classify olives in the feed chain of a pitting machine has been successfully completed. It has been proven that the use of Intel Curie or CM1K chips are suitable for simple sorting (normal, empty, boats), with error rates of less than 5% with the Intel Curie chip and less than 1% with the CM1K chip.
An application developed in Qt Creator has been used that previously detects for its study the cases with small parts of olives or double olives present in the pocket and separates them from those that are sent to the neural chip that correspond to normal, empty pockets and boats.
One limitation of the system currently developed is that it only allows for the diagnosis of the DRR machine so that the operator can recalibrate it. In a future paper, the possibility of ejecting those olives which are badly positioned (“boat” olives or doubles) or which have defects (small pieces) in the feeding chain will be presented. In this way, “boat” olives or doubles being badly de-stoned will be avoided. Likewise, it will assure that the olives which break upon entering in the feed chain causing small pieces do not get to the de-stoning area, thus avoiding, on the one hand, the loss of product (1–2%), and on the other hand, the accelerated deterioration of the punch needles and hats for deboning.

Author Contributions

Conceptualization, A.M.L.; validation, A.M.L., J.M.M.M., A.R.C., M.d.J.L., J.M.M.L., M.J.S. and A.L.P.; supervision: A.M.L., J.M.M.M. and A.R.C.; investigation, writing—original draft preparation, writting—review and editing, conceptualization, software, formal analysis, resources, data curation, visualization, project administration: A.L.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Santos, F.J. Siles New technologies in table olive processing. Grasas Aceites 1999, 502, 131–140. [Google Scholar] [CrossRef]
  2. Madueño, A.; Lineros, M.; Madueño, J. System and Procedure Based on a Synchronism Sensor for the Detection of Malfunctions in Pitting Machines Olive and Filling Machines, Quantification and Optimization of Performance, Signaling, Monitoring and Remote Control. ES2529816A2. Available online: https://patents.google.com/patent/ES2529816A2/en (accessed on 2 February 2020).
  3. Tang, Y.; Li, L.; Wang, C.; Chen, M.; Feng, W.; Zou, X.; Huang, K. Real-time detection of surface deformation and strain in recycled aggregate concrete-filled steel tubular columns via four-ocular vision. Robot. Comput. Integr. Manuf. 2019, 59, 36–46. [Google Scholar] [CrossRef]
  4. Chen, M.; Tan, Y.; Zou, X.; Huang, K.; Li, L.; He, Y. High-accuracy multi-camera reconstruction enhanced by adaptive point cloud correction algorithm. Opt. Lasers Eng. 2019, 122, 170–183. [Google Scholar] [CrossRef]
  5. Nie, M.; Zhao, Q.; Xu, Y.; Shen, T. Machine Vision-based Apple External Quality Grading. In Proceedings of the Chinese Control and Decision Conference, Nanchang, China, 3–5 June 2019. [Google Scholar]
  6. Lucas, A.; Madueño, A.; De Jódar, M.; Molina, J.; Ruiz, A. Characterization of the percentage of poorly positioned olives in pitting, rolling and filling machines for table olives (DRR). In Proceedings of the X Congresso Ibérico de Agroengenharia, Huesca, Spain, 3–6 September 2019. [Google Scholar]
  7. Lin, G.; Tang, Y.; Zou, X.; Li, J.; Xiong, J. In-field citrus detection and localisation based on RGB-D image analysis. Biosyst. Eng. 2019, 186, 34–44. [Google Scholar] [CrossRef]
  8. Lin, G.; Tang, Y.; Zou, X.; ·Xiong, J.; Fang, Y. Color‑, depth‑, and shape‑based 3D fruit detection. Prec. Agric. 2020, 21, 1–17. [Google Scholar] [CrossRef]
  9. Lin, G.; Tang, Y.; Zou, X.; Xiong, J.; Li, J. Guava. Detection and Pose Estimation Using a Low-Cost RGB-D Sensor in the Field. Sensors 2019, 19, 428. [Google Scholar] [CrossRef] [Green Version]
  10. Yang, F. Classification of apple surface features using machine vision and neural networks. Comput. Electron. Agric. 1993, 9, 1–12. [Google Scholar] [CrossRef]
  11. Nagata, M.; Bato, P.; Mitaria, M.; Cao, Q.; Kitahara, T. Study on Sorting System for Strawberry Using Machine Vision (Part 1). Jap. Soc. Agric. Mach. 2000, 62, 100–110. [Google Scholar]
  12. Behroozi, N.; Tavakoli, T.; Ghassemian, H.; Hadi, M.; Banakar, A. Applied machine vision and artificial neural network for modeling and controlling of the grape drying process. Comput. Electron. Agric. 2013, 98, 205–213. [Google Scholar] [CrossRef]
  13. Gatica, G.; Bestb, S.; Ceronic, J.; Lefranc, G. Olive Fruits Recognition Using Neural Networks. Proc. Comput. Sci. 2013, 17, 412–419. [Google Scholar] [CrossRef] [Green Version]
  14. Mancuso, S.; Nicese, F.P. Identifying Olive (Olea europaea) Cultivars Using Artificial Neural Networks. Am. Soc. Hortic. Sci. 1999, 124, 527–531. [Google Scholar] [CrossRef] [Green Version]
  15. Sun, D. Computer Vision Technology for Food Quality Evaluation, 2nd ed.; Academic Press: Waltham, MA, USA, 2016; pp. 273–350. [Google Scholar]
  16. Diaz, R. Computer Vision Technology for Food Quality Evaluation, 2nd ed.; Academic Press: Waltham, MA, USA, 2016; pp. 351–367. [Google Scholar]
  17. Bottle Inspection. General Visions. 2013. Available online: https://www.general-vision.com/appnotes/AN_BottleInspection.pdf (accessed on 2 February 2020).
  18. Menendez, A.; Paillet, G. Fish Inspection System Using a Parallel Neural Network Chip and the Image Knowledge Builder Application. AI Mag. 2008, 29, 21. [Google Scholar]
  19. Liu, Y.; Wei, D.; Zhang, N. Vehicle-license-plate recognition based on neural networks. In Proceedings of the IEEE on Information and Automation, Shenzhen, China, 6–8 June 2011. [Google Scholar]
  20. Sardar, S.; Tewari, G.; Babu, K.A. A hardware/software co-design model for face recognition using Cognimem Neural Network chip. In Proceeding of the IEEE on Image Information Processing, Shimla, India, 3–5 November 2011. [Google Scholar]
  21. Davies, M.; Srinivasa, N.; Lin, T.; Chinya, G.; Cao, Y.; Choday, S.; Dimou, G.; Joshi, P.; Imam, N.; Jain, S.; et al. Loihi: A Neuromorphic Manycore Processor with On-Chip Learning. IEEE Micro 2018, 38, 82–99. [Google Scholar] [CrossRef]
  22. Moran, S.; Gaonkar, B.; Whitehead, W.; Wolk, A.; Macyszyn, L.S.; Iyer, S. Deep learning for medical image segmentation—Using the IBM TrueNorth neurosynaptic system. In Proceedings of the SPIE Medical Imaging, Houston, TX, USA, 6 March 2018. [Google Scholar]
  23. Moradi, S.; Qiao, N.; Stefanini, F.; Indiveri, G. A scalable multi-core architecture with heterogeneous memory structures for dynamic neuromorphic asynchronous processors (dynaps). IEEE Trans. Biomed. Circuits Syst. 2018, 12, 106–122. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Frenkel, C.; Lefebvre, M.; Legat, J.; Bol, D. A 0.086-mm² 12.7-pJ/SOP 64k-Synapse 256-Neuron Online-Learning Digital Spiking Neuromorphic Processor in 28-nm CMOS. IEEE Trans. Biomed. Circuits. Syst. 2019, 13, 145–158. [Google Scholar]
  25. Fried, L. Making machine learning arduino compatible: A gaming handheld that runs neural networks-[Resources_Hands On]. IEEE Spectr. 2019, 56, 14–15. [Google Scholar] [CrossRef]
  26. Lobachev, I.; Maleryk, R.; Antoschuk, S.; Filiahin, D.; Lobachev, M. Integration of neural networks into smart sensor networks. In Proceedings of the IEEE Xplore, Kiev, Ukraine, 24–27 May 2018. [Google Scholar]
  27. Mittal, S. A Survey on optimized implementation of deep learning models on the NVIDIA Jetson platform. J. Syst. Arch. 2019, 97, 428–442. [Google Scholar] [CrossRef]
  28. Kim, J. New Neuromorphic AI NM500 and Its ADAS Application. In AETA-2018 Recent Advances in Electrical Engineering and Related Sciences: Theory and Application; Lecture Notes in Electrical, Engineering; Zelinka, I., Brandstetter, P., Trong Dao, T., Hoang Duy, V., Kim, S., Eds.; Springer: Cham, Switzerland, 2019; Volume 554, pp. 3–12. [Google Scholar]
  29. CogniPat SDK for Matlab. General Visions. 2018. Available online: https://www.general-vision.com/download/cp_sdk_ml/ (accessed on 2 February 2020).
  30. NeuroMem USB Dongle. General Visions. 2019. Available online: https://www.general-vision.com/hardware/usbdongle/ (accessed on 2 February 2020).
  31. López Riquelme, J.A.; Soto, F.; Suardíaz, J.; Sánchez, P.; Iborra, A.; Vera, J.A. Wireless Sensor Networks for precision horticulture in Southern Spain. Comput. Electron. Agric. 2019, 68, 25–35. [Google Scholar] [CrossRef]
  32. Garcia, L.; Parra, L.; Jimenez, J.M.; Lloret, J.; Lorenz, P. IoT-Based Smart Irrigation Systems: An Overview on the Recent Trends on Sensors and IoT Systems for Irrigation in Precision Agriculture. Sensors 2020, 20, 1042. [Google Scholar] [CrossRef] [Green Version]
  33. Urbano, O.; Perles, A.; Pedraza, C.; Rubio-Arraez, S.; Castelló, M.L.; Ortola, M.D.; Mercado, R. Cost-Eective Implementation of a Temperature Traceability System Based on Smart RFID Tags and IoT Services. Sensors 2020, 20, 1163. [Google Scholar] [CrossRef] [Green Version]
  34. Escolar Díaz, S.; Carretero Pérez, J.; Calderón Mateos, A.; Marinescu, M.C.; Bergua Guerra, B. A novel methodology for the monitoring of the agricultural production process based on wireless sensor networks. Comput. Electron. Agric. 2011, 76, 252–265. [Google Scholar] [CrossRef]
  35. Automated-olive-chain. The internet of Food & Farm. 2020. Available online: https://www.iof2020.eu/trials/fruits/automated-olive-chain (accessed on 20 December 2019).
  36. De Jodar, M.; Madueño, A.; Lucas, A.; Molina, J.M.; Cánales, A.R.; Madueño, J.M.; Justicia, M.; Baena, M. Deep learning in olive pitting machines by computer visión. Comput. Electron. Agric. 2020, 171, 105304. [Google Scholar] [CrossRef]
  37. Hecht-Nielsen, R. Theory of the Backpropagation Neural Network. In Proceedings of the International 1989 Joint Conference on Neural Networks, Washington, DC, USA, 18–22 June 1989. [Google Scholar]
  38. Google Coral Edge TPU. Google LLC. 2020. Available online: https://coral.ai/docs/accelerator/datasheet/ (accessed on 15 November 2019).
  39. Intel® Movidius™ Neural Computer Stick 2. Intel Corporation. 2020. Available online: https://www.intel.es/content/www/es/es/design/products-and-solutions/boards-kits-and-modules/movidius-neural-compute-stick-2/technical-library.html?grouping=rdc%20Content%20Types&sort=title:asc (accessed on 10 December 2019).
  40. Nvidia-Jetson-Nano. Nvidia Corporation. 2020. Available online: https://www.nvidia.com/es-es/autonomous-machines/embedded-systems/jetson-nano/ (accessed on 15 December 2019).
  41. TM TestNeurons SimpleScript. General Visions. Available online: http://www.general-vision.com/documentation/TM_TestNeurons_SimpleScript.pdf (accessed on 2 February 2020).
  42. TM NeuroMem Technology Reference Guide. General Visions. 2019. Available online: https://www.general-vision.com/documentation/TM_NeuroMem_Technology_Reference_Guide.pdf (accessed on 2 February 2020).
  43. TM_CM1K_Hardware_Manual. General Visions. 2017. Available online: https://www.general-vision.com/documentation/TM_CM1K_Hardware_Manual.pdf (accessed on 2 February 2020).
  44. Halgamuge, S.; Poechmueller, W.; Glesner, M. An Alternative Approach for Generation of Membership Functions and Fuzzy Rules Based on Radial and Cubic Basis Function Networks. Int. J. Approx. Reason. 1995, 12, 279–298. [Google Scholar] [CrossRef] [Green Version]
  45. DS_CM1K. General Visions. 2014. Available online: https://www.general-vision.com/datasheet/DS_CM1K.pdf (accessed on 2 February 2020).
  46. Neural-network. The MathWorks, Inc. 1994–2017. Available online: https://es.mathworks.com/solutions/deep-learning/convolutional-neural-network.html?s_tid=srchtitle (accessed on 2 February 2020).
  47. Train Autoencoder. The MathWorks, Inc. 1994–2017. Available online: http://es.mathworks.com/help/nnet/ref/trainautoencoder.html (accessed on 2 February 2020).
  48. Train Stacked Autoencoders for Image Classification. The MathWorks Inc. 1994–2019. Available online: https://es.mathworks.com/help/deeplearning/examples/train-stacked-autoencoders-for-image-classification.html (accessed on 2 February 2020).
  49. Image Set Repository. Available online: https://github.com/Torras86/Olive-image-set (accessed on 18 December 2019).
  50. IDS Imaging Development Systems GmbH. 2016. Available online: https://es.ids-imaging.com/store/ui-1220se.html (accessed on 2 February 2020).
  51. QT Creator. The Qt Company. 2020. Available online: https://doc.qt.io/ (accessed on 18 December 2019).
  52. Dropbox. Dropbox Inc. 2020. Available online: https://www.dropbox.com/developers/documentation (accessed on 2 February 2020).
  53. Braincard. General Visions. 2017. Available online: https://www.general-vision.com/documentation/TM_BrainCard.pdf (accessed on 2 February 2020).
  54. NM500 Chip. General Visión. 2019. Available online: https://www.general-vision.com/documentation/TM_NeuroShield_GettingStarted.pdf (accessed on 2 February 2020).
  55. I2C (Inter Integrated Circuit), Phillips Semiconductor, 1982. Available online: https://en.wikipedia.org/wiki/I%C2%B2C (accessed on 2 February 2020).
  56. FT232RL USB UART IC. Future Technology Devices International Limited. 2018. Available online: https://www.ftdichip.com/Support/Documents/DataSheets/ICs/DS_FT232R.pdf (accessed on 20 December 2019).
  57. USB to Serial Chip CH340. SparkFun Electronics. 2015. Available online: https://cdn.sparkfun.com/datasheets/Dev/Arduino/Other/CH340DS1.PDF (accessed on 19 December 2019).
  58. Intel Curie Module. Intel Corporation. 2020. Available online: https://ark.intel.com/content/www/es/es/ark/products/96282/intel-curie-module-intel-quark-se-soc.html (accessed on 21 December 2019).
  59. Teensy 4.0 USB Development Board. PJRC Electronics Projects Components Available Worldwide. Available online: https://www.pjrc.com/teensy/ (accessed on 20 December 2019).
  60. Google Remote Desktop. Google LLC. 2020. Available online: https://support.google.com/chrome/answer/1649523?co=GENIE.Platform%3DDesktop&hl=es (accessed on 20 December 2019).
Figure 1. Example of the use of region of interest (ROI) in an olive in a normal position. The red point corresponds to the origin of coordinates for pixels. The green rectangle comprises the pocket within its walls and depth and limited by the previous and the next pocket of the feed chain. The red rectangle is the ROI.
Figure 1. Example of the use of region of interest (ROI) in an olive in a normal position. The red point corresponds to the origin of coordinates for pixels. The green rectangle comprises the pocket within its walls and depth and limited by the previous and the next pocket of the feed chain. The red rectangle is the ROI.
Sensors 20 01541 g001
Figure 2. Example of the use of ROI in an olive in a “boat” position between [80°, 90°] and [−80°, −90°].
Figure 2. Example of the use of ROI in an olive in a “boat” position between [80°, 90°] and [−80°, −90°].
Sensors 20 01541 g002
Figure 3. Selected ROI.
Figure 3. Selected ROI.
Sensors 20 01541 g003
Figure 4. Punch needles in an olive pitting machine.
Figure 4. Punch needles in an olive pitting machine.
Sensors 20 01541 g004
Figure 5. Magnetic sensor used to detect the passage of the pockets in the chain.
Figure 5. Magnetic sensor used to detect the passage of the pockets in the chain.
Sensors 20 01541 g005
Figure 6. Electronic circuit of the external trigger and LED.
Figure 6. Electronic circuit of the external trigger and LED.
Sensors 20 01541 g006
Figure 7. Electronic circuit of the PC with CM1K chip.
Figure 7. Electronic circuit of the PC with CM1K chip.
Sensors 20 01541 g007
Figure 8. Qt Creator application for IoT management of olive pitting, slicing and stuffing (DRR) machines.
Figure 8. Qt Creator application for IoT management of olive pitting, slicing and stuffing (DRR) machines.
Sensors 20 01541 g008
Figure 9. The system implemented on a “DRR machine”.
Figure 9. The system implemented on a “DRR machine”.
Sensors 20 01541 g009
Figure 10. PC-CM1K communication system.
Figure 10. PC-CM1K communication system.
Sensors 20 01541 g010
Figure 11. IoT control system.
Figure 11. IoT control system.
Sensors 20 01541 g011
Figure 12. Confusion matrix using 10 × 10 resolution images.
Figure 12. Confusion matrix using 10 × 10 resolution images.
Sensors 20 01541 g012
Figure 13. Matrix of the obtained results using MATLAB (11 × 11 pixels resolution).
Figure 13. Matrix of the obtained results using MATLAB (11 × 11 pixels resolution).
Sensors 20 01541 g013
Figure 14. Matrix of the obtained results using MATLAB (16 × 16 pixels resolution).
Figure 14. Matrix of the obtained results using MATLAB (16 × 16 pixels resolution).
Sensors 20 01541 g014
Figure 15. Graphical User Interface (GUI) created in Matlab to analyze data from the “DRR machines”.
Figure 15. Graphical User Interface (GUI) created in Matlab to analyze data from the “DRR machines”.
Sensors 20 01541 g015
Figure 16. Settings dialogue box of the GUI from Matlab.
Figure 16. Settings dialogue box of the GUI from Matlab.
Sensors 20 01541 g016
Figure 17. GUI of “boat” olives and doubles for the “DRR machine” diagnosis.
Figure 17. GUI of “boat” olives and doubles for the “DRR machine” diagnosis.
Sensors 20 01541 g017
Figure 18. Multiple accumulated values (boats, empty, doubles and small parts).
Figure 18. Multiple accumulated values (boats, empty, doubles and small parts).
Sensors 20 01541 g018
Figure 19. Instantaneous values for boats, doubles, small parts and empty.
Figure 19. Instantaneous values for boats, doubles, small parts and empty.
Sensors 20 01541 g019
Figure 20. Instantaneous values of speed and production.
Figure 20. Instantaneous values of speed and production.
Sensors 20 01541 g020
Table 1. Images (both original and binarized) classified according to angular intervals.
Table 1. Images (both original and binarized) classified according to angular intervals.
ANGLE [0°, 10º]ANGLE [0°,−10°]ANGLE [10°,20°]ANGLE [−10°,−20°]
Sensors 20 01541 i001 Sensors 20 01541 i002 Sensors 20 01541 i003 Sensors 20 01541 i004
ANGLE [20°,30°]ANGLE [−20°,−30°]ANGLE [30°,40°]ANGLE [-30°,-40°]
Sensors 20 01541 i005 Sensors 20 01541 i006 Sensors 20 01541 i007 Sensors 20 01541 i008
ANGLE [40°,50°]ANGLE [−40°,−50°]ANGLE [50°,60°]ANGLE [−50°,−60°]
Sensors 20 01541 i009 Sensors 20 01541 i010 Sensors 20 01541 i011 Sensors 20 01541 i012
ANGLE [60°,70°]ANGLE [−60°,−70°]ANGLE [70°,80°]ANGLE [−70°,−80°]
Sensors 20 01541 i013 Sensors 20 01541 i014 Sensors 20 01541 i015 Sensors 20 01541 i016
ANGLE [80°,90°] or [−80°,−90°]
Sensors 20 01541 i017
Table 2. Average communication times in ms during the delivery of a 16 × 16 bytes image to be classified.
Table 2. Average communication times in ms during the delivery of a 16 × 16 bytes image to be classified.
CH340GCURIETEENSY
1 BYTE366.48258.08152.26
258 BYTES22.415.040.78
Table 3. Results of the training patterns using a resolution of 11 × 11 pixels.
Table 3. Results of the training patterns using a resolution of 11 × 11 pixels.
Type of TestDegreesIntermediate PositionSimple
Neuron96Neurons24Neurons4
BoardVectorRepeatabilityErrorRepeatabilityErrorRepeatabilityError
Intel
Curie
121119%17%16%
215%211%29%
37%37%34%
411%410%46%
513%59%56%
614%610%66%
78%76%73%
87%89%82%
912%910%93%
107%107%103%
Totals 11.30% 8.47% 4.80%
Table 4. Results of the training patterns using a resolution of 1 6 × 16 pixels.
Table 4. Results of the training patterns using a resolution of 1 6 × 16 pixels.
Type of TestDegreesIntermediate PositionSimple
Neuron66Neurons24Neurons4
BoardVectorRepeatabilityErrorRepeatabilityErrorRepeatabilityError
Braincard
(CM1K)
25618%13%11%
29%24%20%
310%34%30%
412%43%42%
513%54%51%
613%65%61%
715%72%70%
816%85%81%
916%98%91%
1013%102%100%
Totals 12.43% 4% 0.63%

Share and Cite

MDPI and ACS Style

Lucas Pascual, A.; Madueño Luna, A.; de Jódar Lázaro, M.; Molina Martínez, J.M.; Ruiz Canales, A.; Madueño Luna, J.M.; Justicia Segovia, M. Analysis of the Functionality of the Feed Chain in Olive Pitting, Slicing and Stuffing Machines by IoT, Computer Vision and Neural Network Diagnosis. Sensors 2020, 20, 1541. https://doi.org/10.3390/s20051541

AMA Style

Lucas Pascual A, Madueño Luna A, de Jódar Lázaro M, Molina Martínez JM, Ruiz Canales A, Madueño Luna JM, Justicia Segovia M. Analysis of the Functionality of the Feed Chain in Olive Pitting, Slicing and Stuffing Machines by IoT, Computer Vision and Neural Network Diagnosis. Sensors. 2020; 20(5):1541. https://doi.org/10.3390/s20051541

Chicago/Turabian Style

Lucas Pascual, Alberto, Antonio Madueño Luna, Manuel de Jódar Lázaro, José Miguel Molina Martínez, Antonio Ruiz Canales, José Miguel Madueño Luna, and Meritxell Justicia Segovia. 2020. "Analysis of the Functionality of the Feed Chain in Olive Pitting, Slicing and Stuffing Machines by IoT, Computer Vision and Neural Network Diagnosis" Sensors 20, no. 5: 1541. https://doi.org/10.3390/s20051541

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop