Machine vision system for automatic quality grading of fruit
Derechos de accesoopenAccess
MetadatosMostrar el registro completo del ítem
Cita bibliográficaBlasco, J., Aleixos, N., Moltó, E. (2003). Machine vision system for automatic quality grading of fruit. Biosystems Engineering, 85(4), 415-423.
Fruit and vegetables are normally presented to consumers in batches. The homogeneity and appearance of these have significant effect on consumer decision. For this reason, the presentation of agricultural produce is manipulated at various stages from the field to the final consumer and is generally oriented towards the cleaning of the product and sorting by homogeneous categories. The project ESPRIT 3, reference 9230 'Integrated system for handling, inspection and packing of fruit and vegetable (SHIVA)' developed a robotic system for the automatic, non-destructive inspection and handling of fruit. The aim of this paper is to report on the machine vision techniques developed at the Instituto Valenciano de Investigaciones Agrarias for the on-line estimation of the quality of oranges, peaches and apples, and to evaluate the efficiency of these techniques regarding the following quality attributes: size, colour, stem location and detection of external blemishes. The segmentation procedure used, based on a Bayesian discriminant analysis, allowed fruits to be precisely distinguished from the background. Thus, determination of size was properly solved. The colours of the fruits estimated by the system were well correlated with the colorimetric index values that are currently used as standards. Good results were obtained in the location of the stem and the detection of blemishes. The classification system was tested on-line with apples obtaining a good performance when classifying the fruit in batches, and a repeatability in blemish detection and size estimation of 86 and 93% respectively. The precision and repeatability of the system, was found to be similar to those of manual grading. (C) 2003 Silsoe Research Institute. All rights reserved. Published by Elsevier Science Ltd.