DEVELOPMENT OF A GRAPHICAL USER INTERFACE TO SUPPORT THE SEMI-AUTOMATIC SEMANTIC SEGMENTATION OF UAS-IMAGES
Keywords: Semantic segmentation, Graphical User Interface, Machine Learning, Multispectral Imagery
Abstract. The development of remote sensing techniques dramatically improved the human knowledge of natural phenomena and the real time monitoring and interpretation of the events happening in the environment. The recently developed terrestrial, aerial and satellite remote sensors caused the availability of huge amount of data. The large size of such data is leading the research community to the search for efficient methods for real time information extraction, and, more in general, understanding the collected data. Nowadays, this is typically done by means of artificial intelligence-based methods, and, more specifically, usually by means of machine learning tools. Focusing on semantic segmentation, which is clearly related to a proper interpretation of the acquired remote sensing data, supervised machine learning is often used: it is based on the availability of a set of ground truth labeled data, which are used in order to properly train a machine learning classifier. Despite the latter, after a proper training phase, usually allows to obtain quite effective segmentation results, the ground truth labeled data production is usually a very laborious and time consuming task, performed by a human operator. Motivated by the latter consideration, this work aims at introducing a graphical interface developed in order to support semi-automatic semantic segmentation of images acquired by a UAS. Certain of the potentialities of the proposed graphical are shown in the specific case of plastic litter detection in multi-spectral images.