Listing 1 - 7 of 7 |
Sort by
|
Choose an application
In several fields of application, multi-gigapixel images must be analysed to gather information and take decision. This analysis is often performed manually, which is a tedious task given the volume of data to process. For instance, in cytology, branch of medical sciences which focuses on study of cells, cytopathologists analyse cell samples microscope slides in order to diagnose diseases such as cancers. Typically, malignancy is assessed by the presence or absence of cells with given characteristics. In geology, climate variations can be analysed by studying the concentration of micro-organisms in core samples. The concentration is usually evaluated by smearing the samples onto microscope glass slides and counting those micro-organisms. In those situations, computer sciences and, especially, machine learning and image processing provide a great alternative to a pure-human approach as they can be used to extract relevant information automatically. Especially, those kind of problems can be expressed as object detection and classification problems. This thesis presents the elaboration and assessment of a generic framework, extit{SLDC}, for object detection and classification in multi-gigapixel images. Especially, this framework provides implementers with a concise way of formulating problem dependent-components of their algorithm (i.e. segmentation and classification) while it takes care of problem-independent concerns such as parallelization and large image handling. The performances of the framework are then assessed on a real-world problem, thyroid nodule malignancy. Especially, a workflow is built to detect malignant cells in thyroid cell samples whole-slides. Results are promising: the effective processing time for an image containing 8 gigapixels is less than 10 minutes. In order, to further reduce this execution time, some improvements are proposed. The framework implementation can be found on GitHub: https://github.com/waliens/sldc.
machine learning --- cytomine --- image processing --- cytology --- object detection --- Ingénierie, informatique & technologie > Sciences informatiques
Choose an application
Detecting cracks is a very interesting field of research used to address a large number of issues, including the detection of road and other structural cracks to strengthen the prevention of damages and planning of maintenance and repairs. It is also used to facilitate the authentication of paintings. In most cases, the detection is only performed on simple images while multi-spectral images are increasingly used in many research areas, including Medicine, given their high potential for providing more accurate information. This work responds to a request by the "Lumiere Technology Multispectral Institute" which indicated an interest in using multi-spectral images to detect cracks in paintings. Its objective is to analyse if applying detections on multi-spectral images can help improve the quality and efficiency of maintenance and restoration of paintings. To meet this objective, we work with the Cytomine application, an "Open-source rich Internet application for collaborative analysis of multi-gigapixel images", developed by researchers at the Montefiore Institute of the University of Liège. We more specifically use the recently-developed extension supporting the utilisation of multi-spectral images to apply machine learning algorithms on wide-banded multi-spectral images. The "Musées royaux des beaux-arts de Belgique" have provided the paintings which are analysed in this work. To carry out our analysis: i) we implement the Cytomine’s Extractor and Spectral Reader and complement the application’s new extension by adding ways to optimise the extraction, storage and use of data from multi-spectral image, ii) based on this implementation, we present three softwares which we designed to enable us to make a feature analysis of the data, to fit a model using data from extit{Cytomine}, or to use a fitted model to make predictions, iii) we test two dimension reduction methods, PCA (Principal Components Analysis) and TSNE (T-distributed Stochastic Neighbour Embedding) and three feature importance measures to help reducing the large size of datasets created by multi-spectral images, and iv) we conduct experiments on three datasets to find how machine learning performs on multi-spectral images: one small dataset based on a multi-spectral image of biological cells, one small dataset based on a multi-spectral image where we have extracted two different tones of red, and a third dataset based on a multi-spectral image where we have labelled cracks and undamaged parts on the painting extit{Portrait of Flautist François Devienne} of Jacques-Louis David. These experiments evidence that: i) TSNE is a possible solution to reduce the dimension of multi-spectral images if the necessary resources, i.e. memory and computational power, are available, and ii) using multi-spectral images has an advantage over simply using standard RGB images, even if the increase in computation time implies that a pre-processing is needed to reduce the number of bands. We therefore encourage further research based on the new tools developed in this work, for instance to try Boosting methods or Deep Neural Networks.
painting --- cytomine --- multi spectral --- image --- crack --- Ingénierie, informatique & technologie > Sciences informatiques
Choose an application
In many biomedical applications, manual annotations of whole slide images take a tremendous amount of time. In the computer vision literature, semi-automatic tools using deep learning, known as deep interactive learning, have emerged to speed up the annotation process. These semi-automatic tools exploit the interactions of the annotators in various forms to produce the annotations more rapidly. In recent years, deep interactive learning seems to gain more attention for its performance. However, do the additional information provided by the annotators help to improve the results of automatic tools? An exploration in the literature was made, resulting in the finding of a promising architecture, named NuClick, which uses the scribbles of the annotators in combination with the images to produce decent annotations more quickly. In this thesis, results of the conducted experiments on various datasets show that the additional information provided by the scribbles improve drastically the performance of the segmentation for tissues, such as bronchi, glands, or infiltrations. However, this interactive approach fails to produce accurate segmentation for more complex tissues, such as tumours or inflammations. Also, results indicate that the quality of the scribbles highly influences the produced segmentation. Therefore, care should be taken when the annotators scribble the objects of interest. These results tend to support the benefit that can be gain from the interactions of the annotators, although this thesis shows that there is room for improvements with these semi-automatic tools.
Cytomine --- Pathology --- Deep Learning --- Computer Vision --- Image Segmentation --- Ingénierie, informatique & technologie > Sciences informatiques
Choose an application
Digital microscopy and radiology generate growing amounts of imagery data. To help practitioners find the information crucial to establish the most accurate possible diagnoses, Artificial Intelligence tools need to be developed. This master thesis, based on the study of existing literature and open-source code, proposes a distributed deep learning architecture that allows a user, by using a fast approximate nearest neighbour search, to retrieve similar histopathology images to a query image. The retained Deep Learning architecture, ResNet50 with some modifications, was distributed on different servers in order to allow the handling of up to million or billion images. It was trained on a large-scale dataset of 67 classes of annotated medical images and the obtained results are quite promising, as well for the visual similarity of the retrieved images as for the search time. This research also analyses the generalisation to classes on which the system was not trained, and the impact of the approximated search on the accuracy and the retrieval time. Nevertheless, even though the results are positive, this system might present some limitations as it was tested on only one dataset and was not reviewed by medical practitioners.
Deep Learning --- Cytomine --- Computer Vision --- Histopathology --- Image Retrieval --- Ingénierie, informatique & technologie > Sciences informatiques
Choose an application
In the biomedical research field, and more precisely the field of morphometric studies, the detection of anatomical landmarks is a crucial step in order to quantify the shape and size of an object under study. The annotations of theses landmarks is quite laborious and often requires dedicated human expertise. Thus the use of automatic landmark detection techniques using artificial intelligence began to gain importance. This thesis explores different kinds of approach in order to tackle the problem on butterflies from the Morpho genus. Each butterfly possesses two types of landmarks: true landmarks and semi-landmarks, and has two views: ventral and dorsal, resulting in a total of 4 distinct datasets. Moreover, the Morpho genus contain about 30 species which induces a great variety between the images. Several approaches have been experimented, namely computer vision, machine learning and deep learning. The results shows that the deep learning approach outperforms the others in most cases. Nevertheless the machine learning approach has proven its performance on smaller part of the data, extit{i.e.} datasets composed of one specie only. Unfortunately, the computer vision approach did not lead to any convincing results. Finally, this work presents the Cytomine application that has been built along with the most consistent model from the experimentations. This application provides a user-friendly interface for both training and predicting with a new model.
Machine Learning --- Deep Learning --- Automatic Landmark Detection --- Landmark --- Morpho --- Cytomine --- Ingénierie, informatique & technologie > Sciences informatiques
Choose an application
In the medical field, doctors and researchers should be able to observe and interpret cell samples. The most widely spread method is to observe samples with a microscope. New methods allow for scanning samples into very large and detailed images. These images can be uploaded to the Cytomine web application and doctors can go through the images at ease. The application includes many functionalities, namely the ability to annotate regions of interest. Currently the ULiege MOOC server is open to the general public (see https//www.fun-mooc.fr). Anyone can take part in the MOOC, but it's mostly used for the bene t of the students studying the medical field at the University of Liege. The students study particular images and are then evaluated at the end of the year. Meanwhile, the Cytomine app has been collecting data on the students' time spent on the website. The main objective is to develop IT tools to analyze data collected. The bulk of the data collected consists of where the students decided to look in the images (Gaze data). Consequently, a tool was implemented to find correlations between students' behavior and the results they obtain during exams. Using Machine Learning techniques, the goal is to predict a student's grade based on how they used the application. This allows for the analysis of specific viewing patterns that can give more insight on how the student used Cytomine. Currently, the model contains 395 students' data with over 2000 features. Extra Trees and Random Forest learning techniques have been applied to attempt to predict grades. Otherwise, other tools were implemented to better visualize these patterns. The idea would be to generate Heatmaps of the students' gaze data (Gazemap). These Gazemaps would be included in the application and users can be given access to this information. This could give teachers the ability to keep track of the students' work.
Cytomine --- machine learning --- pathology --- data analysis --- gaze data --- whole slide images --- behavioral analytics --- Ingénierie, informatique & technologie > Sciences informatiques
Choose an application
Multimodal imaging analyses are large scale work, combining experience from many professionals in different disciplines, providing different modalities (i.e. data produced by an experiment) linked together. The growth in multimodal analyses induces a demand for software to make some workflows possible, or help automate some other workflows, at lest partially. Mass Spectrometry Imaging (MSI), although more than 50 years old, continues to see some development in the data processing domain, particularly with machine learning and deep learning applications, where some approaches tackle the preprocessing and the analysis of MSI datasets. The analysis performed on MSI data greatly contributes from multimodal studies, providing a spatial distribution for the molecular content of the sample, thus adding valuable information to the study. Multimodal analyses currently lack an open, collaborative web platform : such tools would allow for a greater share of experience thanks to the collaborative aspect, enable reproducibility because the analyses would run in the cloud, always on the same hardware, and the results would be available to all. Such tools are being developed : Cytomine aims to add more effective multimodal tools to improve its capabilities, but integrating MSI data is not trivial. The analysis of MSI data is not an easy task : file formats for this kind of data are abundant, but often vendor specific. imzML is an open effort to unify all these formats, which is supported by many pieces of software already. However, imzML is not the most appropriate format as its structure is very different from most imaging data format, making it ill-suited for visualization applications such as in Cytomine. This master's thesis introduce a new, versatile and open format based on OME-Zarr, which is suitable for many modalities, including MSI. This file format is benchmarked against imzML to show its potential in server applications, such as Cytomine. In addition to the new file format, the developed pieces of software includes a convertor from imzML and some preprocessing tools designed for the file format. Using the developed file format, a machine learning workflow classifies spectra from a multimodal dataset with label coming from other modalities, and provide a list of important features as a mean of interpretation. While these pieces of software are currently developed to be run on a local machine, they lay the ground for cloud based application that can be integrated with Cytomine.
multimodal --- mass spectrometry imaging --- msi --- cytomine --- template matching --- machine learning --- image registration --- bioimaging --- Ingénierie, informatique & technologie > Sciences informatiques
Listing 1 - 7 of 7 |
Sort by
|