Ciencias Exactas y Ciencias de la Salud
Permanent URI for this collectionhttps://hdl.handle.net/11285/551039
Pertenecen a esta colección Tesis y Trabajos de grado de las Maestrías correspondientes a las Escuelas de Ingeniería y Ciencias así como a Medicina y Ciencias de la Salud.
Browse
Search Results
- Smart camera FPGA hardware implementation for semantic segmentation of wildfire imagery(Instituto Tecnológico y de Estudios Superiores de Monterrey, 2024-06-13) Garduño Martínez, Eduardo; Rodriguez Hernández, Gerardo; mtyahinojosa, emipsanchez; Gonzalez Mendoza, Miguel; Hinojosa Cervantes, Salvador Miguel; School of Engineering and Sciences; Campus Monterrey; Ochoa Ruiz, GilbertoIn the past few years, the more frequent occurrence of wildfires, which are a result of climate change, has devastated society and the environment. Researchers have explored various technologies to address this issue, including deep learning and computer vision solutions. These techniques have yielded promising results in semantic segmentation for detecting fire using visible and infrared images. However, implementing deep learning neural network models can be challenging, as it often requires energy-intensive hardware such as a GPU or a CPU with large cooling systems to achieve high image processing speeds, making it difficult to use in mobile applications such as drone surveillance. Therefore, to solve the portability problem, an FPGA hardware implementation is proposed to satisfy low power consumption requirements, achieve high accuracy, and enable fast image segmentation using convolutional neural network models for fire detection. This thesis employs a modified UNET model as the base model for fire segmentation. Subsequently, compression techniques reduce the number of operations performed by the model by removing filters from the convolutional layers and reducing the arithmetic precision of the CNN, decreasing inference time and storage requirements and allowing the Vitis AI framework to map the model architecture and parameters onto the FPGA. Finally, the model was evaluated using metrics utilized in prior studies to assess the performance of fire detection segmentation models. Additionally, two fire datasets are used to compare different data types for fire segmentation models, including visible images, a fusion of visible and infrared images generated by a GAN model, fine-tuning of the fusion GAN weights, and the use of visible and infrared images independently to evaluate the impact of visible-infrared information on segmentation performance.
- Deep Learning Approach for Alzheimer’s Disease Classification: Integrating Multimodal MRI and FDG- PET Imaging Through Dual Feature Extractors and Shared Neural Network Processing(Instituto Tecnológico y de Estudios Superiores de Monterrey, 2024) Vega Guzmán, Sergio Eduardo; Alfaro Ponce, Mariel; emimmayorquin; Ochoa Ruíz, Gilberto; Chairez Oria, Jorge Isaac; Hernandez Sanchez, Alejandra; School of Engineering and Sciences; Campus Monterrey; Ramírez Nava, Gerardo JuliánAlzheimer’s disease (AD) is a progressive neurodegenerative disorder whose incidence is expected to grow in the coming years. Traditional diagnostic methods, such as MRI and FDG-PET, each provide valuable but limited insights into the disease’s pathology. This thesis researches the potential of a multimodal deep learning classifier to improve the diagnostic accuracy of AD by integrating MRI and FDG-PET imaging data in comparison to single modality implementations. The study proposes a lightweight neural architecture that uses the strengths of both imaging modalities, aiming to reduce computational costs while maintaining state-of-the-art diagnostic performance. The proposed model utilizes two pre-trained feature extractors, one for each imaging modality, fine-tuned to capture the relevant features from the dataset. The outputs of these extractors are fused into a single vector to form an enriched feature map that better describes the brain. Experimental results demonstrate that the multimodal classifier outperforms single modality classifiers, achieving an overall accuracy of 90% on the test dataset. The VGG19 model was the best feature extractor for both MRI and PET data since it showed superior performance when compared to the other experimental models, with an accuracy of 71.9% for MRI and 80.3% for PET images. The multimodal implementation also exhibited higher precision, recall, and F1 scores than the single-modality implementations. For instance, it achieved a precision of 0.90, recall of 0.94, and F1-score of 0.92 for the AD class and a precision of 0.89, recall of 0.82, and F1-score of 0.86 for the CN class. Furthermore, explainable AI techniques provided insights into the model’s decisionmaking process, revealing that it effectively utilizes both structural and metabolic information to distinguish between AD and cognitively normal (CN) subjects. This research adds supporting evidence into the potential of multimodal imaging and machine learning to enhance early detection and diagnosis of Alzheimer’s disease, offering a cost-effective solution suitable for widespread clinical applications.

