Veuillez utiliser cette adresse pour citer ce document :
http://dspace.univ-bouira.dz:8080/jspui/handle/123456789/11810
Affichage complet
Élément Dublin Core | Valeur | Langue |
---|---|---|
dc.contributor.author | Benguerrah, Youcef Anis | - |
dc.contributor.author | Touati, Djamel | - |
dc.date.accessioned | 2021-12-05T09:37:21Z | - |
dc.date.available | 2021-12-05T09:37:21Z | - |
dc.date.issued | 2020 | - |
dc.identifier.uri | http://dspace.univ-bouira.dz:8080/jspui/handle/123456789/11810 | - |
dc.description.abstract | Currently, the detection of coronavirus disease (COVID-19) is one of the main challenges in the world, given the rapid spread of the disease. Recent statistics indicate that the number of people diagnosed with COVID-19 is increasing exponentially, with more than 1.6 million con rmed cases; the disease is spreading to many countries across the world. The objective of this work is to propose a Deep Learning approach in the eld of epidemiology to detect coronavirus. To do this, we have chosen to use the Convolutional Neural networks (CNN), where di erent models have been implemented allowing us to obtain the best results. In this study, we proposed three Convolutional Neural Network approaches and crated three models based on these approaches, the created models were trained and prove their e ciency by achieving high accuracy and F1-Score in their testing using real-world chest CT scans dataset. | en_US |
dc.language.iso | en | en_US |
dc.publisher | université akli mohand oulhadj-bouira | en_US |
dc.subject | COVID-19 ; Coronavirus | en_US |
dc.subject | Chest CT scans ; Medical imaging | en_US |
dc.title | COVID-19 Detection using Deep Learning Based Approaches | en_US |
dc.type | Thesis | en_US |
Collection(s) : | Mémoires Master |
Fichier(s) constituant ce document :
Fichier | Description | Taille | Format | |
---|---|---|---|---|
COVID-19 Detection using DeepLearning Based Approaches final.pdf | 7,84 MB | Adobe PDF | Voir/Ouvrir |
Tous les documents dans DSpace sont protégés par copyright, avec tous droits réservés.