SciELO - Scientific Electronic Library Online

 
vol.28 número1Un modelo de programación entera para la generación de horarios universitarios: Un caso de estudioFramework for Heterogeneous Data Management: An Application Case in a NoSQL Environment from a Climatological Center índice de autoresíndice de materiabúsqueda de artículos
Home Pagelista alfabética de revistas  

Servicios Personalizados

Revista

Articulo

Indicadores

Links relacionados

  • No hay artículos similaresSimilares en SciELO

Compartir


Computación y Sistemas

versión On-line ISSN 2007-9737versión impresa ISSN 1405-5546

Resumen

ADEBANJI, Olaronke O. et al. Adaptation of Transformer-Based Models for Depression Detection. Comp. y Sist. [online]. 2024, vol.28, n.1, pp.151-165.  Epub 10-Jun-2024. ISSN 2007-9737.  https://doi.org/10.13053/cys-28-1-4691.

Pre-trained language models are able to capture a broad range of knowledge and language patterns in text and can be fine-tuned for specific tasks. In this paper, we focus on evaluating the effectiveness of various traditional machine learning and pre-trained language models in identifying depression through the analysis of text from social media. We examined different feature representations with the traditional machine learning models and explored the impact of pre-training on the transformer models and compared their performance. Using BoW, Word2Vec, and GloVe representations, the machine learning models with which we experimented achieved impressive accuracies in the task of detecting depression. However, pre-trained language models exhibited outstanding performance, consistently achieving high accuracy, precision, recall, and F1 scores of approximately 0.98 or higher.

Palabras llave : Depression; bag-of-words; word2vec; GloVe; machine learning; deep learning; transformers; sentiment analysis.

        · texto en Inglés     · Inglés ( pdf )