AlephBERT: Language model pre-training and evaluation from sub-word to sentence level
Abstract Large Pre-trained Language Models (PLMs) have become ubiquitous in the
development of language understanding technology and lie at the heart of many artificial …
development of language understanding technology and lie at the heart of many artificial …
Multitask learning for emotion and personality traits detection
In recent years, deep learning-based automated personality traits detection has received a
lot of attention, especially now, due to the massive digital footprints of an individual …
lot of attention, especially now, due to the massive digital footprints of an individual …
On the frontlines: The impact of the COVID-19 pandemic on social workers' well-being
TM Schwartz Tayri - Social Work, 2023 - academic.oup.com
The crisis created by the spread of COVID-19 brought increasing needs and referrals to
social welfare services in many countries. However, at the same time, social services …
social welfare services in many countries. However, at the same time, social services …
AlephBERT: A Hebrew large pre-trained language model to start-off your Hebrew NLP application with
Large Pre-trained Language Models (PLMs) have become ubiquitous in the development of
language understanding technology and lie at the heart of many artificial intelligence …
language understanding technology and lie at the heart of many artificial intelligence …
Quality at a glance: An audit of web-crawled multilingual datasets
With the success of large-scale pre-training and multilingual modeling in Natural Language
Processing (NLP), recent years have seen a proliferation of large, Web-mined text datasets …
Processing (NLP), recent years have seen a proliferation of large, Web-mined text datasets …
Predicting Client Emotions and Therapist Interventions in Psychotherapy Dialogues
T Mayer, N Warikoo, A Eliassaf… - Proceedings of the …, 2024 - aclanthology.org
Abstract Natural Language Processing (NLP) can advance psychotherapy research by
scaling up therapy dialogue analysis as well as by allowing researchers to examine client …
scaling up therapy dialogue analysis as well as by allowing researchers to examine client …
Large pre-trained models with extra-large vocabularies: A contrastive analysis of hebrew bert models and a new one to outperform them all
E Gueta, A Shmidman, S Shmidman… - arXiv preprint arXiv …, 2022 - arxiv.org
We present a new pre-trained language model (PLM) for modern Hebrew, termed
AlephBERTGimmel, which employs a much larger vocabulary (128K items) than standard …
AlephBERTGimmel, which employs a much larger vocabulary (128K items) than standard …
ParaShoot: A Hebrew question answering dataset
O Keren, O Levy - arXiv preprint arXiv:2109.11314, 2021 - arxiv.org
NLP research in Hebrew has largely focused on morphology and syntax, where rich
annotated datasets in the spirit of Universal Dependencies are available. Semantic datasets …
annotated datasets in the spirit of Universal Dependencies are available. Semantic datasets …
Offensive Hebrew corpus and detection using bert
N Hamad, M Jarrar, M Khalilia… - 2023 20th ACS/IEEE …, 2023 - ieeexplore.ieee.org
Offensive language detection has been well studied in many languages, but it is lagging
behind in low-resource languages, such as Hebrew. In this paper, we present a new …
behind in low-resource languages, such as Hebrew. In this paper, we present a new …
Hero: Roberta and longformer hebrew language models
V Shalumov, H Haskey - arXiv preprint arXiv:2304.11077, 2023 - arxiv.org
In this paper, we fill in an existing gap in resources available to the Hebrew NLP community
by providing it with the largest so far pre-train dataset HeDC4, a state-of-the-art pre-trained …
by providing it with the largest so far pre-train dataset HeDC4, a state-of-the-art pre-trained …