Filtering by
- All Subjects: Computer Science
- Creators: Computer Science and Engineering Program
American Sign Language (ASL) is used for Deaf and Hard of Hearing (DHH) individuals to communicate and learn in a classroom setting. In ASL, fingerspelling and gestures are two primary components used for communication. Fingerspelling is commonly used for words that do not have a specifically designated sign or gesture. In technical contexts, such as Computer Science curriculum, there are many technical terms that fall under this category. Most of its jargon does not have standardized ASL gestures; therefore, students, educators, and interpreters alike have been reliant on fingerspelling, which poses challenges for all parties. This study investigates the efficacy of both fingerspelling and gestures with fifteen technical terms that do have standardized gestures. The terms’ fingerspelling and gesture are assessed based on preference, ease of use, ease of learning, and time by research subjects who were selected as DHH individuals familiar with ASL.
The data is collected in a series of video recordings by research subjects as well as a post-participation questionnaire. Each research subject has produced thirty total videos, two videos to fingerspell and gesture each technical term. Afterwards, they completed a post-participation questionnaire in which they indicated their preference and how easy it was to learn and use both fingerspelling and gestures. Additionally, the videos have been analyzed to determine the time difference between fingerspelling and gestures. Analysis reveals that gestures are favored over fingerspelling as they are generally preferred, considered easier to learn and use, and faster. These results underscore the significance for standardized gestures in the Computer Science curriculum for accessible learning that enhances communication and promotes inclusion.
2018, Google researchers published the BERT (Bidirectional Encoder Representations from Transformers) model, which has since served as a starting point for hundreds of NLP (Natural Language Processing) related experiments and other derivative models. BERT was trained on masked-language modelling (sentence prediction) but its capabilities extend to more common NLP tasks, such as language inference and text classification. Naralytics is a company that seeks to use natural language in order to be able to categorize users who create text into multiple categories – which is a modified version of classification. However, the text that Naralytics seeks to pull from exceed the maximum token length of 512 tokens that BERT supports – so this report discusses the research towards multiple BERT derivatives that seek to address this problem – and then implements a solution that addresses the multiple concerns that are attached to this kind of model.
The field of quantum computing is an exciting area of research that allows quantum mechanics such as superposition, interference, and entanglement to be utilized in solving complex computing problems. One real world application of quantum computing involves applying it to machine learning problems. In this thesis, I explore the effects of choosing different circuit ansatz and optimizers on the performance of a variational quantum classifier tasked with binary classification.