Lexical Analysis
7 papers with code • 0 benchmarks • 1 datasets
Lexical analysis is the process of converting a sequence of characters into a sequence of tokens (strings with an assigned and thus identified meaning). (Source: Adapted from Wikipedia)
Benchmarks
These leaderboards are used to track progress in Lexical Analysis
Latest papers with no code
Towards Lexical Analysis of Dog Vocalizations via Online Videos
This study presents a data-driven investigation into the semantics of dog vocalizations via correlating different sound types with consistent semantics.
Leveraging Natural Language Processing For Public Health Screening On YouTube: A COVID-19 Case Study
The aim of our study was to investigate the use of Natural Language Processing (NLP) to identify the spoken content of YouTube vlogs related to the diagnosis of Coronavirus disease of 2019 (COVID-19) for public health screening.
Happy or grumpy? A Machine Learning Approach to Analyze the Sentiment of Airline Passengers' Tweets
This study aims to measure customer satisfaction by analyzing sentiments of Tweets that mention airlines using a machine learning approach.
CV4Code: Sourcecode Understanding via Visual Code Representations
We present CV4Code, a compact and effective computer vision method for sourcecode understanding.
Effect and Analysis of Large-scale Language Model Rescoring on Competitive ASR Systems
Large-scale language models (LLMs) such as GPT-2, BERT and RoBERTa have been successfully applied to ASR N-best rescoring.
Evaluating Persian Tokenizers
Natural Language Processing in Persian is challenging due to Persian's exceptional cases, such as half-spaces.
COVID-19 and Misinformation: A Large-Scale Lexical Analysis on Twitter
Social media is often used by individuals and organisations as a platform to spread misinformation.
Visual Exploration and Knowledge Discovery from Biomedical Dark Data
Data visualization techniques proffer efficient means to organize and present data in graphically appealing formats, which not only speeds up the process of decision making and pattern recognition but also enables decision-makers to fully understand data insights and make informed decisions.
Integration of Automatic Sentence Segmentation and Lexical Analysis of Ancient Chinese based on BiLSTM-CRF Model
The basic tasks of ancient Chinese information processing include automatic sentence segmentation, word segmentation, part-of-speech tagging and named entity recognition.
Processing Language Resources of Under-Resourced and Endangered Languages for the Generation of Augmentative Alternative Communication Boards
This paper shows an approach using different annotations of enriched linguistic research data to create communication boards commonly used in Alternative Augmentative Communication (AAC).