Search Results for author: Jake Ryland Williams

Found 16 papers, 3 papers with code

Bit Cipher -- A Simple yet Powerful Word Representation System that Integrates Efficiently with Language Models

no code implementations18 Nov 2023 Haoran Zhao, Jake Ryland Williams

While Large Language Models (LLMs) become ever more dominant, classic pre-trained word embeddings sustain their relevance through computational efficiency and nuanced linguistic interpretation.

Computational Efficiency Dimensionality Reduction +8

Explicit Foundation Model Optimization with Self-Attentive Feed-Forward Neural Units

no code implementations13 Nov 2023 Jake Ryland Williams, Haoran Zhao

We will discuss a general result about feed-forward neural networks and then extend this solution to compositional (mult-layer) networks, which are applied to a simplified transformer block containing feed-forward and self-attention layers.

Model Optimization

Reducing the Need for Backpropagation and Discovering Better Optima With Explicit Optimizations of Neural Networks

no code implementations13 Nov 2023 Jake Ryland Williams, Haoran Zhao

Iterative differential approximation methods that rely upon backpropagation have enabled the optimization of neural networks; however, at present, they remain computationally expensive, especially when training models at scale.

Language Modelling

EigenNoise: A Contrastive Prior to Warm-Start Representations

no code implementations9 May 2022 Hunter Scott Heidenreich, Jake Ryland Williams

In this work, we present a naive initialization scheme for word vectors based on a dense, independent co-occurrence model and provide preliminary results that suggest it is competitive and warrants further investigation.

To Know by the Company Words Keep and What Else Lies in the Vicinity

no code implementations30 Apr 2022 Jake Ryland Williams, Hunter Scott Heidenreich

However, we use the solution to demonstrate a seemingly-universal existence of a property that word vectors exhibit and which allows for the prophylactic discernment of biases in data -- prior to their absorption by DL models.

A general solution to the preferential selection model

no code implementations6 Aug 2020 Jake Ryland Williams, Diana Solano-Oropeza, Jacob R. Hunsberger

We provide a general analytic solution to Herbert Simon's 1955 model for time-evolving novelty functions.

A Computational Framework for Multi-Modal Social Action Identification

no code implementations20 Oct 2017 Jason Anastasopoulos, Jake Ryland Williams

We demonstrate how these methods can be used diagnostically-by researchers, government officials and the public-to understand peaceful and violent collective action at very fine-grained levels of time and geography.

BIG-bench Machine Learning Event Detection

Is space a word, too?

1 code implementation20 Oct 2017 Jake Ryland Williams, Giovanni C. Santia

We offer a resolution to these issues by exhibiting how the dark matter of word segmentation, i. e., space, punctuation, etc., connect the Zipf-Mandelbrot law to Simon's mechanistic process.

Boundary-based MWE segmentation with text partitioning

1 code implementation5 Aug 2016 Jake Ryland Williams

This work presents a fine-grained, text-chunking algorithm designed for the task of multiword expressions (MWEs) segmentation.

Chunking Segmentation

Benchmarking sentiment analysis methods for large-scale texts: A case for using continuum-scored words and word shift graphs

2 code implementations2 Dec 2015 Andrew J. Reagan, Brian Tivnan, Jake Ryland Williams, Christopher M. Danforth, Peter Sheridan Dodds

The emergence and global adoption of social media has rendered possible the real-time estimation of population-scale sentiment, bearing profound implications for our understanding of human behavior.

Benchmarking Sentiment Analysis

Sifting Robotic from Organic Text: A Natural Language Approach for Detecting Automation on Twitter

no code implementations17 May 2015 Eric M. Clark, Jake Ryland Williams, Chris A. Jones, Richard A. Galbraith, Christopher M. Danforth, Peter Sheridan Dodds

Twitter, a popular social media outlet, has evolved into a vast source of linguistic data, rich with opinion, sentiment, and discussion.

Identifying missing dictionary entries with frequency-conserving context models

no code implementations7 Mar 2015 Jake Ryland Williams, Eric M. Clark, James P. Bagrow, Christopher M. Danforth, Peter Sheridan Dodds

With our predictions we then engage the editorial community of the Wiktionary and propose short lists of potential missing entries for definition, developing a breakthrough, lexical extraction technique, and expanding our knowledge of the defined English lexicon of phrases.

Zipf's law holds for phrases, not words

no code implementations19 Jun 2014 Jake Ryland Williams, Paul R. Lessard, Suma Desu, Eric Clark, James P. Bagrow, Christopher M. Danforth, Peter Sheridan Dodds

With Zipf's law being originally and most famously observed for word frequency, it is surprisingly limited in its applicability to human language, holding over no more than three to four orders of magnitude before hitting a clear break in scaling.

Human language reveals a universal positivity bias

no code implementations15 Jun 2014 Peter Sheridan Dodds, Eric M. Clark, Suma Desu, Morgan R. Frank, Andrew J. Reagan, Jake Ryland Williams, Lewis Mitchell, Kameron Decker Harris, Isabel M. Kloumann, James P. Bagrow, Karine Megerdoomian, Matthew T. McMahon, Brian F. Tivnan, Christopher M. Danforth

Using human evaluation of 100, 000 words spread across 24 corpora in 10 languages diverse in origin and culture, we present evidence of a deep imprint of human sociality in language, observing that (1) the words of natural human language possess a universal positivity bias; (2) the estimated emotional content of words is consistent between languages under translation; and (3) this positivity bias is strongly independent of frequency of word usage.

Cultural Vocal Bursts Intensity Prediction Translation

Cannot find the paper you are looking for? You can Submit a new open access paper.