no code implementations • NAACL (ACL) 2022 • Judith Gaspers, Anoop Kumar, Greg Ver Steeg, Aram Galstyan
Spoken Language Understanding (SLU) models in industry applications are usually trained offline on historic data, but have to perform well on incoming user requests after deployment.
1 code implementation • Findings (NAACL) 2022 • Elan Markowitz, Keshav Balasubramanian, Mehrnoosh Mirtaheri, Murali Annavaram, Aram Galstyan, Greg Ver Steeg
Knowledge graphs (KGs) often represent knowledge bases that are incomplete.
no code implementations • ICML 2020 • Rob Brekelmans, Vaden Masrani, Frank Wood, Greg Ver Steeg, Aram Galstyan
While the Evidence Lower Bound (ELBO) has become a ubiquitous objective for variational inference, the recently proposed Thermodynamic Variational Objective (TVO) leverages thermodynamic integration to provide a tighter and more general family of bounds.
no code implementations • 7 Nov 2024 • Neal Lawton, Aram Galstyan, Greg Ver Steeg
A popular method for Neural Architecture Search (NAS) is based on growing networks via small local changes to the network's architecture called network morphisms.
no code implementations • 26 Oct 2024 • Sullam Jeoung, Goeric Huybrechts, Bhavana Ganesh, Aram Galstyan, Sravan Bodapati
Understanding long-form video content presents significant challenges due to its temporal complexity and the substantial computational resources required.
no code implementations • 12 Oct 2024 • Jongwoo Ko, Saket Dingliwal, Bhavana Ganesh, Sailik Sengupta, Sravan Bodapati, Aram Galstyan
Direct alignment algorithms (DAAs), such as direct preference optimization (DPO), have become popular alternatives for Reinforcement Learning from Human Feedback (RLHF) due to their simplicity, efficiency, and stability.
no code implementations • 9 Oct 2024 • Neal Lawton, Aishwarya Padmakumar, Judith Gaspers, Jack FitzGerald, Anoop Kumar, Greg Ver Steeg, Aram Galstyan
In this paper we introduce QuAILoRA, a quantization-aware initialization for LoRA that mitigates this negative impact by decreasing quantization errors at initialization.
no code implementations • 7 Oct 2024 • Tao Meng, Ninareh Mehrabi, Palash Goyal, Anil Ramakrishna, Aram Galstyan, Richard Zemel, Kai-Wei Chang, Rahul Gupta, Charith Peris
We propose a constraint learning schema for fine-tuning Large Language Models (LLMs) with attribute control.
no code implementations • 7 Oct 2024 • Fei Wang, Ninareh Mehrabi, Palash Goyal, Rahul Gupta, Kai-Wei Chang, Aram Galstyan
Starting from a set of pre-defined principles in hand, Data Advisor monitors the status of the generated data, identifies weaknesses in the current dataset, and advises the next iteration of data generation accordingly.
1 code implementation • 31 Jul 2024 • Elan Markowitz, Anil Ramakrishna, Jwala Dhamala, Ninareh Mehrabi, Charith Peris, Rahul Gupta, Kai-Wei Chang, Aram Galstyan
Knowledge graphs (KGs) complement Large Language Models (LLMs) by providing reliable, structured, domain-specific, and up-to-date external knowledge.
no code implementations • 25 Jun 2024 • Jinghan Jia, Abi Komma, Timothy Leffel, Xujun Peng, Ajay Nagesh, Tamer Soliman, Aram Galstyan, Anoop Kumar
In task-oriented conversational AI evaluation, unsupervised methods poorly correlate with human judgments, and supervised approaches lack generalization.
no code implementations • 24 Feb 2024 • Yao Qiang, Subhrangshu Nandi, Ninareh Mehrabi, Greg Ver Steeg, Anoop Kumar, Anna Rumshisky, Aram Galstyan
However, their performance on sequence labeling tasks such as intent classification and slot filling (IC-SF), which is a central component in personal assistant systems, lags significantly behind discriminative models.
1 code implementation • 17 Feb 2024 • Tianyi Lorena Yan, Fei Wang, James Y. Huang, Wenxuan Zhou, Fan Yin, Aram Galstyan, Wenpeng Yin, Muhao Chen
Instruction tuning has been used as a promising approach to improve the performance of large language models (LLMs) on unseen tasks.
no code implementations • 19 Dec 2023 • Anaelia Ovalle, Ninareh Mehrabi, Palash Goyal, Jwala Dhamala, Kai-Wei Chang, Richard Zemel, Aram Galstyan, Yuval Pinter, Rahul Gupta
Our paper is the first to link LLM misgendering to tokenization and deficient neopronoun grammar, indicating that LLMs unable to correctly treat neopronouns as pronouns are more prone to misgender.
no code implementations • 16 Nov 2023 • Ninareh Mehrabi, Palash Goyal, Anil Ramakrishna, Jwala Dhamala, Shalini Ghosh, Richard Zemel, Kai-Wei Chang, Aram Galstyan, Rahul Gupta
With the recent surge of language models in different applications, attention to safety and robustness of these models has gained significant importance.
no code implementations • 8 Nov 2023 • Junyi Li, Ninareh Mehrabi, Charith Peris, Palash Goyal, Kai-Wei Chang, Aram Galstyan, Richard Zemel, Rahul Gupta
Large language models (LLMs) are known to generate biased responses where the opinions of certain groups and populations are underrepresented.
no code implementations • 8 Aug 2023 • Ninareh Mehrabi, Palash Goyal, Christophe Dupuy, Qian Hu, Shalini Ghosh, Richard Zemel, Kai-Wei Chang, Aram Galstyan, Rahul Gupta
In this work, we propose an automatic red teaming framework that evaluates a given black-box model and exposes its vulnerabilities against unsafe and inappropriate content generation.
no code implementations • 15 Jun 2023 • Myrl G. Marmarelis, Greg Ver Steeg, Aram Galstyan, Fred Morstatter
We present a simple approach to partial identification using existing causal sensitivity models and show empirically that Caus-Modens gives tighter outcome intervals, as measured by the necessary interval size to achieve sufficient coverage.
no code implementations • 9 Jun 2023 • Elan Markowitz, Ziyan Jiang, Fan Yang, Xing Fan, Tony Chen, Greg Ver Steeg, Aram Galstyan
We propose in this work to unify these approaches: Using information from interactions in other domains as well as external knowledge graphs to make predictions in a new domain that would be impossible with either information source alone.
no code implementations • 6 Jun 2023 • Abishek Komma, Nagesh Panyam Chandrasekarasastry, Timothy Leffel, Anuj Goyal, Angeliki Metallinou, Spyros Matsoukas, Aram Galstyan
Measurement of interaction quality is a critical task for the improvement of spoken dialog systems.
no code implementations • 30 May 2023 • Mehrnoosh Mirtaheri, Mohammad Rostami, Aram Galstyan
Temporal knowledge graph (TKG) completion models typically rely on having access to the entire graph during training.
1 code implementation • 30 May 2023 • Umang Gupta, Aram Galstyan, Greg Ver Steeg
This can be a drawback for low-resource applications and training with differential-privacy constraints, where excessive noise may be introduced during finetuning.
no code implementations • 26 May 2023 • Neal Lawton, Anoop Kumar, Govind Thattai, Aram Galstyan, Greg Ver Steeg
Parameter-efficient tuning (PET) methods fit pre-trained language models (PLMs) to downstream tasks by either computing a small compressed update for a subset of model parameters, or appending and fine-tuning a small number of new model parameters to the pre-trained network.
1 code implementation • 26 May 2023 • Kuan-Hao Huang, Varun Iyer, I-Hung Hsu, Anoop Kumar, Kai-Wei Chang, Aram Galstyan
Paraphrase generation is a long-standing task in natural language processing (NLP).
no code implementations • 23 May 2023 • Zheng Chen, Ziyan Jiang, Fan Yang, Eunah Cho, Xing Fan, Xiaojiang Huang, Yanbin Lu, Aram Galstyan
This paper presents our "Collaborative Query Rewriting" approach, which specifically addresses the task of rewriting new user interactions that have not been previously observed in the user's history.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +9
no code implementations • 18 May 2023 • Arghya Datta, Subhrangshu Nandi, Jingcheng Xu, Greg Ver Steeg, He Xie, Anoop Kumar, Aram Galstyan
We formulate the model stability problem by studying how the predictions of a model change, even when it is retrained on the same data, as a consequence of stochasticity in the training process.
no code implementations • 17 May 2023 • Anaelia Ovalle, Palash Goyal, Jwala Dhamala, Zachary Jaggers, Kai-Wei Chang, Aram Galstyan, Richard Zemel, Rahul Gupta
Transgender and non-binary (TGNB) individuals disproportionately experience discrimination and exclusion from daily life.
1 code implementation • 12 May 2023 • Sarik Ghazarian, Yijia Shao, Rujun Han, Aram Galstyan, Nanyun Peng
We take the first step by focusing on event commonsense that considers events and their relations, and is crucial in both dialogues and general commonsense reasoning.
no code implementations • 17 Nov 2022 • Ninareh Mehrabi, Palash Goyal, Apurv Verma, Jwala Dhamala, Varun Kumar, Qian Hu, Kai-Wei Chang, Richard Zemel, Aram Galstyan, Rahul Gupta
Natural language often contains ambiguities that can lead to misinterpretation and miscommunication.
no code implementations • 2 Nov 2022 • Kuan-Hao Huang, Varun Iyer, Anoop Kumar, Sriram Venkatapathy, Kai-Wei Chang, Aram Galstyan
In this paper, we demonstrate that leveraging Abstract Meaning Representations (AMR) can greatly improve the performance of unsupervised syntactically controlled paraphrase generation.
no code implementations • 7 Oct 2022 • Jwala Dhamala, Varun Kumar, Rahul Gupta, Kai-Wei Chang, Aram Galstyan
We present a systematic analysis of the impact of decoding algorithms on LM fairness, and analyze the trade-off between fairness, diversity and quality.
no code implementations • 13 May 2022 • Hrayr Harutyunyan, Greg Ver Steeg, Aram Galstyan
Remarkably, PAC-Bayes, single-draw and expected squared generalization gap bounds that depend on information in pairs of examples exist.
1 code implementation • NAACL 2022 • Ninareh Mehrabi, Ahmad Beirami, Fred Morstatter, Aram Galstyan
Existing work to generate such attacks is either based on human-generated attacks which is costly and not scalable or, in case of automatic attacks, the attack vector does not conform to human-like language, which can be detected using a language model loss.
no code implementations • 24 Apr 2022 • Myrl G. Marmarelis, Elizabeth Haddad, Andrew Jesson, Neda Jahanshad, Aram Galstyan, Greg Ver Steeg
Sensitivity analyses provide principled ways to give bounds on causal estimates when confounding variables are hidden.
no code implementations • ACL 2022 • Yang Trista Cao, Yada Pruksachatkun, Kai-Wei Chang, Rahul Gupta, Varun Kumar, Jwala Dhamala, Aram Galstyan
Multiple metrics have been introduced to measure fairness in various natural language processing tasks.
no code implementations • Findings (ACL) 2022 • Umang Gupta, Jwala Dhamala, Varun Kumar, Apurv Verma, Yada Pruksachatkun, Satyapriya Krishna, Rahul Gupta, Kai-Wei Chang, Greg Ver Steeg, Aram Galstyan
Language models excel at generating coherent text, and model compression techniques such as knowledge distillation have enabled their use in resource-constrained settings.
1 code implementation • 19 Mar 2022 • Marcin Abram, Keith Burghardt, Greg Ver Steeg, Aram Galstyan, Remi Dingreville
The identification and classification of transitions in topological and microstructural regimes in pattern-forming processes are critical for understanding and fabricating microstructurally precise novel materials in many application domains.
1 code implementation • ACL 2022 • Sarik Ghazarian, Nuan Wen, Aram Galstyan, Nanyun Peng
We also show that DEAM can distinguish between coherent and incoherent dialogues generated by baseline manipulations, whereas those baseline models cannot detect incoherent examples generated by DEAM.
1 code implementation • CVPR 2022 • Tigran Galstyan, Hrayr Harutyunyan, Hrant Khachatrian, Greg Ver Steeg, Aram Galstyan
On Camelyon-17, domain-invariance degrades the quality of representations on unseen domains.
1 code implementation • NeurIPS 2021 • Sami Abu-El-Haija, Hesham Mostafa, Marcel Nassar, Valentino Crespi, Greg Ver Steeg, Aram Galstyan
Recent improvements in the performance of state-of-the-art (SOTA) methods for Graph Representational Learning (GRL) have come at the cost of significant computational resource requirements for training, e. g., for calculating gradients via backprop over many data epochs.
1 code implementation • NeurIPS 2021 • Greg Ver Steeg, Aram Galstyan
Auxiliary neural models can learn to speed up MCMC, but the overhead for training the extra model can be prohibitive.
no code implementations • 9 Oct 2021 • Mohammad Rostami, Aram Galstyan
Humans continually expand their learned knowledge to new domains and learn new concepts without any interference with past learned experiences.
1 code implementation • NeurIPS 2021 • Hrayr Harutyunyan, Maxim Raginsky, Greg Ver Steeg, Aram Galstyan
We derive information-theoretic generalization bounds for supervised learning algorithms based on the information contained in predictions rather than in the output of the training algorithm.
no code implementations • ICCV 2021 • Jiawei Ma, Hanchen Xie, Guangxing Han, Shih-Fu Chang, Aram Galstyan, Wael Abd-Almageed
In this paper, we focus on the design of training strategy to obtain an elemental representation such that the prototype of each novel class can be estimated from a few labeled samples.
1 code implementation • NAACL (TrustNLP) 2022 • Ninareh Mehrabi, Umang Gupta, Fred Morstatter, Greg Ver Steeg, Aram Galstyan
The widespread use of Artificial Intelligence (AI) in consequential domains, such as healthcare and parole decision-making systems, has drawn intense scrutiny on the fairness of these methods.
no code implementations • 4 Jul 2021 • Mohammad Rostami, Aram Galstyan
We introduce a new domain adaptation method which induces large margins between different classes in an embedding space.
1 code implementation • 1 Jul 2021 • Vaden Masrani, Rob Brekelmans, Thang Bui, Frank Nielsen, Aram Galstyan, Greg Ver Steeg, Frank Wood
Many common machine learning methods involve the geometric annealing path, a sequence of intermediate densities between two distributions of interest constructed using the geometric average.
1 code implementation • NAACL 2021 • Sarik Ghazarian, Zixi Liu, Akash SM, Ralph Weischedel, Aram Galstyan, Nanyun Peng
We propose to tackle these issues by generating a more comprehensive set of implausible stories using {\em plots}, which are structured representations of controllable factors used to generate stories.
no code implementations • EMNLP 2021 • Ninareh Mehrabi, Pei Zhou, Fred Morstatter, Jay Pujara, Xiang Ren, Aram Galstyan
In addition, we analyze two downstream models that use ConceptNet as a source for commonsense knowledge and find the existence of biases in those models as well.
1 code implementation • ICLR Workshop GTRL 2021 • Sami Abu-El-Haija, Valentino Crespi, Greg Ver Steeg, Aram Galstyan
We consider two popular Graph Representation Learning (GRL) methods: message passing for node classification and network embedding for link prediction.
1 code implementation • ICLR 2021 • Elan Markowitz, Keshav Balasubramanian, Mehrnoosh Mirtaheri, Sami Abu-El-Haija, Bryan Perozzi, Greg Ver Steeg, Aram Galstyan
We propose Graph Traversal via Tensor Functionals(GTTF), a unifying meta-algorithm framework for easing the implementation of diverse graph algorithms and enabling transparent and efficient scaling to large graphs.
no code implementations • NAACL 2021 • Sarik Ghazarian, Zixi Liu, Tuhin Chakrabarty, Xuezhe Ma, Aram Galstyan, Nanyun Peng
Having engaging and informative conversations with users is the utmost goal for open-domain conversational systems.
no code implementations • 1 Jan 2021 • Mohammad Rostami, Aram Galstyan
Large margins in the source domain help to reduce the effect of ``domain shift'' on the performance of a trained classifier in the target domain.
no code implementations • NeurIPS Workshop DL-IG 2020 • Rob Brekelmans, Frank Nielsen, Alireza Makhzani, Aram Galstyan, Greg Ver Steeg
The exponential family is well known in machine learning and statistical physics as the maximum entropy distribution subject to a set of observed constraints, while the geometric mixture path is common in MCMC methods such as annealed importance sampling.
1 code implementation • 16 Dec 2020 • Ninareh Mehrabi, Muhammad Naveed, Fred Morstatter, Aram Galstyan
Algorithmic fairness has attracted significant attention in recent years, with many quantitative measures suggested for characterizing the fairness of different machine learning algorithms.
2 code implementations • NeurIPS Workshop DL-IG 2020 • Rob Brekelmans, Vaden Masrani, Thang Bui, Frank Wood, Aram Galstyan, Greg Ver Steeg, Frank Nielsen
Annealed importance sampling (AIS) is the gold standard for estimating partition functions or marginal likelihoods, corresponding to importance sampling over a path of distributions between a tractable base and an unnormalized target.
no code implementations • 30 Nov 2020 • Hanchen Xie, Mohamed E. Hussein, Aram Galstyan, Wael Abd-Almageed
We also show that MUSCLE has the potential to boost the classification performance when used in the fine-tuning phase for a model pre-trained only on unlabeled data.
no code implementations • AKBC 2021 • Mehrnoosh Mirtaheri, Mohammad Rostami, Xiang Ren, Fred Morstatter, Aram Galstyan
Most real-world knowledge graphs are characterized by a long-tail relation frequency distribution where a significant fraction of relations occurs only a handful of times.
no code implementations • 4 Sep 2020 • Akira Matsui, Emilio Ferrara, Fred Morstatter, Andres Abeliuk, Aram Galstyan
In this study, we propose the use of a computational framework to identify clusters of underperforming workers using clickstream trajectories.
no code implementations • 29 Jul 2020 • James O' Neill, Greg Ver Steeg, Aram Galstyan
This paper proposes \textit{layer fusion} - a model compression technique that discovers which weights to combine and then fuses weights of similar fully-connected, convolutional and attention layers.
no code implementations • 10 Jul 2020 • Tigran Galstyan, Hrant Khachatrian, Greg Ver Steeg, Aram Galstyan
Investigation of machine learning algorithms robust to changes between the training and test distributions is an active area of research.
1 code implementation • 1 Jul 2020 • Rob Brekelmans, Vaden Masrani, Frank Wood, Greg Ver Steeg, Aram Galstyan
We propose to choose intermediate distributions using equal spacing in the moment parameters of our exponential family, which matches grid search performance and allows the schedule to adaptively update over the course of training.
1 code implementation • 1 Jul 2020 • Mohammad Rostami, Aram Galstyan
We develop an algorithm to improve the performance of a pre-trained model under concept shift without retraining the model from scratch when only unannotated samples of initial concepts are accessible.
no code implementations • 5 May 2020 • Myrl G. Marmarelis, Greg Ver Steeg, Aram Galstyan
A multivariate Hawkes process enables self- and cross-excitations through a triggering matrix that behaves like an asymmetrical covariance structure, characterizing pairwise interactions between the event types.
no code implementations • ACL 2021 • Woojeong Jin, Rahul Khanna, Suji Kim, Dong-Ho Lee, Fred Morstatter, Aram Galstyan, Xiang Ren
In this work, we aim to formulate a task, construct a dataset, and provide benchmarks for developing methods for event forecasting with large volumes of unstructured text data.
1 code implementation • ICML 2020 • Hrayr Harutyunyan, Kyle Reing, Greg Ver Steeg, Aram Galstyan
In the presence of noisy or incorrect labels, neural networks have the undesirable tendency to memorize information about the noise.
no code implementations • 9 Feb 2020 • Shushan Arakelyan, Sima Arasteh, Christophe Hauser, Erik Kline, Aram Galstyan
We demonstrate the versatility of this approach by using our representations to solve two semantically different binary analysis tasks - functional algorithm classification and vulnerability discovery.
2 code implementations • 4 Nov 2019 • Sarik Ghazarian, Ralph Weischedel, Aram Galstyan, Nanyun Peng
In this paper, we investigate the possibility and efficacy of estimating utterance-level engagement and define a novel metric, {\em predictive engagement}, for automatic evaluation of open-domain dialogue systems.
no code implementations • IJCNLP 2019 • Sahil Garg, Aram Galstyan, Greg Ver Steeg, Guillermo Cecchi
Recently, kernelized locality sensitive hashcodes have been successfully employed as representations of natural language text, especially showing high relevance to biomedical relation extraction tasks.
1 code implementation • 24 Oct 2019 • Ninareh Mehrabi, Thamme Gowda, Fred Morstatter, Nanyun Peng, Aram Galstyan
We study the bias in several state-of-the-art named entity recognition (NER) models---specifically, a difference in the ability to recognize male and female names as PERSON entity types.
1 code implementation • CONLL 2019 • Rujun Han, I-Hung Hsu, Mu Yang, Aram Galstyan, Ralph Weischedel, Nanyun Peng
We propose a novel deep structured learning framework for event temporal relation extraction.
2 code implementations • 17 Sep 2019 • Amir Ghasemian, Homa Hosseinmardi, Aram Galstyan, Edoardo M. Airoldi, Aaron Clauset
These results indicate that the state-of-the-art for link prediction comes from combining individual algorithms, which achieves nearly optimal predictions.
no code implementations • 9 Sep 2019 • Sahil Garg, Aram Galstyan, Greg Ver Steeg, Guillermo Cecchi
Recently, kernelized locality sensitive hashcodes have been successfully employed as representations of natural language text, especially showing high relevance to biomedical relation extraction tasks.
2 code implementations • 23 Aug 2019 • Ninareh Mehrabi, Fred Morstatter, Nripsuta Saxena, Kristina Lerman, Aram Galstyan
With the commercialization of these systems, researchers are becoming aware of the biases that these applications can contain and have attempted to address them.
1 code implementation • WS 2019 • Hrant Khachatrian, Lilit Nersisyan, Karen Hambardzumyan, Tigran Galstyan, Anna Hakobyan, Arsen Arakelyan, Andrey Rzhetsky, Aram Galstyan
Automatic extraction of relations and interactions between biological entities from scientific literature remains an extremely challenging problem in biomedical information extraction and natural language processing in general.
1 code implementation • Nature Scientific Data 2019 • Hrayr Harutyunyan, Hrant Khachatrian, David C. Kale, Greg Ver Steeg, Aram Galstyan
Health care is one of the most exciting frontiers in data mining and machine learning.
2 code implementations • 30 May 2019 • Hrayr Harutyunyan, Daniel Moyer, Hrant Khachatrian, Greg Ver Steeg, Aram Galstyan
Estimating the covariance structure of multivariate time series is a fundamental problem with a wide-range of real-world applications -- from financial modeling to fMRI analysis.
3 code implementations • 30 Apr 2019 • Sami Abu-El-Haija, Bryan Perozzi, Amol Kapoor, Nazanin Alipourfard, Kristina Lerman, Hrayr Harutyunyan, Greg Ver Steeg, Aram Galstyan
Existing popular methods for semi-supervised learning with Graph Neural Networks (such as the Graph Convolutional Network) provably cannot learn a general class of neighborhood mixing relationships.
no code implementations • WS 2019 • Sarik Ghazarian, Johnny Tian-Zheng Wei, Aram Galstyan, Nanyun Peng
Despite advances in open-domain dialogue systems, automatic evaluation of such systems is still a challenging problem.
1 code implementation • NeurIPS 2019 • Rob Brekelmans, Daniel Moyer, Aram Galstyan, Greg Ver Steeg
The noise is constructed in a data-driven fashion that does not require restrictive distributional assumptions.
1 code implementation • 4 Feb 2019 • Mehrnoosh Mirtaheri, Sami Abu-El-Haija, Fred Morstatter, Greg Ver Steeg, Aram Galstyan
Because of the speed and relative anonymity offered by social platforms such as Twitter and Telegram, social media has become a preferred platform for scammers who wish to spread false hype about the cryptocurrency they are trying to pump.
1 code implementation • NeurIPS 2018 • Daniel Moyer, Shuyang Gao, Rob Brekelmans, Greg Ver Steeg, Aram Galstyan
Representations of data that are invariant to changes in specified factors are useful for a wide range of problems: removing potential biases in prediction problems, controlling the effects of covariates, and disentangling meaningful factors of variation.
no code implementations • 17 May 2018 • Neal Lawton, Aram Galstyan, Greg Ver Steeg
Coordinate ascent variational inference is an important algorithm for inference in probabilistic models, but it is slow because it updates only a single variable at a time.
1 code implementation • 8 May 2018 • Palash Goyal, Homa Hosseinmardi, Emilio Ferrara, Aram Galstyan
Here we propose a novel embedding method that uses both network structure and edge attributes to learn better network representations.
Social and Information Networks
no code implementations • 26 Apr 2018 • Sahil Garg, Irina Rish, Guillermo Cecchi, Palash Goyal, Sarik Ghazarian, Shuyang Gao, Greg Ver Steeg, Aram Galstyan
We also derive a novel lower bound on mutual information, used as a model-selection criterion favoring representations with better alignment between the utterances of participants in a collaborative dialogue setting, as well as higher predictability of the generated responses.
no code implementations • 16 Feb 2018 • Shuyang Gao, Rob Brekelmans, Greg Ver Steeg, Aram Galstyan
Advances in unsupervised learning enable reconstruction and generation of samples from complex distributions, but this success is marred by the inscrutability of the representations learned.
no code implementations • 11 Jan 2018 • Sahil Garg, Greg Ver Steeg, Aram Galstyan
Natural language processing often involves computations with semantic or syntactic graphs to facilitate sophisticated reasoning based on structural relationships.
1 code implementation • 10 Nov 2017 • Sahil Garg, Aram Galstyan, Greg Ver Steeg, Irina Rish, Guillermo Cecchi, Shuyang Gao
Here we propose to use random subspaces of KLSH codes for efficiently constructing an explicit representation of NLP structures suitable for general classification methods.
1 code implementation • 10 Oct 2017 • Greg Ver Steeg, Rob Brekelmans, Hrayr Harutyunyan, Aram Galstyan
Scientists often seek simplified representations of complex systems to facilitate prediction and understanding.
no code implementations • 9 Oct 2017 • Wenzhe Li, Dong Guo, Greg Ver Steeg, Aram Galstyan
Many real-world networks are complex dynamical systems, where both local (e. g., changing node attributes) and global (e. g., changing network topology) processes unfold over time.
3 code implementations • NeurIPS 2019 • Greg Ver Steeg, Hrayr Harutyunyan, Daniel Moyer, Aram Galstyan
We also use our approach for estimating covariance structure for a number of real-world datasets and show that it consistently outperforms state-of-the-art estimators at a fraction of the computational cost.
11 code implementations • 22 Mar 2017 • Hrayr Harutyunyan, Hrant Khachatrian, David C. Kale, Greg Ver Steeg, Aram Galstyan
Health care is one of the most exciting frontiers in data mining and machine learning.
no code implementations • 22 Jun 2016 • Kyle Reing, David C. Kale, Greg Ver Steeg, Aram Galstyan
Many predictive tasks, such as diagnosing a patient based on their medical chart, are ultimately defined by the decisions of human experts.
1 code implementation • NeurIPS 2016 • Shuyang Gao, Greg Ver Steeg, Aram Galstyan
We demonstrate that approximations made by existing methods are based on unrealistic assumptions.
1 code implementation • 7 Jun 2016 • Greg Ver Steeg, Shuyang Gao, Kyle Reing, Aram Galstyan
Measuring the relationship between any pair of variables is a rich and active area of research that is central to scientific practice.
no code implementations • 2 May 2016 • Emilio Ferrara, Wen-Qiang Wang, Onur Varol, Alessandro Flammini, Aram Galstyan
We present a machine learning framework that leverages a mixture of metadata, network, and temporal features to detect extremist users, and predict content adopters and interaction reciprocity in social media.
no code implementations • 20 Jan 2016 • V. S. Subrahmanian, Amos Azaria, Skylar Durst, Vadim Kagan, Aram Galstyan, Kristina Lerman, Linhong Zhu, Emilio Ferrara, Alessandro Flammini, Filippo Menczer, Andrew Stevens, Alexander Dekhtyar, Shuyang Gao, Tad Hogg, Farshad Kooti, Yan Liu, Onur Varol, Prashant Shiralkar, Vinod Vydiswaran, Qiaozhu Mei, Tim Hwang
A number of organizations ranging from terrorist groups such as ISIS to politicians and nation states reportedly conduct explicit campaigns to influence opinion on social media, posing a risk to democratic processes.
1 code implementation • 4 Dec 2015 • Sahil Garg, Aram Galstyan, Ulf Hermjakob, Daniel Marcu
We advance the state of the art in biomolecular interaction extraction with three contributions: (i) We show that deep, Abstract Meaning Representations (AMR) significantly improve the accuracy of a biomolecular interaction extraction system when compared to a baseline that relies solely on surface- and syntax-based features; (ii) In contrast with previous approaches that infer relations on a sentence-by-sentence basis, we expand our framework to enable consistent predictions over sets of sentences (documents); (iii) We further modify and expand a graph kernel learning framework to enable concurrent exploitation of automatically induced AMR (semantic) and dependency structure (syntactic) representations.
no code implementations • 18 Oct 2015 • Yoon-Sik Cho, Greg Ver Steeg, Emilio Ferrara, Aram Galstyan
We validate the proposed framework on two problems: prediction of social interactions from user attributes and behaviors, and behavior prediction exploiting network information.
2 code implementations • 8 Jul 2015 • Greg Ver Steeg, Aram Galstyan
Each layer of the sieve recovers a single latent factor that is maximally informative about multivariate dependence in the data.
no code implementations • 1 Dec 2014 • Shuyang Gao, Greg Ver Steeg, Aram Galstyan
We revisit some of the previous studies that reported strong signatures of stylistic accommodation, and find that a significant part of the observed coordination can be attributed to a simple confounding effect - length coordination.
no code implementations • 13 Nov 2014 • Linhong Zhu, Dong Guo, Junming Yin, Greg Ver Steeg, Aram Galstyan
We propose a scalable temporal latent space model for link prediction in dynamic social networks, where the goal is to predict links over time based on a sequence of previous graph snapshots.
4 code implementations • 7 Nov 2014 • Shuyang Gao, Greg Ver Steeg, Aram Galstyan
We demonstrate that a popular class of nonparametric mutual information (MI) estimators based on k-nearest-neighbor graphs requires number of samples that scales exponentially with the true MI.
no code implementations • 3 Nov 2014 • Armen E. Allahverdyan, Aram Galstyan
We consider active maximum a posteriori (MAP) inference problem for Hidden Markov Models (HMM), where, given an initial MAP estimate of the hidden sequence, we select to label certain states in the sequence to improve the estimation accuracy of the remaining states.
3 code implementations • 27 Oct 2014 • Greg Ver Steeg, Aram Galstyan
We present bounds on how informative a representation is about input data.
3 code implementations • NeurIPS 2014 • Greg Ver Steeg, Aram Galstyan
We introduce a method to learn a hierarchy of successively more abstract representations of complex data based on optimizing an information-theoretic objective.
no code implementations • 24 Feb 2014 • Linhong Zhu, Aram Galstyan, James Cheng, Kristina Lerman
We further investigate the evolution of user-level sentiments and latent feature vectors in an online framework and devise an efficient online algorithm to sequentially update the clustering of tweets, users and features with newly arrived data.
no code implementations • NeurIPS 2011 • Armen E. Allahverdyan, Aram Galstyan
We present an asymptotic analysis of Viterbi Training (VT) and contrast it with a more conventional Maximum Likelihood (ML) approach to parameter estimation in Hidden Markov Models.
no code implementations • 2 Dec 2013 • Greg Ver Steeg, Cristopher Moore, Aram Galstyan, Armen E. Allahverdyan
It predicts a first-order detectability transition whenever $q > 2$, while the finite-temperature cavity method shows that this is the case only when $q > 4$.
no code implementations • 15 Oct 2013 • Greg Ver Steeg, Aram Galstyan, Fei Sha, Simon DeDeo
We propose a novel method for clustering data which is grounded in information-theoretic principles and requires no parametric assumptions.
no code implementations • 5 Aug 2013 • Ardeshir Kianercy, Aram Galstyan
This paper presents a model of network formation in repeated games where the players adapt their strategies and network ties simultaneously using a simple reinforcement-learning scheme.
no code implementations • 12 Feb 2013 • Yoon-Sik Cho, Aram Galstyan, P. Jeffrey Brantingham, George Tita
We propose a latent self-exciting point process model that describes geographically distributed interactions between pairs of entities.
3 code implementations • 22 Aug 2012 • Greg Ver Steeg, Aram Galstyan
The fundamental building block of social influence is for one person to elicit a response in another.
Social and Information Networks Physics and Society Applications
1 code implementation • 12 Oct 2011 • Greg Ver Steeg, Aram Galstyan
Recent research has explored the increasingly important role of social media by examining the dynamics of individual and group behavior, characterizing patterns of information diffusion, and identifying influential individuals.
Social and Information Networks Physics and Society Applications