no code implementations • ACL 2022 • Haw-Shiuan Chang, Andrew McCallum
The softmax layer produces the distribution based on the dot products of a single hidden state and the embeddings of words in the vocabulary.
no code implementations • sdp (COLING) 2022 • Kathryn Ricci, Haw-Shiuan Chang, Purujit Goyal, Andrew McCallum
Given a citation in the body of a research paper, cited text identification aims to find the sentences in the cited paper that are most relevant to the citing sentence.
no code implementations • 11 Jun 2024 • Haw-Shiuan Chang, Nanyun Peng, Mohit Bansal, Anil Ramakrishna, Tagyoung Chung
If a LLM's entropy is higher than the asymptotic entropy (i. e., the LLM is more uncertain than it should be), the THF model predicts a high hallucination hazard, which leads to a lower p threshold in REAL sampling.
1 code implementation • 21 Oct 2023 • Haw-Shiuan Chang, Nikhil Agarwal, Andrew McCallum
Specifically, the similarity structure of the global item embeddings in the softmax layer sometimes forces the single hidden state embedding to be close to new items when copying is a better choice, while sometimes forcing the hidden state to be close to the items from the input inappropriately.
1 code implementation • 8 Sep 2023 • Ronald Seoh, Haw-Shiuan Chang, Andrew McCallum
Many useful tasks on scientific documents, such as topic classification and citation prediction, involve corpora that span multiple scientific domains.
1 code implementation • 20 May 2023 • Haw-Shiuan Chang, Zonghai Yao, Alolika Gon, Hong Yu, Andrew McCallum
Is the output softmax layer, which is adopted by most language models (LMs), always the best way to compute the next word probability?
1 code implementation • 10 Oct 2022 • Haw-Shiuan Chang, Ruei-Yao Sun, Kathryn Ricci, Andrew McCallum
Ensembling BERT models often significantly improves accuracy, but at the cost of significantly more computation and memory footprint.
1 code implementation • 2 Jun 2022 • Hyeonsu B. Kang, Sheshera Mysore, Kevin Huang, Haw-Shiuan Chang, Thorben Prein, Andrew McCallum, Aniket Kittur, Elsa Olivetti
Exposure to ideas in domains outside a scientist's own may benefit her in reformulating existing research problems in novel ways and discovering new application domains for existing solution ideas.
1 code implementation • EMNLP 2021 • Ronald Seoh, Ian Birle, Mrinal Tak, Haw-Shiuan Chang, Brian Pinette, Alfred Hough
For many business applications, we often seek to analyze sentiments associated with any arbitrary aspects of commercial products, despite having a very limited amount of labels or even without any labels at all.
no code implementations • 29 Mar 2021 • Haw-Shiuan Chang, Amol Agrawal, Andrew McCallum
Most unsupervised NLP models represent each word with a single point or single region in semantic space, while the existing multi-sense word embeddings cannot represent longer word sequences like phrases or sentences.
1 code implementation • EACL 2021 • Haw-Shiuan Chang, Jiaming Yuan, Mohit Iyyer, Andrew McCallum
Our framework consists of two components: (1) a method that produces a set of candidate topics by predicting the centers of word clusters in the possible continuations, and (2) a text generation model whose output adheres to the chosen topics.
no code implementations • EACL 2021 • Rohan Paul, Haw-Shiuan Chang, Andrew McCallum
To address the violation of the USchema assumption, we propose multi-facet universal schema that uses a neural model to represent each sentence pattern as multiple facet embeddings and encourage one of these facet embeddings to be close to that of another sentence pattern if they co-occur with the same entity pair.
no code implementations • 24 Jun 2020 • Xin Luna Dong, Xiang He, Andrey Kan, Xi-An Li, Yan Liang, Jun Ma, Yifan Ethan Xu, Chenwei Zhang, Tong Zhao, Gabriel Blanco Saldana, Saurabh Deshpande, Alexandre Michetti Manduca, Jay Ren, Surender Pal Singh, Fan Xiao, Haw-Shiuan Chang, Giannis Karamanolakis, Yuning Mao, Yaqing Wang, Christos Faloutsos, Andrew McCallum, Jiawei Han
Can one build a knowledge graph (KG) for all products in the world?
no code implementations • 17 Nov 2019 • Haw-Shiuan Chang, Shankar Vembu, Sunil Mohan, Rheeya Uppaal, Andrew McCallum
Existing deep active learning algorithms achieve impressive sampling efficiency on natural language processing tasks.
no code implementations • WS 2019 • Sheshera Mysore, Zach Jensen, Edward Kim, Kevin Huang, Haw-Shiuan Chang, Emma Strubell, Jeffrey Flanigan, Andrew McCallum, Elsa Olivetti
Materials science literature contains millions of materials synthesis procedures described in unstructured natural language text.
1 code implementation • 31 Dec 2018 • Edward Kim, Zach Jensen, Alexander van Grootel, Kevin Huang, Matthew Staib, Sheshera Mysore, Haw-Shiuan Chang, Emma Strubell, Andrew McCallum, Stefanie Jegelka, Elsa Olivetti
Leveraging new data sources is a key step in accelerating the pace of materials design and discovery.
no code implementations • WS 2018 • Haw-Shiuan Chang, Amol Agrawal, Ananya Ganesh, Anirudha Desai, Vinayak Mathur, Alfred Hough, Andrew McCallum
Word sense induction (WSI), which addresses polysemy by unsupervised discovery of multiple word senses, resolves ambiguities for downstream NLP tasks and also makes word representations more interpretable.
no code implementations • 18 Nov 2017 • Sheshera Mysore, Edward Kim, Emma Strubell, Ao Liu, Haw-Shiuan Chang, Srikrishna Kompella, Kevin Huang, Andrew McCallum, Elsa Olivetti
In this work, we present a system for automatically extracting structured representations of synthesis procedures from the texts of materials science journal articles that describe explicit, experimental syntheses of inorganic compounds.
no code implementations • NAACL 2018 • Haw-Shiuan Chang, ZiYun Wang, Luke Vilnis, Andrew McCallum
Modeling hypernymy, such as poodle is-a dog, is an important generalization aid to many NLP tasks, such as entailment, coreference, relation extraction, and question answering.
1 code implementation • NeurIPS 2017 • Haw-Shiuan Chang, Erik Learned-Miller, Andrew McCallum
Self-paced learning and hard example mining re-weight training instances to improve learning accuracy.