Search Results for author: Pritom Saha Akash

Found 8 papers, 1 papers with code

Domain Representative Keywords Selection: A Probabilistic Approach

1 code implementation Findings (ACL) 2022 Pritom Saha Akash, Jie Huang, Kevin Chen-Chuan Chang, Yunyao Li, Lucian Popa, ChengXiang Zhai

We propose a probabilistic approach to select a subset of a \textit{target domain representative keywords} from a candidate set, contrasting with a context domain.

Exploring Variational Graph Auto-Encoders for Extract Class Refactoring Recommendation

no code implementations16 Mar 2022 Pritom Saha Akash, Kevin Chen-Chuan Chang

Then, the variational graph auto-encoder is used to learn a vector representation for each method.

Coordinated Topic Modeling

no code implementations16 Oct 2022 Pritom Saha Akash, Jie Huang, Kevin Chen-Chuan Chang

It then uses the axes to model a corpus for easily understandable representation.

Unsupervised Open-domain Keyphrase Generation

no code implementations19 Jun 2023 Lam Thanh Do, Pritom Saha Akash, Kevin Chen-Chuan Chang

To solve this problem, we propose a seq2seq model that consists of two modules, namely \textit{phraseness} and \textit{informativeness} module, both of which can be built in an unsupervised and open-domain fashion.

Informativeness Keyphrase Generation

TopicAdapt- An Inter-Corpora Topics Adaptation Approach

no code implementations8 Oct 2023 Pritom Saha Akash, Trisha Das, Kevin Chen-Chuan Chang

Topic models are popular statistical tools for detecting latent semantic topics in a text corpus.

Topic Models

Let the Pretrained Language Models "Imagine" for Short Texts Topic Modeling

no code implementations24 Oct 2023 Pritom Saha Akash, Jie Huang, Kevin Chen-Chuan Chang

Besides, we provide a simple solution extending a neural topic model to reduce the effect of noisy out-of-topics text generation from PLMs.

Text Generation Topic Models

Long-form Question Answering: An Iterative Planning-Retrieval-Generation Approach

no code implementations15 Nov 2023 Pritom Saha Akash, Kashob Kumar Roy, Lucian Popa, Kevin Chen-Chuan Chang

From an extensive experiment on both an open domain and a technical domain QA dataset, we find that our model outperforms the state-of-the-art models on various textual and factual metrics for the LFQA task.

Long Form Question Answering Retrieval

Cannot find the paper you are looking for? You can Submit a new open access paper.