Inspired by these findings, we conduct supervised learning tasks to estimate the usefulness of non-click results with brain signals and conventional information (i. e., content and context factors).
Reading comprehension is a complex cognitive process involving many human brain activities.
Compared with previous DR models that use brute-force search, JPQ almost matches the best retrieval performance with 30x compression on index size.
To the best of our knowledge, this is the largest real-world interaction dataset for personalized recommendation.
ADORE replaces the widely-adopted static hard negative sampling method with a dynamic one to directly optimize the ranking performance.
We participated in the two case law tasks, i. e., the legal case retrieval task and the legal case entailment task.
Through this process, it teaches the DR model how to retrieve relevant documents from the entire corpus instead of how to rerank a potentially biased sample of documents.
Although exact term match between queries and documents is the dominant method to perform first-stage retrieval, we propose a different approach, called RepBERT, to represent documents and queries with fixed-length contextualized embeddings.
The framework encourages two modules to complement each other in generating effective and explainable recommendation: 1) inductive rules, mined from item-centric knowledge graphs, summarize common multi-hop relational patterns for inferring different item associations and provide human-readable explanation for model prediction; 2) recommendation module can be augmented by induced rules and thus have better generalization ability dealing with the cold-start issue.
In this paper, we focus on the problem of phrase-level sentiment polarity labelling and attempt to bridge the gap between phrase-level and review-level sentiment analysis.