no code implementations • CVPR 2024 • Zhaohe Liao, Jiangtong Li, Li Niu, Liqing Zhang
The video aligner hierarchically selects the relevant video clips based on the question, while the answer aggregator deduces the answer to the question based on its sub-questions, with compositional consistency ensured by the information flow along question decomposition graph and the contrastive learning strategy.
1 code implementation • 7 Feb 2024 • Yuxuan Bian, Xuan Ju, Jiangtong Li, Zhijian Xu, Dawei Cheng, Qiang Xu
In this study, we present aLLM4TS, an innovative framework that adapts Large Language Models (LLMs) for time-series representation learning.
1 code implementation • 10 Nov 2023 • Yang Lei, Jiangtong Li, Dawei Cheng, Zhijun Ding, Changjun Jiang
Large language models (LLMs) have demonstrated great potential in the financial domain.
no code implementations • 27 Sep 2023 • Lingxiao Lu, Jiangtong Li, Bo Zhang, Li Niu
The goal of image composition is merging a foreground object into a background image to obtain a realistic composite image.
1 code implementation • 19 Sep 2023 • Jiangtong Li, Yuxuan Bian, Guoxuan Wang, Yang Lei, Dawei Cheng, Zhijun Ding, Changjun Jiang
The CFAPP is centered on large language models (LLMs) and augmented with additional modules to ensure multifaceted functionality in real-world application.
1 code implementation • 5 Aug 2023 • Linfeng Tan, Jiangtong Li, Li Niu, Liqing Zhang
The network comprises a $RGB$ harmonization backbone, an $Lab$ encoding module, and an $Lab$ control module.
Ranked #1 on Image Harmonization on HAdobe5k(1024$\times$1024)
1 code implementation • 4 Aug 2023 • Lingxiao Lu, Jiangtong Li, Junyan Cao, Li Niu, Liqing Zhang
Painterly image harmonization aims to insert photographic objects into paintings and obtain artistically coherent composite images.
no code implementations • ICCV 2023 • Jiangtong Li, Li Niu, Liqing Zhang
To tackle the challenge that the confounder in VideoQA is unobserved and non-enumerable in general, we propose a model-agnostic framework called Knowledge Proxy Intervention (KPI), which introduces an extra knowledge proxy variable in the causal graph to cut the backdoor path and remove the confounder.
1 code implementation • CVPR 2022 • Jiangtong Li, Li Niu, Liqing Zhang
We hope that Causal-VidQA can guide the research of video understanding from representation learning to deeper reasoning.
2 code implementations • 28 May 2022 • Li Niu, Qingyang Liu, Zhenchen Liu, Jiangtong Li
However, given a pair of scaled foreground and background, to enumerate all the reasonable locations, existing OPA model needs to place the foreground at each location on the background and pass the obtained composite image through the model one at a time, which is very time-consuming.
4 code implementations • 5 Jul 2021 • Liu Liu, Zhenchen Liu, Bo Zhang, Jiangtong Li, Li Niu, Qingyang Liu, Liqing Zhang
Image composition aims to generate realistic composite image by inserting an object from one image into another background image, where the placement (e. g., location, size, occlusion) of inserted object may be unreasonable, which would significantly degrade the quality of the composite image.
no code implementations • 29 Nov 2019 • Jiangtong Li, Zhixin Ling, Li Niu, Liqing Zhang
The goal of Sketch-Based Image Retrieval (SBIR) is using free-hand sketches to retrieve images of the same category from a natural image gallery.
no code implementations • 18 Sep 2019 • Jiangtong Li, Hai Zhao, Zuchao Li, Wei Bi, Xiaojiang Liu
Embedding from Language Models (ELMo) has shown to be effective for improving many natural language processing (NLP) tasks, and ELMo takes character information to compose word representation to train language models. However, the character is an insufficient and unnatural linguistic unit for word representation. Thus we introduce Embedding from Subword-aware Language Models (ESuLMo) which learns word representation from subwords using unsupervised segmentation over words. We show that ESuLMo can enhance four benchmark NLP tasks more effectively than ELMo, including syntactic dependency parsing, semantic role labeling, implicit discourse relation recognition and textual entailment, which brings a meaningful improvement over ELMo.
no code implementations • ACL 2019 • Fengshun Xiao, Jiangtong Li, Hai Zhao, Rui Wang, Kehai Chen
To integrate different segmentations with the state-of-the-art NMT model, Transformer, we propose lattice-based encoders to explore effective word or subword representation in an automatic way during training.
no code implementations • 22 Apr 2019 • Shu Jiang, Zhuosheng Zhang, Hai Zhao, Jiangtong Li, Yang Yang, Bao-liang Lu, Ning Xia
Chemical reaction practicality is the core task among all symbol intelligence based chemical information processing, for example, it provides indispensable clue for further automatic synthesis route inference.
no code implementations • 6 Nov 2018 • Sufeng Duan, Jiangtong Li, Hai Zhao
Rapidly developed neural models have achieved competitive performance in Chinese word segmentation (CWS) as their traditional counterparts.
1 code implementation • 6 Nov 2018 • Zhuosheng Zhang, Hai Zhao, Kangwei Ling, Jiangtong Li, Zuchao Li, Shexia He, Guohong Fu
Representation learning is the foundation of machine reading comprehension and inference.
no code implementations • COLING 2018 • Pengfei Zhu, Zhuosheng Zhang, Jiangtong Li, Yafang Huang, Hai Zhao
Traditional chatbots usually need a mass of human dialogue data, especially when using supervised machine learning method.
1 code implementation • COLING 2018 • Zhuosheng Zhang, Jiangtong Li, Pengfei Zhu, Hai Zhao, Gongshen Liu
In this paper, we formulate previous utterances into context using a proposed deep utterance aggregation model to form a fine-grained context representation.
Ranked #14 on Conversational Response Selection on E-commerce
no code implementations • SEMEVAL 2018 • Zhuosheng Zhang, Jiangtong Li, Hai Zhao, Bingjie Tang
This paper describes a hypernym discovery system for our participation in the SemEval-2018 Task 9, which aims to discover the best (set of) candidate hypernyms for input concepts or entities, given the search space of a pre-defined vocabulary.
Ranked #5 on Hypernym Discovery on Music domain