Search Results for author: Danyang Zhang

Found 11 papers, 7 papers with code

Rotation-robust Intersection over Union for 3D Object Detection

no code implementations ECCV 2020 Yu Zheng, Danyang Zhang, Sinan Xie, Jiwen Lu, Jie zhou

In this paper, we propose a Rotation-robust Intersection over Union ($ extit{RIoU}$) for 3D object detection, which aims to jointly learn the overlap of rotated bounding boxes.

3D Object Detection Object +1

Improving Academic Skills Assessment with NLP and Ensemble Learning

no code implementations23 Sep 2024 Xinyi Huang, Yingyi Wu, Danyang Zhang, Jiacheng Hu, Yujian Long

This study addresses the critical challenges of assessing foundational academic skills by leveraging advancements in natural language processing (NLP).

Ensemble Learning Pseudo Label +1

ChemDFM-X: Towards Large Multimodal Model for Chemistry

no code implementations20 Sep 2024 Zihan Zhao, Bo Chen, Jingpiao Li, Lu Chen, Liyang Wen, Pengyu Wang, Zichen Zhu, Danyang Zhang, Ziping Wan, Yansi Li, Zhongyang Dai, Xin Chen, Kai Yu

Rapid developments of AI tools are expected to offer unprecedented assistance to the research of natural science including chemistry.

Spider2-V: How Far Are Multimodal Agents From Automating Data Science and Engineering Workflows?

1 code implementation15 Jul 2024 Ruisheng Cao, Fangyu Lei, Haoyuan Wu, Jixuan Chen, Yeqiao Fu, Hongcheng Gao, Xinzhuang Xiong, Hanchong Zhang, Yuchen Mao, Wenjing Hu, Tianbao Xie, Hongshen Xu, Danyang Zhang, Sida Wang, Ruoxi Sun, Pengcheng Yin, Caiming Xiong, Ansong Ni, Qian Liu, Victor Zhong, Lu Chen, Kai Yu, Tao Yu

These tasks, derived from real-world use cases, evaluate the ability of a multimodal agent to perform data-related tasks by writing code and managing the GUI in enterprise data software systems.

Code Generation

OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments

1 code implementation11 Apr 2024 Tianbao Xie, Danyang Zhang, Jixuan Chen, Xiaochuan Li, Siheng Zhao, Ruisheng Cao, Toh Jing Hua, Zhoujun Cheng, Dongchan Shin, Fangyu Lei, Yitao Liu, Yiheng Xu, Shuyan Zhou, Silvio Savarese, Caiming Xiong, Victor Zhong, Tao Yu

Autonomous agents that accomplish complex computer tasks with minimal human interventions have the potential to transform human-computer interaction, significantly enhancing accessibility and productivity.

Benchmarking

Large Language Models Are Semi-Parametric Reinforcement Learning Agents

1 code implementation NeurIPS 2023 Danyang Zhang, Lu Chen, Situo Zhang, Hongshen Xu, Zihan Zhao, Kai Yu

By equipping the LLM with a long-term experience memory, REMEMBERER is capable of exploiting the experiences from the past episodes even for different task goals, which excels an LLM-based agent with fixed exemplars or equipped with a transient working memory.

Language Modeling Language Modelling +3

Mobile-Env: Building Qualified Evaluation Benchmarks for LLM-GUI Interaction

2 code implementations14 May 2023 Danyang Zhang, Zhennan Shen, Rui Xie, Situo Zhang, Tianbao Xie, Zihan Zhao, Siyuan Chen, Lu Chen, Hongshen Xu, Ruisheng Cao, Kai Yu

The Graphical User Interface (GUI) is pivotal for human interaction with the digital world, enabling efficient device control and the completion of complex tasks.

Language Modelling

Learning from Temporal Spatial Cubism for Cross-Dataset Skeleton-based Action Recognition

1 code implementation17 Jul 2022 Yansong Tang, Xingyu Liu, Xumin Yu, Danyang Zhang, Jiwen Lu, Jie zhou

Different from the conventional adversarial learning-based approaches for UDA, we utilize a self-supervision scheme to reduce the domain shift between two skeleton-based action datasets.

Action Recognition Self-Supervised Learning +2

COIN: A Large-scale Dataset for Comprehensive Instructional Video Analysis

no code implementations CVPR 2019 Yansong Tang, Dajun Ding, Yongming Rao, Yu Zheng, Danyang Zhang, Lili Zhao, Jiwen Lu, Jie zhou

There are substantial instructional videos on the Internet, which enables us to acquire knowledge for completing various tasks.

Action Detection

Cannot find the paper you are looking for? You can Submit a new open access paper.