Search Results for author: Lalithkumar Seenivasan

Found 9 papers, 9 papers with code

Surgical-VQLA: Transformer with Gated Vision-Language Embedding for Visual Question Localized-Answering in Robotic Surgery

1 code implementation19 May 2023 Long Bai, Mobarakol Islam, Lalithkumar Seenivasan, Hongliang Ren

In this paper, we propose Visual Question Localized-Answering in Robotic Surgery (Surgical-VQLA) to localize the specific surgical area during the answer prediction.

Answer Generation object-detection +3

SurgicalGPT: End-to-End Language-Vision GPT for Visual Question Answering in Surgery

1 code implementation19 Apr 2023 Lalithkumar Seenivasan, Mobarakol Islam, Gokul Kannan, Hongliang Ren

Given the limitations of unidirectional attention in GPT models and their ability to generate coherent long paragraphs, we carefully sequence the word tokens before vision tokens, mimicking the human thought process of understanding the question to infer an answer from an image.

Question Answering Scene Segmentation +1

Paced-Curriculum Distillation with Prediction and Label Uncertainty for Image Segmentation

1 code implementation2 Feb 2023 Mobarakol Islam, Lalithkumar Seenivasan, S. P. Sharan, V. K. Viekash, Bhavesh Gupta, Ben Glocker, Hongliang Ren

Purpose: In curriculum learning, the idea is to train on easier samples first and gradually increase the difficulty, while in self-paced learning, a pacing function defines the speed to adapt the training progress.

Image Segmentation Medical Image Segmentation +3

Task-Aware Asynchronous Multi-Task Model with Class Incremental Contrastive Learning for Surgical Scene Understanding

1 code implementation28 Nov 2022 Lalithkumar Seenivasan, Mobarakol Islam, Mengya Xu, Chwee Ming Lim, Hongliang Ren

Conclusion: The proposed multi-task model was able to adapt to domain shifts, incorporate novel instruments in the target domain, and perform tool-tissue interaction detection and report generation on par with single-task models.

Contrastive Learning Decision Making +4

Surgical-VQA: Visual Question Answering in Surgical Scenes using Transformer

2 code implementations22 Jun 2022 Lalithkumar Seenivasan, Mobarakol Islam, Adithya K Krishna, Hongliang Ren

This overload often limits their time answering questionnaires from patients, medical students or junior residents related to surgical procedures.

Question Answering Sentence +1

Global-Reasoned Multi-Task Learning Model for Surgical Scene Understanding

2 code implementations28 Jan 2022 Lalithkumar Seenivasan, Sai Mitheran, Mobarakol Islam, Hongliang Ren

Global and local relational reasoning enable scene understanding models to perform human-like scene analysis and understanding.

Graph Attention Knowledge Distillation +5

Class-Distribution-Aware Calibration for Long-Tailed Visual Recognition

1 code implementation11 Sep 2021 Mobarakol Islam, Lalithkumar Seenivasan, Hongliang Ren, Ben Glocker

In CDA-TS, the scalar temperature value is replaced with the CDA temperature vector encoded with class frequency to compensate for the over-confidence.

Learning and Reasoning with the Graph Structure Representation in Robotic Surgery

2 code implementations7 Jul 2020 Mobarakol Islam, Lalithkumar Seenivasan, Lim Chwee Ming, Hongliang Ren

Learning to infer graph representations and performing spatial reasoning in a complex surgical environment can play a vital role in surgical scene understanding in robotic surgery.

Edge Classification Graph Generation +3

Cannot find the paper you are looking for? You can Submit a new open access paper.