no code implementations • 3 May 2022 • Danilo Vucetic, Mohammadreza Tayaranian, Maryam Ziaeefard, James J. Clark, Brett H. Meyer, Warren J. Gross
FAR reduces fine-tuning time on the DistilBERT model and CoLA dataset by 30%, and time spent on memory operations by 47%.
1 code implementation • CVPR 2022 • Samrudhdhi B. Rangrej, Chetan L. Srinidhi, James J. Clark
Most hard attention models initially observe a complete scene to locate and sense informative glimpses, and predict class-label of a scene based on glimpses.
no code implementations • 24 Feb 2022 • Amir Ardakani, Arash Ardakani, Brett Meyer, James J. Clark, Warren J. Gross
Quantization of deep neural networks is a promising approach that reduces the inference cost, making it feasible to run deep networks on resource-restricted devices.
1 code implementation • 15 Nov 2021 • Samrudhdhi B. Rangrej, James J. Clark
A visual hard attention model actively selects and observes a sequence of subregions in an image to make a prediction.
no code implementations • ACL 2022 • Ali Edalati, Marzieh Tahaei, Ahmad Rashid, Vahid Partovi Nia, James J. Clark, Mehdi Rezagholizadeh
GPT is an auto-regressive Transformer-based pre-trained language model which has attracted a lot of attention in the natural language processing (NLP) domain due to its state-of-the-art performance in several downstream tasks.
no code implementations • 1 Apr 2021 • Samrudhdhi B. Rangrej, James J. Clark
The next fixation is planned using uncertainty in the content of the imagined scenes.
1 code implementation • 30 Mar 2021 • Saverio Vadacchino, Raghav Mehta, Nazanin Mohammadi Sepahvand, Brennan Nichyporuk, James J. Clark, Tal Arbel
The proposed network is trained and tested on the BraTS 2019 brain tumour segmentation challenge dataset, where it achieves performance improvements in the ranges of 16% - 26% over (a) recent modality-agnostic segmentation methods (U-HeMIS, U-HVED), (b) KD-Net adapted to this problem, (c) the pre-trained student network and (d) a non-hierarchical version of the network (AD-Net), in terms of Dice scores for enhancing tumour (ET).
no code implementations • 1 Jan 2021 • Samrudhdhi Bharatkumar Rangrej, James J. Clark
Unlike in the deep convolution network, in hard attention it is explainable which regions of the image contributed to the prediction.
no code implementations • 29 Sep 2020 • Qing Tian, Tal Arbel, James J. Clark
We also show that our grown Inception nets (without hard-coded dimension alignment) clearly outperform residual nets of similar complexities.
1 code implementation • 10 Apr 2019 • Guanqing Hu, James J. Clark
In order to achieve automatic compositing in natural scenes, we propose a fully automated method that integrates instance segmentation and image matting processes to generate high-quality semantic mattes that can be used for image editing task.
no code implementations • CVPR 2018 • Siavash Gorji, James J. Clark
We evaluate our model by comparing the performance of several augmented static saliency models with state-of-the-art in spatiotemporal saliency on three largest dynamic eye tracking datasets, HOLLYWOOD2, UCF-Sport and DIEM.
1 code implementation • 21 Mar 2018 • Qing Tian, Tal Arbel, James J. Clark
Moreover, we examine our approach's potential in network architecture search for specific tasks and analyze the influence of our pruning on model robustness to noises and adversarial attacks.
no code implementations • 21 Nov 2017 • Bingqing Yu, James J. Clark
The WAYLA approach is based on the Conditional Generative Adversarial Network (Conditional GAN) image-to-image translation technique of Isola et al. We consider two specific applications - the first, of reconstructing newspaper images from gaze heat maps, and the second, of detailed reconstruction of images containing only text.
no code implementations • 21 Nov 2017 • Bingqing Yu, James J. Clark
The discriminator also has the observer label as an input, which contributes to the personalization ability of our approach.
no code implementations • CVPR 2017 • Siavash Gorji, James J. Clark
The Attentional Push CNN is then fine-tuned along with the augmented saliency CNN to minimize the Euclidean distance between the augmented saliency and ground truth fixations using an eye-tracking dataset, annotated with the head and the gaze location of the scene actors.
no code implementations • 20 Apr 2017 • Qing Tian, Tal Arbel, James J. Clark
Many real-time tasks, such as human-computer interaction, require fast and efficient facial gender classification.
no code implementations • 1 Sep 2016 • Siavash Gorji, James J. Clark
We present a novel visual attention tracking technique based on Shared Attention modeling.