no code implementations • 31 Dec 2024 • Weijia Xu, Nebojsa Jojic, Sudha Rao, Chris Brockett, Bill Dolan
We examine two state-of-the-art LLMs, GPT-4 and LLaMA-3, on story generation and discover that LLM-generated stories often consist of plot elements that are echoed across a number of generations.
1 code implementation • 5 Nov 2024 • Seyed Hossein Alavi, Weijia Xu, Nebojsa Jojic, Daniel Kennett, Raymond T. Ng, Sudha Rao, Haiyan Zhang, Bill Dolan, Vered Shwartz
We introduce GamePlot, an LLM-powered assistant that supports game designers in crafting immersive narratives for turn-based games, and allows them to test these games through a collaborative game play and refine the plot throughout the process.
no code implementations • 24 Oct 2024 • Alexandros Graikos, Nebojsa Jojic, Dimitris Samaras
Diffusion models have dominated the field of large, generative image models, with the prime examples of Stable Diffusion and DALL-E 3 being widely adopted.
no code implementations • 7 Oct 2024 • Weijia Xu, Nebojsa Jojic, Nicolas Le Roux
These tags can be used to model the structure of given documents and for hierarchical sampling of new texts.
no code implementations • 25 Apr 2024 • Xiangyu Peng, Jessica Quaye, Sudha Rao, Weijia Xu, Portia Botchway, Chris Brockett, Nebojsa Jojic, Gabriel DesGarennes, Ken Lobb, Michael Xu, Jorge Leandro, Claire Jin, Bill Dolan
We explore how interaction with large language models (LLMs) can give rise to emergent behaviors, empowering players to participate in the evolution of game narratives.
2 code implementations • 25 Oct 2023 • Xinyuan Wang, Chenxi Li, Zhen Wang, Fan Bai, Haotian Luo, Jiayou Zhang, Nebojsa Jojic, Eric P. Xing, Zhiting Hu
Highly effective, task-specific prompts are often heavily engineered by experts to integrate detailed instructions and domain insights based on a deep understanding of both instincts of large language models (LLMs) and the intricacies of the target task.
no code implementations • 22 May 2023 • ASHISH SHARMA, Sudha Rao, Chris Brockett, Akanksha Malhotra, Nebojsa Jojic, Bill Dolan
While LLMs are being developed to simulate human behavior and serve as human-like agents, little attention has been given to the Agency that these models should possess in order to proactively manage the direction of interaction and collaboration.
no code implementations • 17 May 2023 • Weijia Xu, Andrzej Banburski-Fahey, Nebojsa Jojic
We introduce Reprompting, an iterative sampling algorithm that automatically learns the Chain-of-Thought (CoT) recipes for a given task without human intervention.
no code implementations • 25 Mar 2023 • Ana Jojic, Zhen Wang, Nebojsa Jojic
We demonstrate that, through appropriate prompting, GPT-3 family of models can be triggered to perform iterative behaviours necessary to execute (rather than just write or recall) programs that involve loops, including several popular algorithms found in computer science curricula or software developer interviews.
no code implementations • 4 Oct 2022 • Batu Ozturkler, Nikolay Malkin, Zhen Wang, Nebojsa Jojic
Our results suggest that because the probabilistic inference in ThinkSum is performed outside of calls to the LLM, ThinkSum is less sensitive to prompt design, yields more interpretable predictions, and can be flexibly combined with latent variable models to extract structured knowledge from LLMs.
1 code implementation • 17 Jun 2022 • Alexandros Graikos, Nikolay Malkin, Nebojsa Jojic, Dimitris Samaras
We consider the problem of inferring high-dimensional data $\mathbf{x}$ in a model that consists of a prior $p(\mathbf{x})$ and an auxiliary differentiable constraint $c(\mathbf{x},\mathbf{y})$ on $x$ given some additional information $\mathbf{y}$.
1 code implementation • 28 Feb 2022 • Esther Rolf, Nikolay Malkin, Alexandros Graikos, Ana Jojic, Caleb Robinson, Nebojsa Jojic
We propose a method for jointly inferring labels across a collection of data samples, where each sample consists of an observation and a prior belief about the label.
1 code implementation • ACL 2022 • Nikolay Malkin, Zhen Wang, Nebojsa Jojic
Long-range semantic coherence remains a challenge in automatic language generation and understanding.
no code implementations • 29 Sep 2021 • Esther Rolf, Nikolay Malkin, Alexandros Graikos, Ana Jojic, Caleb Robinson, Nebojsa Jojic
In prediction problems, coarse and imprecise sources of input can provide rich information about labels, but are not readily used by discriminative learners.
1 code implementation • EMNLP 2021 • Nikolay Malkin, Sameera Lanka, Pranav Goel, Nebojsa Jojic
As neural language models approach human performance on NLP benchmark tasks, their advances are widely seen as evidence of an increasingly complex understanding of syntax.
2 code implementations • CVPR 2021 • Elijah Cole, Oisin Mac Aodha, Titouan Lorieul, Pietro Perona, Dan Morris, Nebojsa Jojic
When the number of potential labels is large, human annotators find it difficult to mention all applicable labels for each training image.
no code implementations • 7 Jun 2021 • Zhanning Gao, Le Wang, Nebojsa Jojic, Zhenxing Niu, Nanning Zheng, Gang Hua
In the proposed framework, a dedicated feature alignment module is incorporated for redundancy removal across frames to produce the tensor representation, i. e., the video imprint.
no code implementations • NAACL 2021 • Nikolay Malkin, Sameera Lanka, Pranav Goel, Sudha Rao, Nebojsa Jojic
Human innovation in language, such as inventing new words, is a challenge for pretrained language models.
1 code implementation • 19 May 2021 • Jacob Russin, Roland Fernandez, Hamid Palangi, Eric Rosen, Nebojsa Jojic, Paul Smolensky, Jianfeng Gao
A longstanding question in cognitive science concerns the learning mechanisms underlying compositionality in human cognition.
1 code implementation • 4 Jan 2021 • Nikolay Malkin, Caleb Robinson, Nebojsa Jojic
We present simple algorithms for land cover change detection in the 2021 IEEE GRSS Data Fusion Contest.
no code implementations • ACL 2020 • Shashank Srivastava, Oleks Polozov, R, Nebojsa Jojic, Christopher Meek
We explore learning web-based tasks from a human teacher through natural language explanations and a single demonstration.
1 code implementation • ECCV 2020 • Nikolay Malkin, Anthony Ortiz, Caleb Robinson, Nebojsa Jojic
We show that simple patch-based models, such as epitomes, can have superior performance to the current state of the art in semantic segmentation and label super-resolution, which uses deep convolutional neural networks.
1 code implementation • 8 Apr 2020 • Elijah Cole, Benjamin Deneu, Titouan Lorieul, Maximilien Servajean, Christophe Botella, Dan Morris, Nebojsa Jojic, Pierre Bonnet, Alexis Joly
Understanding the geographic distribution of species is a key concern in conservation.
1 code implementation • CVPR 2020 • Anthony Ortiz, Caleb Robinson, Dan Morris, Olac Fuentes, Christopher Kiekintveld, Md Mahmudulla Hassan, Nebojsa Jojic
In many vision applications the local spatial context of the features is important, but most common normalization schemes including Group Normalization (GN), Instance Normalization (IN), and Layer Normalization (LN) normalize over the entire spatial dimension of a feature.
1 code implementation • 22 Oct 2019 • Mohammad Sadegh Norouzzadeh, Dan Morris, Sara Beery, Neel Joshi, Nebojsa Jojic, Jeff Clune
However, the accuracy of results depends on the amount, quality, and diversity of the data available to train models, and the literature has focused on projects with millions of relevant, labeled training images.
3 code implementations • 15 Oct 2019 • Imanol Schlag, Paul Smolensky, Roland Fernandez, Nebojsa Jojic, Jürgen Schmidhuber, Jianfeng Gao
We incorporate Tensor-Product Representations within the Transformer in order to better support the explicit representation of relation structure.
Ranked #1 on
Question Answering
on Mathematics Dataset
no code implementations • 10 Jun 2019 • Caleb Robinson, Anthony Ortiz, Kolya Malkin, Blake Elias, Andi Peng, Dan Morris, Bistra Dilkina, Nebojsa Jojic
This bi-directional feedback loop allows humans to learn how the model responds to new data.
no code implementations • ICLR 2019 • Kolya Malkin, Caleb Robinson, Le Hou, Nebojsa Jojic
We present a deep learning-based method for super-resolving coarse (low-resolution) labels assigned to groups of image pixels into pixel-level (high-resolution) labels, given the joint distribution between those low- and high-resolution labels.
no code implementations • 9 Apr 2019 • Maozheng Zhao, Le Hou, Han Le, Dimitris Samaras, Nebojsa Jojic, Danielle Fassler, Tahsin Kurc, Rajarsi Gupta, Kolya Malkin, Shroyer Kenneth, Joel Saltz
On the other hand, collecting low resolution labels (labels for a block of pixels) for these high resolution images is much more cost efficient.
no code implementations • ICLR 2020 • Yingzhen Yang, Jiahui Yu, Nebojsa Jojic, Jun Huan, Thomas S. Huang
FSNet has the same architecture as that of the baseline CNN to be compressed, and each convolution layer of FSNet has the same number of filters from FS as that of the basline CNN in the forward process.
no code implementations • ACL 2018 • Shashank Srivastava, Nebojsa Jojic
We present a generative probabilistic model of documents as sequences of sentences, and show that inference in it can lead to extraction of long-range latent discourse structure from a collection of documents.
no code implementations • ICLR 2018 • Yao-Hung Hubert Tsai, Han Zhao, Nebojsa Jojic, Ruslan Salakhutdinov
The assumption that data samples are independently identically distributed is the backbone of many learning algorithms.
no code implementations • ICML 2018 • Xiaojie Jin, Yingzhen Yang, Ning Xu, Jianchao Yang, Nebojsa Jojic, Jiashi Feng, Shuicheng Yan
We present a new approach and a novel architecture, termed WSNet, for learning compact and efficient deep neural networks.
no code implementations • ICLR 2018 • Yao-Hung Hubert Tsai, Han Zhao, Ruslan Salakhutdinov, Nebojsa Jojic
In this technical report, we introduce OrderNet that can be used to extract the order of data instances in an unsupervised way.
no code implementations • ICCV 2017 • Alessandro Perina, Sadegh Mohammadi, Nebojsa Jojic, Vittorio Murino
In particular, we use constrained Markov walks over a counting grid for modeling image sequences, which not only yield good latent representations, but allow for excellent classification with only a handful of labeled training examples of the new scenes or objects, a scenario typical in lifelogging applications.
1 code implementation • EMNLP 2017 • Di Wang, Nebojsa Jojic, Chris Brockett, Eric Nyberg
We propose simple and flexible training and decoding methods for influencing output style and topic in neural encoder-decoder based language generation.
no code implementations • 5 Sep 2017 • Yingzhen Yang, Feng Liang, Nebojsa Jojic, Shuicheng Yan, Jiashi Feng, Thomas S. Huang
By generalization analysis via Rademacher complexity, the generalization error bound for the kernel classifier learned from hypothetical labeling is expressed as the sum of pairwise similarity between the data from different classes, parameterized by the weights of the kernel classifier.
no code implementations • 5 Sep 2017 • Yingzhen Yang, Jiashi Feng, Nebojsa Jojic, Jianchao Yang, Thomas S. Huang
We study the proximal gradient descent (PGD) method for $\ell^{0}$ sparse approximation problem as well as its accelerated optimization with randomized algorithms in this paper.
no code implementations • CVPR 2017 • Zhanning Gao, Gang Hua, Dong-Qing Zhang, Nebojsa Jojic, Le Wang, Jianru Xue, Nanning Zheng
We develop a unified framework for complex event retrieval, recognition and recounting.
1 code implementation • NeurIPS 2016 • R. Devon Hjelm, Kyunghyun Cho, Junyoung Chung, Russ Salakhutdinov, Vince Calhoun, Nebojsa Jojic
Variational methods that rely on a recognition network to approximate the posterior of directed graphical models offer better inference and learning than previous methods.
no code implementations • 12 Mar 2015 • Nebojsa Jojic, Alessandro Perina, Dongwoo Kim
The counting grid is a grid of microtopics, sparse word/feature distributions.
no code implementations • 23 Oct 2014 • Alessandro Perina, Nebojsa Jojic
The space of all possible feature count combinations is constrained both by the properties of the larger scene and the size and the location of the window into it.
no code implementations • NeurIPS 2013 • Alessandro Perina, Nebojsa Jojic, Manuele Bicego, Andrzej Truski
The counting grid \cite{cgUai} models this spatial metaphor literally: it is multidimensional grid of word distributions learned in such a way that a document's own distribution of features can be modeled as the sum of the histograms found in a window into the grid.
no code implementations • NeurIPS 2013 • Fabian L. Wauthier, Nebojsa Jojic, Michael. I. Jordan
The Lasso is a cornerstone of modern multivariate data analysis, yet its performance suffers in the common situation in which covariates are correlated.
no code implementations • CVPR 2013 • Alessandro Perina, Nebojsa Jojic
Recently, the Counting Grid (CG) model [5] was developed to represent each input image as a point in a large grid of feature counts.
no code implementations • 26 Apr 2013 • Alessandro Perina, Nebojsa Jojic
We introduce and we analyze a new dataset which resembles the input to biological vision systems much more than most previously published ones.
no code implementations • NeurIPS 2010 • Nebojsa Jojic, Alessandro Perina, Vittorio Murino
In order to study the properties of total visual input in humans, a single subject wore a camera for two weeks capturing, on average, an image every 20 seconds (www. research. microsoft. com/~jojic/aihs).
no code implementations • NeurIPS 2010 • Nebojsa Jojic, Chris Meek, Jim C. Huang
We compare the performance of the resulting algorithm to Mathematica and D*, and we apply our method to learning models for rainfall and H1N1 data, where we show that CDNs with cycles are able to provide a significantly better fits to the data as compared to tree-structured and unstructured CDNs and other heavy-tailed multivariate distributions such as the multivariate copula and logistic models.
no code implementations • NeurIPS 2009 • Alessandro Perina, Marco Cristani, Umberto Castellani, Vittorio Murino, Nebojsa Jojic
Score functions induced by generative models extract fixed-dimension feature vectors from different-length data observations by subsuming the process of data generation, projecting them in highly informative spaces called score spaces.