no code implementations • 24 Apr 2023 • Yingtao Tian
Computational creativity has contributed heavily to abstract art in modern era, allowing artists to create high quality, abstract two dimension (2D) arts with a high level of controllability and expressibility.
1 code implementation • 21 Apr 2023 • Shanchuan Wan, Yujin Tang, Yingtao Tian, Tomoyuki Kaneko
Exploration is a fundamental aspect of reinforcement learning (RL), and its effectiveness is a deciding factor in the performance of RL algorithms, especially when facing sparse extrinsic rewards.
no code implementations • 18 Apr 2022 • Yingtao Tian, Marco Cuturi, David Ha
Recent advances in deep learning, such as powerful generative models and joint text-image embeddings, have provided the computational creativity community with new tools, opening new perspectives for artistic pursuits.
1 code implementation • 10 Feb 2022 • Yujin Tang, Yingtao Tian, David Ha
Evolutionary computation has been shown to be a highly effective method for training neural networks, particularly when employed at scale on CPU clusters.
1 code implementation • 28 Jan 2022 • Marco Cuturi, Laetitia Meng-Papaxanthos, Yingtao Tian, Charlotte Bunne, Geoff Davis, Olivier Teboul
Optimal transport tools (OTT-JAX) is a Python toolbox that can solve optimal transport problems between point clouds and histograms.
1 code implementation • 18 Sep 2021 • Yingtao Tian, David Ha
Evolutionary algorithms have been used in the digital art scene since the 1970s.
1 code implementation • 4 Jun 2021 • Yingtao Tian, Tarin Clanuwat, Chikahiko Suzuki, Asanobu Kitamoto
The study of Ukiyo-e, an important genre of pre-modern Japanese art, focuses on the object and style like other artwork researches.
no code implementations • 14 Oct 2020 • Ştefan Postăvaru, Anton Tsitsulin, Filipe Miguel Gonçalves de Almeida, Yingtao Tian, Silvio Lattanzi, Bryan Perozzi
In this paper, we introduce InstantEmbedding, an efficient method for generating single-node representations using local PageRank computations.
1 code implementation • 20 Feb 2020 • Yingtao Tian, Chikahiko Suzuki, Tarin Clanuwat, Mikel Bober-Irizar, Alex Lamb, Asanobu Kitamoto
From classifying handwritten digits to generating strings of text, the datasets which have received long-time focus from the machine learning community vary greatly in their subject matter.
1 code implementation • 30 Aug 2019 • Haochen Chen, Syed Fahad Sultan, Yingtao Tian, Muhao Chen, Steven Skiena
Two key features of FastRP are: 1) it explicitly constructs a node similarity matrix that captures transitive relationships in a graph and normalizes matrix entries based on node degrees; 2) it utilizes very sparse random projection, which is a scalable optimization-free method for dimension reduction.
no code implementations • 28 Aug 2019 • Jingjing Li, Wenlu Wang, Wei-Shinn Ku, Yingtao Tian, Haixun Wang
A natural language interface (NLI) to databases is an interface that translates a natural language question to a structured query that is executable by database management systems (DBMS).
no code implementations • WS 2019 • Weijia Shi, Muhao Chen, Yingtao Tian, Kai-Wei Chang
Bilingual word embeddings, which representlexicons of different languages in a shared em-bedding space, are essential for supporting se-mantic and knowledge transfers in a variety ofcross-lingual NLP tasks.
no code implementations • ICLR 2019 • Yingtao Tian, Jesse Engel
We find that a simple variational autoencoder is able to learn a shared latent space to bridge between two generative models in an unsupervised fashion, and even between different types of models (ex.
no code implementations • 21 Feb 2019 • Yingtao Tian, Jesse Engel
We compare to state-of-the-art techniques and find that a straight-forward variational autoencoder is able to best bridge the two generative models through learning a shared latent space.
1 code implementation • 13 Sep 2018 • Haochen Chen, Xiaofei Sun, Yingtao Tian, Bryan Perozzi, Muhao Chen, Steven Skiena
Network embedding methods aim at learning low-dimensional latent representation of nodes in a network.
Social and Information Networks Physics and Society
2 code implementations • 7 Sep 2018 • Wenlu Wang, Yingtao Tian, Hongyu Xiong, Haixun Wang, Wei-Shinn Ku
In this work, we introduce a general purpose transfer-learnable NLI with the goal of learning one model that can be used as NLI for any relational database.
no code implementations • 7 Sep 2018 • Muhao Chen, Yingtao Tian, Xuelu Chen, Zijun Xue, Carlo Zaniolo
Recent advances in translation-based graph embedding methods for populating instance-level knowledge graphs lead to promising new approaching for the ontology population problem.
1 code implementation • CONLL 2019 • Muhao Chen, Yingtao Tian, Haochen Chen, Kai-Wei Chang, Steven Skiena, Carlo Zaniolo
Bilingual word embeddings have been widely used to capture the similarity of lexical semantics in different human languages.
no code implementations • COLING 2018 • Vivek Kulkarni, Yingtao Tian, D, Parth iwala, Steve Skiena
We present domain independent models to date documents based only on neologism usage patterns.
no code implementations • 18 Jun 2018 • Muhao Chen, Yingtao Tian, Kai-Wei Chang, Steven Skiena, Carlo Zaniolo
Since many multilingual KGs also provide literal descriptions of entities, in this paper, we introduce an embedding-based approach which leverages a weakly aligned multilingual KG for semi-supervised cross-lingual learning using entity descriptions.
1 code implementation • ICLR 2018 • Hanjun Dai, Yingtao Tian, Bo Dai, Steven Skiena, Le Song
Deep generative models have been enjoying success in modeling continuous data.
7 code implementations • 18 Aug 2017 • Yanghua Jin, Jiakai Zhang, Minjun Li, Yingtao Tian, Huachun Zhu, Zhihao Fang
With quantitative analysis and case studies we demonstrate that our efforts lead to a stable and high-quality model.
2 code implementations • 12 Nov 2016 • Muhao Chen, Yingtao Tian, Mohan Yang, Carlo Zaniolo
Many recent works have demonstrated the benefits of knowledge graph embeddings in completing monolingual knowledge graphs.
Ranked #38 on
Entity Alignment
on DBP15k zh-en
no code implementations • 12 May 2016 • Yingtao Tian, Vivek Kulkarni, Bryan Perozzi, Steven Skiena
Do word embeddings converge to learn similar things over different initializations?