no code implementations • 6 Jun 2017 • Azad Naik, Huzefa Rangwala
Our experimental evaluation on text and image datasets with varying distribution of features, classes and instances shows upto 3x order of speed-up on massive datasets and upto 45% less memory requirements for storing the weight vectors of learned model without any significant loss (improvement for some datasets) in the classification accuracy.
no code implementations • 6 Jun 2017 • Azad Naik, Anveshi Charuvaka, Huzefa Rangwala
Multi-task learning (MTL) is a supervised learning paradigm in which the prediction models for several related tasks are learned jointly to achieve better generalization performance.
no code implementations • 5 Jun 2017 • Azad Naik, Huzefa Rangwala
In this paper, we propose two different data-driven approaches (local and global) for hierarchical structure modification that identifies and flattens inconsistent nodes present within the hierarchy.
no code implementations • 2 Mar 2016 • Azad Naik, Huzefa Rangwala
Experimental comparisons of top-down HC with our modified hierarchy, on a wide range of datasets shows classification performance improvement over the baseline hierarchy (i:e:, defined by expert), clustered hierarchy and flattening based hierarchy modification approaches.