Paper

Separability is not the best goal for machine learning

Neural networks use their hidden layers to transform input data into linearly separable data clusters, with a linear or a perceptron type output layer making the final projection on the line perpendicular to the discriminating hyperplane. For complex data with multimodal distributions this transformation is difficult to learn. Projection on $k\geq 2$ line segments is the simplest extension of linear separability, defining much easier goal for the learning process. Simple problems are 2-separable, but problems with inherent complex logic may be solved in a simple way by $k$-separable projections. The difficulty of learning non-linear data distributions is shifted to separation of line intervals, simplifying the transformation of data by hidden network layers. For classification of difficult Boolean problems, such as the parity problem, linear projection combined with \ksep is sufficient and provides a powerful new target for learning. More complex targets may also be defined, changing the goal of learning from linear discrimination to creation of data distributions that can easily be handled by specialized models selected to analyze output distributions. This approach can replace many layers of transformation required by deep learning models.

Results in Papers With Code
(↓ scroll down to see all results)