We present a family of verifiable random functions which are provably secure for exponentially-large input spaces under a non-interactive complexity assumption. Prior construction...
Multiple kernel learning (MKL) uses a weighted combination of kernels where the weight of each kernel is optimized during training. However, MKL assigns the same weight to a kerne...
Abstract. In this paper we compare two methods for intrinsic dimensionality (ID) estimation based on optimally topology preserving maps (OTPMs). The rst one is a direct approach, w...
We introduce an algorithm for learning a local metric to a continuous input space that measures distances in terms of relevance to the processing task. The relevance is defined a...
We provide several enhancements to our previously introduced algorithm for a sequential construction of a hybrid network of radial and perceptron hidden units [6]. At each stage, ...
In this paper we propose a novel classification algorithm that fits models of different complexity on separate regions of the input space. The goal is to achieve a balance betwee...
Ricardo Vilalta, Murali-Krishna Achari, Christoph ...
—This paper considers the task of software test case generation from a large space of values of input parameters. The purpose of the paper is to create a model of software input ...
Sergiy A. Vilkomir, W. Thomas Swain, Jesse H. Poor...
This paper presents a novel technique, called input space adaptive software synthesis, for the energy and performance optimization of embedded software. The proposed technique is ...
Intuitively, learning should be easier when the data points lie on a low-dimensional submanifold of the input space. Recently there has been a growing interest in algorithms that ...
Dimensionality reduction involves mapping a set of high dimensional input points onto a low dimensional manifold so that "similar" points in input space are mapped to ne...