We formulate and study search algorithms that consider a user’s prior interactions with a wide variety of content to personalize that user’s current Web search. Rather than re...
A classical measure of similarity between strings is the length of the longest common subsequence(LCS) between the two given strings. The search for efficient algorithms for findi...
We present a probabilistic model for a document corpus that combines many of the desirable features of previous models. The model is called “GaP” for Gamma-Poisson, the distri...
Let us consider an ordered vector A[1 : n]. If the cost of testing each position is similar, then the standard binary search is the best strategy to search the vector. This is true...
We design a randomized polynomial time algorithm which, given a 3-tensor of real numbers A = {aijk}n i,j,k=1 such that for all i, j, k ∈ {1, . . . , n} we have ai jk = aik j = a...