A website can regulate search engine crawler access to its content using the robots exclusion protocol, specified in its robots.txt file. The rules in the protocol enable the site...
Regularized Least Squares (RLS) algorithms have the ability to avoid over-fitting problems and to express solutions as kernel expansions. However, we observe that the current RLS ...
The paper discusses ITS architectures in terms of patterns that exist within them. The concept of patterns has received surprisingly little attention so far from researchers in the...
The uncertainty may be divides it into two major groups, "objective uncertainty" and "subjective uncertainty". The objective uncertainty has already been exten...
We present GoGetIt!, a tool for generating structure-driven crawlers that requires a minimum effort from the users. The tool takes as input a sample page and an entry point to a W...
Altigran Soares da Silva, Edleno Silva de Moura, J...