Abstract. Evaluation is one of the hardest tasks in automatic text summarization. It is perhaps even harder to determine how much a particular component of a summarization system c...
Abstract—Automatic summarization evaluation is very important to the development of summarization systems. In text summarization, ROUGE has been shown to correlate well with huma...
Tabletop groupware systems have natural advantages for collaboration, but they present a challenge for application designers because shared work and interaction progress in differe...
Information retrieval (IR) evaluation scores are generally designed to measure the effectiveness with which relevant documents are identified and retrieved. Many scores have been ...
This paper gives an overview of the evaluation method used for the Web Retrieval Task in the Third NTCIR Workshop, which is currently in progress. In the Web Retrieval Task, we tr...
Koji Eguchi, Keizo Oyama, Emi Ishida, Kazuko Kuriy...