Inferring users' actions and intentions forms an integral part of design and development of any human-computer interface. The presence of noisy and at times ambiguous sensory data makes this problem challenging. We formulate a framework for temporal fusion of multiple sensors using input							
						
							
					 															
					Vladimir Pavlovic, Ashutosh Garg, James M. Rehg