Mover’s Distance (EMD) to explicitly align the volumes. In contrast to the fixed volume-to-volume matching used in [25], the space-time volumes of two videos across different space-time locations can be matched using our ASTPM method, making it better at coping with the large intraclass variations within the same type of events (., moving objects in consumer videos can appear at different space- time locations, and the background within two different videos, even captured from the same scene, may be shifted due to considerable camera motion). The second is ourmain contribution. In order to copewith the considerable variation between feature distributions of videos from the web video domain and consumer video domain, we propose.