It's very interesting!!!
Can we recognize similiar human actions captured from different view angles? Firstly, let's check whether this kind of action recognition across multple views is necessary and useful. In multiple camera system, we may be given visual data from multiple views, so it is necessary to equip computer vison system with the ability of recognizing actions invariant to view. Also in the application of query by action should find not only similiar action segment from the same view angle but also those from different view angles.

Now what's the open issues of this problem!!!!
1. How to design similarity measure on space-time segment that is invariant to view?
Already have a rough idea extended from CVPR 2005 paper ("Space-time Behavior Based Correlation").
2. How to make the similarity measure robust in the noisy environment (error in space-time gradient, error by occlusion ...)? Generalize the binary measure to continuous one!!!
3. How to decide the recognization granularity?? Given a long action segment, the global action may be different from local action? How to find whether global action and/or local action are consistent?