[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Action representations in robotics: : A taxonomy and systematic classification

Published: 01 April 2019 Publication History

Abstract

Understanding and defining the meaning of “action” is substantial for robotics research. This becomes utterly evident when aiming at equipping autonomous robots with robust manipulation skills for action execution. Unfortunately, to this day we still lack both a clear understanding of the concept of an action and a set of established criteria that ultimately characterize an action. In this survey, we thus first review existing ideas and theories on the notion and meaning of action. Subsequently, we discuss the role of action in robotics and attempt to give a seminal definition of action in accordance with its use in robotics research. Given this definition we then introduce a taxonomy for categorizing action representations in robotics along various dimensions. Finally, we provide a meticulous literature survey on action representations in robotics where we categorize relevant literature along our taxonomy. After discussing the current state of the art we conclude with an outlook towards promising research directions.

References

[1]
Acosta Calderon CA, Mohan RE, and Zhou C (2010) Teaching new tricks to a robot learning to solve a task by imitation. In: 2010 IEEE Conference on Robotics, Automation and Mechatronics. IEEE.
[2]
Aein MJ, Aksoy EE, Tamosiunaite M, Papon J, Ude A, and Worgotter F (2013) Toward a library of manipulation actions based on semantic object–action relations. In: 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE.
[3]
Ahad MAR, Tan J, Kim H, and Ishikawa S (2010) Action recognition by employing combined directional motion history and energy images. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops. IEEE.
[4]
Ahmad M and Lee SW (2010) Variable silhouette energy image representations for recognizing human actions. Image and Vision Computing 28(5): 814–824.
[5]
Ahmadzadeh SR, Paikan A, Mastrogiovanni F, Natale L, Kormushev P, and Caldwell DG (2015) Learning symbolic representations of actions from human demonstrations. In: 2015 IEEE International Conference on Robotics and Automation (ICRA). IEEE.
[6]
Aksoy EE, Orhan A, and Wörgötter F (2016a) Semantic decomposition and recognition of long and complex manipulation action sequences. International Journal of Computer Vision 122(1): 84–115.
[7]
Aksoy EE, Tamosiunaite M, and Vuga Ret al. (2013) Structural bootstrapping at the sensorimotor level for the fast acquisition of action knowledge for cognitive robots. In: 2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL). IEEE.
[8]
Aksoy EE, Tamosiunaite M, and Wörgötter F (2015) Model-free incremental learning of the semantics of manipulation actions. Robotics and Autonomous Systems 71: 118–133.
[9]
Aksoy EE, Zhou Y, Wachter M, and Asfour T (2016b) Enriched manipulation action semantics for robot execution of time constrained tasks. In: 2016 IEEE-RAS 16th International Conference on Humanoid Robots (Humanoids). IEEE.
[10]
Alameda-Pineda X, Sanchez-Riera J, and Franch Vet al. (2011) The RAVEL data set. PhD Thesis, INRIA.
[11]
Altahhan A (2015) Deep feature-action processing with mixture of updates. In: Neural Information Processing. Berlin: Springer, pp. 1–10.
[12]
Andry P, Gaussier P, Nadel J, and Hirsbrunner B (2004) Learning invariant sensorimotor behaviors: A developmental approach to imitation mechanisms. Adaptive Behavior 12(2): 117–140.
[13]
Aristotle (1934) Nicomachean Ethics. Loeb Classical Library. Cambridge, MA: Harvard University Press.
[14]
Asfour T, Welke K, Ude A, Azad P, and Dillmann R (2008) Perceiving objects and movements to generate actions on a humanoid robot. In: Unifying Perspectives in Computational and Robot Vision (Lecture Notes in Electrical Engineering, vol. 8). New York: Springer, pp. 41–55.
[15]
Babič J, Hale JG, and Oztop E (2011) Human sensorimotor learning for humanoid robot skill synthesis. Adaptive Behavior 19(4): 250–263.
[16]
Bamert L and Mast FW (2009) Action Representation. Berlin: Springer, pp. 32–34.
[17]
Bartels G, Kresse I, and Beetz M (2013) Constraint-based movement representation grounded in geometric features. In: 2013 13th IEEE-RAS International Conference on Humanoid Robots (Humanoids). IEEE.
[18]
Bernstein NA (1996) On Dexterity and Its Development. Lawrence Erlbaum Associates.
[19]
Bessiere P, Dedieu E, and Mazer E (1994) Representing robot/environment interactions using probabilities: the “beam in the bin" experiment. In: Proceedings of PerAc ’94. From Perception to Action. Los Alamitos, CA: IEEE Computer Society Press.
[20]
Bhat AA and Mohan V (2015) How iCub learns to imitate use of a tool quickly by recycling the past knowledge learnt during drawing. In: Biomimetic and Biohybrid Systems. Berlin: Springer, pp. 339–347.
[21]
Blank M, Gorelick L, Shechtman E, Irani M, and Basri R (2005) Actions as space-time shapes. In: The Tenth IEEE International Conference on Computer Vision (ICCV’05). pp. 1395–1402.
[22]
Bloom V, Makris D, and Argyriou V (2012) G3D: A gaming action dataset and real-time action recognition evaluation framework. In: 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, pp. 7–12.
[23]
Botvinick MM (2008) Hierarchical models of behavior and prefrontal function. Trends in Cognitive Sciences 12(5): 201–208.
[24]
Buonamente M, Dindo H, and Johnsson M (2013) Recognizing actions with the associative self-organizing map. In: 2013 XXIV International Conference on Information, Communication and Automation Technologies (ICAT). IEEE.
[25]
Cantrell R, Schermerhorn P, and Scheutz M (2011) Learning actions from human-robot dialogues. In: 2011 IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN). IEEE.
[26]
Cao L, Liu Z, and Huang TS (2010) Cross-dataset action detection. In: 2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, pp. 1998–2005.
[27]
Chaaraoui AA, Climent-Pérez P, and Flórez-Revuelta F (2012) An efficient approach for multi-view human action recognition based on bag-of-key-poses. In: Human Behavior Understanding. Berlin: Springer, pp. 29–40.
[28]
Chaquet JM, Carmona EJ, and Fernández-Caballero A (2013) A survey of video datasets for human action and activity recognition. Computer Vision and Image Understanding 117(6): 633–659.
[29]
Chella A, Frixione M, and Gaglio S (2000) Towards a conceptual representation of actions. In: AI*IA 99: Advances in Artificial Intelligence (Lecture Notes in Computer Science, vol. 1792) Berlin: Springer, pp. 333–344.
[30]
Chen L, Wei H, and Ferryman J (2014) ReadingAct RGB-D action dataset and human action recognition from local features. Pattern Recognition Letters 50: 159–169.
[31]
Chuang LW, Lin CY, and Cangelosi A (2012) Learning of composite actions and visual categories via grounded linguistic instructions: Humanoid robot simulations. In: The 2012 International Joint Conference on Neural Networks (IJCNN). IEEE.
[32]
Cisek P and Kalaska JF (2010) Neural mechanisms for interacting with a world full of action choices. Annual Review of Neuroscience 33(1): 269–298.
[33]
Claßen J, Röger G, Lakemeyer G, and Nebel B (2011) Platas—integrating planning and the action language GOLOG. KI - Künstliche Intelligenz 26(1): 61–67.
[34]
CMU (2003) Graphics lab motion capture. URL http://mocap.cs.cmu.edu/.
[35]
Cooper RP and Shallice T (2006) Hierarchical schemas and goals in the control of sequential behavior. Psychology Review 113(4): 887–916; discussion 917-31.
[36]
Davidson D (2001) Essays on Actions and Events: Philosophical Essays. Oxford: Clarendon Press.
[37]
De Kleijn R, Kachergis G, and Hommel B (2014) Everyday robotic action: Lessons from human action control. Frontiers in Neurorobotics 8: 1–9.
[38]
Desmurget M and Grafton S (2000) Forward modeling allows feedback control for fast reaching movements. Trends in Cognitive Sciences 4(11): 423–431.
[39]
Dindo H and Chella A (2013) What will you do next? A cognitive model for understanding others’ intentions based on shared representations. In: Virtual Augmented and Mixed Reality. Designing and Developing Augmented and Virtual Environments. Berlin: Springer, pp. 253–266.
[40]
Dindo H, Presti LL, Cascia ML, Chella A, and Dedić R (2017) Hankelet-based action classification for motor intention recognition. Robotics and Autonomous Systems 94: 120–133.
[41]
Do M, Schill J, Ernesti J, and Asfour T (2014) Learn to wipe: A case study of structural bootstrapping from sensorimotor experience. In: 2014 IEEE International Conference on Robotics and Automation (ICRA). IEEE.
[42]
Donnarumma F, Prevete R, de Giorgio A, Montone G, and Pezzulo G (2015) Learning programs is better than learning dynamics: A programmable neural network hierarchical architecture in a multi-task scenario. Adaptive Behavior 24(1): 27–51.
[43]
Droniou A, Ivaldi S, and Sigaud O (2014) Learning a repertoire of actions with deep neural networks. In: 4th International Conference on Development and Learning and on Epigenetic Robotics. IEEE.
[44]
Dum R and Strick P (1991) The origin of corticospinal projections from the premotor areas in the frontal lobe. Journal of Neuroscience 11(3): 667–689.
[45]
Dum RP and Strick PL (1996) Spinal cord terminations of the medial wall motor areas in macaque monkeys. Journal of Neuroscience 16(20): 6513–6525.
[46]
Elsner B and Hommel B (2001) Effect Anticipation and Action Control. Journal of Experimental Psychology: Human Perception and Performance 27(1): 229.
[47]
Endres D, Chiovetto E, and Giese MA (2015) Bayesian approaches for learning of primitive-based compact representations of complex human activities. In: Dance Notations and Robot Motion. New York: Springer, pp. 117–137.
[48]
Englert P, Paraschos A, Deisenroth MP, and Peters J (2013) Probabilistic model-based imitation learning. Adaptive Behavior 21(5): 388–403.
[49]
Farhadi A and Tabrizi MK (2008) Learning to recognize activities from the wrong view point. In: European Conference on Computer Vision (ECCV 2008) (Lecture Notes in Computer Science, vol. 5302). Berlin: Springer, pp. 154–166.
[50]
Fihl P, Holte MB, Moeslund TB, and Reng L (2006) Action recognition using motion primitives and probabilistic edit distance. In: Articulated Motion and Deformable Objects. Berlin: Springer, pp. 375–384.
[51]
Fothergill S, Mentis H, Kohli P, and Nowozin S (2012) Instructing people for training gestural interactive systems. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. New York: ACM Press, pp. 1737–1746.
[52]
Francis BA and Wonham WM (1976) The internal model principle of control theory. Automatica 12(5): 457–465.
[53]
Fuster J (1999) Memory in the Cerebral Cortex. 2nd Ed. Cambridge, MA: MIT Press.
[54]
Gao Y, Vedula SS, and Reiley CEet al. (2014) Jhu-Isi gesture and skill assessment working set (jigsaws): A surgical activity dataset for human motion modeling. In: MICCAI Workshop: M2CAI, vol. 3, p. 3.
[55]
Gerlach C, Law I, Gade A, and Paulson O (2002) The role of action knowledge in the comprehension of artefacts—a PET study. NeuroImage 15(1): 143–152.
[56]
Gibson JJ (1966) The Senses Considered as Perceptual Systems. Houghton Mifflin.
[57]
Gibson JJ (1979) The Ecological Approach to Visual Perception. Psychology Press.
[58]
Gorelick L, Blank M, Shechtman E, Irani M, and Basri R (2007) Actions as space-time shapes. IEEE Transactions on Pattern Analysis and Machine Intelligence 29(12): 2247–2253.
[59]
Grafton S, Aziz-Zadeh L, and Ivry R (2009) Relative hierarchies and the representation of action. In: The Cognitive Neurosciences. Cambridge, MA: MIT Press, pp. 641–652.
[60]
Grave K and Behnke S (2012) Incremental action recognition and generalizing motion generation based on goal-directed features. In: 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE.
[61]
Grinke E, Tetzlaff C, Wörgötter F, and Manoonpong P (2015) Synaptic plasticity in a recurrent neural network for versatile and adaptive behaviors of a walking robot. Frontiers in Neurorobotics 9: 11.
[62]
Gritai A, Sheikh Y, Rao C, and Shah M (2009) Matching trajectories of anatomical landmarks under viewpoint, anthropometric and temporal transforms. International Journal of Computer Vision 84(3): 325–343.
[63]
Gross R (2015) Psychology: The Science of Mind and Behaviour. Hodder Education.
[64]
Guerra-Filho G and Aloimonos Y (2007) A language for human action. Computer 40(5): 42–51.
[65]
Guha A, Yang Y, Fermuuller C, and Aloimonos Y (2013) Minimalist plans for interpreting manipulation actions. In: 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE.
[66]
Guyon I, Athitsos V, Jangyodsuk P, Hamner B, and Escalante HJ (2012) Chalearn gesture challenge: Design and first results. In: 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE, pp. 1–6.
[67]
Hajimirsadeghi H, Ahmadabadi MN, and Araabi BN (2013) Conceptual imitation learning based on perceptual and functional characteristics of action. IEEE Transactions on Autonomous Mental Development 5(4): 311–325.
[68]
Hamilton A and Grafton ST (2007) The motor hierarchy: From kinematics to goals and intentions. Sensorimotor Foundations of Higher Cognition 22: 381–408.
[69]
Hamilton AFdC and Grafton ST (2006) Goal representation in human anterior intraparietal sulcus. Journal of Neuroscience 26(4): 1133–1137.
[70]
Hamilton AFdC and Grafton ST (2008) Action outcomes are represented in human inferior frontoparietal cortex. Cerebral Cortex 18(5): 1160–1168.
[71]
Haneda A, Okada K, and Inaba M (2008) Realtime manipulation planning system integrating symbolic and geometric planning under interactive dynamics siumlator. In: 2008 IEEE International Conference on Mechatronics and Automation. IEEE.
[72]
Hardwick RM, Caspers S, Eickhoff SB, and Swinnen SP (2017) Neural correlates of motor imagery, action observation, and movement execution: A comparison across quantitative meta-analyses. bioRxiv.
[73]
Harnad S (1990) The symbol grounding problem. Physica D: Nonlinear Phenomena 42(1): 335–346.
[74]
Hasson U, Chen J, and Honey CJ (2015) Hierarchical process memory: Memory as an integral component of information processing. Trends in Cognitive Sciences 19(6): 304–313.
[75]
Herzog DL and Krüger V (2012) Tracking in action space. In: Trends and Topics in Computer Vision. Berlin: Springer, pp. 100–113.
[76]
Hobaiter C (2017) What is a gesture? A meaning-based approach to defining gestural repertoires. Neuroscience and Biobehavioral Reviews 82: 3–12.
[77]
Hofer S and Brock O (2016) Coupled learning of action parameters and forward models for manipulation. In: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE.
[78]
Hongeng S and Wyatt J (2008) Learning causality and intentional actions. In: Towards Affordance-Based Robot Control. Berlin: Springer, pp. 27–46.
[79]
Hourdakis E and Trahanias P (2012) Computational modeling of observational learning inspired by the cortical underpinnings of human primates. Adaptive Behavior 20(4): 237–256.
[80]
Hu JF, Zheng WS, Lai J, and Zhang J (2017) Jointly Learning Heterogeneous Features for RGB-D Activity Recognition. In: IEEE Transactions on Pattern Analysis and Machine Intelligence, 1 November 2017, pp. 2186–2200.
[81]
Hwang BW, Kim S, and Lee SW (2007) A full-body gesture database for human gesture analysis. International Journal of Pattern Recognition and Artificial Intelligence 21(06): 1069–1084.
[82]
Hwang J and Tani J (2018) Seamless integration and coordination of cognitive skills in humanoid robots: A deep learning approach. IEEE Transactions on Cognitive and Developmental Systems 10(2): 345–358.
[83]
Ijjina EP and Krishna Mohan C (2016) Classification of human actions using pose-based features and stacked auto encoder. Pattern Recognition Letters 83: 268–277.
[84]
Jeannerod M (1984) The timing of natural prehension movements. Journal of Motor Behavior 16(3): 235–254.
[85]
Jeannerod M (1986) The formation of finger grip during prehension. A cortically mediated visuomotor pattern. Behavioural Brain Research 19(2): 99–116.
[86]
Jeannerod M (2006) Motor Cognition: What Actions tell the Self. Oxford: Oxford University Press.
[87]
Jeon Y, Sandhan T, and Choi JY (2015) Robust feature extraction for shift and direction invariant action recognition. In: Advances in Multimedia Information Processing (PCM 2015) (Lecture Notes in Computer Science, vol. 9315). New York: Springer, pp. 321–329.
[88]
Ji X and Liu H (2009) View-invariant human action recognition using exemplar-based hidden Markov models. In: Intelligent Robotics and Applications. Berlin: Springer, pp. 78–89.
[89]
Ji X, Liu H, and Li Y (2010) A new framework for view-invariant human action recognition. In: Advanced Information and Knowledge Processing. London: Springer, pp. 71–93.
[90]
Ji XF, Wu QQ, Ju ZJ, and Wang YY (2014) Study of human action recognition based on improved spatio-temporal features. International Journal of Automation and Computing 11(5): 500–509.
[91]
Ji Y, Yang Y, Xu X, and Shen HT (2018) One-shot learning based pattern transition map for action early recognition. Signal Processing 143: 364–370.
[92]
Jiang H and Martin DR (2008) Finding actions using shape flows. In: European Conference on Computer Vision (ECCV 2008) (Lecture Notes in Computer Science, vol. 5303). Berlin: Springer, pp. 278–292.
[93]
Johnson SH and Grafton ST (2003) From “acting on” to “acting with”: the functional anatomy of object-oriented action schemata. In: Neural Control of Space Coding and Action Production (Progress in Brain Research, vol. 142). Amsterdam: Elsevier, pp. 127–139.
[94]
Junejo IN, Dexter E, Laptev I, and Pérez P (2008) Cross-view action recognition from temporal self-similarities. In: European Conference on Computer Vision (ECCV 2008) (Lecture Notes in Computer Science, vol. 5303). Berlin: Springer, pp. 293–306.
[95]
Kaiser M (1997) Transfer of elementary skills via human–robot interaction. Adaptive Behavior 5(3-4): 249–280.
[96]
Karn NK and Jiang F (2016) Improved GLOH approach for one-shot learning human gesture recognition. In: Biometric Recognition. New York: Springer, pp. 441–452.
[97]
Keele SW and Jennings PJ (1992) Attention in the representation of sequence: Experiment and theory. Human Movement Science 11(1): 125–138.
[98]
Kemke C (2006) Natural language communication between human and artificial agents. In: Agent Computing and Multi-Agent Systems. Berlin: Springer, pp. 84–93.
[99]
Kitchenham B (2004) Procedures for performing systematic reviews. Technical Report TR/SE-0401, Keele University.
[100]
Kjellström H, Romero J, and Kragić D (2011) Visual object–action recognition: Inferring object affordances from human demonstration. Computer Vision and Image Understanding 115(1): 81–90.
[101]
Koniusz P, Cherian A, and Porikli F (2016) Tensor representations via kernel linearization for action recognition from 3D skeletons. In: Computer Vision – ECCV 2016. New York: Springer, pp. 37–53.
[102]
Krüger V (2006) Recognizing action primitives in complex actions using hidden Markov models. In: Advances in Visual Computing. Berlin: Springer, pp. 538–547.
[103]
Krüger V and Grest D (2007) Using hidden Markov models for recognizing action primitives in complex actions. In: Image Analysis. Berlin: Springer, pp. 203–212.
[104]
Krüger V, Herzog DL, Baby S, Ude A, and Kragić D (2010) Learning actions from observations. IEEE Robotics & Automation Magazine 17(2): 30–43.
[105]
Krüger V, Kragić D, Ude A, and Geib C (2007) The meaning of action: A review on action recognition and mapping. Advanced Robotics 21(13): 1473–1501.
[106]
Krüger N, Geib C, and Piater Jet al. (2011) Object–action complexes: Grounded abstractions of sensory–motor processes. Robotics and Autonomous Systems 59(10): 740–757.
[107]
Krüger V and Herzog D (2013) Tracking in object action space. Computer Vision and Image Understanding 117(7): 764–789.
[108]
Kuehne H, Jhuang H, Garrote E, Poggio T, and Serre T (2011) HMDB: A large video database for human motion recognition. In: 2011 IEEE International Conference on Computer Vision (ICCV). IEEE, pp. 2556–2563.
[109]
Kulkarni K, Boyer E, Horaud R, and Kale A (2011) An unsupervised framework for action recognition using actemes. In: Computer Vision – ACCV 2010. Berlin: Springer, pp. 592–605.
[110]
Kulkarni O, Parameswaran N, and Nagarajan R (1989) Action representation for planning using truth maintenance system. In: Fourth IEEE Region 10 International Conference TENCON. IEEE.
[111]
Kumar SH and Sivaprakash P (2013) New approach for action recognition using motion based features. In: 2013 IEEE Conference on Information and Communication Technologies. IEEE.
[112]
Laaksonen J, Felip J, Morales A, and Kyrki V (2010) Embodiment independent manipulation through action abstraction. In: 2010 IEEE International Conference on Robotics and Automation. IEEE.
[113]
Lallee (2010) Linking language with embodied and teleological representations of action for humanoid cognition. Frontiers in Neurorobotics 4: 8.
[114]
Layher G, Brosch T, and Neumann H (2017) Real-time biologically inspired action recognition from key poses using a neuromorphic architecture. Frontiers in Neurorobotics 11: 13.
[115]
Lea C, Vidal R, and Hager GD (2016) Learning convolutional action primitives for fine-grained action recognition. In: 2016 IEEE International Conference on Robotics and Automation (ICRA). IEEE.
[116]
Lee S and Chen J (1996) Robot skill discovery based on observed data(003) 5335512. In: Proceedings of IEEE International Conference on Robotics and Automation. IEEE.
[117]
Li W, Zhang Z, and Liu Z (2010) Action recognition based on a bag of 3D points. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE, pp. 9–14.
[118]
Liu J, Luo J, and Shah M (2009) Recognizing realistic actions from videos “in the wild”. In: IEEE Conference on Computer Vision and Pattern Recognition, 2009 (CVPR 2009). IEEE, pp. 1996–2003.
[119]
Liu M, Liu H, Sun Q, Zhang T, and Ding R (2016) Salient pairwise spatio-temporal interest points for real-time activity recognition. CAAI Transactions on Intelligence Technology 1(1): 14–29.
[120]
Mandik P (2005) Action-oriented representation. In: Brook A and Akins K (eds.) Cognition and the Brain: The Philosophy and Neuroscience Movement. Cambridge: Cambridge University Press, pp. 284–305.
[121]
Mansur A, Makihara Y, and Yagi Y (2011) Action recognition using dynamics features. In: 2011 IEEE International Conference on Robotics and Automation. IEEE.
[122]
Markievicz I, Vitkute-Adzgauskiene D, and Tamosiunaite M (2013) Semi-supervised learning of action ontology from domain-specific corpora. In: Communications in Computer and Information Science. Berlin: Springer, pp. 173–185.
[123]
Marocco D (2010) Grounding action words in the sensorimotor interaction with the world: Experiments with a simulated iCub humanoid robot. Frontiers in Neurorobotics 4: 7.
[124]
Marszalek M, Laptev I, and Schmid C (2009) Actions in context. In: IEEE Conference on Computer Vision and Pattern Recognition, 2009 (CVPR 2009). IEEE, pp. 2929–2936.
[125]
Martinet LE, Fouque B, Passot JB, Meyer JA, and Arleo A (2008) Modelling the cortical columnar organisation for topological state-space representation, and action planning. In: International Conference on Simulation of Adaptive Behavior (SAB 2008): From Animals to Animats 10 (Lecture Notes in Computer Science, vol. 5040). Berlin: Springer, pp. 137–147.
[126]
Maye A and Engel AK (2013) Extending sensorimotor contingency theory: prediction, planning, and action generation. Adaptive Behavior 21(6): 423–436.
[127]
Mikolajczyk K and Uemura H (2011) Action recognition with appearance–motion features and fast search trees. Computer Vision and Image Understanding 115(3): 426–438.
[128]
Mohan V, Metta G, Zenzeri J, and Morasso P (2010) Teaching humanoids to imitate ‘shapes’ of movements. In: Artificial Neural Networks – ICANN 2010. Berlin: Springer, pp. 234–244.
[129]
Mohan V, Morasso P, Sandini G, and Kasderidis S (2013) Inference through embodied simulation in cognitive robots. Cognitive Computation 5(3): 355–382.
[130]
Mugan J and Kuipers B (2012) Autonomous learning of high-level states and actions in continuous environments. IEEE Transactions on Autonomous Mental Development 4(1): 70–86.
[131]
Mukovskiy A, Vassallo C, Naveau M, Stasse O, Souères P, and Giese MA (2017) Adaptive synthesis of dynamically feasible full-body movements for the humanoid robot HRP-2 by flexible combination of learned dynamic movement primitives. Robotics and Autonomous Systems 91: 270–283.
[132]
Müller M, Röder T, Clausen M, Eberhardt B, Krüger B, and Weber A (2007) Documentation mocap database HDM05. Technical Report CG-2007-2, Universität Bonn. Available at: URL http://resources.mpi-inf.mpg.de/HDM05/.
[133]
Mülling K, Kober J, and Peters J (2011) A biomimetic approach to robot table tennis. Adaptive Behavior 19(5): 359–376.
[134]
Naito E, Morita T, and Amemiya K (2016) Body representations in the human brain revealed by kinesthetic illusions and their essential contributions to motor control and corporeal awareness. Neuroscience Research 104: 16–30.
[135]
Nakajo R, Murata S, Arie H, and Ogata T (2015) Acquisition of viewpoint representation in imitative learning from own sensory-motor experiences. In: 2015 Joint IEEE International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob). IEEE.
[136]
Natarajan P, Banerjee P, Khan FM, and Nevatia R (2009) Graphical framework for action recognition using temporally dense STIPs. In: 2009 Workshop on Motion and Video Computing (WMVC). IEEE.
[137]
Newton NW (2017) Understanding and self-organization. Frontiers in Systems Neuroscience 11(8): 1–9.
[138]
Ni B, Wang G, and Moulin P (2013) RGBD-HUDAACT: A color-depth video database for human daily activity recognition. In: Consumer Depth Cameras for Computer Vision. Berlin: Springer, pp. 193–208.
[139]
Niebles JC, Chen CW, and Fei-Fei L (2010) Modeling temporal structure of decomposable motion segments for activity classification. In: Daniilidis K, Maragos P, and Paragios N (eds.) Computer Vision – ECCV 2010. Berlin: Springer, pp. 392–405.
[140]
Nishimoto R, Namikawa J, and Tani J (2008) Learning multiple goal-directed actions through self-organization of a dynamic neural network model: A humanoid robot experiment. Adaptive Behavior 16(2-3): 166–181.
[141]
Nishimoto R and Tani J (2009) Development process of functional hierarchy for actions and motor imagery. In: 2009 IEEE 8th International Conference on Development and Learning. IEEE.
[142]
Noda K, Kawamoto K, Hasuo T, and Sabe K (2011) A generative model for developmental understanding of visuomotor experience. In: 2011 IEEE International Conference on Development and Learning (ICDL). IEEE.
[143]
M Nussbaum (ed.) (1985) Aristotle’s De Motu Animalium: Text with Translation, Commentary, and Interpretive Essays. Princeton Paperbacks. Princeton, NJ: Princeton University Press.
[144]
Ofli F, Chaudhry R, Kurillo G, Vidal R, and Bajcsy R (2013) Berkeley MHAD: A comprehensive multimodal human action database. In: 2013 IEEE Workshop on Applications of Computer Vision (WACV), pp. 53–60.
[145]
Ogawara K, Iba S, Tanuki T, Kimura H, and Ikeuchi K (2001) Acquiring hand-action models by attention point analysis. In: Proceedings 2001 IEEE International Conference on Robotics and Automation. IEEE.
[146]
Ognibene D, Volpi NC, Pezzulo G, and Baldassare G (2013a) Learning epistemic actions in model-free memory-free reinforcement learning: Experiments with a neuro-robotic model. In: Biomimetic and Biohybrid Systems. Berlin: Springer, pp. 191–203.
[147]
Ognibene D, Wu Y, Lee K, and Demiris Y (2013b) Hierarchies for embodied action perception. In: Computational and Robotic Models of the Hierarchical Organization of Behavior. Berlin: Springer, pp. 81–98.
[148]
Oreifej O and Liu Z (2013) HON4d: Histogram of oriented 4d normals for activity recognition from depth sequences. In: 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, pp. 716–723.
[149]
O’Shaughnessy B (1997) Trying (as the mental ’pineal gland’). In: Mele AR (ed.) The Philosophy of Action. Oxford: Oxford University Press, pp. 365–386.
[150]
Panchev C (2005) A spiking neural network model of multi-modal language processing of robot instructions. In: Biomimetic Neural Learning for Intelligent Robots. Berlin: Springer, pp. 182–210.
[151]
Panzner M and Cimiano P (2016) Comparing hidden Markov models and long short term memory neural networks for learning action representations. In: International Workshop on Machine Learning, Optimization, and Big Data (Lecture Notes in Computer Science, vol. 10122). New York: Springer, pp. 94–105.
[152]
Parisi GI, Weber C, and Wermter S (2015) Self-organizing neural integration of pose-motion features for human action recognition. Frontiers in Neurorobotics 9: 3.
[153]
Park JC, Kim DS, and Nagai Y (2017) Learning for goal-directed actions using RNNPB: Developmental change of “what to imitate”. IEEE Transactions on Cognitive and Developmental Systems 10(3): 545–556.
[154]
Patel M, Miro JV, Kragić D, Ek CH, and Dissanayake G (2014) Learning object, grasping and manipulation activities using hierarchical HMMs. Autonomous Robots 37(3): 317–331.
[155]
Paulus M, van Dam W, Hunnius S, Lindemann O, and Bekkering H (2011) Action–effect binding by observational learning. Psychonomic Bulletin and Review 18(5): 1022.
[156]
Paxton C, Jonathan F, Kobilarov M, and Hager GD (2016) Do what I want, not what I did: Imitation of skills by planning sequences of actions. In: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE.
[157]
Pazhoumand-Dar H, Lam CP, and Masek M (2015) Joint movement similarities for robust 3d action recognition using skeletal data. Journal of Visual Communication and Image Representation 30: 10–21.
[158]
Pezzulo G and Dindo H (2011) What should I do next? Using shared representations to solve interaction problems. Experimental Brain Research 211(3–4): 613–630.
[159]
Pierobon M, Marcon M, Sarti A, and Tubaro S (2005) Clustering of human actions using invariant body shape descriptor and dynamic time warping. In: Proceedings IEEE Conference on Advanced Video and Signal Based Surveillance, 2005. IEEE.
[160]
Pieropan A, Salvi G, Pauwels K, and Kjellstrom H (2014) Audio-visual classification and detection of human manipulation actions. In: 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE.
[161]
Pirsiavash H and Ramanan D (2012) Detecting activities of daily living in first-person camera views. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, pp. 2847–2854.
[162]
Pitti A, Alirezaei H, and Kuniyoshi Y (2009) Cross-modal and scale-free action representations through enaction. Neural Networks 22(2): 144–154.
[163]
Rahman SA, Song I, Leung M, Lee I, and Lee K (2014) Fast action recognition using negative space features. Expert Systems with Applications 41(2): 574–587.
[164]
Rahman SA, Song I, and Leung MKH (2012) Negative space template: A novel feature to describe activities in video. In: The 2012 International Joint Conference on Neural Networks (IJCNN), pp. 1–7.
[165]
Ramirez-Amaro K, Beetz M, and Cheng G (2017) Transferring skills to humanoid robots by extracting semantic representations from observations of human activities. Artificial Intelligence 247: 95–118.
[166]
Razzaghi P, Palhang M, and Gheissari N (2012) A new invariant descriptor for action recognition based on spherical harmonics. Pattern Analysis and Applications 16(4): 507–518.
[167]
Regneri M, Rohrbach M, Wetzel D, Thater S, Schiele B, and Pinkal M (2013) Grounding action descriptions in videos. Transactions of the Association of Computational Linguistics 1: 25–36.
[168]
Reng L, Moeslund TB, and Granum E (2005) Finding motion primitives in human body gestures. In: International Gesture Workshop. Berlin: Springer, pp. 133–144.
[169]
Rizzolatti G and Craighero L (2004) The mirror-neuron system. Annual Reviews in Neuroscience 27: 169–192.
[170]
Rizzolatti G and Luppino G (2001) The cortical motor system. Neuron 31(6): 889–901.
[171]
Rizzolatti G and Matelli M (2003) Two different streams form the dorsal visual system: Anatomy and functions. Experimental Brain Research 153(2): 146–157.
[172]
Rodriguez MD, Ahmed J, and Shah M (2008) Action mach a spatio-temporal maximum average correlation height filter for action recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, 2008 (CVPR 2008). IEEE, pp. 1–8.
[173]
Roh MC, Shin HK, and Lee SW (2010) View-independent human action recognition with volume motion template on single stereo camera. Pattern Recognition Letters 31(7): 639–647.
[174]
Roland PE, Larsen B, Lassen NA, and Skinhoj E (1980a) Supplementary motor area and other cortical areas in organization of voluntary movements in man. Journal of Neurophysiology 43(1): 118–136.
[175]
Roland PE, Skinhoj E, Lassen NA, and Larsen B (1980b) Different cortical areas in man in organization of voluntary movements in extrapersonal space. Journal of Neurophysiology 43(1): 137–150.
[176]
Rosales R and Sclaroff S (2003) A framework for heading-guided recognition of human activity. Computer Vision and Image Understanding 91(3): 335–367.
[177]
Rosenbaum DA, Meulenbroek RG, and Vaughan J (2001) Planning reaching and grasping movements: Theoretical premises and practical implications. Motor Control 5(2): 99–115.
[178]
Rosenbaum DA, Vaughan J, Barnes HJ, and Jorgensen MJ (1992) Time course of movement planning: Selection of handgrips for object manipulation. Journal of Experimental Psychology: Learning, Memory, and Cognition 18(5): 1058.
[179]
Rosenfeld A and Ullman S (2016) Hand–object interaction and precise localization in transitive action recognition. In: 2016 13th Conference on Computer and Robot Vision (CRV). IEEE.
[180]
Roshtkhari MJ and Levine MD (2012) A multi-scale hierarchical codebook method for human action recognition in videos using a single example. In: 2012 Ninth Conference on Computer and Robot Vision. IEEE.
[181]
Roshtkhari MJ and Levine MD (2013) Human activity recognition in videos using a single example. Image and Vision Computing 31(11): 864–876.
[182]
Rudolph M, Muhlig M, Gienger M, and Bohme HJ (2010) Learning the consequences of actions: Representing effects as feature changes. In: 2010 International Conference on Emerging Security Technologies. IEEE.
[183]
Russakovsky O, Deng J, and Su Het al. (2015) ImageNet large scale visual recognition challenge. International Journal of Computer Vision 115(3): 211–252.
[184]
Russell SJ and Norvig P (2016) Artificial Intelligence: A Modern Approach. 3rd Ed. New York: Pearson Education.
[185]
Ryan RM and Deci EL (2000) Intrinsic and extrinsic motivations: Classic definitions and new directions. Contemporary Educational Psychology 25(1): 54–67.
[186]
Ryan S and Andreae J (1993) Learning sequential and continuous control. In: Proceedings 1993 The First New Zealand International Two-Stream Conference on Artificial Neural Networks and Expert Systems. Los Alamitos, CA: IEEE Computer Society Press.
[187]
Ryoo MS and Aggarwal JK (2010) UT-Interaction dataset, ICPR contest on Semantic Description of Human Activities (SDHA). Available at: http://cvrc.ece.utexas.edu/SDHA2010/Human_Interaction.html.
[188]
Salih AAA and Youssef C (2016) Spatiotemporal representation of 3D skeleton joints-based action recognition using modified spherical harmonics. Pattern Recognition Letters 83: 32–41.
[189]
Sanchez-Riera J, Čech J, and Horaud R (2012) Action recognition robust to background clutter by using stereo vision. In: Computer Vision – ECCV 2012. Workshops and Demonstrations. Berlin: Springer, pp. 332–341.
[190]
Sanmohan and Krüger V (2009) Primitive based action representation and recognition. In: Image Analysis. Berlin: Springer, pp. 31–40.
[191]
Sapienza M, Cuzzolin F, and Torr PH (2013) Learning discriminative space–time action parts from weakly labelled videos. International Journal of Computer Vision 110(1): 30–47.
[192]
Schenatti M, Lorenzo Natale G, Metta, and Sandini G (2003) Object grasping data-set. University of Genova, Italy. Available at: URL http://www.lira.dist.unige.it/
[193]
Schüldt C, Laptev I, and Caputo B (2004) Recognizing human actions: A local SVM approach. In: Proceedings of the 17th International Conference on Pattern Recognition, 2004 (ICPR 2004), vol. 3. IEEE, pp. 32–36.
[194]
Searle JR, Dennett DC, and Chalmers DJ (1997) The Mystery of Consciousness. New York Review of Books.
[195]
Seidenari L, Varano V, Berretti S, Bimbo A, and Pala P (2013) Recognizing actions from depth cameras as weakly aligned multi-part bag-of-poses. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 479–485.
[196]
Shah D, Falco P, Saveriano M, and Lee D (2016) Encoding human actions with a frequency domain approach. In: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE.
[197]
Shan Y, Zhang Z, and Huang K (2015) Learning skeleton stream patterns with slow feature analysis for action recognition. In: Computer Vision - ECCV 2014 Workshops. New York: Springer, pp. 111–121.
[198]
She L, Cheng Y, Chai JY, Jia Y, Yang S, and Xi N (2014) Teaching robots new actions through natural language instructions. In: The 23rd IEEE International Symposium on Robot and Human Interactive Communication. IEEE.
[199]
Shimozaki M and Kuniyoshi Y (2003) Integration of spatial and temporal contexts for action recognition by self organizing neural networks. In: Proceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003). IEEE.
[200]
Silva C and Ribeiro B (2003) Navigating mobile robots with a modular neural architecture. Neural Computing and Applications 12(3–4): 200–211.
[201]
Singh S, Velastin SA, and Ragheb H (2010) MUHAVI: A multicamera human action video dataset for the evaluation of action recognition methods. In: 2010 Seventh IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS). IEEE, pp. 48–55.
[202]
Sokolov AA, Gharabaghi A, Tatagiba MS, and Pavlova M (2010) Cerebellar engagement in an action observation network. Cerebral Cortex 20(2): 486–491.
[203]
Soomro K, Zamir AR, and Shah M (2012) UCF101: A dataset of 101 human actions classes from videos in the wild. CoRR abs/1212.0402.
[204]
Sorokin I, Seleznev A, Pavlov M, Fedorov A, and Ignateva A (2015) Deep attention recurrent Q-network. In: NIPS Workshop on Deep Reinforcement Learning. Montreal, Canada: Curran Associates, Inc.
[205]
Steels L (2003) Intelligence with representation. Philosophical Transactions of the Royal Society of London A: Mathematical, Physical and Engineering Sciences 361(1811): 2381–2395.
[206]
Stein S and McKenna SJ (2013) Combining embedded accelerometers with computer vision for recognizing food preparation activities. In: Proceedings of the 2013 ACM International Joint Conference on Pervasive and Ubiquitous Computing. New York: ACM Press, pp. 729–738.
[207]
Stenmark M and Malec J (2015) Knowledge-based instruction of manipulation tasks for industrial robotics. Robotics and Computer-Integrated Manufacturing 33: 56–67.
[208]
Stenmark M and Nugues P (2013) Natural language programming of industrial robots. In: IEEE ISR 2013, pp. 1–5.
[209]
Su YC and Grauman K (2016) Leaving some stones unturned: Dynamic feature prioritization for activity detection in streaming video. In: Computer Vision – ECCV 2016. Berlin: Springer, pp. 783–800.
[210]
Sugita Y and Tani J (2008) A sub-symbolic process underlying the usage-based acquisition of a compositional representation: Results of robotic learning experiments of goal-directed actions. In: 2008 7th IEEE International Conference on Development and Learning. IEEE.
[211]
Sun Q and Liu H (2013) Action disambiguation analysis using normalized google-like distance correlogram. In: Computer Vision – ACCV 2012. Berlin: Springer, pp. 425–437.
[212]
Sun Q, Liu H, Ma L, and Zhang T (2016) A novel hierarchical bag-of-words model for compact action representation. Neurocomputing 174: 722–732.
[213]
Sung J, Ponce C, Selman B, and Saxena A (2012) Unstructured human activity detection from RGBD images. In: 2012 IEEE International Conference on Robotics and Automation (ICRA). IEEE, pp. 842–849.
[214]
Tenorth M and Beetz M (2012) A unified representation for reasoning about robot actions, processes, and their effects on objects. In: 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE.
[215]
Tessitore G, Prevete R, Catanzariti E, and Tamburrini G (2010) From motor to sensory processing in mirror neuron computational modelling. Biological Cybernetics 103(6): 471–485.
[216]
Theodoridis T, Agapitos A, Hu H, and Lucas SM (2008) Ubiquitous robotics in physical human action recognition: A comparison between dynamic ANNs and GP. In: 2008 IEEE International Conference on Robotics and Automation. IEEE.
[217]
Thrun S and Mitchell TM (1995) Lifelong robot learning. In: Steels L (ed.) The Biology and Technology of Intelligent Autonomous Agents. Berlin: Springer, pp. 165–196.
[218]
Thurau C (2007) Behavior histograms for action recognition and human detection. In: Human Motion – Understanding, Modeling, Capture and Animation. Berlin: Springer, pp. 299–312.
[219]
Thurau C and Hlaváč V (2007) n-grams of action primitives for recognizing human behavior. In: Computer Analysis of Images and Patterns. Berlin: Springer, pp. 93–100.
[220]
Thurau C and Hlaváč V (2009) Recognizing human actions by their pose. In: Statistical and Geometrical Approaches to Visual Motion Analysis (Lecture Notes in Computer Science, vol. 5604). Berlin: Springer, pp. 169–192.
[221]
Tucker M and Ellis R (1998) On the relations between seen objects and components of potential actions. Journal of Experimental Psychology: Human Perception and Performance 24(3): 830–846.
[222]
Tunik E, Frey S, and Grafton ST (2005) Virtual lesions of the anterior intraparietal area disrupt goal-dependent on-line adjustments of grasp. Nature Neuroscience 8: 505–11.
[223]
Vafeias E and Ramamoorthy S (2014) Joint classification of actions and object state changes with a latent variable discriminative model. In: 2014 IEEE International Conference on Robotics and Automation (ICRA). IEEE.
[224]
Vanderelst D and Winfield A (2017) Rational imitation for robots: the cost difference model. Adaptive Behavior 25(2): 60–71.
[225]
Veeraraghavan A, Chellappa R, and Roy-Chowdhury AK (2006) The function space of an activity. In: 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 1. IEEE, pp. 959–968.
[226]
Vieira AW, Nascimento ER, Oliveira GL, Liu Z, and Campos MF (2014) On the improvement of human action recognition from depth map sequences using space–time occupancy patterns. Pattern Recognition Letters 36: 221–227.
[227]
Vieira AW, Nascimento ER, Oliveira GL, Liu Z, and Campos MFM (2012) STOP: Space-time occupancy patterns for 3D action recognition from depth map sequences. In: Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications. Berlin: Springer, pp. 252–259.
[228]
Viet VH, Ngoc LQ, Son TT, and Hoang PM (2015) Multiple kernel learning and optical flow for action recognition in RGB-d video. In: 2015 Seventh International Conference on Knowledge and Systems Engineering (KSE). IEEE.
[229]
Vitkute-Adzgauskiene D, Markievicz I, and Krilavicius Tet al. (2014) Chemlab corpus. EU-FP7-STREP (600578) ACAT, Learning and Execution of Action Categories, D1.1: Text corpora and image databases.
[230]
von Goethe JW (1808) Faust, Eine Tragödie. Tübingen: Cotta.
[231]
Vosgerau G (2009) Mental Representation and Self-Consciousness: From Basic Self-Representation to Self-Related Cognition. Mentis.
[232]
Vuga R, Aksoy EE, Wörgötter F, and Ude A (2015) Probabilistic semantic models for manipulation action representation and extraction. Robotics and Autonomous Systems 65: 40–56.
[233]
Wang J, Liu Z, Wu Y, and Yuan J (2012) Mining actionlet ensemble for action recognition with depth cameras. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, pp. 1290–1297.
[234]
Wang L, Li R, and Fang Y (2017) Power difference template for action recognition. Machine Vision and Applications 28(5-6): 463–473.
[235]
Webb TW, Kean HH, and Graziano MSA (2016) Effects of awareness on the control of attention. Journal of Cognitive Neuroscience 28(6): 842–851.
[236]
Wechsler H, Duric Z, and Li F (2002) Hierarchical interpretation of human activities using competitive learning. In: Object Recognition Supported by User Interaction for Service Robots. Los Alamitos, CA: IEEE Computer Society Press.
[237]
Weiller D (2010) Unsupervised learning of reflexive and action-based affordances to model adaptive navigational behavior. Frontiers in Neurorobotics 4: 2.
[238]
Weinland D, Ronfard R, and Boyer E (2006) Free viewpoint action recognition using motion history volumes. Computer Vision and Image Understanding 104(2–3): 249–257.
[239]
Weinland D, Ronfard R, and Boyer E (2011) A survey of vision-based methods for action representation, segmentation and recognition. Computer Vision and Image Understanding 115(2): 224–241.
[240]
Wermter S, Weber C, Elshaw M, Gallese V, and Pulvermüller F (2005) Grounding neural robot language in action. In: Biomimetic Neural Learning for Intelligent Robots. Berlin: Springer, pp. 162–181.
[241]
Whiten C, Laganiere R, and Bilodeau GA (2013) Efficient action recognition with MoFREAK. In: 2013 International Conference on Computer and Robot Vision. IEEE.
[242]
Wolpert DM and Kawato M (1998) Multiple Paired Forward and Inverse Models for Motor Control. Neural networks 11(7-8): 1317–1329.
[243]
Worgotter F, Aksoy EE, Kruger N, Piater J, Ude A, and Tamosiunaite M (2013) A simple ontology of manipulation actions based on hand–object relations. IEEE Transactions on Autonomous Mental Development 5(2): 117–134.
[244]
Wu Z, Song S, and Khosla Aet al. (2015) 3D ShapeNets: A deep representation for volumetric shapes. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1912–1920.
[245]
Xi N and Tarn TJ (1999) Journal of Intelligent and Robotic Systems 25(4): 281–293.
[246]
Xia L, Chen CC, and Aggarwal JK (2012) View invariant human action recognition using histograms of 3D joints. In: 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE, pp. 20–27.
[247]
Xiao Q and Cheng J (2013) Human action recognition framework by fusing multiple features. In: 2013 IEEE International Conference on Information and Automation (ICIA). IEEE.
[248]
Yang J, Xu Y, and Chen C (1997) Human action learning via hidden Markov model. IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans 27(1): 34–44.
[249]
Yang Y, Guha A, Fermuller C, and Aloimonos Y (2014) Manipulation action tree bank: A knowledge resource for humanoids. In: 2014 IEEE-RAS International Conference on Humanoid Robots. IEEE.
[250]
Yao B, Jiang X, Khosla A, Lin AL, Guibas L, and Fei-Fei L (2011) Human action recognition by learning bases of action attributes and parts. In: 2011 IEEE International Conference on Computer Vision (ICCV). IEEE, pp. 1331–1338.
[251]
Zambelli M and Demiris Y (2017) Online multimodal ensemble learning using self-learned sensorimotor representations. IEEE Transactions on Cognitive and Developmental Systems 9(2): 113–126.
[252]
Zampogiannis K, Yang Y, Fermuller C, and Aloimonos Y (2015) Learning the spatial semantics of manipulation actions through preposition grounding. In: 2015 IEEE International Conference on Robotics and Automation (ICRA). IEEE.
[253]
Zech P, Haller S, Rezapour Lakani S, Ridge B, Ugur E, and Piater J (2017) Computational models of affordance in robotics: A taxonomy and systematic classification. Adaptive Behavior 25(5): 235–271.
[254]
Zhang C, Zhang H, Guo R, and Parker LE (2016) Unified robot learning of action labels and motion trajectories from 3D human skeletal data. In: 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN). IEEE.
[255]
Zhang J and Zhuang Y (2007) View-independent human action recognition by action hypersphere in nonlinear subspace. In: Advances in Multimedia Information Processing PCM 2007. Berlin: Springer, pp. 108–117.
[256]
Zhang Z, Hu Y, Chan S, and Chia LT (2008) Motion context: A new representation for human action recognition. In: European Conference on Computer Vision (ECCV 2008) (Lecture Notes in Computer Science, vol. 5305). Berlin: Springer, pp. 817–829.
[257]
Zhou Z, Song M, Zhang L, Tao D, Bu J, and Chen C (2011) kPose: A new representation for action recognition. In: Computer Vision – ACCV 2010. Berlin: Springer, pp. 436–447.
[258]
Zhu G, Zhang L, Shen P, and Song J (2016) Human action recognition using multi-layer codebooks of key poses and atomic motions. Signal Processing: Image Communication 42: 19–30.
[259]
Zimmer M and Doncieux S (2018) Bootstrapping $q$-learning for robotics from neuro-evolution results. IEEE Transactions on Cognitive and Developmental Systems 10(1): 102–119.

Cited By

View all

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image International Journal of Robotics Research
International Journal of Robotics Research  Volume 38, Issue 5
Apr 2019
123 pages
This article is distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 License (http://www.creativecommons.org/licenses/by-nc/4.0/) which permits non-commercial use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/open-access-at-sage).

Publisher

Sage Publications, Inc.

United States

Publication History

Published: 01 April 2019

Author Tags

  1. action representations
  2. robotics
  3. taxonomy

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 05 Jan 2025

Other Metrics

Citations

Cited By

View all

View Options

View options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media