Next Article in Journal
Characteristics of Electroencephalogram in the Prefrontal Cortex during Deep Brain Stimulation of Subthalamic Nucleus in Parkinson’s Disease under Propofol General Anesthesia
Next Article in Special Issue
Impact of Physical Exercise Alone or in Combination with Cognitive Remediation on Cognitive Functions in People with Schizophrenia: A Qualitative Critical Review
Previous Article in Journal
Water Drinking Behavior Associated with Aversive Arousal in Rats: An Integrative Approach
Previous Article in Special Issue
Hand and Foot Selection in Mental Body Rotations Involves Motor-Cognitive Interactions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Large Video Set of Natural Human Actions for Visual and Cognitive Neuroscience Studies and Its Validation with fMRI

1
Department of Psychology, Bilkent University, 06800 Ankara, Türkiye
2
Interdisciplinary Neuroscience Graduate Program, Bilkent University, 06800 Ankara, Türkiye
3
Aysel Sabuncu Brain Research Center and National Magnetic Resonance Research Center (UMRAM), Bilkent University, 06800 Ankara, Türkiye
4
Department of Psychology, Justus Liebig University in Giessen, 35394 Giessen, Germany
5
Department of Medicine and Surgery, University of Parma, 43125 Parma, Italy
*
Author to whom correspondence should be addressed.
Brain Sci. 2023, 13(1), 61; https://doi.org/10.3390/brainsci13010061
Submission received: 30 November 2022 / Revised: 14 December 2022 / Accepted: 22 December 2022 / Published: 29 December 2022
(This article belongs to the Special Issue From Bench to Bedside: Motor-Cognitive Interactions)

Abstract

:
The investigation of the perception of others’ actions and underlying neural mechanisms has been hampered by the lack of a comprehensive stimulus set covering the human behavioral repertoire. To fill this void, we present a video set showing 100 human actions recorded in natural settings, covering the human repertoire except for emotion-driven (e.g., sexual) actions and those involving implements (e.g., tools). We validated the set using fMRI and showed that observation of the 100 actions activated the well-established action observation network. We also quantified the videos’ low-level visual features (luminance, optic flow, and edges). Thus, this comprehensive video set is a valuable resource for perceptual and neuronal studies.

Graphical Abstract

1. Introduction

One of the most important skills social organisms possess is the ability to perceive the actions of conspecifics in their environment. This skill has a survival value since it allows the organisms to take the appropriate action based on what they perceive. For instance, if somebody smiles at you, you probably smile back but if that person is about to attack you, you will probably want to flee. Therefore, the perceptual and neural mechanisms of action perception have received great interest from psychologists, systems, and cognitive neuroscientists.
Although there is a growing body of research in observed action processing [1], the range of actions used in such empirical studies has been limited. Since the beginning of 1990s, the majority of neuroscience studies have used grasping as the exemplary action to study action observation [1]. More recent work has extended this work by introducing other action categories, such as locomotion actions (e.g., walking), communicative actions (e.g., gestures), self-directed actions (e.g., scratching one’s own body), interaction actions (e.g., hugging), vocal actions (e.g., singing), or other manipulative actions (e.g., pushing an object) [2,3,4,5]. This body of work has demonstrated the significance of studying action categories other than grasping as the neural activations for different actions were localized in different brain regions especially at the level of parietal cortex. However, each individual study created its own limited set of action videos, and it has been difficult to make comparisons across studies due to variations in terms of actors, scenes, video durations, and video quality (e.g., frame per seconds or resolution). Given the richness of the human action repertoire, the aim of the current study is to introduce a large video set of actions performed in natural settings, covering the entire human repertoire and make it available to researchers working primarily in the fields of visual, cognitive neuroscience and psychology. We believe that such an effort is necessary and timely given recent work, indicated above, that shows differences in neural representations of different actions [2,3,4,5], the move to use more naturalistic stimuli in neuroscience studies [6,7], and the availability of multivariate pattern analysis techniques to analyze experimental data that has a rich set of conditions, i.e., a wide variety of actions [8], instead of a few of them, as is traditionally done in previous studies. It is important to note that in creating this database, we aim to address the shortcomings of earlier attempts, such as those that provided action stimuli in the form of point-light displays (See Table 1 in [9], as they do not have high ecological validity, and the ones that come from the computer vision community ([10,11]), as they are usually unconstrained in terms of actors, context, camera angle, and movements.
The action set presented in the current paper has two major features. It represents, as far as we know, the first attempt in visual and cognitive neuroscience to systematically encompasses the entire human behavioral repertoire, with the exception of emotion-driven actions (e.g., sexual) and those involving implements (e.g., using a tool or driving a car). Second, it aims to concentrate on actions that are evolutionarily old, i.e., actions to which the human brain would have adapted during evolution. Thus, all videos were recorded in natural settings (beaches, parks, riverbanks) avoiding artificial structures in the background, and using natural objects as targets of the actions (e.g., we used stones, fruits, and pieces of wood instead of man-made items). These specifications define a homogeneous group of action exemplars and distinguish our set from the large set of 80 atomic actions, collected by the Google research group from the internet [12].
Crucially, the action exemplars in this set are unrelated to any a-priori categories. Such a stimulus-driven approach, inspired by an fMRI study of voxel-level selectivity for the meanings of words [13], can be considered complementary to the earlier studies in which the action exemplars were selected a priori as part of a single class (e.g., dragging, grasping, dropping, and pushing considered to be manipulative actions in earlier studies) [3,5,14]. On the other hand, it is important to note that, unlike the studies that use continuous natural movies (e.g., [13,15]), our stimuli set includes human actions without the clutter of other movements such as those of objects or other actors. Therefore, it constitutes a more suitable dataset for researchers who would like to study visual action perception and processing in humans.

2. Materials and Methods

In this section, we describe the stimulus set and the post-processing of the videos. The stimulus set is freely available and can be downloaded from https://osf.io/u62bp/?view_only=393a2924aa05461394fe9f3171863b94 (accessed on 29 December 2022). We also carried out a validation study with fMRI to show that our stimuli drive the regions established to be associated with the processing of actions, also known as the Action Observation Network [1,7,8]. The fMRI data can also be downloaded from the same link above.

2.1. Stimulus Set

Actors: Four actors performed the actions (2 males, 2 females). One additional female actor accompanied some actors in videos portraying actions that involved two individuals. They were undergraduate students at the University of Parma. We did not choose professional actors for two reasons. First, we wanted to record actions that were as natural as possible, without stylized movements that could be introduced by professional actors. Second, we wanted to have variability in the body movements which reflected individual differences. Before the recording of each action video, all actors were directed concerning how to perform the action and instructed to perform the action as naturally as possible within 3 s. For each action, several recordings were taken one after the other to make sure that the action was performed as intended in terms of timing and naturalness, and the best recording was chosen during the post-processing of videos. The actors were paid for their participation in the recordings and gave informed consent for their videos to be published in scientific journals.
Actions: One hundred different actions were recorded for several seconds (at least 3 s) and each action was performed by 3 of the 4 actors. So, in total, we recorded 300 videos with a fixed camera. Actions involved various effectors including the fingers, hand, arm, foot, mouth, upper body, or full body. One or 2 actors were portrayed in each video. When two actors were present, one could target the other with his action, or the two could interact. The actions were recorded from a lateral viewpoint. A sample frame from each video is presented in Figure 1. Table 1 lists the actions shown in the 100 videos and the actors who perform each of them. The different actions are described in the Supplementary Information.

2.2. Post-Processing of the Videos

The videos were recorded using Panasonic HCX 900 camcorders. After recording, the videos were edited using Final Cut Pro software and 3-s clips were made. The frame rate of these videos was 50 fps, so each video consisted of 150 frames. The size (height and width) of the frames was set to 314 × 410 pixels. The 3-s videos were then exported in .avi format and compressed using MPlayer’s mencoder command (http://www.mplayerhq.hu/, accessed on 29 December 2022).
The videos portray different action exemplars in natural settings, which entails variations in low-level features such as luminance, motion, or edges. We quantified those variables for each video. Hence, they can be used as variables of no interest in the experimental designs to minimize the effects of such low-level factors.

2.3. Data Validation

We validated our stimuli set with an fMRI experiment. Four human subjects participated in our study (2 females and 2 males; Mean Age: 26.5). Ethical approval was received from the Human Research Ethics Committee of Bilkent University.

2.3.1. fMRI Experiment

Each participant underwent two fMRI sessions, each having 8 runs. In each session, the 100 action exemplar videos were split across odd and even runs. In each of the odd runs (i.e., runs one, three, five, and seven), the first 50 action exemplars were shown in a random order as mini-blocks of three video versions of the same action presented consecutively. Each video lasted 3 s, and thus, each mini-block was shown for a total of 9 s. In each of the even runs (i.e., runs two, four, six, and eight), the other 50 action exemplars were shown as randomly ordered mini-blocks of three video versions as in the odd runs. So, in total, each of the 100 action exemplars was presented 24 times across the two sessions. The order of the mini-blocks and video versions in each mini-block was randomized across different runs. An inter-stimulus interval ranging between 1–2 s was included in between the mini-blocks. The total duration of each run was 553.36 s.
In order to keep their attention throughout the runs, a question was asked in each repetition cycle about the video that was just presented (e.g., “Was it climbing up a tree?”) with a simple yes or no button-press response time period of 3 s. The periods of question were not included in the analysis.

2.3.2. fMRI Data Acquisition

Participants were scanned at National Magnetic Resonance Research Center (UMRAM) in Bilkent University by using a 3T Siemens TimTrio MR scanner with a 32-channel phase array head coil. In order to minimize head movement, relevant foam paddings were put under their skull, around their neck, and under their legs. Stimuli were presented on an MR-compatible LCD screen (TELEMED, 60 Hz refresh rate, 800 × 600 pixel, 32 inches) and seen through a mirror system mounted on top of the head coil that is 168 cm away.
A high-resolution T1-weighted anatomical image covering the entire brain was acquired before the functional scans using the following acquisition parameters: TE = 2.92 ms, TR = 2.6 s, flip angle = 12°, Acceleration factor = 2, 176 sagittal slices with 1 mm × 1 mm × 1 mm resolution). Later on, for each of the eight experimental runs, functional images were acquired using echo-planar imaging (EPI) sequence (TR = 3 s, TE = 30 ms, flip angle = 90°, 96 × 96 matrix with FOV 240, 49 horizontal slices with 2.5 mm slice thickness). Each run started with the collection of 5 dummy scans to ensure that MR signal reached a steady state.

2.3.3. fMRI Data Preprocessing

Results included in this paper are based on fMRI data preprocessed using fMRIPrep 20.1.1 ([9]; RRID:SCR_016216), which is based on Nipype 1.5.0 ([10]; RRID:SCR_002502).

Anatomical Data Preprocessing

A total of 2 T1-weighted (T1w) images were found within the input BIDS dataset. All of them were corrected for intensity non-uniformity (INU) with N4BiasFieldCorrection [16], distributed with ANTs 2.2.0 ([17], RRID:SCR_004757). The T1w-reference was then skull-stripped with a Nipype implementation of the antsBrainExtraction.sh workflow (from ANTs), using OASIS30ANTs as target template. Brain tissue segmentation of cerebrospinal fluid (CSF), white-matter (WM), and gray-matter (GM) was performed on the brain-extracted T1w using fast (FSL 5.0.9, RRID:SCR_002823, [18]). A T1w-reference map was computed after registration of 2 T1w images (after INU-correction) using mri_robust_template (FreeSurfer 6.0.1, [19]). Brain surfaces were reconstructed using recon-all (FreeSurfer 6.0.1, RRID:SCR_001847, [20]), and the brain mask estimated previously was refined with a custom variation of the method to reconcile ANTs-derived and FreeSurfer-derived segmentations of the cortical gray-matter of Mindboggle (RRID:SCR_002438, [21]). Volume-based spatial normalization to one standard space (MNI152NLin2009cAsym) was performed through nonlinear registration with antsRegistration (ANTs 2.2.0), using brain-extracted versions of both T1w reference and the T1w template. The following template was selected for spatial normalization: ICBM 152 Nonlinear Asymmetrical template version 2009c ([22], RRID:SCR_008796; TemplateFlow ID: MNI152NLin2009cAsym).

Functional Data Preprocessing

For each of the 16 BOLD runs recorded per subject (across all sessions), the following preprocessing was performed. First, a reference volume and its skull-stripped version were generated using a custom methodology of fMRIPrep. The BOLD reference was then co-registered to the T1w reference using bbregister (FreeSurfer) which implements boundary-based registration [23]. Co-registration was configured with six degrees of freedom. Head-motion parameters with respect to the BOLD reference (transformation matrices, and six corresponding rotation and translation parameters) are estimated before any spatiotemporal filtering using mcflirt (FSL 5.0.9, [24]). BOLD runs were slice-time corrected using 3dTshift from AFNI 20160207 ([25], RRID:SCR_005927). The BOLD time-series were resampled to surfaces on the following spaces: fsaverage5. The BOLD time-series (including slice-timing correction when applied) were resampled onto their original, native space by applying a single, composite transform to correct for head-motion and susceptibility distortions. These resampled BOLD time-series will be referred to as preprocessed BOLD in original space, or just preprocessed BOLD. The BOLD time-series were resampled into standard space, generating a preprocessed BOLD run in [‘MNI152NLin2009cAsym’] space. First, a reference volume and its skull-stripped version were generated using a custom methodology of fMRIPrep. Several confounding time-series were calculated based on the preprocessed BOLD: framewise displacement (FD), DVARS and three region-wise global signals. FD and DVARS are calculated for each functional run, both using their implementations in Nipype (following the definitions by [26]). The three global signals are extracted within the CSF, the WM, and the whole-brain masks. Additionally, a set of physiological regressors were extracted to allow for component-based noise correction (CompCor, [27]). Principal components are estimated after high-pass filtering the preprocessed BOLD time-series (using a discrete cosine filter with 128s cut-off) for the two CompCor variants: temporal (tCompCor) and anatomical (aCompCor). tCompCor components are then calculated from the top 5% variable voxels within a mask covering the subcortical regions. This subcortical mask is obtained by heavily eroding the brain mask, which ensures it does not include cortical GM regions. For aCompCor, components are calculated within the intersection of the aforementioned mask and the union of CSF and WM masks calculated in T1w space, after their projection to the native space of each functional run (using the inverse BOLD-to-T1w transformation). Components are also calculated separately within the WM and CSF masks. For each CompCor decomposition, the k components with the largest singular values are retained, such that the retained components’ time series are sufficient to explain 50 percent of variance across the nuisance mask (CSF, WM, combined, or temporal). The remaining components are dropped from consideration. The head-motion estimates calculated in the correction step were also placed within the corresponding confounds file. The confound time series derived from head motion estimates and global signals were expanded with the inclusion of temporal derivatives and quadratic terms for each [28]. Frames that exceeded a threshold of 0.5 mm FD or 1.5 standardized DVARS were annotated as motion outliers. All resamplings can be performed with a single interpolation step by composing all the pertinent transformations (i.e., head-motion transform matrices, susceptibility distortion correction when available, and co-registrations to anatomical and output spaces). Gridded (volumetric) resamplings were performed using antsApplyTransforms (ANTs), configured with Lanczos interpolation to minimize the smoothing effects of other kernels [29]. Non-gridded (surface) resamplings were performed using mri_vol2surf (FreeSurfer).
Many internal operations of fMRIPrep use Nilearn 0.5.2 ([30], RRID:SCR_001362), mostly within the functional processing workflow. For more details of the pipeline, see the section corresponding to workflows in fMRIPrep’s documentation.

2.3.4. Activation Maps for Observed Actions

An important aim of the present study is to show that our novel stimuli set could drive the action observation network as previously identified, and to investigate whether it could be extended given that previous work used a limited number of actions. This is achieved by univariate analysis, which reveals the activation map of actions.
Following pre-processing, we ran a general linear model (GLM) composed of 8 regressors, including 1 regressor for all the action videos, 6 motion regressors (3 translations and 3 rotations), and a constant factor. All regressors were convolved with the default canonical hemodynamic response function in SPM12. The activation map was generated by the beta value corresponding to the action videos.

2.3.5. Definition of ROIs

We defined three a priori ROIs that represent the three levels of the Action Observation Network based on previous work: Lateral occipito-temporal cortex (LOTC), posterior parietal cortex (PPC), and premotor cortex (PMC) (Figure 2). LOTC included the regions of the action observation network based on the activation map (threshold at p < 0.001) of [3]. These regions included (1) the MT cluster [31,32], (2) a portion extending dorsally from the MT cluster onto the middle temporal gyrus referred to as MTG, and (3) a portion extending ventrally from the MT cluster onto occipital temporal sulcus that we refer to as OTS. The MTG and OTS correspond to the upper bank and lower bank of the macaque STS anterior to the MT cluster [33], the two regions which project to the parietal level of the action observation network in the macaque [34]. They are therefore considered to be input regions for the next level, PPC.
The PPC included the cyto-architectonic regions of SPL, IPS, IPL, and parietal operculum [35,36,37,38]. Its posterior boundary coursed between V6 and V6A [39] and extended between V3D and V7 [40].
The PMC included the cyto-architectonic supplementary, dorsal, and ventral premotor areas taken from the Anatomy software, but the ventral part was extended in the rostral direction to include regions that are responsive to observed actions according to [41,42].
All ROIs were defined on flat maps in Caret software [43]. Only ROIs in the left hemisphere were considered, because we now have considerable evidence that the position of the actor in the visual field affects the lateralization of PPC activation in the action observation paradigm [41], with the activation being contralateral to the hemifield in which the actor is shown. In almost all videos (with only a few exceptions represented by the 2 swimming exemplars, diving, and rolling side-way), the body of the actor was either motionless in the right visual field or remained in this field where actions (e.g., walking) implied horizontal motion to the left, as the camera partially followed the action.

3. Results

3.1. Post-Processing of Video Stimuli

We quantified low-level visual features for each video, including luminance, motion, and edges. These features can be used as variables of no interest in fMRI experiments to minimize the effects of those low-level factors. The MATLAB codes that generate these features as well as the output of these codes for each video can be downloaded from https://osf.io/u62bp/?view_only=393a2924aa05461394fe9f3171863b94 (accessed on 29 December 2022).
Luminance: For a given video, we first found the average of the RGB pixel values in all 150 frames (temporal averaging), and then calculated the average of the pixels in the averaged frame (spatial averaging). Thus, we obtained a single luminance value characterizing each video. Figure 3A shows the histogram of the average luminance values over the 100 action exemplars (further averaged over the three versions of an exemplar performed by different actors), and Figure 3B shows the values for the different actions.
Motion: We computed the mean speed in each video using an algorithm by [44]. The local motion vector was computed for each pixel in the image on a frame-by-frame basis. We performed temporal averaging (across frames) and spatial averaging (within a frame) (see Luminance above) to obtain one value for each video. Figure 4A shows the histogram of the average speed values covering the 100 action exemplars (further averaged over the three versions of each exemplar performed by different actors), and Figure 4B shows the values for the different actions.
Edges (Form): We passed each video frame of each action exemplar through a set of Gabor filters [45] to extract the edge information. We used Gabors of 5 scales and 8 orientations using the Gabor filtering algorithm described in [46]. We performed temporal averaging (across frames) and spatial averaging (within a frame) (see Luminance above) to obtain one output value for each video. Figure 5A shows the histogram of the average edge information covering the 100 action exemplars (further averaged over the three versions of an exemplar performed by different actors), and Figure 5B shows the edge information for the different actions.

3.2. fMRI Activation Maps

Observation of the 100 actions used in the present study activated the three levels of the action observation network (p < 0.001 uncorrected): LOTC, PPC, and to a lesser degree PMC in the left hemisphere, as expected (Figure 6). The LOTC level included activations in the MT cluster as well as MTG and to some extent OTS, as previously defined (See Section 2.3.5). The PPC level included activations in functionally defined areas DIPSM in all participants and DIPSA in subjects 1, 2, and 4. Other PPC level activations in cyto-architectonic areas of the inferior parietal lobule (IPL) include PFcm, PGa, and PGp in subject 1; PFcm, PFm, PGa, and PGp in subject2; PFcm and PFt in subject 3, and finally PFop, PGa, and PGp in subject 4. In addition, superior parietal lobule areas in dorsal postcentral gyrus were activated in subjects 2, 3, and 4. The PMC level included areas in the anterior part of the dorsal premotor and ventral premotor cortex in Subject 2 and the posterior part of the ventral premotor cortex in Subject 3.
In addition to the ROIs of the action observation network, several other areas were activated by the observation of actions. These included mainly the early visual cortex, extending into the neighboring parieto-occipital sulcus, including V7 [47], in Subject 1. Additional activations included a medial frontal site in Subjects 1, 2, and 3 and another small cluster neighboring the parieto-occipital sulcus (POS) in Subjects 2 and 4.

4. Discussion

We describe a video stimulus set that consists of 100 different natural actions covering most of the human repertoire. Each action is performed by multiple actors. The low-level features of the videos were quantified, allowing them to be factored out in future experiments. Our fMRI validation study showed that the novel stimuli presented here drove the action observation network. The weak activation of the premotor level likely reflects a combination of two factors: a lower level of activation due to distance from the retina [34], as documented in many studies (e.g., [2]), and an increased selectivity of premotor voxels whereby these voxels are activated by only a small number of observed actions, typical for the present study.
To the best of our knowledge, this is the largest action database to be made available for use in psychology and cognitive neuroscience research. Earlier work used grasping as the exemplary action for a long-time in action observation research [1]. More recent work has introduced different action categories such as locomotion, communicative, self-directed, interaction, and vocal actions [2,3,4,5]. However, each study was constrained by a small set of action videos that was created for the purposes of that study and this made it difficult to compare the results across different studies due to variations in actors, scenes, video durations, and video quality. There are video databases that display actions in the form of point-light displays (See the list in [9]) to overcome the visual differences in the stimuli, but the shortcoming of these databases is that the stimuli are not naturalistic enough and lack ecological validity. There are some action databases with naturalistic actions, such as [48], but it focuses on actions that have emotional content. There are yet other naturalistic action databases, such as UCF50 [10] or HMDB51 [11], but their target is usually the computer vision community. Computer vision research has different constraints such as multiple cameras, camera motion, or clutter in the videos. Therefore, the databases have been created in accordance with the problems needing to be solved from a computer-vision perspective. Given that we are at the early stages of understanding the perceptual and neural mechanisms processing observed actions, it is necessary to initiate such studies using a stimulus set that is natural yet simple and sufficiently controlled to facilitate interpretation. In this respect, they differ from the set of 80 atomic actions [12].
We believe that our database will prove useful for researchers who intend to study the perceptual and neural differences between observing action exemplars such as locomotion or communicative actions, as well as interactions between two individuals, extending ongoing fMRI work in passive subjects with different action classes [2,3,4,5]. Our stimulus set can, however, be used in a much wider set of behavioral and neuroimaging studies, allowing some generic plausible models to be built for action perception. Indeed, our stimulus set can be used in an array of visual tasks. A first set are discrimination tasks, probing the identity of the observed actions, such as identification or same–different tasks [49,50]. If, as has been proposed [51], observed actions of different classes, such as manipulation or locomotion, are processed in different PPC regions, one would expect action discriminability, whether measured perceptually or in neural activity, to depend on the classes involved. A second set of tasks are classification studies probing the semantic categories of observed actions. The classification of static images has received a lot of attention comparing human and deep network performance, in an effort to model object processing in the ventral pathway [52]. This can be extended to the classification of videos to model observed action processing in the dorsal pathway. Yet another set of studies are similarity studies. Subjects have to rate how close two actions are, allowing to derive the distances between observed actions in perceptual space, which can be compared using RSA to the distance between these actions in a neural space derived from single cell recordings or fMRI activations. Such perceptual and neuronal studies would stimulate the computational modelling of observed actions. There have been a few modelling attempts to explain the neural mechanisms of observed actions [53,54,55,56], but so far these have been limited to only a few action exemplars such as locomotion or grasping and such modelling efforts would benefit from testing a larger set of actions. It is noteworthy that in the present stimulus set the actions were performed by several actors, which makes it easy to design control tasks, requiring subjects to discriminate or classify actors. Extending this further, the video set can also be used in studies for person or gender identification from body movements as we have multiple actors performing the same action with the same background.
A limitation of our stimuli set is that all actions take place in an outdoor scene. Therefore, researchers who are interested in contextual effects, such as the scene in which the action takes place, may not find sufficient variability in the stimulus set, although some actions were set on a beach, or in a lake, in addition to the grassy landscape. However, many of these actions can be performed indoors as well, and hence from an action-identity point of view, many of them can still be used in behavioral and neuroimaging experiments probing action observation.
Another limitation of the stimuli set is that we do not systematically control the emotional content of the actions. Most actions can be considered as neutral, such as locomotion, but some of them have an inherently positive valence, such as laughing, or negative valence, such as signing “no”. Therefore, the set may not be optimal for studies aiming to systematically investigate the emotional content of the actions.

5. Conclusions

In summary, we believe that our stimuli set will be beneficial to the scientific community studying action perception from behavioral, neuro-scientific, and computational perspectives, particularly those who wish to move away from mere grasping and reaching as prototypical actions. As the stimulus set can be combined not only with fMRI, but also MEG, EEG, and stereo-EEG, it should find a wide range of applications from the bedside to the laboratory.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/brainsci13010061/s1.

Author Contributions

Conceptualization, B.A.U. and G.A.O.; methodology, B.A.U., H.N., A.E. and G.A.O.; software, B.A.U., H.N. and A.E.; validation, B.A.U., H.N. and A.E.; formal analysis, B.A.U., H.N. and A.E.; investigation, B.A.U., H.N., A.E. and G.A.O.; resources, B.A.U. and G.A.O.; data curation, B.A.U., H.N. and A.E.; writing—original draft preparation, B.A.U.; writing—review and editing, B.A.U., H.N., A.E. and G.A.O.; visualization, B.A.U., H.N. and A.E.; supervision, B.A.U. and G.A.O.; project administration, B.A.U. and G.A.O.; funding acquisition, B.A.U. and G.A.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by ERC Advanced Grant, grant number ERC 2012-ADG 323606 to Guy A. Orban, and TUBITAK 3501 Career Grant, grant number 119K654 to Burcu A. Urgen.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Human Subjects Research Ethics Committee of Bilkent University (approval number 2019_01_29_01; date 29 January 2019).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. The data preprocessing section was used verbatim from the boilerplate text that was automatically generated by fMRIPrep with the express intention that users should copy and paste this text into their manuscripts unchanged. It is released under the CC0 license.

Data Availability Statement

The data and stimuli presented in this study are available on Open Science Framework platform: https://osf.io/u62bp/?view_only=393a2924aa05461394fe9f3171863b94 (accessed on 29 December 2022).

Acknowledgments

The authors would like to thank the actors who participated in the recording of the videos and Stefania Ferri for recording the videos.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Caspers, S.; Zilles, K.; Laird, A.R.; Eickhoff, S.B. ALE meta-analysis of action observation and imitation in the human brain. Neuroimage 2010, 50, 1148–1167. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Abdollahi, R.O.; Jastorff, J.; Orban, G.A. Common and Segregated Processing of Observed Actions in Human SPL. Cereb. Cortex 2012, 23, 2734–2753. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Ferri, S.; Rizzolatti, G.; Orban, G.A. The organization of the posterior parietal cortex devoted to upper limb actions: An fMRI study. Hum. Brain Mapp. 2015, 36, 3845–3866. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Corbo, D.; Orban, G.A. Observing Others Speak or Sing Activates Spt and Neighboring Parietal Cortex. J. Cogn. Neurosci. 2017, 29, 1002–1021. [Google Scholar] [CrossRef]
  5. Urgen, B.A.; Orban, G.A. The unique role of parietal cortex in action observation: Functional organization for communicative and manipulative actions. NeuroImage 2021, 237, 118220. [Google Scholar] [CrossRef]
  6. Sonkusare, S.; Breakspear, M.; Guo, C. Naturalistic Stimuli in Neuroscience: Critically Acclaimed. Trends Cogn. Sci. 2019, 23, 699–714. [Google Scholar] [CrossRef]
  7. Nastase, S.A.; Goldstein, A.; Hasson, U. Keep it real: Rethinking the primacy of experimental control in cognitive neuroscience. Neuroimage 2020, 222, 117254. [Google Scholar] [CrossRef]
  8. Kriegeskorte, N.; Mur, M.; Bandettini, P.A. Representational similarity analysis—Connecting the branches of systems neuroscience. Front. Syst. Neurosci. 2008, 2, 4. [Google Scholar] [CrossRef] [Green Version]
  9. Bidet-Ildei, C.; Francisco, V.; Decatoire, A.; Pylouster, J.; Blandin, Y. PLAViMoP database: A new continuously assessed and collaborative 3D point-light display dataset. Behav. Res. Methods 2022, 1–22. [Google Scholar] [CrossRef]
  10. Reddy, K.K.; Shah, M. Recognizing 50 human action categories of web videos. Mach. Vis. Appl. 2012, 24, 971–981. [Google Scholar] [CrossRef]
  11. Kuehne, H.; Jhuang, H.; Garrote, E.; Poggio, T.; Serre, T. HMDB: A Large Video Database for Human Motion Recognition. In Proceedings of the 2011 International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011. [Google Scholar]
  12. Gu, C.; Sun, C.; Ross, D.A.; Vondrick, C.; Pantofaru, C.; Li, Y.; Vijayanarasimhan, S.; Toderici, G.; Ricco, S.; Sukthankar, R.; et al. AVA: A Video Dataset of Spatio-Temporally Localized Atomic Visual Actions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  13. Huth, A.G.; Nishimoto, S.; Vu, A.T.; Gallant, J.L. A Continuous Semantic Space Describes the Representation of Thousands of Object and Action Categories across the Human Brain. Neuron 2012, 76, 1210–1224. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Jastorff, J.; Begliomini, C.; Destro, M.F.; Rizzolatti, G.; Orban, G. Coding Observed Motor Acts: Different Organizational Principles in the Parietal and Premotor Cortex of Humans. J. Neurophysiol. 2010, 104, 128–140. [Google Scholar] [CrossRef] [PubMed]
  15. Shahdloo, M.; Çelik, E.; Urgen, B.A.; Gallant, J.L.; Çukur, T. Task-Dependent Warping of Semantic Representations during Search for Visual Action Categories. J. Neurosci. 2022, 42, 6782–6799. [Google Scholar] [CrossRef] [PubMed]
  16. Tustison, N.J.; Avants, B.B.; Cook, P.A.; Zheng, Y.; Egan, A.; Yushkevich, P.A.; Gee, J.C. N4ITK: Improved N3 bias correction. IEEE Trans. Med. Imaging 2010, 29, 1310–1320. [Google Scholar] [CrossRef] [Green Version]
  17. Avants, B.; Epstein, C.; Grossman, M.; Gee, J. Symmetric diffeomorphic image registration with cross-correlation: Evaluating automated labeling of elderly and neurodegenerative brain. Med. Image Anal. 2008, 12, 26–41. [Google Scholar] [CrossRef] [Green Version]
  18. Zhang, Y.; Brady, M.; Smith, S. Segmentation of brain MR images through a hidden Markov random field model and the expectation-maximization algorithm. IEEE Trans. Med. Imaging 2001, 20, 45–57. [Google Scholar] [CrossRef]
  19. Reuter, M.; Rosas, H.D.; Fischl, B. Highly accurate inverse consistent registration: A robust approach. Neuroimage 2010, 53, 1181–1196. [Google Scholar] [CrossRef] [Green Version]
  20. Dale, A.M.; Fischl, B.; Sereno, M.I. Cortical surface-based analysis. I. Segmentation and surface reconstruction. NeuroImage 1999, 9, 179–194. [Google Scholar] [CrossRef]
  21. Klein, A.; Ghosh, S.S.; Bao, F.S.; Giard, J.; Häme, Y.; Stavsky, E.; Lee, N.; Rossa, B.; Reuter, M.; Neto, E.C.; et al. Mindboggling morphometry of human brains. PLoS Comput. Biol. 2017, 13, e1005350. [Google Scholar] [CrossRef] [Green Version]
  22. Fonov, V.; Evans, A.; McKinstry, R.; Almli, C.; Collins, D. Unbiased nonlinear average age-appropriate brain templates from birth to adulthood. Neuroimage 2009, 47, S102. [Google Scholar] [CrossRef]
  23. Greve, D.N.; Fischl, B. Accurate and robust brain image alignment using boundary-based registration. Neuroimage 2009, 48, 63–72. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Jenkinson, M.; Bannister, P.; Brady, M.; Smith, S. Improved Optimization for the Robust and Accurate Linear Registration and Motion Correction of Brain Images. Neuroimage 2002, 17, 825–841. [Google Scholar] [CrossRef] [PubMed]
  25. Cox, R.W.; Hyde, J.S. Software Tools for Analysis and Visualization of fMRI Data. NMR Biomed. Int. J. Devoted Dev. Appl. Magn. Reson. In Vivo 1997, 10, 171–178. [Google Scholar] [CrossRef]
  26. Power, J.D.; Mitra, A.; Laumann, T.O.; Snyder, A.Z.; Schlaggar, B.L.; Petersen, S.E. Methods to detect, characterize, and remove motion artifact in resting state fMRI. Neuroimage 2014, 84, 320–341. [Google Scholar] [CrossRef] [Green Version]
  27. Behzadi, Y.; Restom, K.; Liau, J.; Liu, T.T. A component based noise correction method (CompCor) for BOLD and perfusion based fMRI. Neuroimage 2007, 37, 90–101. [Google Scholar] [CrossRef] [Green Version]
  28. Satterthwaite, T.D.; Wolf, D.H.; Ruparel, K.; Erus, G.; Elliott, M.A.; Eickhoff, S.B.; Gennatas, E.D.; Jackson, C.; Prabhakaran, K.; Smith, A.; et al. Heterogeneous impact of motion on fundamental patterns of developmental changes in functional connectivity during youth. Neuroimage 2013, 83, 45–57. [Google Scholar] [CrossRef] [Green Version]
  29. Lanczos, C. Evaluation of noisy data. J. Soc. Ind. Appl. Math. Ser. B Numer. Anal. 1964, 1, 76–85. [Google Scholar] [CrossRef]
  30. Abraham, A.; Pedregosa, F.; Eickenberg, M.; Gervais, P.; Mueller, A.; Kossaifi, J.; Gramfort, A.; Thirion, B.; Varoquaux, G. Machine learning for neuroimaging with scikit-learn. Front. Neuroinform. 2014, 8, 14. [Google Scholar] [CrossRef] [Green Version]
  31. Tootell, R.; Reppas, J.; Kwong, K.; Malach, R.; Born, R.; Brady, T.; Rosen, B.; Belliveau, J. Functional analysis of human MT and related visual cortical areas using magnetic resonance imaging. J. Neurosci. 1995, 15, 3215–3230. [Google Scholar] [CrossRef]
  32. Kolster, H.; Peeters, R.; Orban, G.A. The Retinotopic Organization of the Human Middle Temporal Area MT/V5 and Its Cortical Neighbors. J. Neurosci. 2010, 30, 9801–9820. [Google Scholar] [CrossRef]
  33. Jastorff, J.; Popivanov, I.D.; Vogels, R.; Vanduffel, W.; Orban, G.A. Integration of shape and motion cues in biological motion processing in the monkey STS. Neuroimage 2012, 60, 911–921. [Google Scholar] [CrossRef] [PubMed]
  34. Nelissen, K.; Borra, E.; Gerbella, M.; Rozzi, S.; Luppino, G.; Vanduffel, W.; Rizzolatti, G.; Orban, G. Action Observation Circuits in the Macaque Monkey Cortex. J. Neurosci. 2011, 31, 3743–3756. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Caspers, S.; Geyer, S.; Schleicher, A.; Mohlberg, H.; Amunts, K.; Zilles, K. The human inferior parietal cortex: Cytoarchitectonic parcellation and interindividual variability. Neuroimage 2006, 33, 430–448. [Google Scholar] [CrossRef] [PubMed]
  36. Eickhoff, S.B.; Grefkes, C.; Zilles, K.; Fink, G.R. The Somatotopic Organization of Cytoarchitectonic Areas on the Human Parietal Operculum. Cereb. Cortex 2006, 17, 1800–1811. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Choi, H.-J.; Zilles, K.; Mohlberg, H.; Schleicher, A.; Fink, G.R.; Armstrong, E.; Amunts, K. Cytoarchitectonic identification and probabilistic mapping of two distinct areas within the anterior ventral bank of the human intraparietal sulcus. J. Comp. Neurol. 2006, 495, 53–69. [Google Scholar] [CrossRef] [Green Version]
  38. Scheperjans, F.; Eickhoff, S.B.; Hömke, L.; Mohlberg, H.; Hermann, K.; Amunts, K.; Zilles, K. Probabilistic Maps, Morphometry, and Variability of Cytoarchitectonic Areas in the Human Superior Parietal Cortex. Cereb. Cortex 2008, 18, 2141–2157. [Google Scholar] [CrossRef]
  39. Pitzalis, S.; Sereno, M.I.; Committeri, G.; Fattori, P.; Galati, G.; Patria, F.; Galletti, C. Human V6: The Medial Motion Area. Cereb. Cortex 2010, 20, 411–424. [Google Scholar] [CrossRef] [Green Version]
  40. Abdollahi, R.O.; Kolster, H.; Glasser, M.F.; Robinson, E.C.; Coalson, T.S.; Dierker, D.; Jenkinson, M.; Van Essen, D.C.; Orban, G.A. Correspondences between retinotopic areas and myelin maps in human visual cortex. NeuroImage 2014, 99, 509–524. [Google Scholar] [CrossRef] [Green Version]
  41. Ferri, S.; Pauwels, K.; Rizzolatti, G.; Orban, G.A. Stereoscopically Observing Manipulative Actions. Cereb. Cortex 2016, 26, 3591–3610. [Google Scholar] [CrossRef]
  42. Jastorff, J.; Abdollahi, R.O.; Fasano, F.; Orban, G.A. Seeing biological actions in 3D: An fMRI study. Hum. Brain Mapp. 2016, 37, 213–219. [Google Scholar] [CrossRef]
  43. Van Essen, D.C.; Drury, H.A.; Dickson, J.; Harwell, J.; Hanlon, D.; Anderson, C.H. An integrated software suite for surface-based analyses of cerebral cortex. J. Am. Med. Inform. Assoc. 2001, 8, 443–459. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Pauwels, K.; Van Hulle, M.M. Optic flow from unstable sequences through local velocity constancy maximization. Image Vis. Comput. 2009, 27, 579–587. [Google Scholar] [CrossRef]
  45. Olshausen, B.A.; Field, D.J. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature 1996, 381, 607–609. [Google Scholar] [CrossRef] [PubMed]
  46. Haghighat, M.; Zonouz, S.; Abdel-Mottaleb, M. CloudID: Trustworthy cloud-based and cross-enterprise biometric identification. Expert Syst. Appl. 2015, 42, 7905–7916. [Google Scholar] [CrossRef]
  47. Georgieva, S.; Peeters, R.; Kolster, H.; Todd, J.T.; Orban, G.A. The processing of three-dimensional shape from disparity in the human brain. J. Neurosci. 2009, 29, 727–742. [Google Scholar] [CrossRef] [Green Version]
  48. Keefe, B.D.; Villing, M.; Racey, C.; Strong, S.L.; Wincenciak, J.; Barraclough, N.E. A database of whole-body action videos for the study of action, emotion, and untrustworthiness. Behav. Res. Methods 2014, 46, 1042–1051. [Google Scholar] [CrossRef] [Green Version]
  49. Platonov, A.; Orban, G.A. Action observation: The less-explored part of higher-order vision. Sci. Rep. 2016, 6, 36742. [Google Scholar] [CrossRef] [Green Version]
  50. Platonov, A.; Orban, G.A. Not all observed actions are perceived equally. Sci. Rep. 2017, 7, 17084. [Google Scholar] [CrossRef] [Green Version]
  51. Orban, G.A.; Sepe, A.; Bonini, L. Parietal maps of visual signals for bodily action planning. Anat. Embryol. 2021, 226, 2967–2988. [Google Scholar] [CrossRef]
  52. Zhuang, C.; Yan, S.; Nayebi, A.; Schrimpf, M.; Frank, M.C.; DiCarlo, J.J.; Yamins, D.L. Unsupervised neural network models of the ventral visual stream. Proc. Natl. Acad. Sci. USA 2021, 118, e2014196118. [Google Scholar] [CrossRef]
  53. Giese, M.A.; Poggio, T. Neural mechanisms for the recognition of biological movements. Nat. Rev. Neurosci. 2003, 4, 179–192. [Google Scholar] [CrossRef] [PubMed]
  54. Tan, C.; Singer, J.M.; Serre, T.; Sheinberg, D.; Poggio, T. Neural Representation of Action Sequences: How Far Can a Simple Snippet-Matching Model Take Us? In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–8 December 2013; pp. 593–601. [Google Scholar]
  55. Fleischer, F.; Caggiano, V.; Thier, P.; Giese, M.A. Physiologically Inspired Model for the Visual Recognition of Transitive Hand Actions. J. Neurosci. 2013, 33, 6563–6580. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Theusner, S.; De Lussanet, M.; Lappe, M. Action Recognition by Motion Detection in Posture Space. J. Neurosci. 2014, 34, 909–921. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Sample frames from the stimulus set, one frame from each action exemplar. The first row corresponds to action exemplars 1–10, the second row corresponds to action exemplars 11–20, and so on. The action numbers refer to the numbers in Table 1.
Figure 1. Sample frames from the stimulus set, one frame from each action exemplar. The first row corresponds to action exemplars 1–10, the second row corresponds to action exemplars 11–20, and so on. The action numbers refer to the numbers in Table 1.
Brainsci 13 00061 g001
Figure 2. A priori ROIs of the Action Observation Network: LOTC (yellow line), PPC (blue line), PMC (green line). LOTC consists of three sub-regions including the MT cluster, MTG, and OTS. The anatomical landmarks are indicated in black: CS (Central sulcus), PreSC (Pre-central sulcus), PostCS (Post-central sulcus), IPS (intra-parietal sulcus), POS (parieto-occipital sulcus), STS (superior temporal sulcus), ITS (inferior temporal sulcus), OTS (occipito-temporal sulcus).
Figure 2. A priori ROIs of the Action Observation Network: LOTC (yellow line), PPC (blue line), PMC (green line). LOTC consists of three sub-regions including the MT cluster, MTG, and OTS. The anatomical landmarks are indicated in black: CS (Central sulcus), PreSC (Pre-central sulcus), PostCS (Post-central sulcus), IPS (intra-parietal sulcus), POS (parieto-occipital sulcus), STS (superior temporal sulcus), ITS (inferior temporal sulcus), OTS (occipito-temporal sulcus).
Brainsci 13 00061 g002
Figure 3. Characterization of the average luminance in the videos. (A) Histogram of the average luminance values across 100 action exemplars (averaged over the 3 versions of each action performed by different actors), (B) Average luminance of the 100 action exemplars.
Figure 3. Characterization of the average luminance in the videos. (A) Histogram of the average luminance values across 100 action exemplars (averaged over the 3 versions of each action performed by different actors), (B) Average luminance of the 100 action exemplars.
Brainsci 13 00061 g003
Figure 4. Characterization of the average speed in the videos. (A) Histogram of the average speed values across 100 action exemplars (averaged over the 3 versions of each action performed by different actors), (B) Average speed of the 100 action exemplars.
Figure 4. Characterization of the average speed in the videos. (A) Histogram of the average speed values across 100 action exemplars (averaged over the 3 versions of each action performed by different actors), (B) Average speed of the 100 action exemplars.
Brainsci 13 00061 g004
Figure 5. Characterization of the average edge information in the videos. (A) Histogram of the average edge information across 100 action exemplars (averaged over the 3 versions of each action performed by different actors), (B) Average edge information of the 100 action exemplars.
Figure 5. Characterization of the average edge information in the videos. (A) Histogram of the average edge information across 100 action exemplars (averaged over the 3 versions of each action performed by different actors), (B) Average edge information of the 100 action exemplars.
Brainsci 13 00061 g005
Figure 6. Activation map for observed actions (p < 0.001 uncorrected, k = 10) on the flattened left hemisphere of subjects (A) 1, (B) 2, (C) 3, and (D) 4.
Figure 6. Activation map for observed actions (p < 0.001 uncorrected, k = 10) on the flattened left hemisphere of subjects (A) 1, (B) 2, (C) 3, and (D) 4.
Brainsci 13 00061 g006
Table 1. Number and name of action exemplars, with the actors performing them (M1: male actor 1, M2: male actor 2, F1: female actor 1, F2: female actor 2).
Table 1. Number and name of action exemplars, with the actors performing them (M1: male actor 1, M2: male actor 2, F1: female actor 1, F2: female actor 2).
Action Exemplar NoAction Exemplar NameActors
1Measuring with fingersF1, F2, M1
2ShoutingF1, M1, M2
3Carrying with head and handsF1, M1, M2
4Caressing another personF1, F2, M1
5Free style swimmingF2, M1, M2
6Kicking wood with feetF2, M1, M2
7DraggingF1, F2, M2
8ReachingF1, F2, M2
9Measuring a long distance with feetF1, F2, M1
10Crushing a leaf with fingersF1, F2, M2
11Fanning with leafF1, F2, M2
12Pushing a small stoneF1, F2, M2
13Dropping a small stoneF1, F2, M2
14Ridiculing another personF1, F2, M1
15Massaging own cheekF1, F2, M2
16Scratching own cheekF1, F2, M2
17SwallowingF1, F2, M1
18Yawning with handF1, F2, M2
19Licking an orangeF1, F2, M2
20Gazing at an objectF1, F2, M1
21Peeling a fruitF1, F2, M1
22Filling a hole with handF1, F2, M1
23Hitting own cheekF1, F2, M1
24Swimming back styleF1, M1, M2
25Displacing woodF1, F2, M2
26Weighing an object with one handF1, F2, M2
27Climbing down a treeF1, F2, M1
28WhistlingF2, M1, M2
29Measuring with handsF1, F2, M1
30Picking a fruit from a treeF2, M1, M2
31Kicking horizontallyF2, M1, M2
32Kicking verticallyF2, M1, M2
33Carrying with headF1, M1, M2
34Blowing a leafF1, M1, M2
35Chasing another personF1, M1, M2
36StrugglingF2, M1, M2
37Waving goodbyeF1, M1, M2
38Beating with a piece of woodF1, M1, M2
39Carrying with shoulder and handF1, F2, M2
40Reprimanding a personF2, M1, M2
41Biting a bananaF2, M1, M2
42Fighting with another personF2, M1, M2
43Washing own bodyF2, M1, M2
44ForagingF1, M1, M2
45Stretching own bodyF1, M1, M2
46Writing with fingersF1, M1, M2
47Charging to attackF1, F2, M1
48DivingF2, M1, M2
49Pointing nearbyF2, M1, M2
50Squeezing an orangeF1, F2, M2
51Forbidding with fingersF2, M1, M2
52Wrapping a stoneF1, M1, M2
53GraspingF2, M1, M2
54Carrying on shoulderF1, F2, M1
55RunningF1, F2, M2
56Burying in the sandF2, M1, M2
57Stopping a personF1, F2, M2
58Pushing a personF1, M1, M2
59Kissing a personF1, F2, M1
60Throwing and catching a small piece of woodF1, M1, M2
61Caressing own cheekF2, M1, M2
62Overtaking an obstacleF1, F2, M1
63Touching another person on the shoulderF1, F2, M2
64Building pyramid from sandF1, F2, M1
65Hiding an object behind backF2, M1, M2
66Washing fruitF1, F2, M1
67Carrying with handsF1, F2, M2
68WalkingF1, F2, M2
69MarchingF1, F2, M1
70Rubbing own cheekF1, M1, M2
71Throwing nearbyF1, M1, M2
72MasticatingF2, M1, M2
73People meetingF2, M1, M2
74Climbing up a treeF1, F2, M2
75Dancing with another personF2, M1, M2
76Getting upF1, M1, M2
77CrawlingF1, M1, M2
78Spitting a piece of bananaF1, F2, M1
79Doing gymnastics with both feet and armsF1, F2, M2
80Pushing a large objectF2, M1, M2
81Rolling body sidewiseF1, F2, M1
82Walking on hand and kneesF1, M1, M2
83Laughing together with another personF2, M1, M2
84Carrying with one handF1, F2, M1
85Singing a songF1, F2, M1
86Weighing an object with two handsF1, M1, M2
87Pointing distantlyF1, F2, M2
88Drinking with handsF2, M1, M2
89Massaging another personF1, F2, M1
90Drinking with mouthF1, F2, M2
91Measuring height with own bodyF1, M1, M2
92Pinching off piece of bananaF1, F2, M1
93ErasingF1, F2, M2
94Hugging a person (passive)F1, M1, M2
95Speaking with another personF1, F2, M1
96Rotating a stoneF2, M1, M2
97Measuring a short distance with feetF1, M1, M2
98Hugging each otherF1, M1, M2
99Digging a hole with a handF2, M1, M2
100Throwing farF1, F2, M1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Urgen, B.A.; Nizamoğlu, H.; Eroğlu, A.; Orban, G.A. A Large Video Set of Natural Human Actions for Visual and Cognitive Neuroscience Studies and Its Validation with fMRI. Brain Sci. 2023, 13, 61. https://doi.org/10.3390/brainsci13010061

AMA Style

Urgen BA, Nizamoğlu H, Eroğlu A, Orban GA. A Large Video Set of Natural Human Actions for Visual and Cognitive Neuroscience Studies and Its Validation with fMRI. Brain Sciences. 2023; 13(1):61. https://doi.org/10.3390/brainsci13010061

Chicago/Turabian Style

Urgen, Burcu A., Hilal Nizamoğlu, Aslı Eroğlu, and Guy A. Orban. 2023. "A Large Video Set of Natural Human Actions for Visual and Cognitive Neuroscience Studies and Its Validation with fMRI" Brain Sciences 13, no. 1: 61. https://doi.org/10.3390/brainsci13010061

APA Style

Urgen, B. A., Nizamoğlu, H., Eroğlu, A., & Orban, G. A. (2023). A Large Video Set of Natural Human Actions for Visual and Cognitive Neuroscience Studies and Its Validation with fMRI. Brain Sciences, 13(1), 61. https://doi.org/10.3390/brainsci13010061

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop