Summary

Group Synchronization During Collaborative Drawing Using Functional Near-Infrared Spectroscopy

Published: August 05, 2022
doi:

Summary

The present protocol combines functional near-infrared spectroscopy (fNIRS) and video-based observationto measure interpersonal synchronization in quartets during a collaborative drawing task.

Abstract

Functional near-infrared spectroscopy (fNIRS) is a noninvasive method particularly suitable for measuring cerebral cortex activation in multiple subjects, which is relevant for studying group interpersonal interactions in ecological settings. Although many fNIRS systems technically offer the possibility to monitor more than two individuals simultaneously, establishing easy-to-implement setup procedures and reliable paradigms to track hemodynamic and behavioral responses in group interaction is still required. The present protocol combines fNIRS and video-based observation to measure interpersonal synchronization in quartets during a cooperative task.This protocol provides practical recommendations for data acquisition and paradigm design, as well as guiding principles for an illustrative data analysis example. The procedure is designed to assess differences in brain and behavior interpersonal responses between social and non-social conditions inspired by a well-known ice-breaker activity, the Collaborative Face Drawing Task. The described procedures can guide future studies to adapt group naturalistic social interaction activities to the fNIRS environment.

Introduction

Interpersonal interaction behavior is an important component of the process of connecting and creating empathic bonds. Previous research indicates that this behavior can be expressed in the occurrence of synchronicity, when biological and behavioral signals align during social contact. Evidence shows that synchronicity can occur between people interacting for the first time1,2,3. Most studies on social interactions and their underlying neural mechanisms use a single person or second person approach2,4, and little is known about transposing this knowledge to group social dynamics. Evaluating interpersonal responses in groups of three or more individuals is still a challenge for scientific research. This leads to the necessity of bringing to the laboratory the complex environment of social interactions in everyday human beings under naturalistic conditions5.

In this context, the functional near-infrared spectroscopy (fNIRS) technique is a promising tool for assessing the relationships between interpersonal interaction in naturalistic contexts and its brain correlates. It presents fewer restrictions on participant mobility compared to functional magnetic resonance imaging (fMRI) and is resilient to motion artifacts6,7. The fNIRS technique works by assessing hemodynamic effects in response to brain activation (changes in blood concentration of oxygenated and deoxygenated hemoglobin). These variations can be measured by the amount of diffusion of infrared light through scalp tissue. Previous studies have demonstrated the flexibility and robustness of the technique in ecological hyperscanning experiments and the potential to expand knowledge in applied neuroscience6,8.

The choice of an experimental task for the naturalistic assessment of the neural correlates of social interaction processes in groups is a crucial step in approaching applied neuroscience studies9. Some examples already reported in the literature with the use of fNIRS in group paradigms include music performance10,11,12, classroom interaction8, and communication13,14,15,16,17.

One of the aspects not yet explored by previous studies is the use of drawing games that have as the main feature the manipulation of empathic components to assess social interaction. In this context, one of the games frequently used to induce social interaction in dynamics among strangers is the collaborative drawing game18,19. In this game, sheets of paper are divided into equal parts, and the group participants are challenged to draw shared self-portraits of all members. In the end, each member has their portrait drawn in a collaborative way by several hands.

The objective is to promote quick integration among strangers, provoked by directing visual attention to the faces of the group partners. It can be considered an "ice-breaking" activity due to its ability to support curiosity and consequent empathic processes among the members19.

One of the advantages of using drawing tasks is their simplicity and ease of reproduction20. They also do not require any specific technical training or skills, as seen in the studies using musical performance paradigms21,22,23,24. This simplicity also enables the choice of a more naturalistic stimulus within a social context4,9,25.

Besides being an instrument for inducing social behavior in groups, drawing is also considered a tool for psychological evaluation26. Some graphic-projective psychological tests, such as House-Tree-Person (HTP)27,28,29, Human Figure Drawing – Sisto Scale27, and Kinetic Family Drawing30 are used in a complementary way for qualitative and quantitative diagnoses. Their results usually express unconscious processes, giving clues about the individual's symbolic system and, therefore, their interpretations of the world, experiences, affections, etc.

The practice of drawing makes one think and helps create meaning for experiences and things, adding sensations, feelings, thoughts, and actions31. It gives clues about how to perceive and process these life experiences26. Drawing uses visual codes to allow one to understand and communicate thoughts or feelings, making them accessible to manipulation and, thus, creating the possibility for new ideas and readings31.

In art therapy, drawing is a tool to work on attention, memory, and organization of thoughts and feelings32, and it can be used as means to produce social interaction33.

This study aimed to develop a naturalistic experimental protocol to assess vascular and behavioral brain responses during interpersonal interaction in quartets using a collaborative drawing dynamic. In this protocol, the evaluation of the brain responses of the quartet (individually and the synchronicity between partners) and the possible outcome measures, such as behavioral measures (drawing and gaze behavior) are proposed. The aim is to provide more information on social neuroscience.

Protocol

The methodology was approved by the Hospital Israelita Albert Einstein (HIAE) Ethics Committee and is based on a procedure for collecting neural data (fNIRS), as well as gaze behavior data, with young adults during a collaborative drawing experience. All collected data were managed on the Redcap platform (see Table of Materials). The project was audited by the Scientific Integrity Committee of the Hospital Israelita Albert Einstein (HIAE). Young adults, 18-30 years old, were selected as subjects for the present study. Written informed consent was obtained from all participants.

1. Preparation for the study

  1. Sujetos
    1. Determine the target study sample.
    2. Inform all volunteers about the experimental protocol and their rights, prior to the game. Ensure that they sign an informed consent form and an image use consent form (not mandatory), fill in a registration form, and answer psychological questionnaires and scales.
    3. Control the level of relationship between participants of the quartet (strangers, friends, partners, etc.), as previous knowledge may interfere. In this study, the quartets were composed of strangers.
    4. Compose quartets of same-gender individuals.
      NOTE: This gender criterion avoids social interaction interferences34,35.
  2. Setting
    1. Remove all potential eye distractors from the scene.
    2. To set up, include a square table, four stools (measuring 18.11 in x 14.96 in), and two wire supports (e.g., tripod) (Figure 1).
    3. Turn off all electrical devices such as air conditioning during the experimental condition. Ensure that the room has sufficient lighting for people to observe and draw and that the room temperature is pleasant.
    4. Consider the extent of fNIRS wires (see Table of Materials), position all cables so that they remain stable during the experimental task.
    5. Consider space for two researchers to move along the setting.
    6. Ensure that experimenters follow their scripts and movement schemes.
    7. Position the quartet on the square table, two by two, so that each individual can observe the other three individuals.
    8. Give each quartet participant a tag with a number (1 to 4). Ensure Subject 1 sits across from Subject 3 and next to Subject 2.
      NOTE: The tag number corresponded to subjects position on the table and their previously prepared cap (Table of Materials).
  3. Drawing paradigm
    1. Collaborative face drawing-the social condition
      NOTE: This game's goal is to direct the visual attention of the subjects to their partners' faces, inducing them to more conscious observation among themselves. By connecting feelings and visual perception, the collaborative face drawing technique is a valuable way to activate empathic responses, interpersonal curiosity, and connectivity among participants. It requires theory of mind capacity, which includes imitation and anticipating others' behavior19. Use the following steps:
      1. Instruct the participants about the game rules.
      2. Divide each paper into three horizontal strips, namely drawing strips.
      3. Have each strip correspond to a social drawing condition (e.g., C1, C2.). After every social drawing condition, change papers among the quartet.
      4. Have participants draw the forehead and eye area, on the top strip of all paper sheets.
        NOTE: The middle strip is for depicting the nose and mouth area. The bottom strip is for depicting the chin, neck, and shoulder area.
      5. Include instructions of who to draw (e.g., S1/S3 meaning that participant 1 draws participant 3 and vice versa) in all paper strips.
      6. Have each paper representing a fully drawn portrait of a participant.
        NOTE: Consider different pastel writing paper colors for the different game phases.
      7. Have each participant's face depicted in a collaborative way by their partners. (Figure 2)
    2. Connect the dots game-the non-social condition
      NOTE: The control drawing condition is a game of connecting the dots. Each participant is invited to connect the dots of ascendent serial numbers to form a drawing. The connect the dots game is used as a neuropsychological instrument to measure cognitive domains such as mental flexibility and visual-motor skills36. The game stimulates visuospatial skills, increases mental activity37, and enhances mental abilities38. Use the following steps:
    3. Instruct the participants.
      1. Once the cap is in position, instruct the participants about fNIRS, the equipment, the caps, the wires, and the possible risks or discomforts involving the procedure.
      2. Remind them again about their right to leave the experiment at any time.
      3. Explain the two different drawing tasks.
      4. For the collaborative drawing, explain the horizontal strips and how to know where and who to draw in each strip.
      5. For the connect dots game, explain that they will have to connect the numbers in ascending order until the figure is revealed.
      6. Explain about the resting period and the recorded task commands.
      7. Engage the participants to observe their partners and the details that differentiate them. Indicate that, at the end of the study, the quartet that follows the rules and draws the most detailed figures will be rewarded.

Figure 1
Figure 1: The setting. The setup includes a squared table, four stools, and two wire supports (e.g., tripod), fNIRS equipment, a computer, and the cameras. (A) The setting scheme: Green numbers (1-4) correspond to the participants' labels and their stools/positioning at the table during the experimental run. Yellow numbers: 1 = fNIRS wiring supports, 2 = fNIRS signals' notebook receiver, 3 = NIRSport, 4 = 360° camera, 5 = support cameras. (B) Setting ready for the experimental run. Please click here to view a larger version of this figure.

Figure 2
Figure 2: Collaborative portraits-examples of portraits drawn in a collaborative way. Please click here to view a larger version of this figure.

2. Experimental paradigm

  1. Adapt the game for fNIRS acquisition
    NOTE: Adapt the game so that it is possible to capture the functional image of the brain through fNIRS and that data has a significant quality.
    1. Define the number of blocks.
      NOTE: The conditions must be repeated an adequate number of times to reduce the margin of error in the results. However, many repetitions can lead participants to automate tasks.
    2. Plan the duration of each block.
      NOTE: Consider the hemodynamic caption response time (on average 6 s after the beginning of a task). Also, consider the influence of block sizes to determine the filter for the following steps.
    3. Add a resting state period at the end of each block of both conditions (so that the hemodynamic signal decays before the start of the next block).
    4. Plan the order of blocks and create pseudo-randomized block sequences to reduce anticipatory effects.
    5. Plan the total duration of the game.
      NOTE: Consider participants' possible discomfort regarding the fNIRS tight caps and their proximity to each other. The blocks and conditions used in this protocol were designed as follows: nine blocks of the social condition of collaborative drawing (Table 1) and nine blocks of the non-social condition of connecting dots were created (duration = 40 s each); A resting period of 20 s between each of the blocks; three different sequences (Table 2) to perform the tasks (to avoid a condition being performed more than two times in a row). The experimental task duration was approximately 18 min.
  2. Paradigm programming software
    1. Use a software to assist in creating and organizing paradigm blocks and signaling to the participants when to start a new task.
      NOTE: The NIRStim software (see Table of Materials) was used in this case. Create the block sequences and program their distribution over the time during the experiment.
    2. Define events with visual (text and images) or auditory contents to indicate to the participants when to begin each task. In the Events tab, click on the button Add Event. Name the event in Event Name, select the event type in Stim Type, and define a color to represent the event in a presentation overview on Color-ID. Create markers to send to the acquisition software at the beginning of these tasks on Event Marker.
    3. Determine the task execution order and the number of repetitions of each one of them in the Trials tab. Also, insert rest periods. Determine the duration of both. Randomizing or not randomizing the trials is possible by selecting On/Off on Randomize Presentation; save the settings on the Save button.
    4. During the experimental run, display all the stimuli programmed in a black window (to prevent participant distraction) by pressing Run.

Table 1: Collaborative drawing condition. S1 = Subject 1, S2 = Subject 2, S3 = Subject 3 and S4 = Subject 4. Drawing dyads represents who is drawing who, and the drawing strip represents the writing paper's position for drawing in each condition. For example, for the first Block, use a blue paper sheet. C1, C2, and C3 represent 40 s of the paradigm of drawing social conditions that complete one portrait. C1 (drawing the forehead area, drawing dyads: S2 and S4; S1 and S3), C2 (drawing the nose area, drawing dyads: S1 and S4; S2 and S3) and C3 (drawing the chin area, drawing dyads: S3 and S4; S1 and S2). Follow the diagram for Blocks 2 and 3. This randomization maintains the order of drawing among volunteers (drawing the frontal partner, then the front-side partner, and lastly, the partner sitting next to them) and alters the order of the sheet strips to be drawn. Please click here to download this Table.

Table 2: Sequence 1-task randomization (social, non-social, and resting). Please click here to download this Table.

3. Video setup and data acquisition

  1. Cameras and video recording
    1. Select a commercially available scene camera (360°, see Table of Materials). Position that on the table so that all the participants' eye and head movements can be perceived simultaneously.
    2. Clean and check the memory card and battery. Check the brightness of the image. Test these items before the participants are located.
    3. Check for possible interferences on fNIRS reception. If so, increase the space between the equipment and its receiver.
    4. The equipment receiver must be independent of the fNIRS data receiver. Consider a notebook or a tablet located as far as possible from the table setting.
    5. Start the equipment, check the interface, and set the recording mode before fNIRS calibration.
    6. Consider one or two adjacent or supporting cameras that can be placed after both edges of the table.
  2. Video analysis
    1. For valuable statistical results, select a synchronized view/analysis software or platform that allows the transcription and coding of several video contents simultaneously, like INTERACT (see Table of Materials).
    2. Set parameters that enable the search for patterns/sequences to refine observational data to the research questions, e.g., individual gaze behavior metrics, head and eye movement, hand movement, facial expressions, and talking behavior.
    3. If one plans to record physiological measures, consider a software (see Table of Materials) that allows the integration of measured data from other acquisition systems.
    4. In the analysis process, consider not only the duration of events but also the sequence, their position in time, and how they relate to each other.
  3. Data extraction
    1. Start by downloading the video from all cameras (MP4 format). Load them into INTERACT. Segment the video data for coding and further analysis. For data extraction, mark the video sections manually and provide them with codes.
      NOTE: The purpose of segmenting and coding is to provide data categories so that the researcher can highlight and analyze different target behaviors.
    2. Segmentation
      1. By pressing Code Settings, create a first tier by dividing the block sections into social and non-social conditions and resting period. Create a second tier by dividing participants' behavioral data along with social conditions (face drawing). Align them by using the audio trigger timeline. Manually mark the start and end of each condition. Define the coding scheme following the guidelines (steps 3.3.2.2.-3.3.2.6.).
      2. Ensure that the coding scheme tracks behavior cues (duration and quantity) for each face drawing section (social condition) from all participants individually.
      3. Code for object-related attention-participant's gaze toward the drawing partner.
        NOTE: Gaze behavior has a dual function: gathering information from others (encoding), as well as communicating to others (signaling)39,40.
      4. Code for mutual gaze (when both partners that are drawing one another share visual contact).
        NOTE: Recent studies revealed increased activity in the anterior rostral medial prefrontal cortex (arMPFC) and its coupling with the inferior frontal gyrus (IFG) when partners established mutual gaze41.
      5. Code for associated behaviors during gaze behavior (single or mutual) such as smiling, direct speech, facial expressions, and laughs, indicating higher attentiveness to the drawing partner (Supplementary Figure 1).
      6. Transcribe and subdivide into categories the group participants' gaze behavioral data. Create interaction codes for each participant by labeling them. Make explicit the target behavior and tag number while coding.
    3. Coding and analyses
      NOTE: One of the researchers must undertake the behavioral coding task and analysis, since they are easily identified in the video. Observe the following:
      1. The extraction of the information must occur manually; mark on the timeline of each condition the observed behaviors according to the coding scheme. Mark the duration of each behavior. Do this for each participant separately.
      2. Cross-reference the participants' timelines to look for shared behaviors. Return to the video observation to analyze the quality of sharing (Supplementary Figure 2).
      3. Using the Export key, export the raw data as a text file or table file so that data can be sorted along the timeline, selected, counted, and tabled.
        NOTE: In this protocol, the sequential analysis function was not used due to the small number of coded event sequences42.
  4. Drawing metrics
    NOTE: This protocol uses drawing metrics to study possible correlations between participants' gaze behavior and the applied psychological tests. The following criteria were determined:
    1. Stroke quantity: Manually count the number of drawing strokes made by each participant in every face drawing section.
    2. Line continuity: Subdivide categories of long and short drawn lines. Manually count participants' long and short drawn lines.
      NOTE: Observational drawing results from direct observation of a chosen real object. Some recent studies found a correlation between line length and tracing or drawing tasks. Tracing task lines tend to be longer than drawing task lines43. This protocol associates tracing with memorized images that the individual has made stable and carries as drawing references in his/her symbolic system18.
    3. Drawing patterns: Relate to individual drawing patterns18 (Figure 3).
      NOTE: This protocol considers a binary classification for drawing pattern: 0, when the participant is in observational drawing mode (i.e., when the participant observes his/her drawing object and copies what he/she sees); and 1, when the drawing reflects internal stable memorized images (when there is a pattern of repeating shapes such as eyes, mouth, and hair throughout the drawing conditions).
    4. Observe details, including counting drawn details during the experiment (example.g., wrinkles, spots, eye shape, and eyebrow size, among others).
      NOTE: Drawn details may indicate greater attention to the object of drawing.
  5. Psychological tests
    1. Screen for symptoms of anxiety and depression, attention-deficit/hyperactivity disorder, and social skills when conducting group studies. Use free or commercially available scales.
      NOTE: This protocol suggests using the following: the Hospital Anxiety and Depression Scale44; the Social Skills Inventory45 (an inventory that evaluates the individual's social skills repertoire); and the Adult Self-Report Scale (ASRS-18) for the assessment of attention-deficit/hyperactivity disorder (ADHD) in adults46.

Figure 3
Figure 3: Examples of individual drawing patterns. Please click here to view a larger version of this figure.

4. fNIRS setup and data acquisition

  1. Data acquisition hardware
    1. Ensure to use acquisition hardware for the fNIRS registrations. The recordings must be performed by a combination of systems that can be read in the same recording program, totaling 16 channels.
      NOTE: Data acquisition was carried out using two continuous-wave systems (NIRSport, see Table of Materials) for the present study. Each piece of equipment has eight LED illumination sources emitting two wavelengths of near-infrared light (760 nm and 850 nm) and eight optical detectors (7.91 Hz).
  2. fNIRS optode channel configuration
    1. Use the NIRSite tool to locate the optodes over the PFC regions (see Table of Materials). Configure the distribution of optodes on the caps in a way that the channels are positioned above the regions of interest on all the participants' heads.
    2. Divide the optodes among the four participants for the simultaneous acquisition of signals.
      NOTE: The caps must have a configuration based on the international 10-20 system, and the anatomical areas of interest include the most anterior portion of the bilateral prefrontal cortex. For this protocol, optode placement was guided by the fNIRS Optodes' Location Decider (fOLD) toolbox47. The ICBM 152 head model (see Table of Materials) parcellation generated the montage. The recruitment of the prefrontal cortex region in social interaction tasks has been explained as a correlate of behavior control processes, including self-regulation48. Figure 4 represents the position of the sources and detectors.
  3. Preventing artifacts
    1. Remove distractors from the room where the game will take place.
    2. Advise the volunteers to move only as necessary.
    3. During the experiment, disconnect the NIRSport amplifier and the laptop from the electrical network.
    4. Turn any other equipment that works near the infrared spectrum off, such as air conditioning equipment. Turn off electrical devices present in the environment.
  4. Setting the fNIRS apparatus
    1. Previously, measure the brain perimeters of the four participants as follows: measure the distances between the nasion and the inion around the head to determine each participant's cap size. Always use a cap of a smaller size in relation to the perimeter of the head in order to give more stability to the optodes.
    2. On the day of the acquisition, instruct the participants to sit on the stool and then explain the expected process of placing the cap on the head.
    3. Fit the sources and detectors to the cap according to the predetermined settings. As a matter of organization, follow the pattern of using the optodes 1 to 4 on Subject 1, from 5 to 8 on Subject 2, from 9 to 12 on Subject 3, and from 13 to 16 on Subject 4.
    4. Put the caps on the participants' heads and position them so that the central midline (Cz) is at the top of the head. To check if Cz is at the central position, certify that it is located at half the distance between the nasion and inion.
      1. Also, measure the distance between the left and right ear (Crus of Helix) above the top of the head and Position Cz.
    5. Use overcaps to prevent ambient lights from interfering with the data acquisition.
    6. Connect the wires of the optodes to the amplifiers. As a matter of organization, follow the pattern of connecting optodes 1 to 8 to NIRSport 1 and optodes 9 to 16 to NIRSport2.
    7. Connect both NIRSport 1 and 2 to the computer via a USB cable.
  5. Data acquisition software
    1. After setting up the equipment, enable a software to acquire the fNIRS data. In this study, the NIRStar (see Table of Materials) software was used. On NIRStar, carry out the following steps:
      1. Click on Configure Hardware on the menu bar. Select the option Tandem Mode on the Hardware Specification tab so the hyperscanning can be performed.
      2. On the Configure Hardware tab, select a montage from among the predefined common montages or from the customized ones, and check the settings in Channel Setup and Topo Layout.
      3. Perform an automatic calibration by clicking Calibrate on the Display Panel. The signal quality indicator allows the verification of the integrity of the received data. Assess whether the quality of the data is enough to start the acquisition; that is, see if the channels are signaled as green or yellow.
        NOTE: If the directed channels are represented in red or white, remove them from the cap, check that there is no hair preventing the light from reaching the head, and clean the optodes with a cloth or towel. Connect them again to the cap and repeat the calibration.
      4. When ready to start the procedure, have a preview of how the signals are being received by clicking on Avance. Then, start recording the signals on Record.
      5. Open NIRStim, the blocks programming software (see Table of Materials), and start the presentation of the programmed blocks. The markers must be registered automatically, and their marking must be seen on the fNIRS data acquisition software.
      6. After the end of the procedure, stop recording by clicking on Stop, close the software, and verify if the file is saved in the chosen directory.
  6. fNIRS data analysis
    1. Preprocess the signals using NIRSLAB software49 (see Table of Materials). Follow the steps below:
      1. Apply a band-pass temporal filter (0.01-0.2 Hz) to the raw intensity data to remove cardiac and respiratory frequencies, as well as very low-frequency oscillations.
      2. For signal quality control, determine exclusion criteria for each channel gain above eight and coefficient of variation above 7.5%.
      3. Compute the changes in HbO2 and HHb by applying the modified Beer-Lambert law with the whole time series as a baseline.
        NOTE: In this study, HbO2 and HHb time series were segmented into blocks (social and non-social) and exported as text files for subsequent analysis in the R platform8 for statistical computing (see Table of Materials).
      4. Analyze separately the social and control conditions. Construct a correlation matrix for each of the nine blocks of each condition so that its elements correspond to the correlation (Spearman) between each pair of subjects in the evaluated channel. For the statistical significance of the correlations between individuals across the task, use the t-test8 for a one-sample mean, considering a significance level of 5%.

Figure 4
Figure 4: Distribution of optodes on the Subject 1 cap. The letters S and D represent the sources and detectors, respectively. S1 on AF7 coordinate of the 10-20 system; S2 on AF3; S3 on AF8; S4 on AF4; D1 on Fp1; D2 on F5; D3 on Fp2; and D4 on F6. The channels are placed in the following configuration: channel 1 between S1-D1; 2 between S1-D2; 3 between S2-D1; 4 between S2-D2; 5 between S3-D3; 6 between S3-D4; 7 between S4-D3; and 8 between S4-FD4. Please click here to view a larger version of this figure.

Representative Results

The protocol was applied to a quartet composed of young women (24-27 years old), all of them students on postgraduate programs (Hospital Israelita Albert Einstein, São Paulo, Brasil), with master's or doctorate level education. All participants were right-handed, and only one reported having previous drawing experience. No participants had a reported history of neurological disorders.

For the scales and psychological test results, two participants (2 and 4) showed high scores for anxiety (17 and 15 against the reference value of 9)44 and the cutoff value for depression (9)44. All participants' scale results for attention and hyperactivity showed scores below the cutoff values

The participants' social skills repertoire was also measured. Subjects 2, 3, and 4 obtained scores higher than 70% (well-developed repertoire of social skills). Subject 1 presented a score of 25% (related to a deficit in social skills). This test also analyses specific social skills such as F1, coping and self-assertion with risk; F2, self-assertion in expressing positive feelings; F3, conversation and social resourcefulness; F4, self-exposure to strangers and new situations; and F5, self-control and aggressiveness. For these factors, all participants showed low scores for F1, F2, and F3 (1% to 3%) and high scores for F4 (20% to 65%) and F5 (65% to 100%).

The fNIRs preliminary results (Figure 5) showed typical brain activation for subjects 1, 2, and 3 in both, social and non-social drawing conditions in both channels located in the left and right hemispheres; however, the activation patterns were distinct. Participant 4, on the other hand, showed atypical brain activation.

Figure 5
Figure 5: Results of the group average of fNIRS data. (A) Block average of fNIRS signals over subject 1. Left-side and right-side channels are displayed separately in x-axes for both conditions (social and non-social). (B) Block average of fNIRS signals over Subject 2. Left-side and right-side channels are displayed separately in x-axes for both conditions (social and non-social). (C) Block average of fNIRS signals over Subject 3. Left-side and right-side channels are displayed separately in x-axes for both conditions (social and non-social). (D) Block average of fNIRS signals over Subject 4. Left-side and right-side channels are displayed separately in x-axes for both conditions (social and non-social). Please click here to view a larger version of this figure.

The oxyhemoglobin signal (Figure 6) was significantly synchronized between subjects only during the last part of the task for both conditions, social and control (Figure 6: median correlation of 0.14; t-value = 1.77 and p-value = 0.046) and in the collaborative drawing condition (median correlation of 0.12; t-value = 2.39 and p-value = 0.028).

Figure 6
Figure 6: Boxplot of subjects' brain (oxyhemoglobin) correlations throughout the experiment. Each box contains a horizontal line (that indicates the median). The top edge represents the 75th percentile and the bottom edge the 25th percentile. For error bars, the boxplot is based on the 1,5 IQR value, above the third quartile, and below Q1, the lower quartile. The asterisk (*) indicates a statistically significant difference from zero. Part 1 corresponds to the first one-third of the experimental block (comprising C1, C2, and C3-the first full collaborative design-interspersed with resting periods and the non-social design condition); Part 2: C4-C6; Part 3: C7-C9. Please click here to view a larger version of this figure.

The video analysis showed that participants looked at each other while they were drawing partners and sometimes shared their gazes with one another. As with the amount of total gaze, a greater amount of synchronous gaze was observed in the positions where the partner was in a frontal position. From the middle to the end of the experiment, the synchronous gazes decreased significantly, and in C9, they did not occur.

Regarding drawing analysis, only Subject 3 reported having previous drawing experience (a 6 year course). Subjects 1, 2, and 4 had similar production in quantity and continuity of strokes. Participant 3 showed a drawing manner of short, non-continuous strokes and a greater number of total strokes. All four participants apparently maintained a constant pattern of figures in their drawings (previous drawing patterns), although subjects 3 and 4 reproduced a greater number of observed details. Even when drawing different partners, there were eye, mouth, and nose drawing patterns that the participants repeated in the social condition. For the participant with prior drawing experience, previous drawing patterns (see 3.4.3.) were also observed (e.g., eyebrows and eyes).

Supplementary Figure 1: Code interface and video. This figure represents video codification, segmentation, the timeline of events, and the coding scheme. Please click here to download this File.

Supplementary Figure 2: Intersection of gaze and drawing tasks. Condition Rest: resting period of 20 s; Condition Draw: the social condition of the experiment (40 s); Condition Dots: the control condition of connecting dots (40 s); Interaction Gaze 1: Subject 1 looking to drawing partner; Interaction Talk 1: Subject 1 speaking; Interaction Body 2: Subject 2 moving hands, shoulders, and head in non-verbal communication. Please click here to download this File.

Discussion

This study aimed to create a protocol using hyperscanning on four brains concurrently under naturalistic conditions. The experimental paradigm used different drawing tasks and the correlation of multiple outcome measures, drawing metrics, behaviors, and brain signals. The critical steps within this protocol are the consideration of the challenges arising from its high complexity and the maintenance of its ecological and naturalistic conditions.

Video observation was key to this study. It allowed the coding and segmentation of non-verbal communication behaviors on a timeline50. The video's qualitative analysis allowed the observation of gaze behavior in the search for signs that could indicate the functions of encoding (gathering information from the other) and signaling (communicating with others)39,40. It also helped refine the observational information by highlighting individual and group patterns13. Quantitative measures of gaze alone do not provide an understanding of the quality of gaze. The gaze may represent only an automatic behavior of repetition or thinking in the cognitive mode of comparison and structuring18,19. However, the small number of event sequences did not allow this study to perform sequential analysis and statistical validation42. The results showed a significant decrease in synchronous and asynchronous glances in the social conditions from the middle to the end of the experiment. One of the hypotheses for this behavior may be the very process of visual perception and the formation of symbolic representations inherent to the recognition of the other49,51. This hypothesis is in line with the theory regarding face-to-face observation drawing19 and with the results of studies that have associated longer observation times of an object with difficulty in deciding what and how to draw52. Another possible hypothesis is associated with the acquisition of visual memory of the objects drawn throughout the experimental task18,53. In the results, the gaze behavior measures showed a decreasing quantity associated with the drawing partner position. In frontal drawing positions, the amount of asynchronous and synchronous gaze was greater than in the conditions where the partner was to the side. This result is in line with studies suggesting that stimulus location affects visual perception54. The alignment between attentional focus and saccade behavior seems to suggest a certain "automatic" component to attentional focus regarding proximity and/or that the movement constraint acts as a distractor55. Furthermore, recent studies have shown that the interaction perspective can alter gaze patterns. When participants are engaged in a task, social attention on others decreases; this suggests that people in complex social environments look at each other less, and information gathering does not necessarily use direct gaze25. Differences in stroke, drawing, and gaze behavior were also observed between participants who did and did not have prior drawing experience. The participant with previous drawing experience showed far less pattern repetition than other participants and a greater number of drawn details. The number of drawing strokes was also higher, and this trend was accompanied by the number of dots followed in the tracing game and the number of direct gazes to partners. However, the small number of participants did not allow the resulting sequential data analysis and statistical validation to be performed. Drawing practice involves constant gaze shifts between the figure and the paper and the use of visual memory56. Previous studies suggest that drawing practitioners have an easier time encoding visual shapes than non-practitioners57. Still, the results of the study by Miall et al.52 also suggest that people with drawing training modulate their perception to the experience of observation, while prior knowledge (e.g., stable mental images) seems to direct the perception of non-practitioners. These aspects deserve further study, especially with regard to underlying neural networks and differences in attentional processing and gaze behavior25.

Many challenges arise from performing hyperscanning on a four-person group with fNIRS, especially considering a dynamic paradigm; therefore, modifications and technique troubleshooting were performed while refining the protocol methodology. The first challenge was the adequacy of the target brain area with the limitation of the number of optodes and the difficulty of dealing with the signal capture calibration of fNIRS. This protocol envisaged the investigation of two brain areas, the temporo-parietal junction (TPJ) and the medial prefrontal cortex (mPFC). Signal capture of the TPJ was discarded due to the non-controllable difficulty of the density and color of the participants' hair58 versus the number of caps to be administered simultaneously. There was also a major concern for the participants' comfort and availability of time. The second challenge concerns the recording of the experiment. Initially, the protocol foresaw using only one 360° camera located at the center of the table for the experiment. The use of auxiliary cameras proved essential, though. Another difficulty was addressing drawing technique issues to create a robust protocol. Most participants represented body and clothes, areas that were not foreseen in the game's rules, despite the careful explanation, including exposure to previously drawn examples. Some of the participants verbalized that they had difficulty portraying the sizes of the shapes and continued the drawing where the previous partner had stopped due to the proportions. This verbalization is in line with the results of studies suggesting that visual perception reduces the impact of perspective, causing perceptual distortions52. Other drawing studies focusing on the relationship between visual perception and motor commands have also suggested that distortions due to multiple factors intervene in the gaze/hand process20,43. Gowen's study paradigm43, for example, used two expert drawers, one copying and the other drawing from memory. Their results suggested the use of different neural strategies for each drawing technique. Copying seems to depend on comparisons, visual feedback, and closer tracking of the pen tip.

Concerns about the limitations of the technique also involve the ecological condition of the experiment, as well as the use of a collaborative drawing paradigm. One of them concerns participant proximity (due to fNIRS wiring and stability issues) and its possible interference in gaze measures. Different gaze behavior patterns can be produced in social contexts of forced proximity (such as elevator situations)59. Nevertheless, synchronous gazes were observed, facial expressions and smiles emerged throughout the quartet collection, and they possibly indicate a sense of engagement. These results align with results from experiments where "bottom-up" stimuli, such as looking at the other's face, provoked the creation of shared representations among the participants. Non-verbal communication may result from the conceptual alignment that occurs throughout the interaction4. Using drawing techniques "per se" is challenging since most people stop drawing in the realistic phase (after 11 or 12 years of age). The perception of drawing as not expressing reality generates frustration or the discomfort of self-judgment and, consequently, resistance to the act of drawing18. Face drawing could be yet another factor of discomfort. Despite this, the condition of collaborative drawing done in 40 s sessions proved effective for this protocol. When asked about the experience, all participants responded positively, laughed, and commented on the shared portraits. The structured form for the experimental task seems to have lessened the burden of individual production and facilitated interaction between participants, as in Hass-Cohen et al.19.

The correlation of signals or synchronicity of brain signals occurred most strongly in the last part of the experiment for both conditions, social and non-social. The hypothesis was that the social (collaborative design) and non-social (connecting the dots) conditions would result in distinct brain signal synchronicities at different moments in the timeline. In the non-social condition, the synchronicity was expected to result from the cognitive correspondence of the task common to all participants8. Even if they did not interact directly, the synchronicity of signals was expected to occur earlier in the timeline of the experiment8,10,11,12. In the social condition, on the other hand, synchronicity was expected to occur later due to a possible social interaction between different unknown individuals with diverse personal strategies13,14,16,49.

Although many factors may contribute to the preliminary results, a possible interpretation of them relates to the time participants needed to become familiar with each other and the tasks and, finally, create a sense of group. Drawing in itself may generate reactivity and anxiety through self-judgment or the feeling of being evaluated18. Previous studies have associated negative evaluations with variation in group interpersonal brain synchronization (IBS)17. Also, the impact of proximity between participants in this context is not known yet59. Alternatively, the familiarity acquired individually with the tasks and with the group, albeit late, may have generated a cognitive correspondence of engagement, like an "automatic" engagement that occurred beyond the differences between the two proposed tasks.60 The tasks seem to have worked in a block. Therefore, a longer rest period could possibly result in different brain responses by ensuring decreased brain signals picked up by fNIRS between conditions. Another result that needs further attention is the inverse relationship between the amount of synchronous and asynchronous gaze (which decreased throughout the experimental task) versus individual brain activity (which increased throughout the experimental task). A possible interpretation for this result may lie in the high cognitive demand of the experimental task52, but we cannot fail to consider the neurobiological processes involved in human empathic bonding1.

This protocol has an innovative character, first by applying drawing techniques used in art therapy to provoke empathic bonds between participants; second by the dynamic character of the social-ecological situation; and third by the concomitant measurement of four heads using the fNIRS hyper-scanning technique. The drawing social condition promoted eye contact between the participants, allowing the protocol to explore how gaze behavior supports interpersonal interaction behavior in naturalistic situations and the different personal strategies used to recognize others49. It is also a promising tool to study whether gaze behavior is, indeed, associated with the process of attention18 and engagement between partners61 under the same conditions. Relating all these issues, especially in a paradigm conducted in a group under naturalistic conditions, is a challenge for social neuroscience.

Many factors may have contributed to these preliminary results. All of these different variables deserve further study, and the use of this protocol can provide important clues for better understanding group social relationships in a naturalistic context.

Divulgaciones

The authors have nothing to disclose.

Acknowledgements

The authors thank Instituto do Cérebro (InCe-IIEP) and Hospital Israelita Albert Einstein (HIAE) for this study support. Special thanks to José Belém de Oliveira Neto for the English proofreading of this article.

Materials

2 NIRSport  NIRx Medizintechnik GmbH, Germany Nirsport 88 The equipment belong to InCe ( Instituto do Cérebro – Hospital Israelita Albert Einstein). two continuous-wave systems (NIRSport8x8, NIRx Medical Technologies, Glen Head, NY, USA) with eight LED illumination sources emitting two wavelengths of near-infrared light (760 and 850 nm) and eight optical detectors each. 7.91 Hz. Data were acquired with the NIRStar software version 15.2  (NIRx Medical Technologies, Glen Head, New York) at a sampling rate of 3.472222.
4 fNIRS caps NIRx Medizintechnik GmbH, Germany The blackcaps used in the recordings had a configuration based on the international 10-20
Câmera 360° – Kodak Pix Pro SP360 Kodak Kodak PixPro: https://kodakpixpro.com/cameras/360-vr/sp360
Cameras de suporte – Iphone 8 Apple Iphone 8 Supporting Camera
fOLD toolbox (fNIRS Optodes’ Location Decider) Zimeo Morais, G.A., Balardin, J.B. & Sato, J.R. fNIRS Optodes’ Location Decider (fOLD): a toolbox for probe arrangement guided by brain regions-of-interest. Scientific Reports. 8, 3341 (2018). https://doi.org/10.1038/s41598-018-21716-z Version 2.2 (https://github.com/nirx/fOLD-public) Optodes placement was guided by the fOLD toolbox (fNIRS Optodes’ Location Decider, which allows placement of sources and detectors in the international 10–10 system to maximally cover anatomical regions of interest according to several parcellation atlases. The ICBM 152 head model  parcellation was used to generate the montage, which was designed to provide coverage of the most anterior portion of the bilateral prefrontal cortex
Notebook Microsoft Surface Microsoft Notebook receiver of the fNIRS signals
R platform for statistical computing  https://www.r-project.org  R version 4.2.0 R is a free software environment for statistical computing and graphics. It compiles and runs on a wide variety of UNIX platforms, Windows and MacOS
REDCap REDCap is supported in part by the National Institutes of Health (NIH/NCATS UL1 TR000445) REDCap is a secure web application for building and managing online surveys and databases.
software Mangold Interact Mangold International GmbH, Ed.  interact 5.0 Mangold: https://www.mangold-international.com/en/products/software/behavior-research-with-mangold-interact.html. Allows analysis of videos for behavioral outcomes and of autonomic monitoring for emotionally driven physiological changes (may require additional software, such as DataView). Allow the use of different camera types simultaneously and hundreds of variations of coding methods.
software NIRSite NIRx Medizintechnik GmbH, Germany NIRSite 2.0 For creating the montage and help optode placement and location in the blackcaps.
software nirsLAB-2014 NIRx Medizintechnik GmbH, Germany nirsLAB 2014 fNIRS Data Processing
software NIRStar NIRx Medizintechnik GmbH, Germany version 15.2  for fNIRS data aquisition: NIRStar software version 15.2  at a sampling rate of 3.472222
software NIRStim NIRx Medizintechnik GmbH, Germany  For creation and organization of paradigm blocks

Referencias

  1. Feldman, R. The neurobiology of human attachments. Trends in Cognitive Sciences. 21 (2), 80-99 (2017).
  2. Hove, M. J., Risen, J. L. It’s all in the timing: Interpersonal synchrony increases affiliation. Social Cognition. 27 (6), 949-960 (2009).
  3. Long, M., Verbeke, W., Ein-Dor, T., Vrtička, P. A functional neuro-anatomical model of human attachment (NAMA): Insights from first- and second-person social neuroscience. Cortex. 126, 281-321 (2020).
  4. Redcay, E., Schilbach, L. Using second-person neuroscience to elucidate the mechanisms of social interaction. Nature Reviews Neuroscience. 20 (8), 495-505 (2019).
  5. Babiloni, F., Astolfi, L. Social neuroscience and hyperscanning techniques: Past, present and future. Neuroscience and Biobehavioral Reviews. 44, 76-93 (2014).
  6. Balardin, J. B., et al. Imaging brain function with functional near-infrared spectroscopy in unconstrained environments. Frontiers in Human Neuroscience. 11, 1-7 (2017).
  7. Scholkmann, F., Holper, L., Wolf, U., Wolf, M. A new methodical approach in neuroscience: Assessing inter-personal brain coupling using functional near-infrared imaging (fNIRI) hyperscanning. Frontiers in Human Neuroscience. 7, 1-6 (2013).
  8. Brockington, G., et al. From the laboratory to the classroom: The potential of functional near-infrared spectroscopy in educational neuroscience. Frontiers in Psychology. 9, 1-7 (2018).
  9. Sonkusare, S., Breakspear, M., Guo, C. Naturalistic stimuli in neuroscience: Critically acclaimed. Trends in Cognitive Sciences. 23 (8), 699-714 (2019).
  10. Duan, L., et al. Cluster imaging of multi-brain networks (CIMBN): A general framework for hyperscanning and modeling a group of interacting brains. Frontiers in Neuroscience. 9, 1-8 (2015).
  11. Ikeda, S., et al. Steady beat sound facilitates both coordinated group walking and inter-subject neural synchrony. Frontiers in Human Neuroscience. 11 (147), 1-10 (2017).
  12. Liu, T., Duan, L., Dai, R., Pelowski, M., Zhu, C. Team-work, team-brain: Exploring synchrony and team interdependence in a nine-person drumming task via multiparticipant hyperscanning and inter-brain network topology with fNIRS. NeuroImage. 237, 118147 (2021).
  13. Jiang, J., et al. Leader emergence through interpersonal neural synchronization. Proceedings of the National Academy of Sciences of the United States of America. 112 (14), 4274-4279 (2015).
  14. Nozawa, T., et al. Interpersonal frontopolar neural synchronization in group communication: An exploration toward fNIRS hyperscanning of natural interactions. Neuroimage. 133, 484-497 (2016).
  15. Dai, B., et al. Neural mechanisms for selectively tuning in to the target speaker in a naturalistic noisy situation. Nature Communications. 9 (1), 2405 (2018).
  16. Lu, K., Qiao, X., Hao, N. Praising or keeping silent on partner’s ideas: Leading brainstorming in particular ways. Neuropsychologia. 124, 19-30 (2019).
  17. Lu, K., Hao, N. When do we fall in neural synchrony with others. Social Cognitive and Affective Neuroscience. 14 (3), 253-261 (2019).
  18. Edwards, B. . Drawing on the Right Side of the Brain: The Definitive, 4th Edition. , (2012).
  19. Hass-Cohen, N., Findlay, J. C. . Art Therapy & The Neuroscience of Relationship, Creativity, &Resiliency. Skills and Practices. , (2015).
  20. Maekawa, L. N., de Angelis, M. A. A percepção figura-fundo em paciente com traumatismo crânio-encefálico. Arte-Reabilitação. , 57-68 (2011).
  21. Babiloni, C., et al. Simultaneous recording of electroencephalographic data in musicians playing in ensemble. Cortex. 47 (9), 1082-1090 (2011).
  22. Babiloni, C., et al. Brains "in concert": Frontal oscillatory alpha rhythms and empathy in professional musicians. NeuroImage. 60 (1), 105-116 (2012).
  23. Müller, V., Lindenberger, U. Cardiac and respiratory patterns synchronize between persons during choir singing. PLoS ONE. 6 (9), 24893 (2011).
  24. Greco, A., et al. EEG Hyperconnectivity Study on Saxophone Quartet Playing in Ensemble. Annual International Conference of the IEEE Engineering in Medicine and Biology Society. 2018, 1015-1018 (2018).
  25. Osborne-Crowley, K. Social Cognition in the real world: Reconnecting the study of social cognition with social reality. Review of General Psychology. 24 (2), 144-158 (2020).
  26. Kantrowitz, A., Brew, A., Fava, M. . Proceedings of an interdisciplinary symposium on drawing, cognition and education. , 95-102 (2012).
  27. Petersen, C. S., Wainer, R. . Terapias Cognitivo-Comportamentais para Crianças e Adolescentes. , (2011).
  28. Sheng, L., Yang, G., Pan, Q., Xia, C., Zhao, L. Synthetic house-tree-person drawing test: A new method for screening anxiety in cancer patients. Journal of Oncology. 2019, 5062394 (2019).
  29. Li, C. Y., Chen, T. J., Helfrich, C., Pan, A. W. The development of a scoring system for the kinetic house-tree-person drawing test. Hong Kong Journal of Occupational Therapy. 21 (2), 72-79 (2011).
  30. Ferreira Barros Klumpp, C., Vilar, M., Pereira, M., Siqueirade de Andrade, M. Estudos de fidedignidade para o desenho da família cinética. Revista Avaliação Psicológica. 19 (1), 48-55 (2020).
  31. Adams, E. Drawing to learn learning to draw. TEA: Thinking Expression Action. , (2013).
  32. Bernardo, P. P. . A Prática da Arteterapia. Correlações entre temas e recursos. Vol 1. , (2008).
  33. Cheng, X., Li, X., Hu, Y. Synchronous brain activity during cooperative exchange depends on gender of partner: AfNIRS-based hyperscanning study. Human Brain Mapping. 36 (6), 2039-2048 (2015).
  34. Baker, J., et al. Sex differences in neural and behavioral signatures of cooperation revealed by fNIRS hyperscanning. Scientific Reports. 6, 1-11 (2016).
  35. Bowie, C. R., Harvey, P. D. Administration and interpretation of the Trail Making Test. Nature Protocols. 1 (5), 2277-2281 (2006).
  36. Valenzuela, M. J., Sachdev, P. Brain reserve and dementia: A systematic review. Psychological Medicine. 4 (36), 441-454 (2006).
  37. Johnson, D. K., Storandt, M., Morris, J. C., Galvin, J. E. Longitudinal study of the transition from healthy aging to Alzheimer disease. Archives of Neurology. 66 (10), 1254-1259 (2009).
  38. Risco, E., Richardson, D. C., Kingstone, A. The dual function of gaze. Current Directions in Psychological Science. 25 (1), 70-74 (2016).
  39. Capozzi, F., et al. Tracking the Leader: Gaze Behavior in Group Interactions. iScience. 16, 242-249 (2019).
  40. Cavallo, A., et al. When gaze opens the channel for communication: Integrative role of IFG and MPFC. NeuroImage. 119, 63-69 (2015).
  41. Kauffeld, S., Meyers, R. A. Complaint and solution-oriented circles: Interaction patterns in work group discussions. European Journal of Work and Organizational Psychology. 18 (3), 267-294 (2009).
  42. Gowen, E., Miall, R. C. Eye-hand interactions in tracing and drawing tasks. Human Movement Science. 25 (4-5), 568-585 (2006).
  43. Marcolino, J., Suzuki, F., Alli, L., Gozzani, J., Mathias, L. Medida da ansiedade e da depressão em pacientes no pré-operatório. Estudo comparativo. Revista Brasileira Anestesiologia. 57 (2), 157-166 (2007).
  44. del Prette, Z., del Prette, A., del Prette, Z., del Prette, A. . Inventario de Habilidades Sociais. , (2009).
  45. Mattos, P., et al. Artigo Original: Adaptação transcultural para o português da escala Adult Self-Report Scale para avaliação do transtorno de déficit de atenção/hiperatividade (TDAH) em adultos. Revista de Psiquiatria Clinica. 33 (4), 188-194 (2006).
  46. Zimeo Morais, G. A., Balardin, J. B., Sato, J. R. fNIRS Optodes’ Location Decider (fOLD): A toolbox for probe arrangement guided by brain regions-of-interest. Scientific Reports. 8 (1), 3341 (2018).
  47. Davidson, R. J. What does the prefrontal cortex "do" in affect: Perspectives on frontal EEG asymmetry research. Biological Psychology. 67 (1-2), 219-233 (2004).
  48. Hessels, R. S. How does gaze to faces support face-to-face interaction? A review and perspective. Psychonomic Bulletin and Review. 27 (5), 856-881 (2020).
  49. Mangold, P. Discover the invisible through tool-supported scientific observation: A best practice guide to video-supported behavior observation. Mindful Evolution. Conference Proceedings. , (2018).
  50. Kandel, E. R. . The Age of Insight. The quest to understand the unconscious in art, mind and brain from Vienna 1900 to the present. , (2012).
  51. Miall, R. C., Nam, S. H., Tchalenko, J. The influence of stimulus format on drawing-A functional imaging study of decision making in portrait drawing. Neuroimage. 102, 608-619 (2014).
  52. Gombrich, E. H. . Art and Illusion: A study in the psychology of pictorial representation. 6th ed. , (2002).
  53. Kirsch, W., Kunde, W. The size of attentional focus modulates the perception of object location. Vision Research. 179, 1-8 (2021).
  54. Deubel, H., Schneidert, W. X. Saccade target selection and object recognition: Evidence for a common attentional mechanism. Vision Research. 36 (12), 1827-1837 (1996).
  55. Tchalenko, J. Eye movements in drawing simple lines. Perception. 36 (8), 1152-1167 (2007).
  56. Perdreau, F., Cavanagh, P. The artist’s advantage: Better integration of object information across eye movements. iPerceptions. 4 (6), 380-395 (2013).
  57. Quaresima, V., Bisconti, S., Ferrari, M. A brief review on the use of functional near-infrared spectroscopy (fNIRS) for language imaging studies in human newborns and adults. Brain and Language. 121 (2), 79-89 (2012).
  58. Holleman, G. A., Hessels, R. S., Kemner, C., Hooge, I. T. Implying social interaction and its influence on gaze behavior to the eyes. PLoS One. 15 (2), 0229203 (2020).
  59. Dikker, S., et al. Brain-to-brain synchrony tracks real-world dynamic group interactions in the classroom. Current Biology. 27 (9), 1375-1380 (2017).
  60. Gangopadhyay, N., Schilbach, L. Seeing minds: A neurophilosophical investigation of the role of perception-action coupling in social perception. Social Neuroscience. 7 (4), 410-423 (2012).

Play Video

Citar este artículo
Gonçalves da Cruz Monteiro, V., Antunes Nascimento, J., Bazán, P. R., Silva Lacerda, S., Bisol Balardin, J. Group Synchronization During Collaborative Drawing Using Functional Near-Infrared Spectroscopy. J. Vis. Exp. (186), e63675, doi:10.3791/63675 (2022).

View Video