BIRAFFE2, a multimodal dataset for emotion-based personalization in rich affective game environments

Generic emotion prediction models based on physiological data developed in the field of affective computing apparently are not robust enough. To improve their effectiveness, one needs to personalize them to specific individuals and incorporate broader contextual information. To address the lack of relevant datasets, we propose the 2nd Study in Bio-Reactions and Faces for Emotion-based Personalization for AI Systems (BIRAFFE2) dataset. In addition to the classical procedure in the stimulus-appraisal paradigm, it also contains data from an affective gaming session in which a range of contextual data was collected from the game environment. This is complemented by accelerometer, ECG and EDA signals, participants’ facial expression data, together with personality and game engagement questionnaires. The dataset was collected on 102 participants. Its potential usefulness is presented by validating the correctness of the contextual data and indicating the relationships between personality and participants’ emotions and between personality and physiological signals.


Background & Summary
Affective Computing (AfC) 1 -an interdisciplinary field of study regarding human emotions-is to large extent built upon the assumption that we are able to precisely detect, label and manipulate emotional responses of agents. Therefore the proper understanding and modeling of this complex phenomena 2,3 , as well as maintaining ingenious experimental setup to do so, is a crucial determinant of success in this field. Such setup should expose the participant to specific emotion evoking stimuli and measure the variety of their reactions.
In line with James-Lange 4 and Prinz 5 theories, in our work we assume that the measurement of bodily reactions to stimuli can serve as proper foundation for emotion recognition. This widely used approach considers diverse signals, among which the most important ones are those offering the highest accuracy of emotion prediction: cerebral activity (EEG), heart activity (ECG), and electro-dermal activity (EDA; also GSR -galvanic-skin reaction) 6,7 . Attention is also paid to signals that allow the assessment of emotions without connecting electrodes, such as facial expressions or posture changes 6 . Availability of commercial devices that measure each of these signals, facilitates experimentation and prototype preparation 8 . A rapid growth in the wearables market 9,10 allows the use of these psychophysiological signals not only in AfC but also in other research areas 11 . Wearable devices and other non-intrusive sensors used in our approach make the measurement context as ecological as possible 12,13 .
Several datasets containing emotion-related sensory data are available for research purposes. Most of them were summarized by K-EmoCon dataset authors 14 . Unfortunately, these sets suffer from two main issues. First of all, there are not many people examined, which makes the final datasets too small to attempt to create models for predicting emotions 9,14 . Secondly, as a result, the effectiveness of prediction is only average. Results above 90% are achieved when the set of emotions is limited to 2-3 values, and the larger the emotion set, the poorer the effectiveness 6 .
As the generic models are not so robust, to fully exploit user's emotion detection, the recognition and classification of emotional states must be personalized. Furthermore, the same change in physiological signal may mean different things, depending on the situation. In Prinz's theory 5 , this is called content of the emotion. For 1 instance, an increase in heart rate may be related not only to the intensification of emotions, but also to the start of exercise at the gym or an increase in ambient temperature. In order to take into account the context, an appropriate experimental environment is needed to log and control it 12 . We assume that computer games are suitable for this purpose 13,15 .
Based on the outlined motivation, we created a two-part procedure consisting of an audio-visual emotional stimulus assessment and gaming session. Custom computer games with a mechanism for recording the entire context, i.e., all participant's actions and game environment's changes, were developed. The procedure was complemented by physiological measurements and personality assessment. The whole study resulted in the BIRAFFE1: Bio-Reactions and Faces for Emotion-based Personalization dataset 16,17 , and in a series of analyses regarding the use of personality in emotion prediction 18 . Now, we introduce The 2nd Study in Bio-Reactions and Faces for Emotion-based Personalization for AI Systems (BIRAFFE2) dataset 19 , a follow-up study that complements the previously established framework which introduces a number of improvements. It contains the ECG, EDA, facial expressions and hand movements recorded using portable and affordable devices: BITalino (r)evolution kit, web camera and gamepad. Physiological data are supplemented with detailed game logs, an evaluation of player engagement in the games (GEQ questionnaire), participants' responses from the experimental part, and a personality assessment (NEO-FFI inventory).
Like the ASCERTAIN 20 and AMIGOS 21 datasets, BIRAFFE2 combines physiological data with personality profiles and complex stimuli. However, unlike the other two, that are based on longer movie clips (up to 150 seconds), BIRAFFE2 contains detailed game logs. As a result, context in BIRAFFE2 can be easily broken down into frequently quantified detailed series of events to facilitate further analysis. Also, there is no need for manual data annotation, which is required for movie clips. As such, BIRAFFE2 may be useful in AfC research concerning, but not limited to, multimodal information fusion, personalized prediction of emotional changes, and interface development methods using contextual information.

Methods
Dataset design. The BIRAFFE2 dataset was developed to verify the usefulness of contextual data and personalization in emotion detection, whether in games or in apps. To achieve a balance between a fully personalized and an universal unpersonalized system, we propose a partial emotional adaptation using personality assessment 18 . Research shows, that the use of personality for grouping different types of users enhances the effectiveness of emotion prediction 22 (review of current trends in personalized systems is provided in 23 ). Implementing that kind of context-and personality-based tailoring, requires the appropriate set of information about the user must be collected. To verify these assumptions, we constructed a combined experimental paradigm. First, in the "classical approach" part, we presented stimuli to the participants and collected their answers: both with questionnaires and physiological measures. Secondly, within the "ecological approach", we embedded the reaction measurement in the specific context of some simple computer games. When the player is loosing in the game, and the system detects the increased intensity of the her/his reaction, it is crucial for the model to interpret the context of such change -only by the contextual information coming from the game progression, we are able to tell whether the intensification was a collateral of anger or joy. Knowing its contextual origin, we can also easily prevent these specific changes from happening again, or on the contrary -repeatedly elicit detected and labeled emotions. As such, games offer a perfect opportunity for testing the human emotional reactions in the feedback loop with the computer system (also called affective loop) 24 .
The paradigm presented in this paper is the continuation of our previous work on the BIRAFFE1 experiment 17 . In the current work, a number of extensions and improvements were introduced, drawing on the conclusions and lessons learned from the previous study 18 : 1. We improved the affect assessment widget. 2. We have enhanced the stimulus selection: it currently covers a wider area in the Valence-Arousal space, is more randomized and excludes erotic stimuli. 3. We have developed new custom computer games designed to arouse players' emotions -unlike BIRAFFE1, here the games use only well-defined mechanics to evoke a limited set of emotions, which makes the analysis of game logs easier and clarifies conclusions. 4. We used the GEQ questionnaire to assess the involvement in the game, as well as asked about previous experiences with games to allow more accurate analysis of the emotions in games. 5. We extended the range of the measured responses by the introduction of accelerometer from gamepad. 6. EDA and ECG electrodes placement was changed to overcome issues identified in the previous experiment. 7. Several small improvements were also made, e.g., face photos are now taken with higher frequency, the screencast is recorded during game sessions.
Ethics statement. Described  www.nature.com/scientificdata www.nature.com/scientificdata/ Questionnaires. First, the paper-and-pen Polish adaptation 25 of the NEO Five Factor Inventory 26 consisting of 60 self-descriptive statements evaluated on 1-5 scale (1 -strongly disagree; 5 -strongly agree) was used to measure the Big Five personality traits.
Second, our own paper-and-pen Polish translation of The Game Experience Questionnaire (GEQ) Core Module 27 was used to measure players' feelings during the game session. The module consists of 33 items, ranked on 0-4 scale (0 -not at all; 4 -extremely). Items were arranged in seven components in original version: Competence, Sensory and Imaginative Immersion, Flow, Tension/Annoyance, Challenge, Negative affect and Positive affect. The questionnaire has been used in many game studies 28,29 . However, the 7-factor structure has not been confirmed by anyone. In 28 revised version (GEQ-R) was proposed. Tension/Annoyance, Challenge and Negative affect were merged into Negativity, leading to a 5 component solution.
Finally, our own simple questionnaire was used to measure gaming experience. It consists of two questions: (1) "Over the past year I have played computer / mobile / video games:" (2) "In the period of my most intense interest in computer / mobile / video games, I played:". Both were answered by selecting one of the five possible answers: (a) daily or almost daily, (b) several times a week, (c) several times a month, (d) several times a year, (e) not at all. There was also a space for leaving comments on the experiment. Stimuli selection. Standardized emotionally-evocative images and sounds from IAPS 30 and IADS 31 datasets were used as stimuli, each characterized by its coordinates in the Valence-Arousal space. The analysis of IADS and IAPS scores revealed the following trend: arousal score increases as the valence score strives for it's positive or negative extreme (Fig. 1).
For the purpose of the experiment, we divided the stimuli into three groups according to their arousal and valence index: + (positive valence and high arousal), 0 (neutral valence and medium arousal), -(negative valence and high arousal). Erotic stimuli were excluded from the blind selection, due to the risk of creating weird or disgusting combinations (e.g. picture of child or snake is paired with the erotic sound), not intended by the aim of this study.
The stimuli set for each participant was generated by random sampling without replacement and formed nine conditions, each consisting of 13 stimuli: • three consisting only of pictures: p+, p0, p−, • three consisting only of sounds: s+, s0, s−, • three, where pictures were paired with sounds: + picture with + sound (ps+), 0 picture with 0 sound (ps0),picture with -sound (ps−).
Conditions were mixed during the presentation, which was divided into two sessions (17.5 min each) and separated by the game session.
Emotion evaluation widget. Emotional assessment was carried out using the Valence-arousal faces widget controlled by a left joystick on a gamepad. The widget was adapted from our previous experiment 17 , with the following improvements: • Emoticons placed as hints were moved outside the selection area. Also, the border of the selection area was introduced. In the previous version the participants often chose the location of the smiley as their answer. Now, there is no possibility to put selection marker on them. • The selection marker changes color to indicate that there is only half a second left for the rating. In the previous version there was no information about the remaining time. • The returned valence and arousal scores are now within a range of [1,9], so they are within the same range as the assessments in IAPS and IADS. In the previous version, they were in the [−1,1] range, which required conversion of values before the analysis started.
To compare current and previous version of Valence-arousal faces widget see Fig. 2.
Games. Three affective games developed by our team 15 were used during the study (see Fig. 3). All of them were controlled by a gamepad and produced game log CSV files. They have been designed with the emphasis on differentiating the levels of difficulty: • Room of the Ghosts: The goal: pass through a series of rooms and defeat the arriving ghosts. Difficulty: very easy, achieved by the following implementations: the collider for protagonist is smaller than their real model -removing the feeling of being hit before the projectile hits the player; the protagonist's weapon can shoot more often and faster than the opponents' weapons. • Jump!: The goal: reach the end of the path by jumping on the platforms and avoiding obstacles. Difficulty: hard and frustrating, achieved by the following implementations: colliders are too big -player can get hit by trap before s/he touches it with the model; movement is clunky, and there are several traps, i.e. invisible blocks, which increases the confusion and irritation in player; each time the protagonist dies, the background music is getting less pleasant (the pitch and distort levels of music playing in background increases by 0.07). • Labyrinth: The goal: walk the protagonist through the labyrinth. Difficulty: optimal, achieved by the following implementations: the colliders have been adjusted to not hit the walls too often and to make the movement smooth; the protagonist control is natural and predictable. Both are presented in Polish, as in studies. X axis has labels "negative", "neutral", "positive", while Y axis has labels: "high arousal" and "low arousal". Software. The procedure was running under the Python 3.8, written with PsychoPy 3.2.4 library 32 . Python code controlled the execution of the whole protocol, i.e. stimuli presentation, screencast recording (using OBS Studio software, https://obsproject.com/), photos taking, games' start and end management.
Physiological signals were gathered using BITalino (r)evolution kit, as it is the most promising of cheap mobile hardware platforms (for comparison see 33 ). Electrocardiogram recording was implemented using the classical 3 leads montage with electrodes placed below the collarbones (V-and reference) and below the last rib on the left side of the body (V+). EDA signal was gathered by 2 leads placed on the forehead (placement as good as classical palmar location 34 , with no side effects related to gamepad held by the participants). EDA electrodes were hidden under the body-coloured band, to not interfere with the facial emotion recognition process conducted by the API in the later experimental stages, and to provide tight and stable contact of the EDA electrode and skin. Both signals were probed with 1 kHz sampling rate.
Experimental protocol. The study took place in a staff room at the university. For the duration of the study, the window blinds were drawn and the lights were turned on to reduce external stimuli and to balance the light levels among the participants. During the study, the room was reserved for the experiment only. Two research stands were prepared, located next to the opposite walls, so that the participants sat back to back. Each participant was seated in front of a monitor and provided with the consent and short information about the whole experiment. Throughout the whole procedure, the investigator remained at his desk, rear-facing to the subjects, to minimise their impact on the participant's behaviour.
When registering for the experiment, participants were asked to wear loose clothes to avoid problems with sticking ECG electrodes.
Participants began the procedure by filling out the NEO-FFI inventory. They were then connected to measuring devices, with the headphones set up, and a gamepad given to hand. Computer protocol consisted of five phases:

Fig. 4
Research setup: 23" LCD screen, headphones, external web camera, gamepad, band for EDA electrodes and Bitalino with 3-leads ECG and 2-leads EDA. Keyboard and mice were used only by the researcher to start the protocol.
Sensors are being calibrated. We will begin after one minute" (for the purpose of the paper, the instructions have been translated from Polish; original images used in the procedure can be found in a dedicated repository: https://gitlab.geist.re/pro/biraffe2-supplementary-codes). 2. Instructions and training (approx. 5 min). The entire instruction is divided into 4 screens: (a) "In our study, you will look at pictures and listen to sounds, sometimes occurring separately, sometimes in pairs (sound + picture), and then assess the emotions they evoked in you. You will also play a game consisting of three levels. Sit back and adjust the headphones to feel comfortable. Everything will be controlled by a game pad. If anything is unclear during the study, feel free to ask questions. " (b) "The task of assessing pictures and sounds will have two equal parts, separated by a game. Each trial will begin with a cross, followed by a stimulus (sound, picture, or picture + sound). You will then see our emotion assessment tool -you will have 6 seconds to answer. " (c) "The emotion rating tool is a two-dimensional space -the emoticons are there to help you orient yourself. The horizontal axis is for pleasure and the horizontal axis is for arousal. Select an emotion by moving anywhere in the square using the LEFT JOYSTICK. " (d) "To practice this part, you will now go through a TRAINING consisting of 4 stimuli, evaluate them using our tool. We are not collecting data yet -this is just a practice session. " At the end of the practice session, a message appeared: "End of training. If you have any questions, ask them now". 3. First part of stimuli presentation and rating (17.5 min): each presentation lasted 6 s (as each sound in IADS set lasts 6 s), followed by 6 s for affective rating and 6 s ISI. It results in 18 s between stimuli onset, which is enough for observing reactions in the ECG and EDA signals. 4. Games session (up to 15 min in total) started with on-screen message: "Game time". Each game had a 5 min time limit, after which it turned itself off. After the completion of one game, another automatically switched on. 5. Second part of stimuli presentation and rating (17.5 min).
After the computer protocol, participants filled out three GEQ questionnaires (one for each game) and gaming experience questionnaire.
ECG and EDA signal, as well as gamepad accelerometer and gyroscope readings, were collected continuously during the whole experiment. Facial photos were taken every 250 ms. A screencast was recorded during the game session, in case for the need to fill in the missing information in game logs after the experiment. The whole protocol lasted up to 75 min.

Data Records
The presented BIRAFFE2 dataset is available under the CC BY 4.0 licence at Zenodo 19 (https://doi.org/10.5281/ zenodo.3865859). It consists of data gathered from 102 out of 103 participants. Unfortunately, during the protocol for participant 723 there were problems with the hard drive and all data was lost, except the paper-and-pen NEO-FFI and GEQ questionnaires. Also, some smaller issues occurred for a subset of participants, e.g., game crashed, Bluetooth signal was lost, electrode contact was poor. We have published also these incomplete records, as in many analysis only selected of the subsets will be used and it will not be the problem. Missing values in all files are represented by NaN. Table 1 summarizes the collected data.
The BIRAFFE2 dataset is composed of a metadata file (BIRAFFE2-metadata.csv) and 7 archives described in detail in the subsequent paragraphs. The whole is complemented by sample files (sample-SUB211-…) that allow one to explore the structure and content of individual data types without having to download entire archives. All files have Unix timestamps which can be used for synchronization between different subsets.
The dataset is supplemented by the raw images of participants placed in two Zenodo repositories 35 www.nature.com/scientificdata www.nature.com/scientificdata/ permission from the participants to distribute them, we can only make them available for the purpose of validating the processed data available in the BIRAFFE2-photo.zip and BIRAFFE2-photo-full.zip archives. Interested researchers are invited to submit an access request via the Zenodo platform with a brief description of the research within which validation will be performed.
BIRAFFE2-metadata.csv contains a summary of each participant: age, sex, personality profile, GEQ results and information about subsets available for given person (whether there is a BioSigs, Screencast,… file available for the person). Each line of this file represents one participant and includes the following values: • ID -a randomly assigned participant ID from range {100,999}. It is used to identify all participant-related files as filenames. Filenames are created according to the format SUBxxx-yyyy, where xxx is the ID, and yyyy is the data type identifier (e.g., BioSigs, Gamepad), • NEO-FFI;GEQ;BIOSIGS;GAMEPAD;GAME-1;GAME-2;GAME-3;PHOTOS;PROCE-DURE;SCREENCAST -information about subsets available for given person, i.e., whether there is a Bio-Sigs file, Gamepad file, etc. available for the person (Y or NaN). GAME-X columns inform about availability of game logs for level X. The NEO-FFI and GEQ columns only indicate whether there are questionnaire results in the following columns, The signals were recorded using the Python library for BITalino (http://bitalino.com/pyAPI/). Due to the instability of the bluetooth connection, and in the absence of handling such a situation in the library, the timestamps after series of connection errors (i.e., the interval between consecutive error entries in the Procedure log was less than 0.05 s) cannot be considered as fully reliable. This will be further investigated by our team.
B I R A F F E 2 -g a m e p a d . z i p contains accelerometer and gyroscope recordings. Each SUBxxx-gamepad.csv file represents one participant and consists of one line per each gamepad recording. The values were recorded as quickly as they were transmitted through the USB interface, with an average frequency of 250 Hz. The fields contained in each line are: • GYR-X;GYR-Y;GYR-Z -gyroscope readings: right side of the gamepad upward (GYR-X), buttons and joysticks panel upward (GYR-Y), audio port upward/light bar downward (GYR-Z). Note that the gyroscope values represent the position of the gamepad, not the angular rate, as in the DS4Windows library, which code was used as the base for our Python data acquisition code (https://github.com/Jays2Kings/DS4Windows). BIRAFFE2-games.zip contains logs from the games. Five JSON log files are created for each participant. Note that the file was created when the level has started. If the game has crashed in a given level (which sometimes happened), the subsequent levels were not started and the participant was returned to the stimulus session. This means that in some cases files may be missing, e.g., files for level 3 if the game crashed at the second level. • SUBxxx-Level02_Log.json is a log from the Jump!, as the log from the first level, it is composed of a repeated pattern collected at multiple points in time: • SUBxxx-Level02_BlockEvents.json contains the information about the dynamic blocks of the Jump! game world. The design is analogous to SUBxxx-Level01_MapLog.json: • SUBxxx-Level03_Log.json is a log from the Labyrinth, as the previous logs, it consists of often sampled structure: www.nature.com/scientificdata www.nature.com/scientificdata/ BIRAFFE2-photo.zip contains face emotions description calculated by MS Face API with recogni-tion_02 model. Each SUBxxx-Face.csv file represents one participant and consists of one line per each photo taken. Photos were taken with 4 frequency during games and during stimuli presentation (every 15 frames at 60 fps). Photos were not taken while the participant was responding on the widget. When no face was recognized or two faces were found (the second was the experimenter face) NaN value was used.
• GAME-TIMESTAMP -Unix timestamp available only during the game (NaN value during the stimuli presentation), • FRAME-NUMBER -Index of the photo within the context of the stimuli presentation (NaN value during the games), measured in frames since the beginning of the stimuli presentation: −1 for pre-stimulation photo, 0 for photo in the moment when stimuli appears, 15 for the next photo Raw images of participants were placed in two Zenodo repositories 35,36 , described in the Data Records section, solely for the purpose of validating the processed data.
BIRAFFE2-photo-full.zip contains all information available in BIRAFFE2-photo.zip but also other face-related values recognized by MS Face API, e.g., recognized age, whether the person wears glasses, what is the color of the hair (the full documentation is available at https://azure.microsoft.com/services/ cognitive-services/face/). BIRAFFE2-procedure.zip contains a log of all the stimuli presented to a given user. Each SUBxxx-Procedure.csv file represents one participant and consists of one line per each stimuli presentation. The fields contained in each line are: Personality. The personality profiles inferred with the NEO-FFI questionnaire were inspected. In Fig. 6, it can be seen that for each personality trait the values span the entire possible scale ( [1,10]), but most values represent medium trait levels. The largest deviation from the norm are the higher values on the agreeableness scale. This confirms that the personality profiles in the sample do not deviate much from the expected distributions, thus the sample can be treated as representative with respect to the Polish norms of the NEO-FFI questionnaire 25 .
Emotion-based personalization capabilities. As one of the main motivations for developing the BIRAFFE2 dataset is to verify the usefulness of personality profiles in recognizing emotions from physiological signals, It was examined, on the one hand, whether there are relationships between personality and self-reported emotions on widgets, and on the other hand, it was examined whether the characteristics of physiological responses differ for different scores of personality traits. www.nature.com/scientificdata www.nature.com/scientificdata/ To verify the former, four regression models were fitted with five personality traits-each on ( [1,10]) scale for the first two models and on ( [1,3]) scale (low, medium, high; according to the aforementioned ranges) for the other two models-as predictors, and with valence and arousal as criterion. The results presented in Table 2 indicate significant relationships between the four personality traits and arousal as well as between agreeableness and valence.
To check the relationship between physiological signals and personality, 3 variables were extracted with the heartpy library 37 from the ECG signal for each participant: (a) heart rate (number of beats per minute), (b) mean of successive differences between R-R intervals (MoSD), and (c) breathing rate. Each of these was used as the dependent variable and its relationship with five personality traits was examined using ANOVAs. A summary of each of the three models is included in Table 3. They indicate significant relationships between personality traits and ECG signal characteristics.
As both relationships between personality and participants' emotions and between personality and physiological signals were found, we assume the potential usefulness of the dataset in the proposed research area of emotion personalization.  www.nature.com/scientificdata www.nature.com/scientificdata/ Games validation. The generated game logs were supposed to be complete, i.e., they should allow to recreate everything that happened in the game. In order to verify this, firstly level maps were recreated from static logs. Then the specific events were extracted from the dynamic logs containing the course of each player's gameplay and placed on such prepared maps. The maps thus generated (see Fig. 7) are correct, which was cross-checked by comparing them with the games. Also, the events appear to be correctly logged, as indicated, e.g., by the large number of health events in the middle room on the first level. This is due to players just becoming familiar with the game interface, leading to frequent protagonist deaths in this location, which is the first room in the game. The dense occurrence of events is thus consistent with observations made by the researchers during the experiment.
Finally, due to the assumed variation in the level of difficulty and emotion evoked by the games, the results of the GEQ questionnaire were analyzed, thus confirming the assumptions stated in the Games section for each game. The first level was characterized by high Flow, Competence, and Positive Affect values, indicating that players were engaged in the game, had control over it, and felt positive emotions. The questionnaire results for the second level indicate the highest level of Negativity of all levels, along with low levels of Positive Affect and Competence. This points to negative emotions associated with poor performance and a poor sense of control. What is interesting is that the level of Flow does not differ between the first and second levels, which may mean that emotions are most important for game engagement, regardless of their valence. The third level was characterized by low values of Flow and Negativity, associated with its emotional neutrality, while maintaining a high level of Competence, indicating a high self-assessment of the players' effectiveness at this level. For detailed results of GEQ-related analysis, see 15 .

Usage Notes
The data from all subsets are saved in either CSV or JSON formats, supported by all programming languages and data analysis tools.
Among the potential directions for analysis of the data contained in the BIRAFFE2 dataset, one should consider: • Development of prediction models for emotions represented in Valence-Arousal space based on multimodal data fusion methods (accelerometer signal, biosignals, facial data and personality profiles). • Analyzing the effects of gaming sessions, that is, high cognitive load associated with game challenges and frustration associated with the "Jump!", on affective ratings. • Attempt to extract higher-level context from low-level game log data, for example, converting protagonist and enemy coordinate information into "runs away from enemy" and "attacks enemy" labels. • Conduct process mining on game data aligned with physiological data to discover recurring patterns of bodily changes and game events.
It is important to note the limited demographics of the participants. All of the participants were young people of Polish nationality, and most of them were students in technical fields. This may have introduced bias into the dataset and limited the ability to generalize the results to other groups of people.  Table 3. Three ANOVAs with five personality traits (each on [1,10] scale) as independent variables, and with ECG-related features as dependent variables.

Code availability
The recording of all data was controlled by a procedure written with PsychoPy 3.2.4 that established a connection to the gamepad, BITalino, and camera and recorded the data streams and images. The raw data thus collected was then converted to CSV and JSON formats using custom scripts in Python 3.8 with libraries: pandas for data manipulation, heartpy for ECG signal processing, and neurokit2 for EDA signal processing. All analyses supporting technical validation of the dataset were performed using statsmodels and bioinfokit libraries for