情感计算
此词条暂由彩云小译翻译,翻译字数共5068,未经人工整理和审校,带来阅读不便,请见谅。
Affective computing is the study and development of systems and devices that can recognize, interpret, process, and simulate human affects. It is an interdisciplinary field spanning computer science, psychology, and cognitive science.[1] While some core ideas in the field may be traced as far back as to early philosophical inquiries into emotion,[2] the more modern branch of computer science originated with Rosalind Picard's 1995 paper[3] on affective computing and her book Affective Computing[4] published by MIT Press.[5][6] One of the motivations for the research is the ability to give machines emotional intelligence, including to simulate empathy. The machine should interpret the emotional state of humans and adapt its behavior to them, giving an appropriate response to those emotions.
Affective computing is the study and development of systems and devices that can recognize, interpret, process, and simulate human affects. It is an interdisciplinary field spanning computer science, psychology, and cognitive science. While some core ideas in the field may be traced as far back as to early philosophical inquiries into emotion, Cited by Tao and Tan. the more modern branch of computer science originated with Rosalind Picard's 1995 paper"Affective Computing" MIT Technical Report #321 (Abstract), 1995 on affective computing and her book Affective Computing published by MIT Press.
One of the motivations for the research is the ability to give machines emotional intelligence, including to simulate empathy. The machine should interpret the emotional state of humans and adapt its behavior to them, giving an appropriate response to those emotions.
情感计算是研究和开发能够识别、解释、处理和模拟人类情感的系统和设备。这是一个跨越计算机科学、心理学和认知科学的跨学科领域。而这一领域的一些核心思想可以追溯到早期的情感哲学研究,陶和谭引用了这些思想。更现代的计算机科学分支起源于 Rosalind Picard 1995年的论文“ Affective Computing”MIT 技术报告 # 321(摘要) ,1995年关于 Affective Computing 和她由 MIT 出版社出版的书 Affective Computing。这项研究的动机之一是赋予机器情商的能力,包括模拟移情。机器应该解读人类的情绪状态,并使其行为适应人类的情绪,对这些情绪作出适当的反应。
Areas
Areas
= 面积 =
Detecting and recognizing emotional information
Detecting emotional information usually begins with passive sensors that capture data about the user's physical state or behavior without interpreting the input. The data gathered is analogous to the cues humans use to perceive emotions in others. For example, a video camera might capture facial expressions, body posture, and gestures, while a microphone might capture speech. Other sensors detect emotional cues by directly measuring physiological data, such as skin temperature and galvanic resistance.[7]
Detecting emotional information usually begins with passive sensors that capture data about the user's physical state or behavior without interpreting the input. The data gathered is analogous to the cues humans use to perceive emotions in others. For example, a video camera might capture facial expressions, body posture, and gestures, while a microphone might capture speech. Other sensors detect emotional cues by directly measuring physiological data, such as skin temperature and galvanic resistance.
检测情感信息通常从被动传感器开始,这些传感器捕捉关于用户身体状态或行为的数据,而不解释输入信息。收集的数据类似于人类用来感知他人情感的线索。例如,摄像机可以捕捉面部表情、身体姿势和手势,而麦克风可以捕捉语音。其他传感器通过直接测量生理数据(如皮肤温度和电流电阻)来探测情感信号。
Recognizing emotional information requires the extraction of meaningful patterns from the gathered data. This is done using machine learning techniques that process different modalities, such as speech recognition, natural language processing, or facial expression detection. The goal of most of these techniques is to produce labels that would match the labels a human perceiver would give in the same situation: For example, if a person makes a facial expression furrowing their brow, then the computer vision system might be taught to label their face as appearing "confused" or as "concentrating" or "slightly negative" (as opposed to positive, which it might say if they were smiling in a happy-appearing way). These labels may or may not correspond to what the person is actually feeling.
Recognizing emotional information requires the extraction of meaningful patterns from the gathered data. This is done using machine learning techniques that process different modalities, such as speech recognition, natural language processing, or facial expression detection. The goal of most of these techniques is to produce labels that would match the labels a human perceiver would give in the same situation: For example, if a person makes a facial expression furrowing their brow, then the computer vision system might be taught to label their face as appearing "confused" or as "concentrating" or "slightly negative" (as opposed to positive, which it might say if they were smiling in a happy-appearing way). These labels may or may not correspond to what the person is actually feeling.
识别情感信息需要从收集到的数据中提取出有意义的模式。这是使用机器学习技术,处理不同的模式,如语音识别,自然语言处理,或面部表情检测。大多数这些技术的目标是产生与人类感知者在相同情况下给出的标签相匹配的标签: 例如,如果一个人做出皱眉的面部表情,那么计算机视觉系统可能会被教导将他们的脸标记为看起来“困惑”、“专注”或“轻微消极”(与积极相反,它可能会说,如果他们正在以一种快乐的方式微笑)。这些标签可能与人们的真实感受相符,也可能不相符。
Emotion in machines
Another area within affective computing is the design of computational devices proposed to exhibit either innate emotional capabilities or that are capable of convincingly simulating emotions. A more practical approach, based on current technological capabilities, is the simulation of emotions in conversational agents in order to enrich and facilitate interactivity between human and machine.[8]
Another area within affective computing is the design of computational devices proposed to exhibit either innate emotional capabilities or that are capable of convincingly simulating emotions. A more practical approach, based on current technological capabilities, is the simulation of emotions in conversational agents in order to enrich and facilitate interactivity between human and machine.
情感计算的另一个领域是计算机设备的设计,提出要么展示先天的情感能力或能够令人信服地模拟情绪。基于当前的技术能力,一个更加实用的方法是模拟会话代理中的情绪,以丰富和促进人与机器之间的互动。
Marvin Minsky, one of the pioneering computer scientists in artificial intelligence, relates emotions to the broader issues of machine intelligence stating in The Emotion Machine that emotion is "not especially different from the processes that we call 'thinking.'"[9]
Marvin Minsky, one of the pioneering computer scientists in artificial intelligence, relates emotions to the broader issues of machine intelligence stating in The Emotion Machine that emotion is "not especially different from the processes that we call 'thinking.'"
人工智能领域的计算机科学先驱之一马文•明斯基(Marvin Minsky)在《情绪机器》(The Emotion Machine)一书中将情绪与更广泛的机器智能问题联系起来。他在书中表示,情绪“与我们所谓的‘思考’过程并没有特别的不同。'"
Technologies
In psychology, cognitive science, and in neuroscience, there have been two main approaches for describing how humans perceive and classify emotion: continuous or categorical. The continuous approach tends to use dimensions such as negative vs. positive, calm vs. aroused.
In psychology, cognitive science, and in neuroscience, there have been two main approaches for describing how humans perceive and classify emotion: continuous or categorical. The continuous approach tends to use dimensions such as negative vs. positive, calm vs. aroused.
在心理学、认知科学和神经科学中,描述人类如何感知和分类情绪的方法主要有两种: 连续的和分类的。持续的方法倾向于使用诸如消极与积极、平静与被唤醒之类的维度。
The categorical approach tends to use discrete classes such as happy, sad, angry, fearful, surprise, disgust. Different kinds of machine learning regression and classification models can be used for having machines produce continuous or discrete labels. Sometimes models are also built that allow combinations across the categories, e.g. a happy-surprised face or a fearful-surprised face.[10]
The categorical approach tends to use discrete classes such as happy, sad, angry, fearful, surprise, disgust. Different kinds of machine learning regression and classification models can be used for having machines produce continuous or discrete labels. Sometimes models are also built that allow combinations across the categories, e.g. a happy-surprised face or a fearful-surprised face.
绝对类方法倾向于使用分离的类,比如高兴、悲伤、愤怒、恐惧、惊讶、厌恶。不同类型的机器学习回归和分类模型可以用于让机器产生连续或离散的标签。有时建立的模型也允许跨类别的组合,例如。一张惊喜的脸,还是一张惊恐的脸。
The following sections consider many of the kinds of input data used for the task of emotion recognition.
The following sections consider many of the kinds of input data used for the task of emotion recognition.
接下来的部分将讨论用于情感识别任务的各种输入数据。
Emotional speech
Various changes in the autonomic nervous system can indirectly alter a person's speech, and affective technologies can leverage this information to recognize emotion. For example, speech produced in a state of fear, anger, or joy becomes fast, loud, and precisely enunciated, with a higher and wider range in pitch, whereas emotions such as tiredness, boredom, or sadness tend to generate slow, low-pitched, and slurred speech.[11] Some emotions have been found to be more easily computationally identified, such as anger[12] or approval.[13]
Various changes in the autonomic nervous system can indirectly alter a person's speech, and affective technologies can leverage this information to recognize emotion. For example, speech produced in a state of fear, anger, or joy becomes fast, loud, and precisely enunciated, with a higher and wider range in pitch, whereas emotions such as tiredness, boredom, or sadness tend to generate slow, low-pitched, and slurred speech.Breazeal, C. and Aryananda, L. Recognition of affective communicative intent in robot-directed speech. Autonomous Robots 12 1, 2002. pp. 83–104. Some emotions have been found to be more easily computationally identified, such as anger or approval.
自主神经系统的各种变化可以间接地改变一个人的语言,情感技术可以利用这些信息来识别情绪。例如,在恐惧、愤怒或高兴的状态下发言变得快速、响亮、清晰,音调变得越来越高、越来越宽,而诸如疲倦、厌倦或悲伤等情绪往往会产生缓慢、低沉、含糊不清的发言。机器人导向语音中情感交流意图的识别。Autonomous Robots 12 1, 2002. pp.83–104.一些情绪被发现更容易被计算识别,比如愤怒或认可。
Emotional speech processing technologies recognize the user's emotional state using computational analysis of speech features. Vocal parameters and prosodic features such as pitch variables and speech rate can be analyzed through pattern recognition techniques.[12][14]
Emotional speech processing technologies recognize the user's emotional state using computational analysis of speech features. Vocal parameters and prosodic features such as pitch variables and speech rate can be analyzed through pattern recognition techniques.Dellaert, F., Polizin, t., and Waibel, A., Recognizing Emotion in Speech", In Proc. Of ICSLP 1996, Philadelphia, PA, pp.1970–1973, 1996Lee, C.M.; Narayanan, S.; Pieraccini, R., Recognition of Negative Emotion in the Human Speech Signals, Workshop on Auto. Speech Recognition and Understanding, Dec 2001
情感语音处理技术通过对语音特征的计算分析来识别用户的情感状态。通过模式识别技术可以分析声音参数和韵律特征,如音高变量和语速等。和 Waibel,a,Recognizing Emotion In Speech”,In Proc。1996,Philadelphia,PA,pp. 1970-1973,1996 lee,c.m.; Narayanan,s. ; Pieraccini,r. ,《人类语音信号中负面情绪的识别》 ,《汽车工作室》。语音识别与理解,2001年12月
Speech analysis is an effective method of identifying affective state, having an average reported accuracy of 70 to 80% in recent research.[15][16] These systems tend to outperform average human accuracy (approximately 60%[12]) but are less accurate than systems which employ other modalities for emotion detection, such as physiological states or facial expressions.[17] However, since many speech characteristics are independent of semantics or culture, this technique is considered to be a promising route for further research.[18]
Speech analysis is an effective method of identifying affective state, having an average reported accuracy of 70 to 80% in recent research. These systems tend to outperform average human accuracy (approximately 60%) but are less accurate than systems which employ other modalities for emotion detection, such as physiological states or facial expressions. However, since many speech characteristics are independent of semantics or culture, this technique is considered to be a promising route for further research.
语音分析是一种有效的情感状态识别方法,在最近的研究中,语音分析的平均报告准确率为70%-80% 。这些系统往往比人类的平均准确率(大约60%)更高,但是不如使用其他情绪检测方式的系统准确,比如生理状态或面部表情。然而,由于许多言语特征是独立于语义或文化的,这种技术被认为是一个很有前途的进一步研究路线。
Algorithms
Algorithms
= = 算法 = =
The process of speech/text affect detection requires the creation of a reliable database, knowledge base, or vector space model,[19] broad enough to fit every need for its application, as well as the selection of a successful classifier which will allow for quick and accurate emotion identification.
The process of speech/text affect detection requires the creation of a reliable database, knowledge base, or vector space model,
broad enough to fit every need for its application, as well as the selection of a successful classifier which will allow for quick and accurate emotion identification.
语音/文本影响检测的过程需要创建一个可靠的数据库、知识库或者向量空间模型数据库,这些数据库的范围足以满足其应用的所有需要,同时还需要选择一个成功的分类器,这样才能快速准确地识别情感。
Currently, the most frequently used classifiers are linear discriminant classifiers (LDC), k-nearest neighbor (k-NN), Gaussian mixture model (GMM), support vector machines (SVM), artificial neural networks (ANN), decision tree algorithms and hidden Markov models (HMMs).[20] Various studies showed that choosing the appropriate classifier can significantly enhance the overall performance of the system.[17] The list below gives a brief description of each algorithm:
Currently, the most frequently used classifiers are linear discriminant classifiers (LDC), k-nearest neighbor (k-NN), Gaussian mixture model (GMM), support vector machines (SVM), artificial neural networks (ANN), decision tree algorithms and hidden Markov models (HMMs). Various studies showed that choosing the appropriate classifier can significantly enhance the overall performance of the system. The list below gives a brief description of each algorithm:
目前常用的分类器有线性判别分类器(LDC)、 k- 近邻分类器(k-NN)、高斯混合模型(GMM)、支持向量机(SVM)、人工神经网络(ANN)、决策树算法和隐马尔可夫模型(HMMs)。各种研究表明,选择合适的分类器可以显著提高系统的整体性能。下面的列表给出了每个算法的简要描述:
- LDC – Classification happens based on the value obtained from the linear combination of the feature values, which are usually provided in the form of vector features.
- k-NN – Classification happens by locating the object in the feature space, and comparing it with the k nearest neighbors (training examples). The majority vote decides on the classification.
- GMM – is a probabilistic model used for representing the existence of subpopulations within the overall population. Each sub-population is described using the mixture distribution, which allows for classification of observations into the sub-populations.[21]
- SVM – is a type of (usually binary) linear classifier which decides in which of the two (or more) possible classes, each input may fall into.
- ANN – is a mathematical model, inspired by biological neural networks, that can better grasp possible non-linearities of the feature space.
- Decision tree algorithms – work based on following a decision tree in which leaves represent the classification outcome, and branches represent the conjunction of subsequent features that lead to the classification.
- HMMs – a statistical Markov model in which the states and state transitions are not directly available to observation. Instead, the series of outputs dependent on the states are visible. In the case of affect recognition, the outputs represent the sequence of speech feature vectors, which allow the deduction of states' sequences through which the model progressed. The states can consist of various intermediate steps in the expression of an emotion, and each of them has a probability distribution over the possible output vectors. The states' sequences allow us to predict the affective state which we are trying to classify, and this is one of the most commonly used techniques within the area of speech affect detection.
- LDC – Classification happens based on the value obtained from the linear combination of the feature values, which are usually provided in the form of vector features.
- k-NN – Classification happens by locating the object in the feature space, and comparing it with the k nearest neighbors (training examples). The majority vote decides on the classification.
- GMM – is a probabilistic model used for representing the existence of subpopulations within the overall population. Each sub-population is described using the mixture distribution, which allows for classification of observations into the sub-populations."Gaussian Mixture Model". Connexions – Sharing Knowledge and Building Communities. Retrieved 10 March 2011.
- SVM – is a type of (usually binary) linear classifier which decides in which of the two (or more) possible classes, each input may fall into.
- ANN – is a mathematical model, inspired by biological neural networks, that can better grasp possible non-linearities of the feature space.
- Decision tree algorithms – work based on following a decision tree in which leaves represent the classification outcome, and branches represent the conjunction of subsequent features that lead to the classification.
- HMMs – a statistical Markov model in which the states and state transitions are not directly available to observation. Instead, the series of outputs dependent on the states are visible. In the case of affect recognition, the outputs represent the sequence of speech feature vectors, which allow the deduction of states' sequences through which the model progressed. The states can consist of various intermediate steps in the expression of an emotion, and each of them has a probability distribution over the possible output vectors. The states' sequences allow us to predict the affective state which we are trying to classify, and this is one of the most commonly used techniques within the area of speech affect detection.
- LDC-根据特征值的线性组合值进行分类,特征值通常以矢量特征的形式提供。
- k-NN-分类是通过在特征空间中定位目标,并与 k 个最近邻(训练样本)进行比较来实现的。多数票决定分类。
- GMM-是一个概率模型,用于表示总体种群中存在的子种群。每个子种群使用混合分布来描述,这允许将观测分类为子种群。“高斯混合模型”。知识共享与社区建设。10 March 2011.
- SVM-是一种(通常为二进制)线性分类器,它决定每个输入可能属于两个(或多个)可能类别中的哪一个。人工神经网络是一种受生物神经网络启发的数学模型,能够更好地把握特征空间可能存在的非线性。
- 决策树算法——基于下面的决策树,其中的叶子代表分类结果,分支代表导致分类的后续特征之间的关联。
- HMMs-一个统计马尔可夫模型,其中的状态和状态转变不能直接用于观测。相反,依赖于状态的一系列输出是可见的。在情感识别的情况下,输出表示语音特征向量的序列,这样可以推导出模型所经过的状态序列。这些状态可以由表达情绪的各种中间步骤组成,每个概率分布都有一个可能的输出向量。状态序列允许我们预测我们试图分类的情感状态,这是语音情感检测领域最常用的技术之一。
It is proved that having enough acoustic evidence available the emotional state of a person can be classified by a set of majority voting classifiers. The proposed set of classifiers is based on three main classifiers: kNN, C4.5 and SVM-RBF Kernel. This set achieves better performance than each basic classifier taken separately. It is compared with two other sets of classifiers: one-against-all (OAA) multiclass SVM with Hybrid kernels and the set of classifiers which consists of the following two basic classifiers: C5.0 and Neural Network. The proposed variant achieves better performance than the other two sets of classifiers.[22]
It is proved that having enough acoustic evidence available the emotional state of a person can be classified by a set of majority voting classifiers. The proposed set of classifiers is based on three main classifiers: kNN, C4.5 and SVM-RBF Kernel. This set achieves better performance than each basic classifier taken separately. It is compared with two other sets of classifiers: one-against-all (OAA) multiclass SVM with Hybrid kernels and the set of classifiers which consists of the following two basic classifiers: C5.0 and Neural Network. The proposed variant achieves better performance than the other two sets of classifiers.
证明了一组多数投票分类器可以用足够的声学证据对人的情绪状态进行分类。该分类器集合基于三个主要分类器: kNN、 C4.5和 SVM-RBF 核。该分类器比单独采集的基本分类器具有更好的分类性能。本文提出了一种基于混合核函数的多类支持向量机,并将其与另外两类支持向量机进行了比较。与其他两组分类器相比,该方法具有更好的分类性能。
Databases
Databases
= = = 数据库 = =
The vast majority of present systems are data-dependent. This creates one of the biggest challenges in detecting emotions based on speech, as it implicates choosing an appropriate database used to train the classifier. Most of the currently possessed data was obtained from actors and is thus a representation of archetypal emotions. Those so-called acted databases are usually based on the Basic Emotions theory (by Paul Ekman), which assumes the existence of six basic emotions (anger, fear, disgust, surprise, joy, sadness), the others simply being a mix of the former ones.[23] Nevertheless, these still offer high audio quality and balanced classes (although often too few), which contribute to high success rates in recognizing emotions.
The vast majority of present systems are data-dependent. This creates one of the biggest challenges in detecting emotions based on speech, as it implicates choosing an appropriate database used to train the classifier. Most of the currently possessed data was obtained from actors and is thus a representation of archetypal emotions. Those so-called acted databases are usually based on the Basic Emotions theory (by Paul Ekman), which assumes the existence of six basic emotions (anger, fear, disgust, surprise, joy, sadness), the others simply being a mix of the former ones.Ekman, P. & Friesen, W. V (1969). The repertoire of nonverbal behavior: Categories, origins, usage, and coding. Semiotica, 1, 49–98. Nevertheless, these still offer high audio quality and balanced classes (although often too few), which contribute to high success rates in recognizing emotions.
目前绝大多数系统都依赖于数据。这在基于语音的情绪检测中创造了一个最大的挑战,因为它牵涉到选择一个合适的数据库用于训练分类器。目前拥有的大多数数据是从演员,因此是一个代表的原型情绪。这些所谓的行为数据库通常是基于基本情绪理论(保罗 · 埃克曼) ,该理论假定存在六种基本情绪(愤怒、恐惧、厌恶、惊讶、喜悦、悲伤) ,其他情绪只是前者的混合体。埃克曼,p. & Friesen,w. v (1969)。非语言行为的全部: 分类、起源、用法和编码。Semiotica,1,49-98.然而,这些仍然提供高音质和平衡的课程(虽然通常太少) ,这有助于高成功率识别情绪。
However, for real life application, naturalistic data is preferred. A naturalistic database can be produced by observation and analysis of subjects in their natural context. Ultimately, such database should allow the system to recognize emotions based on their context as well as work out the goals and outcomes of the interaction. The nature of this type of data allows for authentic real life implementation, due to the fact it describes states naturally occurring during the human–computer interaction (HCI).
However, for real life application, naturalistic data is preferred. A naturalistic database can be produced by observation and analysis of subjects in their natural context. Ultimately, such database should allow the system to recognize emotions based on their context as well as work out the goals and outcomes of the interaction. The nature of this type of data allows for authentic real life implementation, due to the fact it describes states naturally occurring during the human–computer interaction (HCI).
然而,对于现实生活应用,自然数据是首选的。一个自然主义的数据库可以通过观察和分析的主题在他们的自然环境。最终,这样的数据库应该允许系统根据情绪的背景识别情绪,并制定出交互的目标和结果。这种类型的数据的性质允许真实的实现,因为它描述了在人机交互(HCI)过程中自然发生的状态。
Despite the numerous advantages which naturalistic data has over acted data, it is difficult to obtain and usually has low emotional intensity. Moreover, data obtained in a natural context has lower signal quality, due to surroundings noise and distance of the subjects from the microphone. The first attempt to produce such database was the FAU Aibo Emotion Corpus for CEICES (Combining Efforts for Improving Automatic Classification of Emotional User States), which was developed based on a realistic context of children (age 10–13) playing with Sony's Aibo robot pet.[24][25] Likewise, producing one standard database for all emotional research would provide a method of evaluating and comparing different affect recognition systems.
Despite the numerous advantages which naturalistic data has over acted data, it is difficult to obtain and usually has low emotional intensity. Moreover, data obtained in a natural context has lower signal quality, due to surroundings noise and distance of the subjects from the microphone. The first attempt to produce such database was the FAU Aibo Emotion Corpus for CEICES (Combining Efforts for Improving Automatic Classification of Emotional User States), which was developed based on a realistic context of children (age 10–13) playing with Sony's Aibo robot pet. Likewise, producing one standard database for all emotional research would provide a method of evaluating and comparing different affect recognition systems.
尽管自然主义数据比实际数据有许多优点,但它很难获得,而且通常情绪强度较低。此外,在自然环境下获得的数据信号质量较低,这是由于周围环境的噪声和被试者距离麦克风的距离。第一个尝试建立这样的数据库的是 FAU Aibo 情感语料库,该语料库是基于一个真实的儿童(10-13岁)与索尼的 Aibo 机器人宠物玩耍的环境而开发的。同样地,为所有情感研究建立一个标准的数据库将提供一种评估和比较不同情感识别系统的方法。
Speech descriptors
Speech descriptors
= = 语言描述符 = =
The complexity of the affect recognition process increases with the number of classes (affects) and speech descriptors used within the classifier. It is, therefore, crucial to select only the most relevant features in order to assure the ability of the model to successfully identify emotions, as well as increasing the performance, which is particularly significant to real-time detection. The range of possible choices is vast, with some studies mentioning the use of over 200 distinct features.[20] It is crucial to identify those that are redundant and undesirable in order to optimize the system and increase the success rate of correct emotion detection. The most common speech characteristics are categorized into the following groups.[24][25]
The complexity of the affect recognition process increases with the number of classes (affects) and speech descriptors used within the classifier. It is, therefore, crucial to select only the most relevant features in order to assure the ability of the model to successfully identify emotions, as well as increasing the performance, which is particularly significant to real-time detection. The range of possible choices is vast, with some studies mentioning the use of over 200 distinct features. It is crucial to identify those that are redundant and undesirable in order to optimize the system and increase the success rate of correct emotion detection. The most common speech characteristics are categorized into the following groups.
情感识别过程的复杂性随着分类器中使用的类(情感)和语音描述符的数量的增加而增加。因此,为了保证模型能够成功地识别情绪,并提高性能,只选择最相关的特征是至关重要的,这对于实时检测尤为重要。可能的选择范围很广,有些研究提到使用了200多种不同的特征。识别冗余和不需要的情感信息对于优化系统、提高情感检测的成功率至关重要。最常见的言语特征可分为以下几类。
- Frequency characteristics[26]
- Accent shape – affected by the rate of change of the fundamental frequency.
- Average pitch – description of how high/low the speaker speaks relative to the normal speech.
- Contour slope – describes the tendency of the frequency change over time, it can be rising, falling or level.
- Final lowering – the amount by which the frequency falls at the end of an utterance.
- Pitch range – measures the spread between the maximum and minimum frequency of an utterance.
- Time-related features:
- Speech rate – describes the rate of words or syllables uttered over a unit of time
- Stress frequency – measures the rate of occurrences of pitch accented utterances
- Voice quality parameters and energy descriptors:
- Breathiness – measures the aspiration noise in speech
- Brilliance – describes the dominance of high Or low frequencies In the speech
- Loudness – measures the amplitude of the speech waveform, translates to the energy of an utterance
- Pause Discontinuity – describes the transitions between sound and silence
- Pitch Discontinuity – describes the transitions of the fundamental frequency.
- Frequency characteristics
- Accent shape – affected by the rate of change of the fundamental frequency.
- Average pitch – description of how high/low the speaker speaks relative to the normal speech.
- Contour slope – describes the tendency of the frequency change over time, it can be rising, falling or level.
- Final lowering – the amount by which the frequency falls at the end of an utterance.
- Pitch range – measures the spread between the maximum and minimum frequency of an utterance.
- Time-related features:
- Speech rate – describes the rate of words or syllables uttered over a unit of time
- Stress frequency – measures the rate of occurrences of pitch accented utterances
- Voice quality parameters and energy descriptors:
- Breathiness – measures the aspiration noise in speech
- Brilliance – describes the dominance of high Or low frequencies In the speech
- Loudness – measures the amplitude of the speech waveform, translates to the energy of an utterance
- Pause Discontinuity – describes the transitions between sound and silence
- Pitch Discontinuity – describes the transitions of the fundamental frequency.
- 频率特性 #
- 重音形状——受基频变化率的影响。
- 平均音调-描述说话者相对于正常语言的音调高低。#
- 等高线斜率-描述频率随时间变化的趋势,可以是上升、下降或水平。#
- 最后降低频率——话语结束时频率下降的幅度。#
- 音高范围-量度一段话语的最高和最低频率之间的差距。# 与时间相关的特征: #
- 语速-描述在一个时间单位内发出的单词或音节的频率 #
- 重音频率-测量出现带有沥青口音的发音的频率 # 语音质量参数和能量描述符: #
- 呼吸质-测量语音中的吸气噪声 #
- 辉度-描述语音中高频或低频的主导地位 #
- 响度-测量语音的振幅,转换为话音的能量 #
- 暂停间断-描述声音和静音之间的转换 #
- 音高间断-描述基本频率的转换。
Facial affect detection
The detection and processing of facial expression are achieved through various methods such as optical flow, hidden Markov models, neural network processing or active appearance models. More than one modalities can be combined or fused (multimodal recognition, e.g. facial expressions and speech prosody,[27] facial expressions and hand gestures,[28] or facial expressions with speech and text for multimodal data and metadata analysis) to provide a more robust estimation of the subject's emotional state. Affectiva is a company (co-founded by Rosalind Picard and Rana El Kaliouby) directly related to affective computing and aims at investigating solutions and software for facial affect detection.
The detection and processing of facial expression are achieved through various methods such as optical flow, hidden Markov models, neural network processing or active appearance models. More than one modalities can be combined or fused (multimodal recognition, e.g. facial expressions and speech prosody, facial expressions and hand gestures, or facial expressions with speech and text for multimodal data and metadata analysis) to provide a more robust estimation of the subject's emotional state. Affectiva is a company (co-founded by Rosalind Picard and Rana El Kaliouby) directly related to affective computing and aims at investigating solutions and software for facial affect detection.
面部表情的检测和处理通过光流、隐马尔可夫模型、神经网络处理或主动外观模型等多种方法实现。多模式识别可以组合或融合多种模式。面部表情和语音韵律,面部表情和手势,或面部表情与语音和文本的多模态数据和元数据分析) ,以提供一个更稳健的估计主题的情绪状态。Affectiva 是一家与情感计算直接相关的公司(由 Rosalind Picard 和 Rana El Kaliouby 共同创办) ,旨在研究面部情感检测的解决方案和软件。
Facial expression databases
Creation of an emotion database is a difficult and time-consuming task. However, database creation is an essential step in the creation of a system that will recognize human emotions. Most of the publicly available emotion databases include posed facial expressions only. In posed expression databases, the participants are asked to display different basic emotional expressions, while in spontaneous expression database, the expressions are natural. Spontaneous emotion elicitation requires significant effort in the selection of proper stimuli which can lead to a rich display of intended emotions. Secondly, the process involves tagging of emotions by trained individuals manually which makes the databases highly reliable. Since perception of expressions and their intensity is subjective in nature, the annotation by experts is essential for the purpose of validation.
Creation of an emotion database is a difficult and time-consuming task. However, database creation is an essential step in the creation of a system that will recognize human emotions. Most of the publicly available emotion databases include posed facial expressions only. In posed expression databases, the participants are asked to display different basic emotional expressions, while in spontaneous expression database, the expressions are natural. Spontaneous emotion elicitation requires significant effort in the selection of proper stimuli which can lead to a rich display of intended emotions. Secondly, the process involves tagging of emotions by trained individuals manually which makes the databases highly reliable. Since perception of expressions and their intensity is subjective in nature, the annotation by experts is essential for the purpose of validation.
情感数据库的建立是一项既困难又耗时的工作。然而,创建数据库是创建识别人类情感的系统的关键步骤。大多数公开的情感数据库只包含摆出的面部表情。在提出的表达式数据库中,要求参与者显示不同的基本情感表达,而在自发表达式数据库中,表达是自然的。自发的情绪诱导需要在选择合适的刺激物时付出巨大的努力,这会导致丰富的预期情绪的展示。其次,这个过程包括由训练有素的人手动标记情绪,使数据库高度可靠。由于对表达式及其强度的感知在本质上是主观的,专家的注释对于验证是必不可少的。
Researchers work with three types of databases, such as a database of peak expression images only, a database of image sequences portraying an emotion from neutral to its peak, and video clips with emotional annotations. Many facial expression databases have been created and made public for expression recognition purpose. Two of the widely used databases are CK+ and JAFFE.
Researchers work with three types of databases, such as a database of peak expression images only, a database of image sequences portraying an emotion from neutral to its peak, and video clips with emotional annotations. Many facial expression databases have been created and made public for expression recognition purpose. Two of the widely used databases are CK+ and JAFFE.
研究人员使用三种类型的数据库进行研究,比如只使用峰值表情图像的数据库,描绘从中性到峰值的情感的图像序列数据库,以及带有情感注释的视频剪辑。面部表情数据库是面部表情识别领域的一个重要研究课题。两个广泛使用的数据库是 CK + 和 JAFFE。
Emotion classification
By doing cross-cultural research in Papua New Guinea, on the Fore Tribesmen, at the end of the 1960s, Paul Ekman proposed the idea that facial expressions of emotion are not culturally determined, but universal. Thus, he suggested that they are biological in origin and can, therefore, be safely and correctly categorized.[23] He therefore officially put forth six basic emotions, in 1972:[29]
By doing cross-cultural research in Papua New Guinea, on the Fore Tribesmen, at the end of the 1960s, Paul Ekman proposed the idea that facial expressions of emotion are not culturally determined, but universal. Thus, he suggested that they are biological in origin and can, therefore, be safely and correctly categorized.
He therefore officially put forth six basic emotions, in 1972:
20世纪60年代末,Paul Ekman 在福尔部落做了跨文化研究,提出面部表情不是由文化决定的,而是普遍存在的巴布亚新几内亚。因此,他认为它们是生物起源的,因此可以安全而正确地归类。因此,他在1972年正式提出了六种基本情感:
- Anger
- Disgust
- Fear
- Happiness
- Sadness
- Surprise
- 愤怒
- 厌恶
- 恐惧
- 快乐
- 悲伤
- 惊喜
However, in the 1990s Ekman expanded his list of basic emotions, including a range of positive and negative emotions not all of which are encoded in facial muscles.[30] The newly included emotions are:
- Amusement
- Contempt
- Contentment
- Embarrassment
- Excitement
- Guilt
- Pride in achievement
- Relief
- Satisfaction
- Sensory pleasure
- Shame
However, in the 1990s Ekman expanded his list of basic emotions, including a range of positive and negative emotions not all of which are encoded in facial muscles.. The newly included emotions are:
- Amusement
- Contempt
- Contentment
- Embarrassment
- Excitement
- Guilt
- Pride in achievement
- Relief
- Satisfaction
- Sensory pleasure
- Shame
然而,在20世纪90年代,埃克曼扩展了他的基本情绪列表,包括一系列积极和消极的情绪,这些情绪并非都编码在面部肌肉中。.新增的情绪是: # 娱乐 # 轻蔑 # 满足 # 尴尬 # 兴奋 # 内疚 # 成就骄傲 # 解脱 # 满足 # 感官愉悦 # 羞耻
Facial Action Coding System
A system has been conceived by psychologists in order to formally categorize the physical expression of emotions on faces. The central concept of the Facial Action Coding System, or FACS, as created by Paul Ekman and Wallace V. Friesen in 1978 based on earlier work by Carl-Herman Hjortsjö[31] are action units (AU). They are, basically, a contraction or a relaxation of one or more muscles. Psychologists have proposed the following classification of six basic emotions, according to their action units ("+" here mean "and"):
A system has been conceived by psychologists in order to formally categorize the physical expression of emotions on faces. The central concept of the Facial Action Coding System, or FACS, as created by Paul Ekman and Wallace V. Friesen in 1978 based on earlier work by Carl-Herman Hjortsjö"Facial Action Coding System (FACS) and the FACS Manual" . A Human Face. Retrieved 21 March 2011. are action units (AU). They are, basically, a contraction or a relaxation of one or more muscles. Psychologists have proposed the following classification of six basic emotions, according to their action units ("+" here mean "and"):
心理学家已经构想出一个系统,用来正式分类脸上情绪的物理表达。1978年,Paul Ekman 和 Wallace v. Friesen 根据 Carl-Herman hjortsjö 的早期作品《面部动作编码系统面部动作编码系统和 FACS 手册》创建了 FACS 的中心概念。一张人类的脸。检索2011年3月21日. 是行动单位(AU) 。它们基本上是一块或多块肌肉的收缩或放松。心理学家根据他们的行为单位,提出了以下六种基本情绪的分类(这里的“ +”是指“和”) :
Emotion | Action units |
---|---|
Happiness | 6+12 |
Sadness | 1+4+15 |
Surprise | 1+2+5B+26 |
Fear | 1+2+4+5+20+26 |
Anger | 4+5+7+23 |
Disgust | 9+15+16 |
Contempt | R12A+R14A |
Emotion | Action units |
---|---|
Happiness | 6+12 |
Sadness | 1+4+15 |
Surprise | 1+2+5B+26 |
Fear | 1+2+4+5+20+26 |
Anger | 4+5+7+23 |
Disgust | 9+15+16 |
Contempt | R12A+R14A |
-| 快乐 | | 6 + 12 |-| 悲伤 | | 1 + 4 + 15 |-| 惊喜 | | 1 + 2 + 5 b + 26 |-| 恐惧 | | 1 + 2 + 4 + 5 + 20 + 26 |-愤怒 | 4 + 5 + 7 + 23 |-| 厌恶 | 9 + 15 + 16 |-| 藐视 | R12A + R14A | }
Challenges in facial detectionAs with every computational practice, in affect detection by facial processing, some obstacles need to be surpassed, in order to fully unlock the hidden potential of the overall algorithm or method employed. In the early days of almost every kind of AI-based detection (speech recognition, face recognition, affect recognition), the accuracy of modeling and tracking has been an issue. As hardware evolves, as more data are collected and as new discoveries are made and new practices introduced, this lack of accuracy fades, leaving behind noise issues. However, methods for noise removal exist including neighborhood averaging, linear Gaussian smoothing, median filtering,[32] or newer methods such as the Bacterial Foraging Optimization Algorithm.[33][34] As with every computational practice, in affect detection by facial processing, some obstacles need to be surpassed, in order to fully unlock the hidden potential of the overall algorithm or method employed. In the early days of almost every kind of AI-based detection (speech recognition, face recognition, affect recognition), the accuracy of modeling and tracking has been an issue. As hardware evolves, as more data are collected and as new discoveries are made and new practices introduced, this lack of accuracy fades, leaving behind noise issues. However, methods for noise removal exist including neighborhood averaging, linear Gaussian smoothing, median filtering, or newer methods such as the Bacterial Foraging Optimization Algorithm.Clever Algorithms. "Bacterial Foraging Optimization Algorithm – Swarm Algorithms – Clever Algorithms" . Clever Algorithms. Retrieved 21 March 2011."Soft Computing". Soft Computing. Retrieved 18 March 2011. 正如每一个计算实践,在人脸处理的情感检测中,一些障碍需要被超越,以便充分释放所使用的整体算法或方法的隐藏潜力。在几乎所有基于人工智能的检测(语音识别、人脸识别、情感识别)的早期,建模和跟踪的准确性一直是个问题。随着硬件的发展,随着更多的数据被收集,随着新的发现和新的实践的引入,这种缺乏准确性的现象逐渐消失,留下了噪音问题。然而,现有的去噪方法包括邻域平均法、线性高斯平滑法、中值滤波法,或者更新的方法如细菌觅食优化算法。聪明的算法。“细菌觅食优化算法-群算法-巧妙算法”。聪明的算法。2011年3月21日。「软电脑」。软计算。2011年3月18日。 Other challenges include
Other challenges include
其他的挑战包括
Body gestureGestures could be efficiently used as a means of detecting a particular emotional state of the user, especially when used in conjunction with speech and face recognition. Depending on the specific action, gestures could be simple reflexive responses, like lifting your shoulders when you don't know the answer to a question, or they could be complex and meaningful as when communicating with sign language. Without making use of any object or surrounding environment, we can wave our hands, clap or beckon. On the other hand, when using objects, we can point at them, move, touch or handle these. A computer should be able to recognize these, analyze the context and respond in a meaningful way, in order to be efficiently used for Human–Computer Interaction.
手势可以有效地作为一种检测用户特定情绪状态的手段,特别是与语音和面部识别结合使用时。根据具体的动作,手势可以是简单的反射性反应,比如当你不知道一个问题的答案时抬起你的肩膀,或者它们可以是复杂和有意义的,比如当与手语交流时。不需要利用任何物体或周围环境,我们可以挥手、拍手或招手。另一方面,当我们使用物体时,我们可以指向它们,移动,触摸或者处理它们。计算机应该能够识别这些,分析上下文,并以一种有意义的方式作出响应,以便有效地用于人机交互。 There are many proposed methods[36] to detect the body gesture. Some literature differentiates 2 different approaches in gesture recognition: a 3D model based and an appearance-based.[37] The foremost method makes use of 3D information of key elements of the body parts in order to obtain several important parameters, like palm position or joint angles. On the other hand, appearance-based systems use images or videos to for direct interpretation. Hand gestures have been a common focus of body gesture detection methods.[37] There are many proposed methodsJ. K. Aggarwal, Q. Cai, Human Motion Analysis: A Review, Computer Vision and Image Understanding, Vol. 73, No. 3, 1999 to detect the body gesture. Some literature differentiates 2 different approaches in gesture recognition: a 3D model based and an appearance-based. The foremost method makes use of 3D information of key elements of the body parts in order to obtain several important parameters, like palm position or joint angles. On the other hand, appearance-based systems use images or videos to for direct interpretation. Hand gestures have been a common focus of body gesture detection methods. 提出了许多方法。人体运动分析: 评论,计算机视觉与图像理解,第卷。73,No.1999年3月3日,用来探测身体姿势。一些文献将手势识别的2种不同方法进行了区分: 基于3 d 模型和基于外观的。最重要的方法是利用人体关键部位的三维信息,获得手掌位置、关节角度等重要参数。另一方面,基于外观的系统使用图像或视频进行直接解释。手势一直是身体姿态检测方法的共同焦点。 Physiological monitoringThis could be used to detect a user's affective state by monitoring and analyzing their physiological signs. These signs range from changes in heart rate and skin conductance to minute contractions of the facial muscles and changes in facial blood flow. This area is gaining momentum and we are now seeing real products that implement the techniques. The four main physiological signs that are usually analyzed are blood volume pulse, galvanic skin response, facial electromyography, and facial color patterns. This could be used to detect a user's affective state by monitoring and analyzing their physiological signs. These signs range from changes in heart rate and skin conductance to minute contractions of the facial muscles and changes in facial blood flow. This area is gaining momentum and we are now seeing real products that implement the techniques. The four main physiological signs that are usually analyzed are blood volume pulse, galvanic skin response, facial electromyography, and facial color patterns. 这可以用来检测用户的情感状态,通过监测和分析他们的生理信号。这些信号包括心率和皮肤导电反应的变化,面部肌肉的微小收缩和面部血流的变化。这个领域的发展势头越来越强劲,我们现在看到了实现这些技术的真正产品。通常被分析的4个主要生理特征是血容量脉搏、皮肤电反应、面部肌电图和面部颜色模式。 Blood volume pulseBlood volume pulse= = 血容量脉搏 = =OverviewOverview= = 概述 = =A subject's blood volume pulse (BVP) can be measured by a process called photoplethysmography, which produces a graph indicating blood flow through the extremities.[38] The peaks of the waves indicate a cardiac cycle where the heart has pumped blood to the extremities. If the subject experiences fear or is startled, their heart usually 'jumps' and beats quickly for some time, causing the amplitude of the cardiac cycle to increase. This can clearly be seen on a photoplethysmograph when the distance between the trough and the peak of the wave has decreased. As the subject calms down, and as the body's inner core expands, allowing more blood to flow back to the extremities, the cycle will return to normal. A subject's blood volume pulse (BVP) can be measured by a process called photoplethysmography, which produces a graph indicating blood flow through the extremities.Picard, Rosalind (1998). Affective Computing. MIT. The peaks of the waves indicate a cardiac cycle where the heart has pumped blood to the extremities. If the subject experiences fear or is startled, their heart usually 'jumps' and beats quickly for some time, causing the amplitude of the cardiac cycle to increase. This can clearly be seen on a photoplethysmograph when the distance between the trough and the peak of the wave has decreased. As the subject calms down, and as the body's inner core expands, allowing more blood to flow back to the extremities, the cycle will return to normal. 一个实验对象的血容量脉搏(BVP)可以通过一个叫做光容血管造影术的过程来测量,这个过程产生一个图表来显示通过四肢的血液流动。皮卡德,罗莎琳德(1998)。情感计算。麻省理工学院。波峰表明心脏将血液泵入四肢的心动周期。如果受试者感到恐惧或受到惊吓,他们的心脏通常会“跳动”并快速跳动一段时间,导致心脏周期的振幅增加。当波谷和波峰之间的距离减小时,可以清楚地看到这一点。当受试者平静下来,身体内核扩张,允许更多的血液回流到四肢,循环将恢复正常。 MethodologyMethodology= = = 方法论 = =Infra-red light is shone on the skin by special sensor hardware, and the amount of light reflected is measured. The amount of reflected and transmitted light correlates to the BVP as light is absorbed by hemoglobin which is found richly in the bloodstream. Infra-red light is shone on the skin by special sensor hardware, and the amount of light reflected is measured. The amount of reflected and transmitted light correlates to the BVP as light is absorbed by hemoglobin which is found richly in the bloodstream. 红外光通过特殊的传感器硬件照射在皮肤上,测量皮肤反射的光量。反射和透射光的数量与 BVP 相关,因为光线被血红蛋白吸收,而血液中的血红蛋白含量丰富。 DisadvantagesDisadvantages= = = 劣势 = = =It can be cumbersome to ensure that the sensor shining an infra-red light and monitoring the reflected light is always pointing at the same extremity, especially seeing as subjects often stretch and readjust their position while using a computer. There are other factors that can affect one's blood volume pulse. As it is a measure of blood flow through the extremities, if the subject feels hot, or particularly cold, then their body may allow more, or less, blood to flow to the extremities, all of this regardless of the subject's emotional state. It can be cumbersome to ensure that the sensor shining an infra-red light and monitoring the reflected light is always pointing at the same extremity, especially seeing as subjects often stretch and readjust their position while using a computer. There are other factors that can affect one's blood volume pulse. As it is a measure of blood flow through the extremities, if the subject feels hot, or particularly cold, then their body may allow more, or less, blood to flow to the extremities, all of this regardless of the subject's emotional state. 要确保传感器发出红外光并监测反射光总是指向同一极端,可能有些麻烦,尤其是在使用计算机时,受试者经常伸展和重新调整自己的位置。还有其他因素可以影响一个人的血容量脉搏。因为这是通过四肢血液流量的测量,如果受试者感到热或特别冷,那么他们的身体可能允许更多或更少的血液流向四肢---- 所有这一切都与受试者的情绪状态无关。 thumb|left| The corrugator supercilii muscle and zygomaticus major muscle are the 2 main muscles used for measuring the electrical activity, in facial electromyography 皱眉肌和颧肌是用来测量面部肌电图电活动的主要肌肉 Facial electromyographyFacial electromyography is a technique used to measure the electrical activity of the facial muscles by amplifying the tiny electrical impulses that are generated by muscle fibers when they contract.[39] The face expresses a great deal of emotion, however, there are two main facial muscle groups that are usually studied to detect emotion: The corrugator supercilii muscle, also known as the 'frowning' muscle, draws the brow down into a frown, and therefore is the best test for negative, unpleasant emotional response.↵The zygomaticus major muscle is responsible for pulling the corners of the mouth back when you smile, and therefore is the muscle used to test for a positive emotional response. Facial electromyography is a technique used to measure the electrical activity of the facial muscles by amplifying the tiny electrical impulses that are generated by muscle fibers when they contract.Larsen JT, Norris CJ, Cacioppo JT, "Effects of positive and negative affect on electromyographic activity over zygomaticus major and corrugator supercilii", (September 2003) The face expresses a great deal of emotion, however, there are two main facial muscle groups that are usually studied to detect emotion: The corrugator supercilii muscle, also known as the 'frowning' muscle, draws the brow down into a frown, and therefore is the best test for negative, unpleasant emotional response.↵The zygomaticus major muscle is responsible for pulling the corners of the mouth back when you smile, and therefore is the muscle used to test for a positive emotional response. 面部肌电图是一种通过放大肌肉收缩时产生的微小电脉冲来测量面部肌肉电活动的技术。O </o < o </o < o </o < > < o </o < > < o </o < > < o </o < > < o </o < > < o </o < > < o </o < > < o </o < > < o </o < > < o </o < > < o </o < > < o < > < o </o < > < o </o < > < o </o < > < o < > < o </o < > < o < > < o < > < o </o < > < o </o < > < o < > < o < > < o < > < o </o < > < o </o < > < o </o < > < o < > < > < o < > < > < o < > < > < > < o < > < > < o </o < >.当你微笑时,颧肌的主要肌肉负责将嘴角向后拉,因此是用来测试积极情绪反应的肌肉。 500px|thumb|Here we can see a plot of skin resistance measured using GSR and time whilst the subject played a video game. There are several peaks that are clear in the graph, which suggests that GSR is a good method of differentiating between an aroused and a non-aroused state. For example, at the start of the game where there is usually not much exciting game play, there is a high level of resistance recorded, which suggests a low level of conductivity and therefore less arousal. This is in clear contrast with the sudden trough where the player is killed as one is usually very stressed and tense as their character is killed in the game 500px | thumb | Here we can see a plot of skin resistance measured using GSR and time while the subject played a video game.在图中有几个明显的峰值,这表明 GSR 是区分性唤起和非性唤起状态的一个很好的方法。例如,在游戏开始的时候,通常没有多少激动人心的游戏,但是有一个高水平的电阻记录,这意味着低水平的电导率,因此唤起较少。这与玩家被杀的突然低谷形成了鲜明的对比,因为他们的角色在游戏中被杀时通常非常紧张和紧张 Galvanic skin responseGalvanic skin response (GSR) is an outdated term for a more general phenomenon known as [Electrodermal Activity] or EDA. EDA is a general phenomena whereby the skin's electrical properties change. The skin is innervated by the [sympathetic nervous system], so measuring its resistance or conductance provides a way to quantify small changes in the sympathetic branch of the autonomic nervous system. As the sweat glands are activated, even before the skin feels sweaty, the level of the EDA can be captured (usually using conductance) and used to discern small changes in autonomic arousal. The more aroused a subject is, the greater the skin conductance tends to be.[38] Galvanic skin response (GSR) is an outdated term for a more general phenomenon known as [Electrodermal Activity] or EDA. EDA is a general phenomena whereby the skin's electrical properties change. The skin is innervated by the [sympathetic nervous system], so measuring its resistance or conductance provides a way to quantify small changes in the sympathetic branch of the autonomic nervous system. As the sweat glands are activated, even before the skin feels sweaty, the level of the EDA can be captured (usually using conductance) and used to discern small changes in autonomic arousal. The more aroused a subject is, the greater the skin conductance tends to be. 皮肤电反应(GSR)是一个过时的术语,更一般的现象称为[皮肤电活动]或 EDA。EDA 是皮肤电特性改变的普遍现象。皮肤受交感神经神经支配,因此测量皮肤的电阻或电导率可以量化自主神经系统交感神经分支的细微变化。当汗腺被激活时,甚至在皮肤出汗之前,EDA 的水平就可以被捕获(通常使用电导) ,并用于辨别自主觉醒的微小变化。一个主体越兴奋,皮肤导电反应就越强烈。 Skin conductance is often measured using two small silver-silver chloride electrodes placed somewhere on the skin and applying a small voltage between them. To maximize comfort and reduce irritation the electrodes can be placed on the wrist, legs, or feet, which leaves the hands fully free for daily activity. Skin conductance is often measured using two small silver-silver chloride electrodes placed somewhere on the skin and applying a small voltage between them. To maximize comfort and reduce irritation the electrodes can be placed on the wrist, legs, or feet, which leaves the hands fully free for daily activity. 皮肤导电反应通常是通过放置在皮肤某处的小型氯化银电极并在两者之间施加一个小电压来测量的。为了最大限度地舒适和减少刺激,电极可以放在手腕、腿上或脚上,这样手就可以完全自由地进行日常活动。 Facial colorFacial color= = 面部表情 =OverviewOverview= = 概述 = =The surface of the human face is innervated with a large network of blood vessels. Blood flow variations in these vessels yield visible color changes on the face. Whether or not facial emotions activate facial muscles, variations in blood flow, blood pressure, glucose levels, and other changes occur. Also, the facial color signal is independent from that provided by facial muscle movements.[40] The surface of the human face is innervated with a large network of blood vessels. Blood flow variations in these vessels yield visible color changes on the face. Whether or not facial emotions activate facial muscles, variations in blood flow, blood pressure, glucose levels, and other changes occur. Also, the facial color signal is independent from that provided by facial muscle movements.Carlos F. Benitez-Quiroz, Ramprakash Srinivasan, Aleix M. Martinez, Facial color is an efficient mechanism to visually transmit emotion, PNAS. April 3, 2018 115 (14) 3581–3586; first published March 19, 2018 https://doi.org/10.1073/pnas.1716084115. 人脸的表面受到大型血管网的支配。这些血管中的血流变化在脸部产生可见的颜色变化。无论面部情绪是否激活面部肌肉,血流量、血压、血糖水平和其他变化都会发生。此外,面部颜色信号独立于面部肌肉运动提供的信号。卡洛斯 · f · 贝尼特斯-奎罗斯,兰帕卡什 · 斯里尼瓦桑,阿莱克斯 · m · 马丁内斯,面部颜色是一种有效的机制,可以在视觉上传达情感,PNAS。2018年4月3日115(14)3581-3586; 首次发表于2018年3月19日 https://doi.org/10.1073/pnas.1716084115。 MethodologyMethodology= = = 方法论 = =Approaches are based on facial color changes. Delaunay triangulation is used to create the triangular local areas. Some of these triangles which define the interior of the mouth and eyes (sclera and iris) are removed. Use the left triangular areas’ pixels to create feature vectors.[40] It shows that converting the pixel color of the standard RGB color space to a color space such as oRGB color space[41] or LMS channels perform better when dealing with faces.[42] So, map the above vector onto the better color space and decompose into red-green and yellow-blue channels. Then use deep learning methods to find equivalent emotions. Approaches are based on facial color changes. Delaunay triangulation is used to create the triangular local areas. Some of these triangles which define the interior of the mouth and eyes (sclera and iris) are removed. Use the left triangular areas’ pixels to create feature vectors. It shows that converting the pixel color of the standard RGB color space to a color space such as oRGB color spaceM. Bratkova, S. Boulos, and P. Shirley, oRGB: a practical opponent color space for computer graphics, IEEE Computer Graphics and Applications, 29(1):42–55, 2009. or LMS channels perform better when dealing with faces.Hadas Shahar, Hagit Hel-Or, Micro Expression Classification using Facial Color and Deep Learning Methods, The IEEE International Conference on Computer Vision (ICCV), 2019, pp. 0–0. So, map the above vector onto the better color space and decompose into red-green and yellow-blue channels. Then use deep learning methods to find equivalent emotions. 方法是基于面部颜色的变化。德劳内三角化被用来创建三角形的局部区域。其中一些定义嘴和眼睛内部的三角形(巩膜和虹膜)被移除。使用左边三角形区域的像素来创建特征向量。它表明将标准 RGB 颜色空间的像素颜色转换为颜色空间,如 oRGB 颜色空间。和 p. Shirley,org b: 一个实用的对手色彩空间,计算机图形学,IEEE 计算机图形学与应用,29(1) : 42-55,2009。或 LMS 通道在处理面孔时表现更好。使用面部颜色和深度学习方法的微表情分类,IEEE 国际计算机视觉会议(ICCV) ,2019,pp。0–0.因此,将上面的矢量映射到较好的颜色空间,并分解为红绿色和黄蓝色通道。然后使用深度学习的方法来找到等效的情绪。 Visual aestheticsAesthetics, in the world of art and photography, refers to the principles of the nature and appreciation of beauty. Judging beauty and other aesthetic qualities is a highly subjective task. Computer scientists at Penn State treat the challenge of automatically inferring the aesthetic quality of pictures using their visual content as a machine learning problem, with a peer-rated on-line photo sharing website as a data source.[43] They extract certain visual features based on the intuition that they can discriminate between aesthetically pleasing and displeasing images. Aesthetics, in the world of art and photography, refers to the principles of the nature and appreciation of beauty. Judging beauty and other aesthetic qualities is a highly subjective task. Computer scientists at Penn State treat the challenge of automatically inferring the aesthetic quality of pictures using their visual content as a machine learning problem, with a peer-rated on-line photo sharing website as a data source.Ritendra Datta, Dhiraj Joshi, Jia Li and James Z. Wang, Studying Aesthetics in Photographic Images Using a Computational Approach, Lecture Notes in Computer Science, vol. 3953, Proceedings of the European Conference on Computer Vision, Part III, pp. 288–301, Graz, Austria, May 2006. They extract certain visual features based on the intuition that they can discriminate between aesthetically pleasing and displeasing images. 美学,在艺术和摄影的世界里,是指自然的原则和审美的原则。判断美和其他审美品质是一项高度主观的任务。宾夕法尼亚州立大学的计算机科学家将利用图片的视觉内容自动推断图片的审美质量的挑战视为一个机器学习问题,同行评级的在线照片共享网站则是一个数据源。利 · 达塔,Dhiraj Joshi,贾力和詹姆斯 · 王,用计算方法研究摄影图像美学,计算机科学讲义,第卷。3953,Proceedings of the European Conference on Computer Vision,Part III,pp.288-301,格拉茨,2006年5月。他们提取某些视觉特征是基于这样一种直觉,即他们可以区分美观的和不美观的图像。 Potential applicationsEducationAffection influences learners' learning state. Using affective computing technology, computers can judge the learners' affection and learning state by recognizing their facial expressions. In education, the teacher can use the analysis result to understand the student's learning and accepting ability, and then formulate reasonable teaching plans. At the same time, they can pay attention to students' inner feelings, which is helpful to students' psychological health. Especially in distance education, due to the separation of time and space, there is no emotional incentive between teachers and students for two-way communication. Without the atmosphere brought by traditional classroom learning, students are easily bored, and affect the learning effect. Applying affective computing in distance education system can effectively improve this situation. [44] Affection influences learners' learning state. Using affective computing technology, computers can judge the learners' affection and learning state by recognizing their facial expressions. In education, the teacher can use the analysis result to understand the student's learning and accepting ability, and then formulate reasonable teaching plans. At the same time, they can pay attention to students' inner feelings, which is helpful to students' psychological health. Especially in distance education, due to the separation of time and space, there is no emotional incentive between teachers and students for two-way communication. Without the atmosphere brought by traditional classroom learning, students are easily bored, and affect the learning effect. Applying affective computing in distance education system can effectively improve this situation. http://www.learntechlib.org/p/173785/ 情感影响学习者的学习状态。利用情感计算技术,计算机可以通过学习者的面部表情识别来判断学习者的情感和学习状态。在教学中,教师可以利用分析结果了解学生的学习和接受能力,制定合理的教学计划。同时关注学生的内心感受,有利于学生的心理健康。特别是在远程教育中,由于时间和空间的分离,师生之间缺乏双向交流的情感激励。没有了传统课堂学习带来的氛围,学生很容易感到无聊,影响学习效果。将情感计算应用于远程教育系统可以有效地改善这种状况。http://www.learntechlib.org/p/173785/ HealthcareSocial robots, as well as a growing number of robots used in health care benefit from emotional awareness because they can better judge users' and patient's emotional states and alter their actions/programming appropriately. This is especially important in those countries with growing aging populations and/or a lack of younger workers to address their needs.[45] Social robots, as well as a growing number of robots used in health care benefit from emotional awareness because they can better judge users' and patient's emotional states and alter their actions/programming appropriately. This is especially important in those countries with growing aging populations and/or a lack of younger workers to address their needs. 社会机器人,以及越来越多的机器人在医疗保健中的应用都受益于情感意识,因为它们可以更好地判断用户和病人的情感状态,并适当地改变他们的行为/编程。在人口老龄化日益严重和/或缺乏年轻工人满足其需要的国家,这一点尤为重要。 Affective computing is also being applied to the development of communicative technologies for use by people with autism.[46] The affective component of a text is also increasingly gaining attention, particularly its role in the so-called emotional or emotive Internet.[47] Affective computing is also being applied to the development of communicative technologies for use by people with autism.Projects in Affective Computing The affective component of a text is also increasingly gaining attention, particularly its role in the so-called emotional or emotive Internet.Shanahan, James; Qu, Yan; Wiebe, Janyce (2006). Computing Attitude and Affect in Text: Theory and Applications. Dordrecht: Springer Science & Business Media. p. 94. 情感计算也被应用于交流技术的发展,以供孤独症患者使用。情感计算项目文本中的情感成分也越来越受到关注,特别是它在所谓的情感或情感互联网中的作用。夏纳汉,詹姆斯; 曲,严; 韦伯,詹尼策(2006)。文本中的计算态度和情感: 理论与应用。多德雷赫特: 斯普林格科学与商业媒体。3月94日。 Video gamesVideo games= = 电子游戏 =Affective video games can access their players' emotional states through biofeedback devices.[48] A particularly simple form of biofeedback is available through gamepads that measure the pressure with which a button is pressed: this has been shown to correlate strongly with the players' level of arousal;[49] at the other end of the scale are brain–computer interfaces.[50][51] Affective games have been used in medical research to support the emotional development of autistic children.[52] Affective video games can access their players' emotional states through biofeedback devices. A particularly simple form of biofeedback is available through gamepads that measure the pressure with which a button is pressed: this has been shown to correlate strongly with the players' level of arousal; at the other end of the scale are brain–computer interfaces. Affective games have been used in medical research to support the emotional development of autistic children. 情感视频游戏可以通过生物反馈设备进入玩家的情绪状态。一种特别简单的生物反馈形式可以通过游戏来测量按钮被按下的压力: 这已被证明与玩家的觉醒水平密切相关; 在天平的另一端是大脑-计算机接口。情感游戏已被用于医学研究,以支持自闭症儿童的情感发展。 Other applicationsOther applications= = 其他应用 = =Other potential applications are centered around social monitoring. For example, a car can monitor the emotion of all occupants and engage in additional safety measures, such as alerting other vehicles if it detects the driver to be angry.[53] Affective computing has potential applications in human–computer interaction, such as affective mirrors allowing the user to see how he or she performs; emotion monitoring agents sending a warning before one sends an angry email; or even music players selecting tracks based on mood.[54] Other potential applications are centered around social monitoring. For example, a car can monitor the emotion of all occupants and engage in additional safety measures, such as alerting other vehicles if it detects the driver to be angry. Affective computing has potential applications in human–computer interaction, such as affective mirrors allowing the user to see how he or she performs; emotion monitoring agents sending a warning before one sends an angry email; or even music players selecting tracks based on mood. 其他潜在的应用主要围绕社会监控。例如,一辆汽车可以监控所有乘客的情绪,并采取额外的安全措施,例如,如果发现司机生气,就向其他车辆发出警报。情感计算在人机交互方面有着潜在的应用,比如情感镜子可以让用户看到自己的表现; 情感监控代理在发送愤怒邮件之前发送警告; 甚至音乐播放器可以根据情绪选择音轨。 One idea put forth by the Romanian researcher Dr. Nicu Sebe in an interview is the analysis of a person's face while they are using a certain product (he mentioned ice cream as an example).[55] Companies would then be able to use such analysis to infer whether their product will or will not be well received by the respective market. One idea put forth by the Romanian researcher Dr. Nicu Sebe in an interview is the analysis of a person's face while they are using a certain product (he mentioned ice cream as an example). Companies would then be able to use such analysis to infer whether their product will or will not be well received by the respective market. 罗马尼亚研究人员尼库 · 塞贝博士在一次采访中提出的一个想法是,当一个人使用某种产品时,对他的脸进行分析(他提到了冰淇淋作为一个例子)。然后,公司就能够利用这种分析来推断他们的产品是否会受到各自市场的欢迎。 One could also use affective state recognition in order to judge the impact of a TV advertisement through a real-time video recording of that person and through the subsequent study of his or her facial expression. Averaging the results obtained on a large group of subjects, one can tell whether that commercial (or movie) has the desired effect and what the elements which interest the watcher most are. One could also use affective state recognition in order to judge the impact of a TV advertisement through a real-time video recording of that person and through the subsequent study of his or her facial expression. Averaging the results obtained on a large group of subjects, one can tell whether that commercial (or movie) has the desired effect and what the elements which interest the watcher most are. 人们也可以利用情感状态识别来判断电视广告的影响,通过实时录像和随后对他或她的面部表情的研究。对一大群受试者的结果进行平均,你就可以知道那个商业广告(或电影)是否有预期的效果,以及观众最感兴趣的元素是什么。 Cognitivist vs. interactional approachesCognitivist vs. interactional approaches= = Cognitivist vs. interactionapproach =Within the field of human–computer interaction, Rosalind Picard's cognitivist or "information model" concept of emotion has been criticized by and contrasted with the "post-cognitivist" or "interactional" pragmatist approach taken by Kirsten Boehner and others which views emotion as inherently social.[56] Within the field of human–computer interaction, Rosalind Picard's cognitivist or "information model" concept of emotion has been criticized by and contrasted with the "post-cognitivist" or "interactional" pragmatist approach taken by Kirsten Boehner and others which views emotion as inherently social. 在人机交互领域,罗莎琳德 · 皮卡德的情绪认知主义或“信息模型”概念受到了后认知主义或“互动”实用主义者柯尔斯滕 · 博纳等人的批判和对比。 Picard's focus is human–computer interaction, and her goal for affective computing is to "give computers the ability to recognize, express, and in some cases, 'have' emotions".[4] In contrast, the interactional approach seeks to help "people to understand and experience their own emotions"[57] and to improve computer-mediated interpersonal communication. It does not necessarily seek to map emotion into an objective mathematical model for machine interpretation, but rather let humans make sense of each other's emotional expressions in open-ended ways that might be ambiguous, subjective, and sensitive to context.[57]:284模板:Example needed Picard's focus is human–computer interaction, and her goal for affective computing is to "give computers the ability to recognize, express, and in some cases, 'have' emotions". In contrast, the interactional approach seeks to help "people to understand and experience their own emotions" and to improve computer-mediated interpersonal communication. It does not necessarily seek to map emotion into an objective mathematical model for machine interpretation, but rather let humans make sense of each other's emotional expressions in open-ended ways that might be ambiguous, subjective, and sensitive to context. 皮卡德的研究重点是人机交互,她研究情感计算的目标是“赋予计算机识别、表达、在某些情况下‘拥有’情感的能力”。相比之下,交互式的方法旨在帮助“人们理解和体验他们自己的情绪”,并改善以电脑为媒介的人际沟通。它并不一定寻求将情绪映射到一个用于机器解释的客观数学模型中,而是让人类以开放的方式理解彼此的情绪表达,这种方式可能是模糊的、主观的,并且对上下文敏感。 Picard's critics describe her concept of emotion as "objective, internal, private, and mechanistic". They say it reduces emotion to a discrete psychological signal occurring inside the body that can be measured and which is an input to cognition, undercutting the complexity of emotional experience.[57]:280[57]:278 Picard's critics describe her concept of emotion as "objective, internal, private, and mechanistic". They say it reduces emotion to a discrete psychological signal occurring inside the body that can be measured and which is an input to cognition, undercutting the complexity of emotional experience. 皮卡德的批评者将她的情感概念描述为“客观的、内在的、私人的和机械的”。他们说,它把情绪简化为发生在身体内部的一个离散的心理信号,这个信号可以被测量,并且是认知的输入,削弱了情绪体验的复杂性。 The interactional approach asserts that though emotion has biophysical aspects, it is "culturally grounded, dynamically experienced, and to some degree constructed in action and interaction".[57]:276 Put another way, it considers "emotion as a social and cultural product experienced through our interactions".[58][57][59] The interactional approach asserts that though emotion has biophysical aspects, it is "culturally grounded, dynamically experienced, and to some degree constructed in action and interaction". Put another way, it considers "emotion as a social and cultural product experienced through our interactions". 交互式教学法认为,情感虽然具有生物物理性,但它是“有文化基础的、动态体验的,并且在一定程度上是在行动和互动中建构起来的”。换句话说,它认为“情感是一种通过我们的互动体验到的社会和文化产品”。 See alsoCitations
General sources
External links
This page was moved from wikipedia:en:Affective computing. Its edit history can be viewed at 情感计算/edithistory |
---|