{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,20]],"date-time":"2026-02-20T18:45:56Z","timestamp":1771613156364,"version":"3.50.1"},"reference-count":72,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2023,8,1]],"date-time":"2023-08-01T00:00:00Z","timestamp":1690848000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2023,8,1]],"date-time":"2023-08-01T00:00:00Z","timestamp":1690848000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2023,8,1]],"date-time":"2023-08-01T00:00:00Z","timestamp":1690848000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2023,8,1]],"date-time":"2023-08-01T00:00:00Z","timestamp":1690848000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2023,8,1]],"date-time":"2023-08-01T00:00:00Z","timestamp":1690848000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2023,8,1]],"date-time":"2023-08-01T00:00:00Z","timestamp":1690848000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,8,1]],"date-time":"2023-08-01T00:00:00Z","timestamp":1690848000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100004826","name":"Natural Science Foundation of Beijing Municipality","doi-asserted-by":"publisher","award":["9222019"],"award-info":[{"award-number":["9222019"]}],"id":[{"id":"10.13039\/501100004826","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62277006"],"award-info":[{"award-number":["62277006"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002726","name":"Beijing Normal University","doi-asserted-by":"publisher","award":["ICER201903"],"award-info":[{"award-number":["ICER201903"]}],"id":[{"id":"10.13039\/501100002726","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Image and Vision Computing"],"published-print":{"date-parts":[[2023,8]]},"DOI":"10.1016\/j.imavis.2023.104739","type":"journal-article","created":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T20:57:43Z","timestamp":1686603463000},"page":"104739","update-policy":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":12,"special_numbering":"C","title":["Spontaneous visual database for detecting learning-centered emotions during online learning"],"prefix":"10.1016","volume":"136","author":[{"given":"Yaping","family":"Xu","sequence":"first","affiliation":[]},{"given":"Yanyan","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yunshan","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Haogang","family":"Bao","sequence":"additional","affiliation":[]},{"given":"Yaqian","family":"Zheng","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.imavis.2023.104739_bb0005","first-page":"1","article-title":"Advantages and constraints of a hybrid model k-12 e-learning assistant chatbot","volume":"99","author":"Wu","year":"2020","journal-title":"IEEE Access"},{"key":"10.1016\/j.imavis.2023.104739_bb0010","series-title":"Emotion in Education","first-page":"37","article-title":"Understanding students' affective processes in the classroom","author":"Boekaerts","year":"2007"},{"issue":"4","key":"10.1016\/j.imavis.2023.104739_bb0015","first-page":"9","article-title":"When emotions enhance students' engagement in e-learning processes","volume":"12","author":"D'Errico","year":"2016","journal-title":"J. E-Learn. Knowl. Soc."},{"key":"10.1016\/j.imavis.2023.104739_bb0020","first-page":"1","article-title":"Online behavior analysis-based student profile for intelligent e-learning","author":"Liang","year":"2017","journal-title":"J. Electr. Comput. Eng."},{"key":"10.1016\/j.imavis.2023.104739_bb0025","series-title":"An Approach to Environmental Psychology","author":"Mehrabian","year":"1974"},{"key":"10.1016\/j.imavis.2023.104739_bb0030","series-title":"Affective Computing and Intelligent Interaction","first-page":"248","article-title":"Interpreting hand-over-face gestures","author":"Mahmoud","year":"2011"},{"issue":"4","key":"10.1016\/j.imavis.2023.104739_bb0035","doi-asserted-by":"crossref","first-page":"53","DOI":"10.1109\/MIS.2007.79","article-title":"Towards an affect-sensitive AutoTutor","volume":"22","author":"D\u2019Mello","year":"2007","journal-title":"IEEE Intell. Syst."},{"key":"10.1016\/j.imavis.2023.104739_bb0040","doi-asserted-by":"crossref","first-page":"168","DOI":"10.1016\/j.image.2017.08.012","article-title":"BNU-LSVED 2.0: spontaneous multimodal student affect database with multi-dimensional labels","volume":"59","author":"Wei","year":"2017","journal-title":"Signal Process. Image Commun."},{"issue":"3","key":"10.1016\/j.imavis.2023.104739_bb0045","doi-asserted-by":"crossref","first-page":"329","DOI":"10.1049\/iet-cvi.2018.5281","article-title":"A spontaneous facial expression database for academic emotion inference in online learning","volume":"13","author":"Bian","year":"2018","journal-title":"IET Comput. Vis."},{"key":"10.1016\/j.imavis.2023.104739_bb0050","doi-asserted-by":"crossref","first-page":"334","DOI":"10.1016\/j.future.2020.02.075","article-title":"Affective database for e-learning and classroom environments using Indian students\u2019 faces, hand gestures and body postures","volume":"108","author":"Ashwin","year":"2020","journal-title":"Futur. Gener. Comput. Syst."},{"issue":"5","key":"10.1016\/j.imavis.2023.104739_bb0055","doi-asserted-by":"crossref","first-page":"377","DOI":"10.1016\/j.learninstruc.2005.07.006","article-title":"Emotional experiences during learning: multiple, situated and dynamic","volume":"15","author":"Efklides","year":"2005","journal-title":"Learn. Instr."},{"issue":"1","key":"10.1016\/j.imavis.2023.104739_bb0060","doi-asserted-by":"crossref","first-page":"19","DOI":"10.1037\/a0023463","article-title":"Internal representations reveal cultural diversity in expectations of facial expressions of emotion","volume":"141","author":"Jack","year":"2012","journal-title":"J. Exp. Psychol. Gen."},{"key":"10.1016\/j.imavis.2023.104739_bb0065","series-title":"Third IEEE International Conference on Automatic Face and Gesture Recognition","first-page":"200","article-title":"Coding facial expressions with Gabor wavelets","author":"Lyons","year":"2002"},{"key":"10.1016\/j.imavis.2023.104739_bb0070","series-title":"2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","first-page":"94","article-title":"The extended cohn-kanade dataset (ck+): a complete dataset for action unit and emotion-specified expression","author":"Lucey","year":"2010"},{"issue":"12","key":"10.1016\/j.imavis.2023.104739_bb0075","first-page":"1615","article-title":"The CMU pose, illumination, and expression database","volume":"25","author":"Sim","year":"2004","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.imavis.2023.104739_bb0080","series-title":"International Conference on Pattern Recognition","first-page":"1148","article-title":"A bimodal face and body gesture database for automatic analysis of human nonverbal affective behavior","author":"Gunes","year":"2006"},{"key":"10.1016\/j.imavis.2023.104739_bb0085","series-title":"IEEE International Conference on Multimedia & Expo","first-page":"317","article-title":"Web-based database for facial expression analysis","author":"Pantic","year":"2005"},{"issue":"1","key":"10.1016\/j.imavis.2023.104739_bb0090","doi-asserted-by":"crossref","first-page":"131","DOI":"10.1109\/TAFFC.2015.2498174","article-title":"The Indian spontaneous expression database for emotion recognition","volume":"8","author":"Happy","year":"2017","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"1","key":"10.1016\/j.imavis.2023.104739_bb0095","doi-asserted-by":"crossref","DOI":"10.1371\/journal.pone.0086041","article-title":"CASME II: an improved spontaneous micro-expression database and the baseline evaluation","volume":"9","author":"Yan","year":"2014","journal-title":"PLoS One"},{"issue":"1","key":"10.1016\/j.imavis.2023.104739_bb0100","doi-asserted-by":"crossref","first-page":"351","DOI":"10.3758\/BRM.42.1.351","article-title":"FACES--a database of facial expressions in young, middle-aged, and older women and men: development and validation","volume":"42","author":"Ebner","year":"2010","journal-title":"Behav. Res. Methods"},{"key":"10.1016\/j.imavis.2023.104739_bb0105","series-title":"10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition","first-page":"1","article-title":"A spontaneous micro-expression database: inducement, collection and baseline","author":"Li","year":"2013"},{"key":"10.1016\/j.imavis.2023.104739_bb0110","series-title":"15th ACM on International Conference on Multimodal Interaction","first-page":"509","article-title":"Emotion recognition in the wild challenge 2013","author":"Dhall","year":"2013"},{"key":"10.1016\/j.imavis.2023.104739_bb0115","doi-asserted-by":"crossref","first-page":"59","DOI":"10.1016\/j.neunet.2014.09.005","article-title":"Challenges in representation learning: a report on three machine learning contests","volume":"64","author":"Goodfellow","year":"2015","journal-title":"Neural Netw."},{"key":"10.1016\/j.imavis.2023.104739_bb0120","series-title":"DAiSEE: Towards user engagement recognition in the wild","first-page":"1","author":"Gupta","year":"2016"},{"issue":"5","key":"10.1016\/j.imavis.2023.104739_bb0125","doi-asserted-by":"crossref","first-page":"807","DOI":"10.1016\/j.imavis.2009.08.002","article-title":"Multi-PIE","volume":"28","author":"Gross","year":"2009","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.imavis.2023.104739_bb0130","series-title":"The AR face database","volume":"24","author":"Martinez","year":"1998"},{"key":"10.1016\/j.imavis.2023.104739_bb0135","series-title":"International Conference on Pattern Recognition","first-page":"153","article-title":"Multimodal database of emotional speech, video and gestures","author":"Sapi\u0144ski","year":"2019"},{"issue":"9","key":"10.1016\/j.imavis.2023.104739_bb0140","doi-asserted-by":"crossref","first-page":"607","DOI":"10.1016\/j.imavis.2011.07.002","article-title":"Facial expression recognition from near-infrared videos","volume":"29","author":"Zhao","year":"2011","journal-title":"Image Vis. Comput."},{"issue":"11","key":"10.1016\/j.imavis.2023.104739_bb0145","doi-asserted-by":"crossref","DOI":"10.1371\/journal.pone.0079131","article-title":"The Dartmouth database of children\u2019s faces: acquisition and validation of a new face stimulus set","volume":"8","author":"Dalrymple","year":"2013","journal-title":"PLoS One"},{"key":"10.1016\/j.imavis.2023.104739_bb0150","series-title":"7th International Conference on Automatic Face and Gesture Recognition","first-page":"211","article-title":"A 3D facial expression database for facial behavior research","author":"Yin","year":"2006"},{"issue":"11","key":"10.1016\/j.imavis.2023.104739_bb0155","doi-asserted-by":"crossref","first-page":"2106","DOI":"10.1109\/TPAMI.2009.42","article-title":"Toward practical smile detection","volume":"31","author":"Whitehill","year":"2009","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.imavis.2023.104739_bb0160","series-title":"IEEE Conference on Computer Vision and Pattern Recognition Workshops","first-page":"881","article-title":"Affectiva-mit facial expression dataset (am-fed): naturalistic and spontaneous facial expressions collected \u201cin-the-wild\u201d","author":"McDuff","year":"2013"},{"issue":"1","key":"10.1016\/j.imavis.2023.104739_bb0165","doi-asserted-by":"crossref","first-page":"51","DOI":"10.1016\/0031-3203(95)00067-4","article-title":"A comparative study of texture measures with classification based on feature distributions","volume":"29","author":"Ojala","year":"1996","journal-title":"Pattern Recogn."},{"key":"10.1016\/j.imavis.2023.104739_bb0170","series-title":"2005 IEEE Computer Society Conference on Computer Vision & Pattern Recognition","first-page":"886","article-title":"Histograms of oriented gradients for human detection","author":"Dalal","year":"2005"},{"issue":"10","key":"10.1016\/j.imavis.2023.104739_bb0175","doi-asserted-by":"crossref","first-page":"959","DOI":"10.1109\/34.541406","article-title":"Image representation using 2D gabor wavelets","volume":"18","author":"Lee","year":"1996","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"6","key":"10.1016\/j.imavis.2023.104739_bb0180","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1145\/3065386","article-title":"ImageNet classification with deep convolutional neural networks","volume":"60","author":"Krizhevsky","year":"2017","journal-title":"Commun. ACM"},{"key":"10.1016\/j.imavis.2023.104739_bb0185","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1109\/TIM.2020.3031835","article-title":"Facial expression recognition using local gravitational force descriptor-based deep convolution neural networks","volume":"70","author":"Mohan","year":"2021","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"4","key":"10.1016\/j.imavis.2023.104739_bb0190","doi-asserted-by":"crossref","first-page":"2058","DOI":"10.1109\/TAFFC.2022.3208309","article-title":"FLEPNet: feature level ensemble parallel network for facial expression recognition","volume":"13","author":"Karnati","year":"2022","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"15","key":"10.1016\/j.imavis.2023.104739_bb0195","doi-asserted-by":"crossref","first-page":"9125","DOI":"10.1007\/s00521-020-05676-y","article-title":"FER-net: Facial expression recognition using deep neural net","volume":"33","author":"Mohan","year":"2021","journal-title":"Neural Comput. Applic."},{"key":"10.1016\/j.imavis.2023.104739_bb0200","first-page":"1","article-title":"Understanding deep learning techniques for recognition of human emotions using facial expressions: a comprehensive survey","volume":"72","author":"Karnati","year":"2023","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.imavis.2023.104739_bb0205","series-title":"Adaptive Technologies for Training and Education","first-page":"169","article-title":"Emotions during learning with AutoTutor","author":"D\u2019Mello","year":"2012"},{"key":"10.1016\/j.imavis.2023.104739_bb0210","doi-asserted-by":"crossref","first-page":"37","DOI":"10.1016\/B978-012064455-1\/50006-3","article-title":"Messages that motivate: how praise molds students\u2019 beliefs, motivation, and performance (in surprising ways)","author":"Dweck","year":"2002","journal-title":"Improv. Acad. Achiev."},{"key":"10.1016\/j.imavis.2023.104739_bb0215","series-title":"An Attributional Theory of Motivation and Emotion","first-page":"159","author":"Weiner","year":"2012"},{"key":"10.1016\/j.imavis.2023.104739_bb0220","series-title":"International Handbook of Metacognition and Learning Technologies","first-page":"669","article-title":"Affect, meta-affect, and affect regulation during complex learning","author":"D\u2019Mello","year":"2013"},{"issue":"4","key":"10.1016\/j.imavis.2023.104739_bb0225","doi-asserted-by":"crossref","first-page":"1082","DOI":"10.1037\/a0032674","article-title":"A selective meta-analysis on the relative incidence of discrete affective states during learning with technology","volume":"105","author":"D\u2019Mello","year":"2013","journal-title":"J. Educ. Psychol."},{"issue":"4","key":"10.1016\/j.imavis.2023.104739_bb0230","doi-asserted-by":"crossref","first-page":"223","DOI":"10.1016\/j.ijhcs.2009.12.003","article-title":"Better to be frustrated than bored: the incidence, persistence, and impact of learners\u2019 cognitive\u2013affective states during interactions with three different computer-based learning environments","volume":"68","author":"Baker","year":"2010","journal-title":"Int. J. Human-Comput. Stud."},{"issue":"3","key":"10.1016\/j.imavis.2023.104739_bb0235","first-page":"78","article-title":"Emotion and e-learning","volume":"7","author":"O\u2019Regan","year":"2003","journal-title":"J. Asynchron. Learn. Netw."},{"issue":"3","key":"10.1016\/j.imavis.2023.104739_bb0240","first-page":"19","article-title":"Development and construct validation of e-learning academic emotions scale(e-AES)","volume":"20","author":"You","year":"2012","journal-title":"J. Yeolin Educ."},{"issue":"12","key":"10.1016\/j.imavis.2023.104739_bb0245","doi-asserted-by":"crossref","DOI":"10.5210\/fm.v4i12.710","article-title":"Student frustrations with a web-based distance education course","volume":"4","author":"Hara","year":"1999","journal-title":"First Monday"},{"issue":"1","key":"10.1016\/j.imavis.2023.104739_bb0250","doi-asserted-by":"crossref","first-page":"86","DOI":"10.1109\/TAFFC.2014.2316163","article-title":"The faces of engagement: automatic recognition of student engagement from facial expressions","volume":"5","author":"Whitehill","year":"2014","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.imavis.2023.104739_bb0255","first-page":"31","article-title":"Mining bodily patterns of affective experience during learning","author":"D'Mello","year":"2010","journal-title":"Educ. Data Min."},{"key":"10.1016\/j.imavis.2023.104739_bb0260","series-title":"14th ACM International Conference on Multimodal Interaction","first-page":"145","article-title":"Multimodal analysis of the implicit affective channel in computer-mediated textual communication","author":"Grafsgaard","year":"2012"},{"key":"10.1016\/j.imavis.2023.104739_bb0265","series-title":"Affective Computing and Intelligent Interaction","first-page":"248","article-title":"Interpreting hand-over-face gestures","author":"Mahmoud","year":"2011"},{"key":"10.1016\/j.imavis.2023.104739_bb0270","series-title":"International Symposium on Quality Education for Teenagers","first-page":"280","article-title":"The influence of mathematical contest in modeling on cultivation of innovation ability of undergraduate","author":"Qu","year":"2017"},{"issue":"1","key":"10.1016\/j.imavis.2023.104739_bb0275","first-page":"62","article-title":"Kappa coefficient: a popular measure of rater agreement","volume":"27","author":"Tang","year":"2015","journal-title":"Shanghai Arch. Psychiatry"},{"key":"10.1016\/j.imavis.2023.104739_bb0280","series-title":"Computer Recognition Systems","first-page":"816","article-title":"The performance of the haar cascade classifiers applied to the face and eyes detection","author":"Schmidt","year":"2007"},{"key":"10.1016\/j.imavis.2023.104739_bb0285","series-title":"You only look once: Unified, real-time object detection","first-page":"779","author":"Redmon","year":"2016"},{"issue":"7","key":"10.1016\/j.imavis.2023.104739_bb0290","doi-asserted-by":"crossref","first-page":"971","DOI":"10.1109\/TPAMI.2002.1017623","article-title":"Multiresolution gray-scale and rotation invariant texture classification with local binary patterns","volume":"24","author":"Ojala","year":"2002","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.imavis.2023.104739_bb0295","series-title":"6th ACM International Conference on Image and Video Retrieval","first-page":"401","article-title":"Representing shape with a spatial pyramid kernel","author":"Bosch","year":"2007"},{"key":"10.1016\/j.imavis.2023.104739_bb0300","series-title":"Environmental Psychology & Nonverbal Behavior","article-title":"Facial action coding system","author":"Ekman","year":"1978"},{"key":"10.1016\/j.imavis.2023.104739_bb0305","doi-asserted-by":"crossref","first-page":"3731","DOI":"10.22214\/ijraset.2022.43205","article-title":"Emotion and gesture detection","volume":"10","author":"Raman","year":"2022","journal-title":"Int. J. Res. Appl. Sci. Eng. Technol."},{"key":"10.1016\/j.imavis.2023.104739_bb0310","series-title":"Data Science: 6th International Conference","first-page":"275","article-title":"A novel video emotion recognition system in the wild using a random forest classifier","author":"Samadiani","year":"2020"},{"key":"10.1016\/j.imavis.2023.104739_bb0315","series-title":"IEEE Conference on Computer Vision and Pattern Recognition","first-page":"1251","article-title":"Xception: Deep learning with depthwise separable convolutions","author":"Chollet","year":"2017"},{"key":"10.1016\/j.imavis.2023.104739_bb0320","series-title":"Real-time convolutional neural networks for emotion and gender classification","author":"Arriaga","year":"2017"},{"key":"10.1016\/j.imavis.2023.104739_bb0325","first-page":"1","article-title":"Adam: a method for stochastic optimization","author":"Kingma","year":"2014","journal-title":"Comput. Therm. Sci."},{"issue":"5","key":"10.1016\/j.imavis.2023.104739_bb0330","doi-asserted-by":"crossref","first-page":"1028","DOI":"10.1111\/bjet.12325","article-title":"3D face model dataset: automatic detection of facial expressions and emotions for educational environments","volume":"46","author":"Chickerur","year":"2015","journal-title":"Br. J. Educ. Technol."},{"key":"10.1016\/j.imavis.2023.104739_bb0335","series-title":"3rd International Conference on Affective Computing and Intelligent Interaction and Workshops","first-page":"1","article-title":"Natural affect data\u2014collection & annotation in a learning context","author":"Afzal","year":"2009"},{"key":"10.1016\/j.imavis.2023.104739_bb0340","series-title":"IEEE\/CVF International Conference on Computer Vision Workshops","first-page":"3628","article-title":"Student engagement dataset","author":"Delgado","year":"2021"},{"key":"10.1016\/j.imavis.2023.104739_bb0345","series-title":"Emotions, Technology, Design, and Learning","first-page":"27","article-title":"A real-time speech emotion recognition system and its application in online learning","author":"Cen","year":"2016"},{"key":"10.1016\/j.imavis.2023.104739_bb0350","doi-asserted-by":"crossref","DOI":"10.1155\/2014\/484873","article-title":"Towards emotion detection in educational scenarios from facial expressions and body movements through multimodal approaches","author":"Saneiro","year":"2014","journal-title":"Sci. World J."},{"key":"10.1016\/j.imavis.2023.104739_bb0355","series-title":"DeLFI\/GMW Workshops","first-page":"1","article-title":"A pilot study of emotion detection using sensors in a learning context: towards an affective learning companion","author":"Yun","year":"2017"},{"key":"10.1016\/j.imavis.2023.104739_bb0360","series-title":"Artificial Intelligence in Education","first-page":"17","article-title":"Emotion sensors go to school","author":"Arroyo","year":"2009"}],"container-title":["Image and Vision Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/api.elsevier.com\/content\/article\/PII:S0262885623001130?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/api.elsevier.com\/content\/article\/PII:S0262885623001130?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,10,2]],"date-time":"2025-10-02T01:41:44Z","timestamp":1759369304000},"score":1,"resource":{"primary":{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/linkinghub.elsevier.com\/retrieve\/pii\/S0262885623001130"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8]]},"references-count":72,"alternative-id":["S0262885623001130"],"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1016\/j.imavis.2023.104739","relation":{},"ISSN":["0262-8856"],"issn-type":[{"value":"0262-8856","type":"print"}],"subject":[],"published":{"date-parts":[[2023,8]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Spontaneous visual database for detecting learning-centered emotions during online learning","name":"articletitle","label":"Article Title"},{"value":"Image and Vision Computing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1016\/j.imavis.2023.104739","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2023 Elsevier B.V. All rights reserved.","name":"copyright","label":"Copyright"}],"article-number":"104739"}}