Publications
Search
Core, Mark G.; Nye, Benjamin D.; Fegley, Brent D.
Trend-Aware Scenario Authoring: Adapting Training Toward Patterns from Real Operations Book Section
In: Sottilare, Robert A.; Schwarz, Jessica (Ed.): Adaptive Instructional Systems, vol. 14727, pp. 15–24, Springer Nature Switzerland, Cham, 2024, ISBN: 978-3-031-60608-3 978-3-031-60609-0, (Series Title: Lecture Notes in Computer Science).
@incollection{sottilare_trend-aware_2024,
title = {Trend-Aware Scenario Authoring: Adapting Training Toward Patterns from Real Operations},
author = {Mark G. Core and Benjamin D. Nye and Brent D. Fegley},
editor = {Robert A. Sottilare and Jessica Schwarz},
url = {https://link.springer.com/10.1007/978-3-031-60609-0_2},
doi = {10.1007/978-3-031-60609-0_2},
isbn = {978-3-031-60608-3 978-3-031-60609-0},
year = {2024},
date = {2024-06-01},
urldate = {2024-06-18},
booktitle = {Adaptive Instructional Systems},
volume = {14727},
pages = {15–24},
publisher = {Springer Nature Switzerland},
address = {Cham},
note = {Series Title: Lecture Notes in Computer Science},
keywords = {},
pubstate = {published},
tppubtype = {incollection}
}
Bohy, Hugo; Tran, Minh; Haddad, Kevin El; Dutoit, Thierry; Soleymani, Mohammad
Social-MAE: A Transformer-Based Multimodal Autoencoder for Face and Voice Proceedings Article
In: 2024 IEEE 18th International Conference on Automatic Face and Gesture Recognition (FG), pp. 1–5, IEEE, Istanbul, Turkiye, 2024, ISBN: 9798350394948.
@inproceedings{bohy_social-mae_2024,
title = {Social-MAE: A Transformer-Based Multimodal Autoencoder for Face and Voice},
author = {Hugo Bohy and Minh Tran and Kevin El Haddad and Thierry Dutoit and Mohammad Soleymani},
url = {https://ieeexplore.ieee.org/document/10581940/},
doi = {10.1109/FG59268.2024.10581940},
isbn = {9798350394948},
year = {2024},
date = {2024-05-01},
urldate = {2024-07-18},
booktitle = {2024 IEEE 18th International Conference on Automatic Face and Gesture Recognition (FG)},
pages = {1–5},
publisher = {IEEE},
address = {Istanbul, Turkiye},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Liu, Rong; Xu, Rui; Hu, Yue; Chen, Meida; Feng, Andrew
AtomGS: Atomizing Gaussian Splatting for High-Fidelity Radiance Field Miscellaneous
2024, (Version Number: 2).
@misc{liu_atomgs_2024,
title = {AtomGS: Atomizing Gaussian Splatting for High-Fidelity Radiance Field},
author = {Rong Liu and Rui Xu and Yue Hu and Meida Chen and Andrew Feng},
url = {https://arxiv.org/abs/2405.12369},
doi = {10.48550/ARXIV.2405.12369},
year = {2024},
date = {2024-05-01},
urldate = {2024-07-11},
publisher = {arXiv},
abstract = {3D Gaussian Splatting (3DGS) has recently advanced radiance field reconstruction by offering superior capabilities for novel view synthesis and real-time rendering speed. However, its strategy of blending optimization and adaptive density control might lead to sub-optimal results; it can sometimes yield noisy geometry and blurry artifacts due to prioritizing optimizing large Gaussians at the cost of adequately densifying smaller ones. To address this, we introduce AtomGS, consisting of Atomized Proliferation and Geometry-Guided Optimization. The Atomized Proliferation constrains ellipsoid Gaussians of various sizes into more uniform-sized Atom Gaussians. The strategy enhances the representation of areas with fine features by placing greater emphasis on densification in accordance with scene details. In addition, we proposed a Geometry-Guided Optimization approach that incorporates an Edge-Aware Normal Loss. This optimization method effectively smooths flat surfaces while preserving intricate details. Our evaluation shows that AtomGS outperforms existing state-of-the-art methods in rendering quality. Additionally, it achieves competitive accuracy in geometry reconstruction and offers a significant improvement in training speed over other SDF-based methods. More interactive demos can be found in our website (https://rongliu-leo.github.io/AtomGS/).},
note = {Version Number: 2},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Chang, Di; Shi, Yichun; Gao, Quankai; Fu, Jessica; Xu, Hongyi; Song, Guoxian; Yan, Qing; Zhu, Yizhe; Yang, Xiao; Soleymani, Mohammad
MagicPose: Realistic Human Poses and Facial Expressions Retargeting with Identity-aware Diffusion Miscellaneous
2024, (arXiv:2311.12052 [cs]).
@misc{chang_magicpose_2024,
title = {MagicPose: Realistic Human Poses and Facial Expressions Retargeting with Identity-aware Diffusion},
author = {Di Chang and Yichun Shi and Quankai Gao and Jessica Fu and Hongyi Xu and Guoxian Song and Qing Yan and Yizhe Zhu and Xiao Yang and Mohammad Soleymani},
url = {http://arxiv.org/abs/2311.12052},
year = {2024},
date = {2024-05-01},
urldate = {2024-07-18},
publisher = {arXiv},
abstract = {In this work, we propose MagicPose, a diffusion-based model for 2D human pose and facial expression retargeting. Specifically, given a reference image, we aim to generate a person's new images by controlling the poses and facial expressions while keeping the identity unchanged. To this end, we propose a two-stage training strategy to disentangle human motions and appearance (e.g., facial expressions, skin tone and dressing), consisting of (1) the pre-training of an appearance-control block and (2) learning appearance-disentangled pose control. Our novel design enables robust appearance control over generated human images, including body, facial attributes, and even background. By leveraging the prior knowledge of image diffusion models, MagicPose generalizes well to unseen human identities and complex poses without the need for additional fine-tuning. Moreover, the proposed model is easy to use and can be considered as a plug-in module/extension to Stable Diffusion. The code is available at: https://github.com/Boese0601/MagicDance},
note = {arXiv:2311.12052 [cs]},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Koresh, Caleb; Ustun, Volkan; Kumar, Rajay; Aris, Tim
Improving Reinforcement Learning Experiments in Unity through Waypoint Utilization Journal Article
In: FLAIRS, vol. 37, 2024, ISSN: 2334-0762.
@article{koresh_improving_2024,
title = {Improving Reinforcement Learning Experiments in Unity through Waypoint Utilization},
author = {Caleb Koresh and Volkan Ustun and Rajay Kumar and Tim Aris},
url = {https://journals.flvc.org/FLAIRS/article/view/135571},
doi = {10.32473/flairs.37.1.135571},
issn = {2334-0762},
year = {2024},
date = {2024-05-01},
urldate = {2024-08-13},
journal = {FLAIRS},
volume = {37},
abstract = {Multi-agent Reinforcement Learning (MARL) models teams of agents that learn by dynamically interacting with an environment and each other, presenting opportunities to train adaptive models for team-based scenarios. However, MARL algorithms pose substantial challenges due to their immense computational requirements. This paper introduces an automatically generated waypoint-based movement system to abstract and simplify complex environments in Unity while allowing agents to learn strategic cooperation. To demonstrate the effectiveness of our approach, we utilized a simple scenario with heterogeneous roles in each team. We trained this scenario on variations of realistic terrains and compared learning between fine-grained (almost) continuous and waypoint-based movement systems. Our results indicate efficiency in learning and improved performance with waypoint-based navigation. Furthermore, our results show that waypoint-based movement systems can effectively learn differentiated behavior policies for heterogeneous roles in these experiments. These early exploratory results point out the potential of waypoint-based navigation for reducing the computational costs of developing and training MARL models in complex environments. The complete project with all scenarios and results is available on GitHub: https://github.com/HATS-ICT/ml-agents-dodgeball-env-ICT.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Aris, Timothy; Ustun, Volkan; Kumar, Rajay
Training Reinforcement Learning Agents to React to an Ambush for Military Simulations Journal Article
In: FLAIRS, vol. 37, 2024, ISSN: 2334-0762.
@article{aris_training_2024,
title = {Training Reinforcement Learning Agents to React to an Ambush for Military Simulations},
author = {Timothy Aris and Volkan Ustun and Rajay Kumar},
url = {https://journals.flvc.org/FLAIRS/article/view/135578},
doi = {10.32473/flairs.37.1.135578},
issn = {2334-0762},
year = {2024},
date = {2024-05-01},
urldate = {2024-08-13},
journal = {FLAIRS},
volume = {37},
abstract = {There is a need for realistic Opposing Forces (OPFOR)behavior in military training simulations. Current trainingsimulations generally only have simple, non-adaptivebehaviors, requiring human instructors to play the role ofOPFOR in any complicated scenario. This poster addressesthis need by focusing on a specific scenario: trainingreinforcement learning agents to react to an ambush. Itproposes a novel way to check for occlusion algorithmically.It shows vector fields showing the agent’s actions throughthe course of a training run. It shows that a single agentswitching between multiple goals is possible, at least in asimplified environment. Such an approach could reduce theneed to develop different agents for different scenarios.Finally, it shows a competent agent trained on a simplifiedReact to Ambush scenario, demonstrating the plausibility ofa scaled-up version.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Liu, Lixing; Ustun, Volkan; Kumar, Rajay
Leveraging Organizational Hierarchy to Simplify Reward Design in Cooperative Multi-agent Reinforcement Learning Journal Article
In: FLAIRS, vol. 37, 2024, ISSN: 2334-0762.
@article{liu_leveraging_2024,
title = {Leveraging Organizational Hierarchy to Simplify Reward Design in Cooperative Multi-agent Reinforcement Learning},
author = {Lixing Liu and Volkan Ustun and Rajay Kumar},
url = {https://journals.flvc.org/FLAIRS/article/view/135588},
doi = {10.32473/flairs.37.1.135588},
issn = {2334-0762},
year = {2024},
date = {2024-05-01},
urldate = {2024-08-13},
journal = {FLAIRS},
volume = {37},
abstract = {The effectiveness of multi-agent reinforcement learning (MARL) hinges largely on the meticulous arrangement of objectives. Yet, conventional MARL methods might not completely harness the inherent structures present in environmental states and agent relationships for goal organization. This study is conducted within the domain of military training simulations, which are typically characterized by complexity, heterogeneity, non-stationary and doctrine-driven environments with a clear organizational hierarchy and a top-down chain of command. This research investigates the approximation and integration of the organizational hierarchy into MARL for cooperative training scenarios, with the goal of streamlining the processes of reward engineering and enhancing team coordination. In the preliminary experiments, we employed two-tiered commander-subordinate feudal hierarchical (CSFH) networks to separate the prioritized team goal and individual goals. The empirical results demonstrate that the proposed framework enhances learning efficiency. It guarantees the learning of a prioritized policy for the commander agent and encourages subordinate agents to explore areas of interest more frequently, guided by appropriate soft constraints imposed by the commander.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Lukin, Stephanie M; Bonial, Claire; Marge, Matthew; Hudson, Taylor; Hayes, Cory J.; Pollard, Kimberly; Baker, Anthony L.; Foots, Ashley; Artstein, Ron; Gervits, Felix; Abrams, Mitchell; Cassidy, Henry; Donatelli, Lucia; Leuski, Anton; Hill, Susan G.; Traum, David; Voss, Clare
SCOUT: A Situated and Multi-Modal Human-Robot Dialogue Corpus Journal Article
In: pp. 14445 - 144458, 2024.
@article{lukin-etal-2024-scout-situated,
title = {SCOUT: A Situated and Multi-Modal Human-Robot Dialogue Corpus},
author = {Stephanie M Lukin and Claire Bonial and Matthew Marge and Taylor Hudson and Cory J. Hayes and Kimberly Pollard and Anthony L. Baker and Ashley Foots and Ron Artstein and Felix Gervits and Mitchell Abrams and Henry Cassidy and Lucia Donatelli and Anton Leuski and Susan G. Hill and David Traum and Clare Voss},
url = {https://aclanthology.org/2024.lrec-main.1259},
year = {2024},
date = {2024-05-01},
pages = {14445 - 144458},
abstract = {We introduce the Situated Corpus Of Understanding Transactions (SCOUT), a multi-modal collection of human-robot dialogue in the task domain of collaborative exploration. The corpus was constructed from multiple Wizard-of-Oz experiments where human participants gave verbal instructions to a remotely-located robot to move and gather information about its surroundings. SCOUT contains 89,056 utterances and 310,095 words from 278 dialogues averaging 320 utterances per dialogue. The dialogues are aligned with the multi-modal data streams available during the experiments: 5,785 images and 30 maps. The corpus has been annotated with Abstract Meaning Representation and Dialogue-AMR to identify the speaker’s intent and meaning within an utterance, and with Transactional Units and Relations to track relationships between utterances to reveal patterns of the Dialogue Structure. We describe how the corpus and its annotations have been used to develop autonomous human-robot systems and enable research in open questions of how humans speak to robots. We release this corpus to accelerate progress in autonomous, situated, human-robot dialogue, especially in the context of navigation tasks where details about the environment need to be discovered.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
West, Taylor Nicole; Prinzing, Michael; Garton, Catherine; Berman, Catherine J.; Zhou, Jieni; Hale, James; Gratch, Jonathan; Fredrickson, Barbara
2024.
@misc{west_improving_2024,
title = {Improving Social Connection with Weak Ties and Strangers: Effects of a New Micro-Intervention on Interaction Quality and Social Behavior},
author = {Taylor Nicole West and Michael Prinzing and Catherine Garton and Catherine J. Berman and Jieni Zhou and James Hale and Jonathan Gratch and Barbara Fredrickson},
url = {https://osf.io/ytjr6},
doi = {10.31234/osf.io/ytjr6},
year = {2024},
date = {2024-05-01},
urldate = {2024-06-25},
abstract = {We propose that the emotional quality of people’s interactions with acquaintances (i.e., weak ties) and strangers contributes to well-being. We test whether a new micro-intervention can raise the quality of these interactions. We randomized young adults (N = 335) to this connectedness micro-intervention or a control intervention. Both interventions were delivered via a psychoeducational video followed by a brief conversation with a virtual human, with whom participants developed if-then plans to carry out their assigned behavioral goal. Pre-intervention, high-quality weak-tie and stranger interactions were associated with lower loneliness and greater mental health independent of strong-tie interaction quality. Experimental data showed the connectedness intervention improved the emotional quality of participants' interactions with weak ties and strangers over two days, evident in participants’ episodic self-reports and faster in-lab conversational response time. Discussion centers on implications for developing scalable behavioral interventions to improve well-being.},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Zhang, Hao; Chang, Di; Li, Fang; Soleymani, Mohammad; Ahuja, Narendra
MagicPose4D: Crafting Articulated Models with Appearance and Motion Control Miscellaneous
2024, (Version Number: 1).
@misc{zhang_magicpose4d_2024,
title = {MagicPose4D: Crafting Articulated Models with Appearance and Motion Control},
author = {Hao Zhang and Di Chang and Fang Li and Mohammad Soleymani and Narendra Ahuja},
url = {https://arxiv.org/abs/2405.14017},
doi = {10.48550/ARXIV.2405.14017},
year = {2024},
date = {2024-05-01},
urldate = {2024-06-25},
publisher = {arXiv},
abstract = {With the success of 2D and 3D visual generative models, there is growing interest in generating 4D content. Existing methods primarily rely on text prompts to produce 4D content, but they often fall short of accurately defining complex or rare motions. To address this limitation, we propose MagicPose4D, a novel framework for refined control over both appearance and motion in 4D generation. Unlike traditional methods, MagicPose4D accepts monocular videos as motion prompts, enabling precise and customizable motion generation. MagicPose4D comprises two key modules:
i) Dual-Phase 4D Reconstruction Modulevphantom which operates in two phases. The first phase focuses on capturing the model's shape using accurate 2D supervision and less accurate but geometrically informative 3D pseudo-supervision without imposing skeleton constraints. The second phase refines the model using more accurate pseudo-3D supervision, obtained in the first phase and introduces kinematic chain-based skeleton constraints to ensure physical plausibility. Additionally, we propose a Global-local Chamfer loss that aligns the overall distribution of predicted mesh vertices with the supervision while maintaining part-level alignment without extra annotations.
ii) Cross-category Motion Transfer Modulevphantom leverages the predictions from the 4D reconstruction module and uses a kinematic-chain-based skeleton to achieve cross-category motion transfer. It ensures smooth transitions between frames through dynamic rigidity, facilitating robust generalization without additional training.
Through extensive experiments, we demonstrate that MagicPose4D significantly improves the accuracy and consistency of 4D content generation, outperforming existing methods in various benchmarks.},
note = {Version Number: 1},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
i) Dual-Phase 4D Reconstruction Modulevphantom which operates in two phases. The first phase focuses on capturing the model's shape using accurate 2D supervision and less accurate but geometrically informative 3D pseudo-supervision without imposing skeleton constraints. The second phase refines the model using more accurate pseudo-3D supervision, obtained in the first phase and introduces kinematic chain-based skeleton constraints to ensure physical plausibility. Additionally, we propose a Global-local Chamfer loss that aligns the overall distribution of predicted mesh vertices with the supervision while maintaining part-level alignment without extra annotations.
ii) Cross-category Motion Transfer Modulevphantom leverages the predictions from the 4D reconstruction module and uses a kinematic-chain-based skeleton to achieve cross-category motion transfer. It ensures smooth transitions between frames through dynamic rigidity, facilitating robust generalization without additional training.
Through extensive experiments, we demonstrate that MagicPose4D significantly improves the accuracy and consistency of 4D content generation, outperforming existing methods in various benchmarks.
Jones, Brennan; Xu, Yan; Li, Qisheng; Scherer, Stefan
Designing a Proactive Context-Aware AI Chatbot for People's Long-Term Goals Proceedings Article
In: Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, pp. 1–7, ACM, Honolulu HI USA, 2024, ISBN: 9798400703317.
@inproceedings{jones_designing_2024,
title = {Designing a Proactive Context-Aware AI Chatbot for People's Long-Term Goals},
author = {Brennan Jones and Yan Xu and Qisheng Li and Stefan Scherer},
url = {https://dl.acm.org/doi/10.1145/3613905.3650912},
doi = {10.1145/3613905.3650912},
isbn = {9798400703317},
year = {2024},
date = {2024-05-01},
urldate = {2024-06-25},
booktitle = {Extended Abstracts of the CHI Conference on Human Factors in Computing Systems},
pages = {1–7},
publisher = {ACM},
address = {Honolulu HI USA},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Chemburkar, Ankur; Gordon, Andrew; Feng, Andrew
Evaluating Vision-Language Models on the TriangleCOPA Benchmark Journal Article
In: FLAIRS-37, vol. 37, 2024.
@article{chemburkar_evaluating_2024,
title = {Evaluating Vision-Language Models on the TriangleCOPA Benchmark},
author = {Ankur Chemburkar and Andrew Gordon and Andrew Feng},
year = {2024},
date = {2024-05-01},
journal = {FLAIRS-37},
volume = {37},
abstract = {The TriangleCOPA benchmark consists of 100 textual questions with videos depicting the movements of simple shapes in the style of the classic social-psychology film created by Fritz Heider and Marianne Simmel in 1944. In our experiments, we investigate the performance of current vision-language models on this challenging benchmark, assessing the capability of these models for visual anthropomorphism and abstract interpretation.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Mozgai, Sharon A; Kaurloto, Cari; Winn, Jade G; Leeds, Andrew; Beland, Sarah; Sookiassian, Arman; Hartholt, Arno
Accelerating Scoping Reviews: A Case Study in the User-Centered Design of an AI-Enabled Interdisciplinary Research Tool Proceedings Article
In: Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, pp. 1–8, ACM, Honolulu HI USA, 2024, ISBN: 9798400703317.
@inproceedings{mozgai_accelerating_2024,
title = {Accelerating Scoping Reviews: A Case Study in the User-Centered Design of an AI-Enabled Interdisciplinary Research Tool},
author = {Sharon A Mozgai and Cari Kaurloto and Jade G Winn and Andrew Leeds and Sarah Beland and Arman Sookiassian and Arno Hartholt},
url = {https://dl.acm.org/doi/10.1145/3613905.3637110},
doi = {10.1145/3613905.3637110},
isbn = {9798400703317},
year = {2024},
date = {2024-05-01},
urldate = {2024-06-18},
booktitle = {Extended Abstracts of the CHI Conference on Human Factors in Computing Systems},
pages = {1–8},
publisher = {ACM},
address = {Honolulu HI USA},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Soleymani, Mohammad; Kumano, Shiro; Provost, Emily Mower; Bianchi-Berthouze, Nadia; Sano, Akane; Suzuki, Kenji
Guest Editorial Best of ACII 2021 Journal Article
In: IEEE Trans. Affective Comput., vol. 15, no. 2, pp. 376–379, 2024, ISSN: 1949-3045, 2371-9850.
@article{soleymani_guest_2024,
title = {Guest Editorial Best of ACII 2021},
author = {Mohammad Soleymani and Shiro Kumano and Emily Mower Provost and Nadia Bianchi-Berthouze and Akane Sano and Kenji Suzuki},
url = {https://ieeexplore.ieee.org/document/10542496/},
doi = {10.1109/TAFFC.2024.3389249},
issn = {1949-3045, 2371-9850},
year = {2024},
date = {2024-04-01},
urldate = {2024-06-25},
journal = {IEEE Trans. Affective Comput.},
volume = {15},
number = {2},
pages = {376–379},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Zhang, Hui; Kuang, Bingran; Zhao, Yajie
Camera Calibration using a Single View of a Symmetric Object Proceedings Article
In: ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2705–2709, IEEE, Seoul, Korea, Republic of, 2024, ISBN: 9798350344851.
@inproceedings{zhang_camera_2024,
title = {Camera Calibration using a Single View of a Symmetric Object},
author = {Hui Zhang and Bingran Kuang and Yajie Zhao},
url = {https://ieeexplore.ieee.org/document/10446005/},
doi = {10.1109/ICASSP48485.2024.10446005},
isbn = {9798350344851},
year = {2024},
date = {2024-04-01},
urldate = {2024-06-25},
booktitle = {ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)},
pages = {2705–2709},
publisher = {IEEE},
address = {Seoul, Korea, Republic of},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Rizzo, Albert Skip; Hartholt, Arno; Mozgai, Sharon
Settling the Score: Virtual Reality as a Tool to Enhance Trauma-Focused Therapy for PTSD Book Section
In: Rich, Grant J.; Kumar, V. K.; Farley, Frank H. (Ed.): Handbook of Media Psychology, pp. 187–213, Springer Nature Switzerland, Cham, 2024, ISBN: 978-3-031-56536-6 978-3-031-56537-3.
@incollection{rich_settling_2024,
title = {Settling the Score: Virtual Reality as a Tool to Enhance Trauma-Focused Therapy for PTSD},
author = {Albert Skip Rizzo and Arno Hartholt and Sharon Mozgai},
editor = {Grant J. Rich and V. K. Kumar and Frank H. Farley},
url = {https://link.springer.com/10.1007/978-3-031-56537-3_14},
doi = {10.1007/978-3-031-56537-3_14},
isbn = {978-3-031-56536-6 978-3-031-56537-3},
year = {2024},
date = {2024-04-01},
urldate = {2024-06-18},
booktitle = {Handbook of Media Psychology},
pages = {187–213},
publisher = {Springer Nature Switzerland},
address = {Cham},
keywords = {},
pubstate = {published},
tppubtype = {incollection}
}
Goh, Crystal; Ma, Yu; Rizzo, Albert
Normative performance data on visual attention in neurotypical children: virtual reality assessment of cognitive and psychomotor development Journal Article
In: Front. Virtual Real., vol. 5, pp. 1309176, 2024, ISSN: 2673-4192.
@article{goh_normative_2024,
title = {Normative performance data on visual attention in neurotypical children: virtual reality assessment of cognitive and psychomotor development},
author = {Crystal Goh and Yu Ma and Albert Rizzo},
url = {https://www.frontiersin.org/articles/10.3389/frvir.2024.1309176/full},
doi = {10.3389/frvir.2024.1309176},
issn = {2673-4192},
year = {2024},
date = {2024-04-01},
urldate = {2024-04-16},
journal = {Front. Virtual Real.},
volume = {5},
pages = {1309176},
abstract = {Introduction:
Virtual Reality (VR) is revolutionizing healthcare research and practice by offering innovative methodologies across various clinical conditions. Advances in VR technology enable the creation of controllable, multisensory 3D environments, making it an appealing tool for capturing and quantifying behavior in realistic scenarios. This paper details the application of VR as a tool for neurocognitive evaluation, specifically in attention process assessment, an area of relevance for informing the diagnosis of childhood health conditions such as Attention Deficit Hyperactivity Disorder (ADHD).
Methods:
The data presented focuses on attention performance results from a large sample (
n = 837) of neurotypical male and female children (ages 6–13) tested on a visual continuous performance task, administered within an immersive VR classroom environment. This data was collected to create a normative baseline database for use to inform comparisons with the performances of children with ADHD to support diagnostic decision-making in this area.
Results:
Results indicate systematic improvements on most metrics across the age span, and sex differences are noted on key variables thought to reflect differential measures of hyperactivity and inattention in children with ADHD. Results support VR technology as a safe and viable option for testing attention processes in children, under stimulus conditions that closely mimic ecologically relevant challenges found in everyday life.
Discussion:
In response to these stimulus conditions, VR can support advanced methods for capturing and quantifying users’ behavioral responses. VR offers a more systematic and objective approach for clinical assessment and intervention and provides conceptual support for its use in a wide variety of healthcare contexts.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Virtual Reality (VR) is revolutionizing healthcare research and practice by offering innovative methodologies across various clinical conditions. Advances in VR technology enable the creation of controllable, multisensory 3D environments, making it an appealing tool for capturing and quantifying behavior in realistic scenarios. This paper details the application of VR as a tool for neurocognitive evaluation, specifically in attention process assessment, an area of relevance for informing the diagnosis of childhood health conditions such as Attention Deficit Hyperactivity Disorder (ADHD).
Methods:
The data presented focuses on attention performance results from a large sample (
n = 837) of neurotypical male and female children (ages 6–13) tested on a visual continuous performance task, administered within an immersive VR classroom environment. This data was collected to create a normative baseline database for use to inform comparisons with the performances of children with ADHD to support diagnostic decision-making in this area.
Results:
Results indicate systematic improvements on most metrics across the age span, and sex differences are noted on key variables thought to reflect differential measures of hyperactivity and inattention in children with ADHD. Results support VR technology as a safe and viable option for testing attention processes in children, under stimulus conditions that closely mimic ecologically relevant challenges found in everyday life.
Discussion:
In response to these stimulus conditions, VR can support advanced methods for capturing and quantifying users’ behavioral responses. VR offers a more systematic and objective approach for clinical assessment and intervention and provides conceptual support for its use in a wide variety of healthcare contexts.
Soleymani, Mohammad; Rahmani, Mehdi; Bigdeli, Nooshin
Robust Tube-Based Reference Tracking Nonlinear Model Predictive Control for Wind Turbines Journal Article
In: IEEE Trans. Automat. Sci. Eng., pp. 1–13, 2024, ISSN: 1545-5955, 1558-3783.
@article{soleymani_robust_2024,
title = {Robust Tube-Based Reference Tracking Nonlinear Model Predictive Control for Wind Turbines},
author = {Mohammad Soleymani and Mehdi Rahmani and Nooshin Bigdeli},
url = {https://ieeexplore.ieee.org/document/10495787/},
doi = {10.1109/TASE.2024.3385714},
issn = {1545-5955, 1558-3783},
year = {2024},
date = {2024-04-01},
urldate = {2024-04-16},
journal = {IEEE Trans. Automat. Sci. Eng.},
pages = {1–13},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Gurney, Nikolos; Loewenstein, George; Chater, Nick
Conversational technology and reactions to withheld information Journal Article
In: PLoS ONE, vol. 19, no. 4, pp. e0301382, 2024, ISSN: 1932-6203.
@article{gurney_conversational_2024,
title = {Conversational technology and reactions to withheld information},
author = {Nikolos Gurney and George Loewenstein and Nick Chater},
editor = {Petre Caraiani},
url = {https://dx.plos.org/10.1371/journal.pone.0301382},
doi = {10.1371/journal.pone.0301382},
issn = {1932-6203},
year = {2024},
date = {2024-04-01},
urldate = {2024-04-16},
journal = {PLoS ONE},
volume = {19},
number = {4},
pages = {e0301382},
abstract = {People frequently face decisions that require making inferences about withheld information. The advent of large language models coupled with conversational technology, e.g., Alexa, Siri, Cortana, and the Google Assistant, is changing the mode in which people make these inferences. We demonstrate that conversational modes of information provision, relative to traditional digital media, result in more critical responses to withheld information, including: (1) a reduction in evaluations of a product or service for which information is withheld and (2) an increased likelihood of recalling that information was withheld. These effects are robust across multiple conversational modes: a recorded phone conversation, an unfolding chat conversation, and a conversation script. We provide further evidence that these effects hold for conversations with the Google Assistant, a prominent conversational technology. The experimental results point to participants’ intuitions about why the information was withheld as the driver of the effect.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Harris, Vera; Braggs, Robert; Traum, David
I’m not sure I heard you right, but I think I know what you mean – investigations into the impact of speech recognition errors on response selection for a virtual human. Proceedings Article
In: Sapporo Japan, 2024.
@inproceedings{harris_im_2024,
title = {I’m not sure I heard you right, but I think I know what you mean – investigations into the impact of speech recognition errors on response selection for a virtual human.},
author = {Vera Harris and Robert Braggs and David Traum},
url = {chrome-extension://efaidnbmnnnibpcajpcglclefindmkaj/https://people.ict.usc.edu/~traum/Papers/23-harris-iwsds2024.pdf},
year = {2024},
date = {2024-03-01},
address = {Sapporo Japan},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Filter
2024
Gurney, Nikolos; Pynadath, David V.; Ustun, Volkan
Spontaneous Theory of Mind for Artificial Intelligence Journal Article
In: 2024, (Publisher: [object Object] Version Number: 1).
Abstract | Links | BibTeX | Tags: AI, DTIC, Social Simulation, UARC
@article{gurney_spontaneous_2024,
title = {Spontaneous Theory of Mind for Artificial Intelligence},
author = {Nikolos Gurney and David V. Pynadath and Volkan Ustun},
url = {https://arxiv.org/abs/2402.13272},
doi = {10.48550/ARXIV.2402.13272},
year = {2024},
date = {2024-02-01},
urldate = {2024-03-14},
abstract = {Existing approaches to Theory of Mind (ToM) in Artificial Intelligence (AI) overemphasize prompted, or cue-based, ToM, which may limit our collective ability to develop Artificial Social Intelligence (ASI). Drawing from research in computer science, cognitive science, and related disciplines, we contrast prompted ToM with what we call spontaneous ToM – reasoning about others' mental states that is grounded in unintentional, possibly uncontrollable cognitive functions. We argue for a principled approach to studying and developing AI ToM and suggest that a robust, or general, ASI will respond to prompts textbackslashtextitand spontaneously engage in social reasoning.},
note = {Publisher: [object Object]
Version Number: 1},
keywords = {AI, DTIC, Social Simulation, UARC},
pubstate = {published},
tppubtype = {article}
}
Kwon, Deuksin; Weiss, Emily; Kulshrestha, Tara; Chawla, Kushal; Lucas, Gale M.; Gratch, Jonathan
Are LLMs Effective Negotiators? Systematic Evaluation of the Multifaceted Capabilities of LLMs in Negotiation Dialogues Miscellaneous
2024, (arXiv:2402.13550 [cs]).
Abstract | Links | BibTeX | Tags: AI, Virtual Humans
@misc{kwon_are_2024,
title = {Are LLMs Effective Negotiators? Systematic Evaluation of the Multifaceted Capabilities of LLMs in Negotiation Dialogues},
author = {Deuksin Kwon and Emily Weiss and Tara Kulshrestha and Kushal Chawla and Gale M. Lucas and Jonathan Gratch},
url = {http://arxiv.org/abs/2402.13550},
year = {2024},
date = {2024-02-01},
urldate = {2024-03-14},
publisher = {arXiv},
abstract = {A successful negotiation demands a deep comprehension of the conversation context, Theory-of-Mind (ToM) skills to infer the partner's motives, as well as strategic reasoning and effective communication, making it challenging for automated systems. Given the remarkable performance of LLMs across a variety of NLP tasks, in this work, we aim to understand how LLMs can advance different aspects of negotiation research, ranging from designing dialogue systems to providing pedagogical feedback and scaling up data collection practices. To this end, we devise a methodology to analyze the multifaceted capabilities of LLMs across diverse dialogue scenarios covering all the time stages of a typical negotiation interaction. Our analysis adds to the increasing evidence for the superiority of GPT-4 across various tasks while also providing insights into specific tasks that remain difficult for LLMs. For instance, the models correlate poorly with human players when making subjective assessments about the negotiation dialogues and often struggle to generate responses that are contextually appropriate as well as strategically advantageous.},
note = {arXiv:2402.13550 [cs]},
keywords = {AI, Virtual Humans},
pubstate = {published},
tppubtype = {misc}
}
Marti, Deniz; Hanrahan, David; Sanchez-Triana, Ernesto; Wells, Mona; Corra, Lilian; Hu, Howard; Breysse, Patrick N.; Laborde, Amalia; Caravanos, Jack; Bertollini, Roberto; Porterfield, Kate; Fuller, Richard
Structured Expert Judgement Approach of the Health Impact of Various Chemicals and Classes of Chemicals Technical Report
Public and Global Health 2024.
Abstract | Links | BibTeX | Tags:
@techreport{marti_structured_2024,
title = {Structured Expert Judgement Approach of the Health Impact of Various Chemicals and Classes of Chemicals},
author = {Deniz Marti and David Hanrahan and Ernesto Sanchez-Triana and Mona Wells and Lilian Corra and Howard Hu and Patrick N. Breysse and Amalia Laborde and Jack Caravanos and Roberto Bertollini and Kate Porterfield and Richard Fuller},
url = {http://medrxiv.org/lookup/doi/10.1101/2024.01.30.24301863},
doi = {10.1101/2024.01.30.24301863},
year = {2024},
date = {2024-02-01},
urldate = {2024-02-21},
institution = {Public and Global Health},
abstract = {ABSTRACT
Introduction
Chemical contamination and pollution are an ongoing threat to human health and the environment. The concern over the consequences of chemical exposures at the global level continues to grow. Because resources are constrained, there is a need to prioritize interventions focused on the greatest health impact. Data, especially related to chemical exposures, are rarely available for most substances of concern, and alternate methods to evaluate their impact are needed.
Structured Expert Judgment (SEJ) Process
A Structured Expert Judgment
3
process was performed to provide plausible estimates of health impacts for 16 commonly found pollutants: asbestos, arsenic, benzene, chromium, cadmium, dioxins, fluoride, highly hazardous pesticides (HHPs), lead, mercury, polycyclic-aromatic hydrocarbons (PAHs), polychlorinated biphenyls (PCBs), Per- and Polyfluorinated Substances (PFAs), phthalates, endocrine disrupting chemicals (EDCs), and brominated flame retardants (BRFs). This process, undertaken by sector experts, weighed individual estimations of the probable global health scale health impacts of each pollutant using objective estimates of the expert opinions’ statistical accuracy and informativeness.
Main Findings
The foremost substances, in terms of mean projected annual total deaths, were lead, asbestos, arsenic, and HHPs. Lead surpasses the others by a large margin, with an estimated median value of 1.7 million deaths annually. The three other substances averaged between 136,000 and 274,000 deaths per year. Of the 12 other chemicals evaluated, none reached an estimated annual death count exceeding 100,000. These findings underscore the importance of prioritizing available resources on reducing and remediating the impacts of these key pollutants.
Range of Health Impacts
Based on the evidence available, experts concluded some of the more notorious chemical pollutants, such as PCBs and dioxin, do not result in high levels of human health impact from a global scale perspective. However, the chemical toxicity of some compounds released in recent decades, such as Endocrine Disrupters and PFAs, cannot be ignored, even if current impacts are limited. Moreover, the impact of some chemicals may be disproportionately large in some geographic areas. Continued research and monitoring are essential; and a preventative approach is needed for chemicals.
Future Directions
These results, and potential similar analyses of other chemicals, are provided as inputs to ongoing discussions about priority setting for global chemicals and pollution management. Furthermore, we suggest that this SEJ process be repeated periodically as new information becomes available.},
keywords = {},
pubstate = {published},
tppubtype = {techreport}
}
Introduction
Chemical contamination and pollution are an ongoing threat to human health and the environment. The concern over the consequences of chemical exposures at the global level continues to grow. Because resources are constrained, there is a need to prioritize interventions focused on the greatest health impact. Data, especially related to chemical exposures, are rarely available for most substances of concern, and alternate methods to evaluate their impact are needed.
Structured Expert Judgment (SEJ) Process
A Structured Expert Judgment
3
process was performed to provide plausible estimates of health impacts for 16 commonly found pollutants: asbestos, arsenic, benzene, chromium, cadmium, dioxins, fluoride, highly hazardous pesticides (HHPs), lead, mercury, polycyclic-aromatic hydrocarbons (PAHs), polychlorinated biphenyls (PCBs), Per- and Polyfluorinated Substances (PFAs), phthalates, endocrine disrupting chemicals (EDCs), and brominated flame retardants (BRFs). This process, undertaken by sector experts, weighed individual estimations of the probable global health scale health impacts of each pollutant using objective estimates of the expert opinions’ statistical accuracy and informativeness.
Main Findings
The foremost substances, in terms of mean projected annual total deaths, were lead, asbestos, arsenic, and HHPs. Lead surpasses the others by a large margin, with an estimated median value of 1.7 million deaths annually. The three other substances averaged between 136,000 and 274,000 deaths per year. Of the 12 other chemicals evaluated, none reached an estimated annual death count exceeding 100,000. These findings underscore the importance of prioritizing available resources on reducing and remediating the impacts of these key pollutants.
Range of Health Impacts
Based on the evidence available, experts concluded some of the more notorious chemical pollutants, such as PCBs and dioxin, do not result in high levels of human health impact from a global scale perspective. However, the chemical toxicity of some compounds released in recent decades, such as Endocrine Disrupters and PFAs, cannot be ignored, even if current impacts are limited. Moreover, the impact of some chemicals may be disproportionately large in some geographic areas. Continued research and monitoring are essential; and a preventative approach is needed for chemicals.
Future Directions
These results, and potential similar analyses of other chemicals, are provided as inputs to ongoing discussions about priority setting for global chemicals and pollution management. Furthermore, we suggest that this SEJ process be repeated periodically as new information becomes available.
Yu, Zifan; Tavakoli, Erfan Bank; Chen, Meida; You, Suya; Rao, Raghuveer; Agarwal, Sanjeev; Ren, Fengbo
TokenMotion: Motion-Guided Vision Transformer for Video Camouflaged Object Detection Via Learnable Token Selection Miscellaneous
2024, (arXiv:2311.02535 [cs]).
Abstract | Links | BibTeX | Tags: Narrative
@misc{yu_tokenmotion_2024,
title = {TokenMotion: Motion-Guided Vision Transformer for Video Camouflaged Object Detection Via Learnable Token Selection},
author = {Zifan Yu and Erfan Bank Tavakoli and Meida Chen and Suya You and Raghuveer Rao and Sanjeev Agarwal and Fengbo Ren},
url = {http://arxiv.org/abs/2311.02535},
year = {2024},
date = {2024-02-01},
urldate = {2024-02-21},
publisher = {arXiv},
abstract = {The area of Video Camouflaged Object Detection (VCOD) presents unique challenges in the field of computer vision due to texture similarities between target objects and their surroundings, as well as irregular motion patterns caused by both objects and camera movement. In this paper, we introduce TokenMotion (TMNet), which employs a transformer-based model to enhance VCOD by extracting motion-guided features using a learnable token selection. Evaluated on the challenging MoCA-Mask dataset, TMNet achieves state-of-the-art performance in VCOD. It outperforms the existing state-of-the-art method by a 12.8% improvement in weighted F-measure, an 8.4% enhancement in S-measure, and a 10.7% boost in mean IoU. The results demonstrate the benefits of utilizing motion-guided features via learnable token selection within a transformer-based framework to tackle the intricate task of VCOD.},
note = {arXiv:2311.02535 [cs]},
keywords = {Narrative},
pubstate = {published},
tppubtype = {misc}
}
Ehsanpour, Mahsa; Reid, Ian; Rezatofighi, Hamid
Social-MAE: Social Masked Autoencoder for Multi-person Motion Representation Learning Miscellaneous
2024, (Version Number: 1).
Abstract | Links | BibTeX | Tags: Social Simulation
@misc{ehsanpour_social-mae_2024,
title = {Social-MAE: Social Masked Autoencoder for Multi-person Motion Representation Learning},
author = {Mahsa Ehsanpour and Ian Reid and Hamid Rezatofighi},
url = {https://arxiv.org/abs/2404.05578},
doi = {10.48550/ARXIV.2404.05578},
year = {2024},
date = {2024-01-01},
urldate = {2024-07-12},
publisher = {arXiv},
abstract = {For a complete comprehension of multi-person scenes, it is essential to go beyond basic tasks like detection and tracking. Higher-level tasks, such as understanding the interactions and social activities among individuals, are also crucial. Progress towards models that can fully understand scenes involving multiple people is hindered by a lack of sufficient annotated data for such high-level tasks. To address this challenge, we introduce Social-MAE, a simple yet effective transformer-based masked autoencoder framework for multi-person human motion data. The framework uses masked modeling to pre-train the encoder to reconstruct masked human joint trajectories, enabling it to learn generalizable and data efficient representations of motion in human crowded scenes. Social-MAE comprises a transformer as the MAE encoder and a lighter-weight transformer as the MAE decoder which operates on multi-person joints' trajectory in the frequency domain. After the reconstruction task, the MAE decoder is replaced with a task-specific decoder and the model is fine-tuned end-to-end for a variety of high-level social tasks. Our proposed model combined with our pre-training approach achieves the state-of-the-art results on various high-level social tasks, including multi-person pose forecasting, social grouping, and social action understanding. These improvements are demonstrated across four popular multi-person datasets encompassing both human 2D and 3D body pose.},
note = {Version Number: 1},
keywords = {Social Simulation},
pubstate = {published},
tppubtype = {misc}
}
Gunasekara, Chulaka; Kim, Seokhwan; D'Haro, Luis Fernando; Rastogi, Abhinav; Chen, Yun-Nung; Eric, Mihail; Hedayatnia, Behnam; Gopalakrishnan, Karthik; Liu, Yang; Huang, Chao-Wei; Hakkani-Tür, Dilek; Li, Jinchao; Zhu, Qi; Luo, Lingxiao; Liden, Lars; Huang, Kaili; Shayandeh, Shahin; Liang, Runze; Peng, Baolin; Zhang, Zheng; Shukla, Swadheen; Huang, Minlie; Gao, Jianfeng; Mehri, Shikib; Feng, Yulan; Gordon, Carla; Alavi, Seyed Hossein; Traum, David; Eskenazi, Maxine; Beirami, Ahmad; Cho, Eunjoon; Crook, Paul A.; De, Ankita; Geramifard, Alborz; Kottur, Satwik; Moon, Seungwhan; Poddar, Shivani; Subba, Rajen
Overview of the Ninth Dialog System Technology Challenge: DSTC9 Journal Article
In: IEEE/ACM Trans. Audio Speech Lang. Process., pp. 1–10, 2024, ISSN: 2329-9290, 2329-9304.
@article{gunasekara_overview_2024,
title = {Overview of the Ninth Dialog System Technology Challenge: DSTC9},
author = {Chulaka Gunasekara and Seokhwan Kim and Luis Fernando D'Haro and Abhinav Rastogi and Yun-Nung Chen and Mihail Eric and Behnam Hedayatnia and Karthik Gopalakrishnan and Yang Liu and Chao-Wei Huang and Dilek Hakkani-Tür and Jinchao Li and Qi Zhu and Lingxiao Luo and Lars Liden and Kaili Huang and Shahin Shayandeh and Runze Liang and Baolin Peng and Zheng Zhang and Swadheen Shukla and Minlie Huang and Jianfeng Gao and Shikib Mehri and Yulan Feng and Carla Gordon and Seyed Hossein Alavi and David Traum and Maxine Eskenazi and Ahmad Beirami and Eunjoon Cho and Paul A. Crook and Ankita De and Alborz Geramifard and Satwik Kottur and Seungwhan Moon and Shivani Poddar and Rajen Subba},
url = {https://ieeexplore.ieee.org/document/10595468/},
doi = {10.1109/TASLP.2024.3426331},
issn = {2329-9290, 2329-9304},
year = {2024},
date = {2024-01-01},
urldate = {2024-08-15},
journal = {IEEE/ACM Trans. Audio Speech Lang. Process.},
pages = {1–10},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Awada, Mohamad; Seyedrezaei, Mirmahdi; Becerik-Gerber, Burcin; Lucas, Gale
Investigating the Interplay between Indoor Environmental Quality and Workers’ Health and Productivity: Preliminary Results Proceedings Article
In: Computing in Civil Engineering 2023, pp. 614–622, American Society of Civil Engineers, Corvallis, Oregon, 2024, ISBN: 978-0-7844-8524-8.
Links | BibTeX | Tags: Virtual Humans
@inproceedings{awada_investigating_2024,
title = {Investigating the Interplay between Indoor Environmental Quality and Workers’ Health and Productivity: Preliminary Results},
author = {Mohamad Awada and Mirmahdi Seyedrezaei and Burcin Becerik-Gerber and Gale Lucas},
url = {https://ascelibrary.org/doi/10.1061/9780784485248.074},
doi = {10.1061/9780784485248.074},
isbn = {978-0-7844-8524-8},
year = {2024},
date = {2024-01-01},
urldate = {2024-04-16},
booktitle = {Computing in Civil Engineering 2023},
pages = {614–622},
publisher = {American Society of Civil Engineers},
address = {Corvallis, Oregon},
keywords = {Virtual Humans},
pubstate = {published},
tppubtype = {inproceedings}
}
Liu, Ruying; Becerik-Gerber, Burçin; Lucas, Gale M.; Busta, Kelly
Development of a VR Training Platform for Active Shooter Incident Preparedness in Healthcare Environments via a Stakeholder-Engaged Process Proceedings Article
In: Computing in Civil Engineering 2023, pp. 45–53, American Society of Civil Engineers, Corvallis, Oregon, 2024, ISBN: 978-0-7844-8523-1.
Links | BibTeX | Tags: Virtual Humans, VR
@inproceedings{liu_development_2024,
title = {Development of a VR Training Platform for Active Shooter Incident Preparedness in Healthcare Environments via a Stakeholder-Engaged Process},
author = {Ruying Liu and Burçin Becerik-Gerber and Gale M. Lucas and Kelly Busta},
url = {https://ascelibrary.org/doi/10.1061/9780784485231.006},
doi = {10.1061/9780784485231.006},
isbn = {978-0-7844-8523-1},
year = {2024},
date = {2024-01-01},
urldate = {2024-04-16},
booktitle = {Computing in Civil Engineering 2023},
pages = {45–53},
publisher = {American Society of Civil Engineers},
address = {Corvallis, Oregon},
keywords = {Virtual Humans, VR},
pubstate = {published},
tppubtype = {inproceedings}
}
Rodrigues, Patrick B.; Becerik-Gerber, Burcin; Soibelman, Lucio; Lucas, Gale M.; Roll, Shawn C.
Virtual Environment for Studying the Effects of Operational and Environmental Sounds on Teleoperated Demolition Proceedings Article
In: Computing in Civil Engineering 2023, pp. 54–61, American Society of Civil Engineers, Corvallis, Oregon, 2024, ISBN: 978-0-7844-8523-1.
Links | BibTeX | Tags: DTIC, Virtual Humans, VR
@inproceedings{rodrigues_virtual_2024,
title = {Virtual Environment for Studying the Effects of Operational and Environmental Sounds on Teleoperated Demolition},
author = {Patrick B. Rodrigues and Burcin Becerik-Gerber and Lucio Soibelman and Gale M. Lucas and Shawn C. Roll},
url = {https://ascelibrary.org/doi/10.1061/9780784485231.007},
doi = {10.1061/9780784485231.007},
isbn = {978-0-7844-8523-1},
year = {2024},
date = {2024-01-01},
urldate = {2024-04-16},
booktitle = {Computing in Civil Engineering 2023},
pages = {54–61},
publisher = {American Society of Civil Engineers},
address = {Corvallis, Oregon},
keywords = {DTIC, Virtual Humans, VR},
pubstate = {published},
tppubtype = {inproceedings}
}
Hartholt, Arno; Leeds, Andrew; Fast, Ed; Sookiassian, Edwin; Kim, Kevin; Beland, Sarah; Kulkarni, Pranav; Mozgai, Sharon
Multidisciplinary Research & Development of Multi-Agents and Virtual Humans Leveraging Integrated Middleware Platforms Proceedings Article
In: 2024.
Abstract | Links | BibTeX | Tags: DTIC, UARC, Virtual Humans
@inproceedings{hartholt_multidisciplinary_2024,
title = {Multidisciplinary Research & Development of Multi-Agents and Virtual Humans Leveraging Integrated Middleware Platforms},
author = {Arno Hartholt and Andrew Leeds and Ed Fast and Edwin Sookiassian and Kevin Kim and Sarah Beland and Pranav Kulkarni and Sharon Mozgai},
url = {https://openaccess.cms-conferences.org/publications/book/978-1-958651-95-7/article/978-1-958651-95-7_33},
doi = {10.54941/ahfe1004497},
year = {2024},
date = {2024-01-01},
urldate = {2024-04-16},
abstract = {The current pace of technological advancements has led to an ever-increasing availability of technologies to investigate and help address the challenges that contemporary society faces today. However, while this trend increases the potential for creating more relevant, effective, and efficient solutions, it also inherently increases the complexity of realizing that potential. Our work aims to manage this complexity through the creation and dissemination of integrated middleware platforms that enable researchers and developers to rapidly prototype novel solutions within the areas of modelling & simulation, virtual humans, and virtual worlds. In this paper, we discuss two related platforms: the Rapid Integration & Development Environment (RIDE) and the Virtual Human Toolkit (VHToolkit). Specifically, we explore two use cases: 1) the development of an authoring tool aimed at domain experts to rapidly create low-echelon military training scenarios, and 2) the development of a virtual human led mHealth wellness and suicide prevention app for veterans.},
keywords = {DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {inproceedings}
}
Joshi, Himanshu; Ustun, Volkan
Augmenting Cognitive Architectures with Large Language Models Journal Article
In: AAAI-SS, vol. 2, no. 1, pp. 281–285, 2024, ISSN: 2994-4317.
Abstract | Links | BibTeX | Tags: Cognitive Architecture
@article{joshi_augmenting_2024,
title = {Augmenting Cognitive Architectures with Large Language Models},
author = {Himanshu Joshi and Volkan Ustun},
url = {https://ojs.aaai.org/index.php/AAAI-SS/article/view/27689},
doi = {10.1609/aaaiss.v2i1.27689},
issn = {2994-4317},
year = {2024},
date = {2024-01-01},
urldate = {2024-04-16},
journal = {AAAI-SS},
volume = {2},
number = {1},
pages = {281–285},
abstract = {A particular fusion of generative models and cognitive architectures is discussed with the help of the Soar and Sigma cognitive architectures. After a brief introduction to cognitive architecture concepts and Large Language Models as exemplar generative AI models, one approach towards their fusion is discussed. This is then analyzed with a summary of potential benefits and extensions needed to existing cognitive architecture that is closest to the proposal.},
keywords = {Cognitive Architecture},
pubstate = {published},
tppubtype = {article}
}
Liu, Ziming; Suen, Christine Wun Ki; Zou, Zhengbo; Chen, Meida; Shi, Yangming
Assessing Workers’ Operational Postures via Egocentric Camera Mapping Proceedings Article
In: Computing in Civil Engineering 2023, pp. 17–24, American Society of Civil Engineers, Corvallis, Oregon, 2024, ISBN: 978-0-7844-8522-4.
Links | BibTeX | Tags: Narrative, STG
@inproceedings{liu_assessing_2024,
title = {Assessing Workers’ Operational Postures via Egocentric Camera Mapping},
author = {Ziming Liu and Christine Wun Ki Suen and Zhengbo Zou and Meida Chen and Yangming Shi},
url = {https://ascelibrary.org/doi/10.1061/9780784485224.003},
doi = {10.1061/9780784485224.003},
isbn = {978-0-7844-8522-4},
year = {2024},
date = {2024-01-01},
urldate = {2024-03-19},
booktitle = {Computing in Civil Engineering 2023},
pages = {17–24},
publisher = {American Society of Civil Engineers},
address = {Corvallis, Oregon},
keywords = {Narrative, STG},
pubstate = {published},
tppubtype = {inproceedings}
}
Gratch, Jonathan; Greene, Gretchen; Picard, Rosalind; Urquhart, Lachlan; Valstar, Michel
Guest Editorial: Ethics in Affective Computing Journal Article
In: IEEE Trans. Affective Comput., vol. 15, no. 1, pp. 1–3, 2024, ISSN: 1949-3045, 2371-9850.
Links | BibTeX | Tags: Virtual Humans
@article{gratch_guest_2024,
title = {Guest Editorial: Ethics in Affective Computing},
author = {Jonathan Gratch and Gretchen Greene and Rosalind Picard and Lachlan Urquhart and Michel Valstar},
url = {https://ieeexplore.ieee.org/document/10454111/},
doi = {10.1109/TAFFC.2023.3322918},
issn = {1949-3045, 2371-9850},
year = {2024},
date = {2024-01-01},
urldate = {2024-03-14},
journal = {IEEE Trans. Affective Comput.},
volume = {15},
number = {1},
pages = {1–3},
keywords = {Virtual Humans},
pubstate = {published},
tppubtype = {article}
}
Spiegel, Brennan M. R.; Rizzo, Albert; Persky, Susan; Liran, Omer; Wiederhold, Brenda; Woods, Susan; Donovan, Kate; Sarkar, Korak; Xiang, Henry; Joo, Sun; Jotwani, Rohan; Lang, Min; Paul, Margot; Senter-Zapata, Mike; Widmeier, Keith; Zhang, Haipeng
What Is Medical Extended Reality? A Taxonomy Defining the Current Breadth and Depth of an Evolving Field Journal Article
In: Journal of Medical Extended Reality, vol. 1, no. 1, pp. 4–12, 2024, ISSN: 2994-1520.
Links | BibTeX | Tags: DTIC, MedVR, UARC
@article{spiegel_what_2024,
title = {What Is Medical Extended Reality? A Taxonomy Defining the Current Breadth and Depth of an Evolving Field},
author = {Brennan M. R. Spiegel and Albert Rizzo and Susan Persky and Omer Liran and Brenda Wiederhold and Susan Woods and Kate Donovan and Korak Sarkar and Henry Xiang and Sun Joo and Rohan Jotwani and Min Lang and Margot Paul and Mike Senter-Zapata and Keith Widmeier and Haipeng Zhang},
url = {https://www.liebertpub.com/doi/10.1089/jmxr.2023.0012},
doi = {10.1089/jmxr.2023.0012},
issn = {2994-1520},
year = {2024},
date = {2024-01-01},
urldate = {2024-02-20},
journal = {Journal of Medical Extended Reality},
volume = {1},
number = {1},
pages = {4–12},
keywords = {DTIC, MedVR, UARC},
pubstate = {published},
tppubtype = {article}
}
Awada, Mohamad; Gerber, Burcin Becerik; Lucas, Gale M.; Roll, Shawn C.
Stress appraisal in the workplace and its associations with productivity and mood: Insights from a multimodal machine learning analysis Journal Article
In: PLoS ONE, vol. 19, no. 1, pp. e0296468, 2024, ISSN: 1932-6203.
Abstract | Links | BibTeX | Tags: DTIC, Machine Learning, UARC
@article{awada_stress_2024,
title = {Stress appraisal in the workplace and its associations with productivity and mood: Insights from a multimodal machine learning analysis},
author = {Mohamad Awada and Burcin Becerik Gerber and Gale M. Lucas and Shawn C. Roll},
editor = {Iftikhar Ahmed Khan},
url = {https://dx.plos.org/10.1371/journal.pone.0296468},
doi = {10.1371/journal.pone.0296468},
issn = {1932-6203},
year = {2024},
date = {2024-01-01},
urldate = {2024-02-21},
journal = {PLoS ONE},
volume = {19},
number = {1},
pages = {e0296468},
abstract = {Previous studies have primarily focused on predicting stress arousal, encompassing physiological, behavioral, and psychological responses to stressors, while neglecting the examination of stress appraisal. Stress appraisal involves the cognitive evaluation of a situation as stressful or non-stressful, and as a threat/pressure or a challenge/opportunity. In this study, we investigated several research questions related to the association between states of stress appraisal (i.e., boredom, eustress, coexisting eustress-distress, distress) and various factors such as stress levels, mood, productivity, physiological and behavioral responses, as well as the most effective ML algorithms and data signals for predicting stress appraisal. The results support the Yerkes-Dodson law, showing that a moderate stress level is associated with increased productivity and positive mood, while low and high levels of stress are related to decreased productivity and negative mood, with distress overpowering eustress when they coexist. Changes in stress appraisal relative to physiological and behavioral features were examined through the lenses of stress arousal, activity engagement, and performance. An XGBOOST model achieved the best prediction accuracies of stress appraisal, reaching 82.78% when combining physiological and behavioral features and 79.55% using only the physiological dataset. The small accuracy difference of 3% indicates that physiological data alone may be adequate to accurately predict stress appraisal, and the feature importance results identified electrodermal activity, skin temperature, and blood volume pulse as the most useful physiologic features. Implementing these models within work environments can serve as a foundation for designing workplace policies, practices, and stress management strategies that prioritize the promotion of eustress while reducing distress and boredom. Such efforts can foster a supportive work environment to enhance employee well-being and productivity.},
keywords = {DTIC, Machine Learning, UARC},
pubstate = {published},
tppubtype = {article}
}
Jago, Arthur S.; Raveendhran, Roshni; Fast, Nathanael; Gratch, Jonathan
Algorithmic management diminishes status: An unintended consequence of using machines to perform social roles Journal Article
In: Journal of Experimental Social Psychology, vol. 110, pp. 104553, 2024, ISSN: 00221031.
Links | BibTeX | Tags: Virtual Humans
@article{jago_algorithmic_2024,
title = {Algorithmic management diminishes status: An unintended consequence of using machines to perform social roles},
author = {Arthur S. Jago and Roshni Raveendhran and Nathanael Fast and Jonathan Gratch},
url = {https://linkinghub.elsevier.com/retrieve/pii/S0022103123001105},
doi = {10.1016/j.jesp.2023.104553},
issn = {00221031},
year = {2024},
date = {2024-01-01},
urldate = {2024-02-21},
journal = {Journal of Experimental Social Psychology},
volume = {110},
pages = {104553},
keywords = {Virtual Humans},
pubstate = {published},
tppubtype = {article}
}
Shi, Zhonghao; O'Connell, Allison; Li, Zongjian; Liu, Siqi; Ayissi, Jennifer; Hoffman, Guy; Soleymani, Mohammad; Matarić, Maja J.
Build Your Own Robot Friend: An Open-Source Learning Module for Accessible and Engaging AI Education Miscellaneous
2024, (arXiv:2402.01647 [cs]).
Abstract | Links | BibTeX | Tags: Virtual Humans
@misc{shi_build_2024,
title = {Build Your Own Robot Friend: An Open-Source Learning Module for Accessible and Engaging AI Education},
author = {Zhonghao Shi and Allison O'Connell and Zongjian Li and Siqi Liu and Jennifer Ayissi and Guy Hoffman and Mohammad Soleymani and Maja J. Matarić},
url = {http://arxiv.org/abs/2402.01647},
year = {2024},
date = {2024-01-01},
urldate = {2024-02-21},
publisher = {arXiv},
abstract = {As artificial intelligence (AI) is playing an increasingly important role in our society and global economy, AI education and literacy have become necessary components in college and K-12 education to prepare students for an AI-powered society. However, current AI curricula have not yet been made accessible and engaging enough for students and schools from all socio-economic backgrounds with different educational goals. In this work, we developed an open-source learning module for college and high school students, which allows students to build their own robot companion from the ground up. This open platform can be used to provide hands-on experience and introductory knowledge about various aspects of AI, including robotics, machine learning (ML), software engineering, and mechanical engineering. Because of the social and personal nature of a socially assistive robot companion, this module also puts a special emphasis on human-centered AI, enabling students to develop a better understanding of human-AI interaction and AI ethics through hands-on learning activities. With open-source documentation, assembling manuals and affordable materials, students from different socio-economic backgrounds can personalize their learning experience based on their individual educational goals. To evaluate the student-perceived quality of our module, we conducted a usability testing workshop with 15 college students recruited from a minority-serving institution. Our results indicate that our AI module is effective, easy-to-follow, and engaging, and it increases student interest in studying AI/ML and robotics in the future. We hope that this work will contribute toward accessible and engaging AI education in human-AI interaction for college and high school students.},
note = {arXiv:2402.01647 [cs]},
keywords = {Virtual Humans},
pubstate = {published},
tppubtype = {misc}
}
Murawski, Alaine; Ramirez‐Zohfeld, Vanessa; Mell, Johnathan; Tschoe, Marianne; Schierer, Allison; Olvera, Charles; Brett, Jeanne; Gratch, Jonathan; Lindquist, Lee A.
Development and pilot testing of an artificial intelligence‐based family caregiver negotiation program Journal Article
In: J American Geriatrics Society, pp. jgs.18775, 2024, ISSN: 0002-8614, 1532-5415.
Abstract | Links | BibTeX | Tags: AI, Virtual Humans
@article{murawski_development_2024,
title = {Development and pilot testing of an artificial intelligence‐based family caregiver negotiation program},
author = {Alaine Murawski and Vanessa Ramirez‐Zohfeld and Johnathan Mell and Marianne Tschoe and Allison Schierer and Charles Olvera and Jeanne Brett and Jonathan Gratch and Lee A. Lindquist},
url = {https://agsjournals.onlinelibrary.wiley.com/doi/10.1111/jgs.18775},
doi = {10.1111/jgs.18775},
issn = {0002-8614, 1532-5415},
year = {2024},
date = {2024-01-01},
urldate = {2024-02-21},
journal = {J American Geriatrics Society},
pages = {jgs.18775},
abstract = {Abstract
Background
Family caregivers of people with Alzheimer's disease experience conflicts as they navigate health care but lack training to resolve these disputes. We sought to develop and pilot test an artificial‐intelligence negotiation training program, NegotiAge, for family caregivers.
Methods
We convened negotiation experts, a geriatrician, a social worker, and community‐based family caregivers. Content matter experts created short videos to teach negotiation skills. Caregivers generated dialogue surrounding conflicts. Computer scientists utilized the dialogue with the Interactive Arbitration Guide Online (IAGO) platform to develop avatar‐based agents (e.g., sibling, older adult, physician) for caregivers to practice negotiating. Pilot testing was conducted with family caregivers to assess usability (USE) and satisfaction (open‐ended questions with thematic analysis).
Results
Development: With NegotiAge, caregivers progress through didactic material, then receive scenarios to negotiate (e.g., physician recommends gastric tube, sibling disagrees with home support, older adult refusing support). Caregivers negotiate in real‐time with avatars who are designed to act like humans, including emotional tactics and irrational behaviors. Caregivers send/receive offers, using tactics until either mutual agreement or time expires. Immediate feedback is generated for the user to improve skills training. Pilot testing: Family caregivers (
n = 12) completed the program and survey. USE questionnaire (Likert scale 1–7) subset scores revealed: (1) Useful—Mean 5.69 (SD 0.76); (2) Ease—Mean 5.24 (SD 0.96); (3) Learn—Mean 5.69 (SD 0.74); (4) Satisfy—Mean 5.62 (SD 1.10). Items that received over 80% agreements were: It helps me be more effective; It helps me be more productive; It is useful; It gives me more control over the activities in my life; It makes the things I want to accomplish easier to get done. Participants were highly satisfied and found NegotiAge fun to use (91.7%), with 100% who would recommend it to a friend.
Conclusion
NegotiAge is an Artificial‐Intelligent Caregiver Negotiation Program, that is usable and feasible for family caregivers to become familiar with negotiating conflicts commonly seen in health care.},
keywords = {AI, Virtual Humans},
pubstate = {published},
tppubtype = {article}
}
Background
Family caregivers of people with Alzheimer's disease experience conflicts as they navigate health care but lack training to resolve these disputes. We sought to develop and pilot test an artificial‐intelligence negotiation training program, NegotiAge, for family caregivers.
Methods
We convened negotiation experts, a geriatrician, a social worker, and community‐based family caregivers. Content matter experts created short videos to teach negotiation skills. Caregivers generated dialogue surrounding conflicts. Computer scientists utilized the dialogue with the Interactive Arbitration Guide Online (IAGO) platform to develop avatar‐based agents (e.g., sibling, older adult, physician) for caregivers to practice negotiating. Pilot testing was conducted with family caregivers to assess usability (USE) and satisfaction (open‐ended questions with thematic analysis).
Results
Development: With NegotiAge, caregivers progress through didactic material, then receive scenarios to negotiate (e.g., physician recommends gastric tube, sibling disagrees with home support, older adult refusing support). Caregivers negotiate in real‐time with avatars who are designed to act like humans, including emotional tactics and irrational behaviors. Caregivers send/receive offers, using tactics until either mutual agreement or time expires. Immediate feedback is generated for the user to improve skills training. Pilot testing: Family caregivers (
n = 12) completed the program and survey. USE questionnaire (Likert scale 1–7) subset scores revealed: (1) Useful—Mean 5.69 (SD 0.76); (2) Ease—Mean 5.24 (SD 0.96); (3) Learn—Mean 5.69 (SD 0.74); (4) Satisfy—Mean 5.62 (SD 1.10). Items that received over 80% agreements were: It helps me be more effective; It helps me be more productive; It is useful; It gives me more control over the activities in my life; It makes the things I want to accomplish easier to get done. Participants were highly satisfied and found NegotiAge fun to use (91.7%), with 100% who would recommend it to a friend.
Conclusion
NegotiAge is an Artificial‐Intelligent Caregiver Negotiation Program, that is usable and feasible for family caregivers to become familiar with negotiating conflicts commonly seen in health care.
Barrett, Trevor; Faulk, Robert; Sergeant, Army Master; Boberg, Jill; Bartels, Matthew; Colonel, Marine Lieutenant; Saxon, Leslie A.
Force plate assessments in reconnaissance marine training company Journal Article
In: BMC Sports Sci Med Rehabil, vol. 16, no. 1, pp. 16, 2024, ISSN: 2052-1847.
Abstract | Links | BibTeX | Tags: DTIC, MedVR, UARC
@article{barrett_force_2024,
title = {Force plate assessments in reconnaissance marine training company},
author = {Trevor Barrett and Robert Faulk and Army Master Sergeant and Jill Boberg and Matthew Bartels and Marine Lieutenant Colonel and Leslie A. Saxon},
url = {https://bmcsportsscimedrehabil.biomedcentral.com/articles/10.1186/s13102-023-00796-z},
doi = {10.1186/s13102-023-00796-z},
issn = {2052-1847},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-22},
journal = {BMC Sports Sci Med Rehabil},
volume = {16},
number = {1},
pages = {16},
abstract = {Abstract
The ability to obtain dynamic movement assessments using force plate technology holds the promise of providing more detailed knowledge of the strength, balance and forces generated by active-duty military personnel. To date, there are not well-defined use cases for implementation of force plate assessments in military training environments. We sought to determine if force plate technology assessments could provide additional insights, related to the likelihood of graduation, beyond that provided by traditional physical fitness tests (PFT’s), in an elite Marine training school. Serial force plate measures were also obtained on those Marines successfully completing training to determine if consistent measures reflecting the effects of training on muscle skeletal load-over-time could be accurately measured. A pre-training force plate assessment performed in 112 Marines did not predict graduation rates. For Marines who successfully completed the course, serial measures obtained throughout training were highly variable for each individual and no firm conclusions could be drawn related to load imposed or the fitness attained during training.},
keywords = {DTIC, MedVR, UARC},
pubstate = {published},
tppubtype = {article}
}
The ability to obtain dynamic movement assessments using force plate technology holds the promise of providing more detailed knowledge of the strength, balance and forces generated by active-duty military personnel. To date, there are not well-defined use cases for implementation of force plate assessments in military training environments. We sought to determine if force plate technology assessments could provide additional insights, related to the likelihood of graduation, beyond that provided by traditional physical fitness tests (PFT’s), in an elite Marine training school. Serial force plate measures were also obtained on those Marines successfully completing training to determine if consistent measures reflecting the effects of training on muscle skeletal load-over-time could be accurately measured. A pre-training force plate assessment performed in 112 Marines did not predict graduation rates. For Marines who successfully completed the course, serial measures obtained throughout training were highly variable for each individual and no firm conclusions could be drawn related to load imposed or the fitness attained during training.
2023
Zhou, Emily; Soleymani, Mohammad; Matarić, Maja J.
Investigating the Generalizability of Physiological Characteristics of Anxiety Proceedings Article
In: 2023 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 4848–4855, IEEE, Istanbul, Turkiye, 2023, ISBN: 9798350337488.
Links | BibTeX | Tags: Machine Learning
@inproceedings{zhou_investigating_2023,
title = {Investigating the Generalizability of Physiological Characteristics of Anxiety},
author = {Emily Zhou and Mohammad Soleymani and Maja J. Matarić},
url = {https://ieeexplore.ieee.org/document/10385292/},
doi = {10.1109/BIBM58861.2023.10385292},
isbn = {9798350337488},
year = {2023},
date = {2023-12-01},
urldate = {2024-04-16},
booktitle = {2023 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)},
pages = {4848–4855},
publisher = {IEEE},
address = {Istanbul, Turkiye},
keywords = {Machine Learning},
pubstate = {published},
tppubtype = {inproceedings}
}
Melo, Celso M. De; Santos, Francisco C.; Terada, Kazunori
Emotion expression and cooperation under collective risks Journal Article
In: iScience, vol. 26, no. 11, pp. 108063, 2023, ISSN: 25890042.
@article{de_melo_emotion_2023,
title = {Emotion expression and cooperation under collective risks},
author = {Celso M. De Melo and Francisco C. Santos and Kazunori Terada},
url = {https://linkinghub.elsevier.com/retrieve/pii/S2589004223021405},
doi = {10.1016/j.isci.2023.108063},
issn = {25890042},
year = {2023},
date = {2023-11-01},
urldate = {2024-07-12},
journal = {iScience},
volume = {26},
number = {11},
pages = {108063},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Tak, Ala Nekouvaght; Becerik-Gerber, Burçin; Soibelman, Lucio; Lucas, Gale
A framework for investigating the acceptance of smart home technologies: Findings for residential smart HVAC systems Journal Article
In: Building and Environment, vol. 245, pp. 110935, 2023, ISSN: 03601323.
Links | BibTeX | Tags: DTIC, UARC, Virtual Humans
@article{tak_framework_2023,
title = {A framework for investigating the acceptance of smart home technologies: Findings for residential smart HVAC systems},
author = {Ala Nekouvaght Tak and Burçin Becerik-Gerber and Lucio Soibelman and Gale Lucas},
url = {https://linkinghub.elsevier.com/retrieve/pii/S0360132323009629},
doi = {10.1016/j.buildenv.2023.110935},
issn = {03601323},
year = {2023},
date = {2023-11-01},
urldate = {2023-12-07},
journal = {Building and Environment},
volume = {245},
pages = {110935},
keywords = {DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {article}
}
Cho, Hyundong; Liu, Shuai; Shi, Taiwei; Jain, Darpan; Rizk, Basem; Huang, Yuyang; Lu, Zixun; Wen, Nuan; Gratch, Jonathan; Ferrara, Emilio; May, Jonathan
Can Language Model Moderators Improve the Health of Online Discourse? Miscellaneous
2023, (arXiv:2311.10781 [cs]).
Abstract | Links | BibTeX | Tags: AI, Dialogue, DTIC, UARC, Virtual Humans
@misc{cho_can_2023,
title = {Can Language Model Moderators Improve the Health of Online Discourse?},
author = {Hyundong Cho and Shuai Liu and Taiwei Shi and Darpan Jain and Basem Rizk and Yuyang Huang and Zixun Lu and Nuan Wen and Jonathan Gratch and Emilio Ferrara and Jonathan May},
url = {http://arxiv.org/abs/2311.10781},
year = {2023},
date = {2023-11-01},
urldate = {2023-12-07},
publisher = {arXiv},
abstract = {Human moderation of online conversation is essential to maintaining civility and focus in a dialogue, but is challenging to scale and harmful to moderators. The inclusion of sophisticated natural language generation modules as a force multiplier aid moderators is a tantalizing prospect, but adequate evaluation approaches have so far been elusive. In this paper, we establish a systematic definition of conversational moderation effectiveness through a multidisciplinary lens that incorporates insights from social science. We then propose a comprehensive evaluation framework that uses this definition to asses models' moderation capabilities independently of human intervention. With our framework, we conduct the first known study of conversational dialogue models as moderators, finding that appropriately prompted models can provide specific and fair feedback on toxic behavior but struggle to influence users to increase their levels of respect and cooperation.},
note = {arXiv:2311.10781 [cs]},
keywords = {AI, Dialogue, DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {misc}
}
Yang, Daniel; Kommineni, Aditya; Alshehri, Mohammad; Mohanty, Nilamadhab; Modi, Vedant; Gratch, Jonathan; Narayanan, Shrikanth
Context Unlocks Emotions: Text-based Emotion Classification Dataset Auditing with Large Language Models Miscellaneous
2023, (arXiv:2311.03551 [cs]).
Abstract | Links | BibTeX | Tags: AI, DTIC, UARC, Virtual Humans
@misc{yang_context_2023,
title = {Context Unlocks Emotions: Text-based Emotion Classification Dataset Auditing with Large Language Models},
author = {Daniel Yang and Aditya Kommineni and Mohammad Alshehri and Nilamadhab Mohanty and Vedant Modi and Jonathan Gratch and Shrikanth Narayanan},
url = {http://arxiv.org/abs/2311.03551},
year = {2023},
date = {2023-11-01},
urldate = {2023-12-07},
publisher = {arXiv},
abstract = {The lack of contextual information in text data can make the annotation process of text-based emotion classification datasets challenging. As a result, such datasets often contain labels that fail to consider all the relevant emotions in the vocabulary. This misalignment between text inputs and labels can degrade the performance of machine learning models trained on top of them. As re-annotating entire datasets is a costly and time-consuming task that cannot be done at scale, we propose to use the expressive capabilities of large language models to synthesize additional context for input text to increase its alignment with the annotated emotional labels. In this work, we propose a formal definition of textual context to motivate a prompting strategy to enhance such contextual information. We provide both human and empirical evaluation to demonstrate the efficacy of the enhanced context. Our method improves alignment between inputs and their human-annotated labels from both an empirical and human-evaluated standpoint.},
note = {arXiv:2311.03551 [cs]},
keywords = {AI, DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {misc}
}
Chang, Di; Shi, Yichun; Gao, Quankai; Fu, Jessica; Xu, Hongyi; Song, Guoxian; Yan, Qing; Yang, Xiao; Soleymani, Mohammad
MagicDance: Realistic Human Dance Video Generation with Motions & Facial Expressions Transfer Miscellaneous
2023, (arXiv:2311.12052 [cs]).
Abstract | Links | BibTeX | Tags: DTIC, UARC, Virtual Humans
@misc{chang_magicdance_2023,
title = {MagicDance: Realistic Human Dance Video Generation with Motions & Facial Expressions Transfer},
author = {Di Chang and Yichun Shi and Quankai Gao and Jessica Fu and Hongyi Xu and Guoxian Song and Qing Yan and Xiao Yang and Mohammad Soleymani},
url = {http://arxiv.org/abs/2311.12052},
year = {2023},
date = {2023-11-01},
urldate = {2023-12-07},
publisher = {arXiv},
abstract = {In this work, we propose MagicDance, a diffusion-based model for 2D human motion and facial expression transfer on challenging human dance videos. Specifically, we aim to generate human dance videos of any target identity driven by novel pose sequences while keeping the identity unchanged. To this end, we propose a two-stage training strategy to disentangle human motions and appearance (e.g., facial expressions, skin tone and dressing), consisting of the pretraining of an appearance-control block and fine-tuning of an appearance-pose-joint-control block over human dance poses of the same dataset. Our novel design enables robust appearance control with temporally consistent upper body, facial attributes, and even background. The model also generalizes well on unseen human identities and complex motion sequences without the need for any fine-tuning with additional data with diverse human attributes by leveraging the prior knowledge of image diffusion models. Moreover, the proposed model is easy to use and can be considered as a plug-in module/extension to Stable Diffusion. We also demonstrate the model's ability for zero-shot 2D animation generation, enabling not only the appearance transfer from one identity to another but also allowing for cartoon-like stylization given only pose inputs. Extensive experiments demonstrate our superior performance on the TikTok dataset.},
note = {arXiv:2311.12052 [cs]},
keywords = {DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {misc}
}
Liu, Ruying; Awada, Mohamad; Gerber, Burcin Becerik; Lucas, Gale M.; Roll, Shawn C.
Gender moderates the effects of ambient bergamot scent on stress restoration in offices Journal Article
In: Journal of Environmental Psychology, vol. 91, pp. 102135, 2023, ISSN: 02724944.
Links | BibTeX | Tags: DTIC, UARC, Virtual Humans
@article{liu_gender_2023,
title = {Gender moderates the effects of ambient bergamot scent on stress restoration in offices},
author = {Ruying Liu and Mohamad Awada and Burcin Becerik Gerber and Gale M. Lucas and Shawn C. Roll},
url = {https://linkinghub.elsevier.com/retrieve/pii/S0272494423001834},
doi = {10.1016/j.jenvp.2023.102135},
issn = {02724944},
year = {2023},
date = {2023-11-01},
urldate = {2023-09-20},
journal = {Journal of Environmental Psychology},
volume = {91},
pages = {102135},
keywords = {DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {article}
}
Prinzing, Michael; Garton, Catherine; Berman, Catherine J.; Zhou, Jieni; West, Taylor Nicole; Gratch, Jonathan; Fredrickson, Barbara
Can AI Agents Help Humans to Connect? Miscellaneous
2023.
Abstract | Links | BibTeX | Tags: AI
@misc{prinzing_can_2023-1,
title = {Can AI Agents Help Humans to Connect?},
author = {Michael Prinzing and Catherine Garton and Catherine J. Berman and Jieni Zhou and Taylor Nicole West and Jonathan Gratch and Barbara Fredrickson},
url = {https://osf.io/muq6s},
doi = {10.31234/osf.io/muq6s},
year = {2023},
date = {2023-10-01},
urldate = {2024-08-13},
abstract = {This paper reports on a pre-registered experiment designed to test whether artificial agents can help people to create more moments of high-quality connection with other humans. Of four pre-registered hypotheses, we found (partial) support for only one.},
keywords = {AI},
pubstate = {published},
tppubtype = {misc}
}
Chemburkar, Ankur; Lu, Shuhong; Feng, Andrew
Discrete Diffusion for Co-Speech Gesture Synthesis Proceedings Article
In: International Cconference on Multimodal Interaction, pp. 186–192, ACM, Paris France, 2023, ISBN: 9798400703218.
Links | BibTeX | Tags: DTIC, Natural Language
@inproceedings{chemburkar_discrete_2023,
title = {Discrete Diffusion for Co-Speech Gesture Synthesis},
author = {Ankur Chemburkar and Shuhong Lu and Andrew Feng},
url = {https://dl.acm.org/doi/10.1145/3610661.3616556},
doi = {10.1145/3610661.3616556},
isbn = {9798400703218},
year = {2023},
date = {2023-10-01},
urldate = {2024-07-09},
booktitle = {International Cconference on Multimodal Interaction},
pages = {186–192},
publisher = {ACM},
address = {Paris France},
keywords = {DTIC, Natural Language},
pubstate = {published},
tppubtype = {inproceedings}
}
Liu, Rong; Zhao, Enyu; Liu, Zhiyuan; Feng, Andrew; Easley, Scott John
Instant Photorealistic Style Transfer: A Lightweight and Adaptive Approach Miscellaneous
2023, (arXiv:2309.10011 [cs, eess]).
Abstract | Links | BibTeX | Tags: DTIC, UARC
@misc{liu_instant_2023,
title = {Instant Photorealistic Style Transfer: A Lightweight and Adaptive Approach},
author = {Rong Liu and Enyu Zhao and Zhiyuan Liu and Andrew Feng and Scott John Easley},
url = {http://arxiv.org/abs/2309.10011},
year = {2023},
date = {2023-10-01},
urldate = {2024-05-14},
publisher = {arXiv},
abstract = {In this paper, we propose an Instant Photorealistic Style Transfer (IPST) approach, designed to achieve instant photorealistic style transfer on super-resolution inputs without the need for pre-training on pair-wise datasets or imposing extra constraints. Our method utilizes a lightweight StyleNet to enable style transfer from a style image to a content image while preserving non-color information. To further enhance the style transfer process, we introduce an instance-adaptive optimization to prioritize the photorealism of outputs and accelerate the convergence of the style network, leading to a rapid training completion within seconds. Moreover, IPST is well-suited for multi-frame style transfer tasks, as it retains temporal and multi-view consistency of the multi-frame inputs such as video and Neural Radiance Field (NeRF). Experimental results demonstrate that IPST requires less GPU memory usage, offers faster multi-frame transfer speed, and generates photorealistic outputs, making it a promising solution for various photorealistic transfer applications.},
note = {arXiv:2309.10011 [cs, eess]},
keywords = {DTIC, UARC},
pubstate = {published},
tppubtype = {misc}
}
Wang, Timothy S.; Gordon, Andrew S.
Playing Story Creation Games with Large Language Models: Experiments with GPT-3.5 Book Section
In: Holloway-Attaway, Lissa; Murray, John T. (Ed.): Interactive Storytelling, vol. 14384, pp. 297–305, Springer Nature Switzerland, Cham, 2023, ISBN: 978-3-031-47657-0 978-3-031-47658-7, (Series Title: Lecture Notes in Computer Science).
Links | BibTeX | Tags: DTIC, Narrative, UARC
@incollection{holloway-attaway_playing_2023,
title = {Playing Story Creation Games with Large Language Models: Experiments with GPT-3.5},
author = {Timothy S. Wang and Andrew S. Gordon},
editor = {Lissa Holloway-Attaway and John T. Murray},
url = {https://link.springer.com/10.1007/978-3-031-47658-7_28},
doi = {10.1007/978-3-031-47658-7_28},
isbn = {978-3-031-47657-0 978-3-031-47658-7},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
booktitle = {Interactive Storytelling},
volume = {14384},
pages = {297–305},
publisher = {Springer Nature Switzerland},
address = {Cham},
note = {Series Title: Lecture Notes in Computer Science},
keywords = {DTIC, Narrative, UARC},
pubstate = {published},
tppubtype = {incollection}
}
Lukin, Stephanie M.; Pollard, Kimberly A.; Bonial, Claire; Hudson, Taylor; Arstein, Ron; Voss, Clare; Traum, David
Navigating to Success in Multi-Modal Human-Robot Collaboration: Analysis and Corpus Release Miscellaneous
2023, (arXiv:2310.17568 [cs]).
Abstract | Links | BibTeX | Tags: DTIC, Natural Language, UARC
@misc{lukin_navigating_2023,
title = {Navigating to Success in Multi-Modal Human-Robot Collaboration: Analysis and Corpus Release},
author = {Stephanie M. Lukin and Kimberly A. Pollard and Claire Bonial and Taylor Hudson and Ron Arstein and Clare Voss and David Traum},
url = {http://arxiv.org/abs/2310.17568},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
publisher = {arXiv},
abstract = {Human-guided robotic exploration is a useful approach to gathering information at remote locations, especially those that might be too risky, inhospitable, or inaccessible for humans. Maintaining common ground between the remotely-located partners is a challenge, one that can be facilitated by multi-modal communication. In this paper, we explore how participants utilized multiple modalities to investigate a remote location with the help of a robotic partner. Participants issued spoken natural language instructions and received from the robot: text-based feedback, continuous 2D LIDAR mapping, and upon-request static photographs. We noticed that different strategies were adopted in terms of use of the modalities, and hypothesize that these differences may be correlated with success at several exploration sub-tasks. We found that requesting photos may have improved the identification and counting of some key entities (doorways in particular) and that this strategy did not hinder the amount of overall area exploration. Future work with larger samples may reveal the effects of more nuanced photo and dialogue strategies, which can inform the training of robotic agents. Additionally, we announce the release of our unique multi-modal corpus of human-robot communication in an exploration context: SCOUT, the Situated Corpus on Understanding Transactions.},
note = {arXiv:2310.17568 [cs]},
keywords = {DTIC, Natural Language, UARC},
pubstate = {published},
tppubtype = {misc}
}
Gilani, Setareh Nasihati; Pollard, Kimberly; Traum, David
Multimodal Prediction of User's Performance in High-Stress Dialogue Interactions Proceedings Article
In: International Cconference on Multimodal Interaction, pp. 71–75, ACM, Paris France, 2023, ISBN: 9798400703218.
Links | BibTeX | Tags: DTIC, Natural Language, UARC
@inproceedings{nasihati_gilani_multimodal_2023,
title = {Multimodal Prediction of User's Performance in High-Stress Dialogue Interactions},
author = {Setareh Nasihati Gilani and Kimberly Pollard and David Traum},
url = {https://dl.acm.org/doi/10.1145/3610661.3617166},
doi = {10.1145/3610661.3617166},
isbn = {9798400703218},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
booktitle = {International Cconference on Multimodal Interaction},
pages = {71–75},
publisher = {ACM},
address = {Paris France},
keywords = {DTIC, Natural Language, UARC},
pubstate = {published},
tppubtype = {inproceedings}
}
Awada, Mohamad; Becerik-Gerber, Burcin; Lucas, Gale; Roll, Shawn C.
Predicting Office Workers’ Productivity: A Machine Learning Approach Integrating Physiological, Behavioral, and Psychological Indicators Journal Article
In: Sensors, vol. 23, no. 21, pp. 8694, 2023, ISSN: 1424-8220.
Abstract | Links | BibTeX | Tags: DTIC, Machine Learning, UARC, Virtual Humans
@article{awada_predicting_2023,
title = {Predicting Office Workers’ Productivity: A Machine Learning Approach Integrating Physiological, Behavioral, and Psychological Indicators},
author = {Mohamad Awada and Burcin Becerik-Gerber and Gale Lucas and Shawn C. Roll},
url = {https://www.mdpi.com/1424-8220/23/21/8694},
doi = {10.3390/s23218694},
issn = {1424-8220},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
journal = {Sensors},
volume = {23},
number = {21},
pages = {8694},
abstract = {This research pioneers the application of a machine learning framework to predict the perceived productivity of office workers using physiological, behavioral, and psychological features. Two approaches were compared: the baseline model, predicting productivity based on physiological and behavioral characteristics, and the extended model, incorporating predictions of psychological states such as stress, eustress, distress, and mood. Various machine learning models were utilized and compared to assess their predictive accuracy for psychological states and productivity, with XGBoost emerging as the top performer. The extended model outperformed the baseline model, achieving an R2 of 0.60 and a lower MAE of 10.52, compared to the baseline model’s R2 of 0.48 and MAE of 16.62. The extended model’s feature importance analysis revealed valuable insights into the key predictors of productivity, shedding light on the role of psychological states in the prediction process. Notably, mood and eustress emerged as significant predictors of productivity. Physiological and behavioral features, including skin temperature, electrodermal activity, facial movements, and wrist acceleration, were also identified. Lastly, a comparative analysis revealed that wearable devices (Empatica E4 and H10 Polar) outperformed workstation addons (Kinect camera and computer-usage monitoring application) in predicting productivity, emphasizing the potential utility of wearable devices as an independent tool for assessment of productivity. Implementing the model within smart workstations allows for adaptable environments that boost productivity and overall well-being among office workers.},
keywords = {DTIC, Machine Learning, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {article}
}
Chawla, Kushal; Wu, Ian; Rong, Yu; Lucas, Gale M.; Gratch, Jonathan
Be Selfish, But Wisely: Investigating the Impact of Agent Personality in Mixed-Motive Human-Agent Interactions Miscellaneous
2023, (arXiv:2310.14404 [cs]).
Abstract | Links | BibTeX | Tags: Dialogue, DTIC, UARC, Virtual Humans
@misc{chawla_be_2023,
title = {Be Selfish, But Wisely: Investigating the Impact of Agent Personality in Mixed-Motive Human-Agent Interactions},
author = {Kushal Chawla and Ian Wu and Yu Rong and Gale M. Lucas and Jonathan Gratch},
url = {http://arxiv.org/abs/2310.14404},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
publisher = {arXiv},
abstract = {A natural way to design a negotiation dialogue system is via self-play RL: train an agent that learns to maximize its performance by interacting with a simulated user that has been designed to imitate human-human dialogue data. Although this procedure has been adopted in prior work, we find that it results in a fundamentally flawed system that fails to learn the value of compromise in a negotiation, which can often lead to no agreements (i.e., the partner walking away without a deal), ultimately hurting the model's overall performance. We investigate this observation in the context of the DealOrNoDeal task, a multi-issue negotiation over books, hats, and balls. Grounded in negotiation theory from Economics, we modify the training procedure in two novel ways to design agents with diverse personalities and analyze their performance with human partners. We find that although both techniques show promise, a selfish agent, which maximizes its own performance while also avoiding walkaways, performs superior to other variants by implicitly learning to generate value for both itself and the negotiation partner. We discuss the implications of our findings for what it means to be a successful negotiation dialogue system and how these systems should be designed in the future.},
note = {arXiv:2310.14404 [cs]},
keywords = {Dialogue, DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {misc}
}
Awada, Mohamad; Becerik-Gerber, Burcin; Lucas, Gale; Roll, Shawn; Liu, Ruying
A New Perspective on Stress Detection: An Automated Approach for Detecting Eustress and Distress Journal Article
In: IEEE Trans. Affective Comput., pp. 1–15, 2023, ISSN: 1949-3045, 2371-9850.
Links | BibTeX | Tags: DTIC, Machine Learning, UARC
@article{awada_new_2023,
title = {A New Perspective on Stress Detection: An Automated Approach for Detecting Eustress and Distress},
author = {Mohamad Awada and Burcin Becerik-Gerber and Gale Lucas and Shawn Roll and Ruying Liu},
url = {https://ieeexplore.ieee.org/document/10286408/},
doi = {10.1109/TAFFC.2023.3324910},
issn = {1949-3045, 2371-9850},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
journal = {IEEE Trans. Affective Comput.},
pages = {1–15},
keywords = {DTIC, Machine Learning, UARC},
pubstate = {published},
tppubtype = {article}
}
Prinzing, Michael; Garton, Catherine; Berman, Catherine J.; Zhou, Jieni; West, Taylor Nicole; Gratch, Jonathan; Fredrickson, Barbara
Can AI Agents Help Humans to Connect? Technical Report
PsyArXiv 2023.
Abstract | Links | BibTeX | Tags: AI, DTIC, UARC, Virtual Humans
@techreport{prinzing_can_2023,
title = {Can AI Agents Help Humans to Connect?},
author = {Michael Prinzing and Catherine Garton and Catherine J. Berman and Jieni Zhou and Taylor Nicole West and Jonathan Gratch and Barbara Fredrickson},
url = {https://osf.io/muq6s},
doi = {10.31234/osf.io/muq6s},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
institution = {PsyArXiv},
abstract = {This paper reports on a pre-registered experiment designed to test whether artificial agents can help people to create more moments of high-quality connection with other humans. Of four pre-registered hypotheses, we found (partial) support for only one.},
keywords = {AI, DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {techreport}
}
Lin, Eleanor; Hale, James; Gratch, Jonathan
Toward a Better Understanding of the Emotional Dynamics of Negotiation with Large Language Models Proceedings Article
In: Proceedings of the Twenty-fourth International Symposium on Theory, Algorithmic Foundations, and Protocol Design for Mobile Networks and Mobile Computing, pp. 545–550, ACM, Washington DC USA, 2023, ISBN: 978-1-4503-9926-5.
Links | BibTeX | Tags: DTIC, UARC, Virtual Humans
@inproceedings{lin_toward_2023,
title = {Toward a Better Understanding of the Emotional Dynamics of Negotiation with Large Language Models},
author = {Eleanor Lin and James Hale and Jonathan Gratch},
url = {https://dl.acm.org/doi/10.1145/3565287.3617637},
doi = {10.1145/3565287.3617637},
isbn = {978-1-4503-9926-5},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
booktitle = {Proceedings of the Twenty-fourth International Symposium on Theory, Algorithmic Foundations, and Protocol Design for Mobile Networks and Mobile Computing},
pages = {545–550},
publisher = {ACM},
address = {Washington DC USA},
keywords = {DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {inproceedings}
}
Tran, Minh; Soleymani, Mohammad
Privacy-preserving Representation Learning for Speech Understanding Miscellaneous
2023, (arXiv:2310.17194 [eess]).
Abstract | Links | BibTeX | Tags: UARC, Virtual Humans
@misc{tran_privacy-preserving_2023,
title = {Privacy-preserving Representation Learning for Speech Understanding},
author = {Minh Tran and Mohammad Soleymani},
url = {http://arxiv.org/abs/2310.17194},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
publisher = {arXiv},
abstract = {Existing privacy-preserving speech representation learning methods target a single application domain. In this paper, we present a novel framework to anonymize utterance-level speech embeddings generated by pre-trained encoders and show its effectiveness for a range of speech classification tasks. Specifically, given the representations from a pre-trained encoder, we train a Transformer to estimate the representations for the same utterances spoken by other speakers. During inference, the extracted representations can be converted into different identities to preserve privacy. We compare the results with the voice anonymization baselines from the VoicePrivacy 2022 challenge. We evaluate our framework on speaker identification for privacy and emotion recognition, depression classification, and intent classification for utility. Our method outperforms the baselines on privacy and utility in paralinguistic tasks and achieves comparable performance for intent classification.},
note = {arXiv:2310.17194 [eess]},
keywords = {UARC, Virtual Humans},
pubstate = {published},
tppubtype = {misc}
}
Ahmed, Tamim; Rikakis, Thanassis; Kelliher, Aisling; Soleymani, Mohammad
ASAR Dataset and Computational Model for Affective State Recognition During ARAT Assessment for Upper Extremity Stroke Survivors Proceedings Article
In: International Cconference on Multimodal Interaction, pp. 11–15, ACM, Paris France, 2023, ISBN: 9798400703218.
Links | BibTeX | Tags: DTIC, UARC, Virtual Humans
@inproceedings{ahmed_asar_2023,
title = {ASAR Dataset and Computational Model for Affective State Recognition During ARAT Assessment for Upper Extremity Stroke Survivors},
author = {Tamim Ahmed and Thanassis Rikakis and Aisling Kelliher and Mohammad Soleymani},
url = {https://dl.acm.org/doi/10.1145/3610661.3617154},
doi = {10.1145/3610661.3617154},
isbn = {9798400703218},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
booktitle = {International Cconference on Multimodal Interaction},
pages = {11–15},
publisher = {ACM},
address = {Paris France},
keywords = {DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {inproceedings}
}
Andrist, Sean; Bohus, Dan; Li, Zongjian; Soleymani, Mohammad
Platform for Situated Intelligence and OpenSense: A Tutorial on Building Multimodal Interactive Applications for Research Proceedings Article
In: International Cconference on Multimodal Interaction, pp. 105–106, ACM, Paris France, 2023, ISBN: 9798400703218.
Links | BibTeX | Tags: AI, UARC, Virtual Humans
@inproceedings{andrist_platform_2023,
title = {Platform for Situated Intelligence and OpenSense: A Tutorial on Building Multimodal Interactive Applications for Research},
author = {Sean Andrist and Dan Bohus and Zongjian Li and Mohammad Soleymani},
url = {https://dl.acm.org/doi/10.1145/3610661.3617603},
doi = {10.1145/3610661.3617603},
isbn = {9798400703218},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
booktitle = {International Cconference on Multimodal Interaction},
pages = {105–106},
publisher = {ACM},
address = {Paris France},
keywords = {AI, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {inproceedings}
}
Tran, Trang; Yin, Yufeng; Tavabi, Leili; Delacruz, Joannalyn; Borsari, Brian; Woolley, Joshua D; Scherer, Stefan; Soleymani, Mohammad
Multimodal Analysis and Assessment of Therapist Empathy in Motivational Interviews Proceedings Article
In: INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, pp. 406–415, ACM, Paris France, 2023, ISBN: 9798400700552.
Links | BibTeX | Tags: DTIC, UARC, Virtual Humans
@inproceedings{tran_multimodal_2023,
title = {Multimodal Analysis and Assessment of Therapist Empathy in Motivational Interviews},
author = {Trang Tran and Yufeng Yin and Leili Tavabi and Joannalyn Delacruz and Brian Borsari and Joshua D Woolley and Stefan Scherer and Mohammad Soleymani},
url = {https://dl.acm.org/doi/10.1145/3577190.3614105},
doi = {10.1145/3577190.3614105},
isbn = {9798400700552},
year = {2023},
date = {2023-10-01},
urldate = {2023-12-07},
booktitle = {INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION},
pages = {406–415},
publisher = {ACM},
address = {Paris France},
keywords = {DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {inproceedings}
}
Seyedrezaei, Mirmahdi; Awada, Mohamad; Becerik-Gerber, Burcin; Lucas, Gale; Roll, Shawn
In: Building and Environment, vol. 244, pp. 110743, 2023, ISSN: 03601323.
Links | BibTeX | Tags: DTIC, UARC, Virtual Humans
@article{seyedrezaei_interaction_2023,
title = {Interaction effects of indoor environmental quality factors on cognitive performance and perceived comfort of young adults in open plan offices in North American Mediterranean climate},
author = {Mirmahdi Seyedrezaei and Mohamad Awada and Burcin Becerik-Gerber and Gale Lucas and Shawn Roll},
url = {https://linkinghub.elsevier.com/retrieve/pii/S0360132323007709},
doi = {10.1016/j.buildenv.2023.110743},
issn = {03601323},
year = {2023},
date = {2023-10-01},
urldate = {2023-09-20},
journal = {Building and Environment},
volume = {244},
pages = {110743},
keywords = {DTIC, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {article}
}
Lei, Su; Gratch, Jonathan
Sources of Facial Expression Synchrony Proceedings Article
In: 2023 11th International Conference on Affective Computing and Intelligent Interaction (ACII), pp. 1–8, IEEE, Cambridge, MA, USA, 2023, ISBN: 9798350327434.
@inproceedings{lei_sources_2023,
title = {Sources of Facial Expression Synchrony},
author = {Su Lei and Jonathan Gratch},
url = {https://ieeexplore.ieee.org/document/10388107/},
doi = {10.1109/ACII59096.2023.10388107},
isbn = {9798350327434},
year = {2023},
date = {2023-09-01},
urldate = {2024-07-16},
booktitle = {2023 11th International Conference on Affective Computing and Intelligent Interaction (ACII)},
pages = {1–8},
publisher = {IEEE},
address = {Cambridge, MA, USA},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Yang, Daniel; Kommineni, Aditya; Alshehri, Mohammad; Mohanty, Nilamadhab; Modi, Vedant; Gratch, Jonathan; Narayanan, Shrikanth
Context Unlocks Emotions: Text-based Emotion Classification Dataset Auditing with Large Language Models Proceedings Article
In: 2023 11th International Conference on Affective Computing and Intelligent Interaction (ACII), pp. 1–8, IEEE, Cambridge, MA, USA, 2023, ISBN: 9798350327434.
Links | BibTeX | Tags: Emotions
@inproceedings{yang_context_2023-1,
title = {Context Unlocks Emotions: Text-based Emotion Classification Dataset Auditing with Large Language Models},
author = {Daniel Yang and Aditya Kommineni and Mohammad Alshehri and Nilamadhab Mohanty and Vedant Modi and Jonathan Gratch and Shrikanth Narayanan},
url = {https://ieeexplore.ieee.org/document/10388131/},
doi = {10.1109/ACII59096.2023.10388131},
isbn = {9798350327434},
year = {2023},
date = {2023-09-01},
urldate = {2024-08-13},
booktitle = {2023 11th International Conference on Affective Computing and Intelligent Interaction (ACII)},
pages = {1–8},
publisher = {IEEE},
address = {Cambridge, MA, USA},
keywords = {Emotions},
pubstate = {published},
tppubtype = {inproceedings}
}
Junghaenel, Doerte U.; Schneider, Stefan; Lucas, Gale; Boberg, Jill; Weinstein, Faye M.; Richeimer, Steven H.; Stone, Arthur A.; Lumley, Mark A.
In: Psychosom Med, vol. 85, no. 7, pp. 627–638, 2023, ISSN: 1534-7796, 0033-3174.
Abstract | Links | BibTeX | Tags: MedVR, VR
@article{junghaenel_virtual_2023,
title = {Virtual Human–Delivered Interviews for Patients With Chronic Pain: Feasibility, Acceptability, and a Pilot Randomized Trial of Standard Medical, Psychosocial, and Educational Interviews},
author = {Doerte U. Junghaenel and Stefan Schneider and Gale Lucas and Jill Boberg and Faye M. Weinstein and Steven H. Richeimer and Arthur A. Stone and Mark A. Lumley},
url = {https://journals.lww.com/10.1097/PSY.0000000000001228},
doi = {10.1097/PSY.0000000000001228},
issn = {1534-7796, 0033-3174},
year = {2023},
date = {2023-09-01},
urldate = {2024-07-11},
journal = {Psychosom Med},
volume = {85},
number = {7},
pages = {627–638},
abstract = {ABSTRACT
Objective
Seminal advances in virtual human (VH) technology have introduced highly interactive, computer-animated VH interviewers. Their utility for aiding in chronic pain care is unknown. We developed three interactive telehealth VH interviews—a
standard
pain-focused, a
psychosocial
risk factor, and a pain psychology and neuroscience
educational
interview. We then conducted a preliminary investigation of their feasibility, acceptability, and efficacy. We also experimentally compared a human and a computer-generated VH voice.
Methods
Patients (
N = 94},
keywords = {MedVR, VR},
pubstate = {published},
tppubtype = {article}
}
Objective
Seminal advances in virtual human (VH) technology have introduced highly interactive, computer-animated VH interviewers. Their utility for aiding in chronic pain care is unknown. We developed three interactive telehealth VH interviews—a
standard
pain-focused, a
psychosocial
risk factor, and a pain psychology and neuroscience
educational
interview. We then conducted a preliminary investigation of their feasibility, acceptability, and efficacy. We also experimentally compared a human and a computer-generated VH voice.
Methods
Patients (
N = 94
Hoegen, Jessie; Lucas, Gale; Shore, Danielle; Parkinson, Brian; Gratch, Jonathan
How Expression and Context Determine Second-person Judgments of Emotion Proceedings Article
In: 2023 11th International Conference on Affective Computing and Intelligent Interaction (ACII), pp. 1–7, IEEE, Cambridge, MA, USA, 2023, ISBN: 9798350327434.
Links | BibTeX | Tags: DTIC, Emotion
@inproceedings{hoegen_how_2023,
title = {How Expression and Context Determine Second-person Judgments of Emotion},
author = {Jessie Hoegen and Gale Lucas and Danielle Shore and Brian Parkinson and Jonathan Gratch},
url = {https://ieeexplore.ieee.org/document/10388189/},
doi = {10.1109/ACII59096.2023.10388189},
isbn = {9798350327434},
year = {2023},
date = {2023-09-01},
urldate = {2024-07-09},
booktitle = {2023 11th International Conference on Affective Computing and Intelligent Interaction (ACII)},
pages = {1–7},
publisher = {IEEE},
address = {Cambridge, MA, USA},
keywords = {DTIC, Emotion},
pubstate = {published},
tppubtype = {inproceedings}
}
Gainer, Alesia; Aptaker, Allison; Artstein, Ron; Cobbins, David; Core, Mark; Gordon, Carla; Leuski, Anton; Li, Zongjian; Merchant, Chirag; Nelson, David; Soleymani, Mohammad; Traum, David
DIVIS: Digital Interactive Victim Intake Simulator Proceedings Article
In: Proceedings of the 23rd ACM International Conference on Intelligent Virtual Agents, pp. 1–2, ACM, Würzburg Germany, 2023, ISBN: 978-1-4503-9994-4.
Links | BibTeX | Tags: DTIC, MxR, UARC, Virtual Humans
@inproceedings{gainer_divis_2023,
title = {DIVIS: Digital Interactive Victim Intake Simulator},
author = {Alesia Gainer and Allison Aptaker and Ron Artstein and David Cobbins and Mark Core and Carla Gordon and Anton Leuski and Zongjian Li and Chirag Merchant and David Nelson and Mohammad Soleymani and David Traum},
url = {https://dl.acm.org/doi/10.1145/3570945.3607328},
doi = {10.1145/3570945.3607328},
isbn = {978-1-4503-9994-4},
year = {2023},
date = {2023-09-01},
urldate = {2024-02-20},
booktitle = {Proceedings of the 23rd ACM International Conference on Intelligent Virtual Agents},
pages = {1–2},
publisher = {ACM},
address = {Würzburg Germany},
keywords = {DTIC, MxR, UARC, Virtual Humans},
pubstate = {published},
tppubtype = {inproceedings}
}
Mozgai, Sharon; Kaurloto, Cari; Winn, Jade; Leeds, Andrew; Heylen, Dirk; Hartholt, Arno; Scherer, Stefan
Machine learning for semi-automated scoping reviews Journal Article
In: Intelligent Systems with Applications, vol. 19, pp. 200249, 2023, ISSN: 26673053.
Links | BibTeX | Tags: DTIC, UARC, VHTL, Virtual Humans
@article{mozgai_machine_2023,
title = {Machine learning for semi-automated scoping reviews},
author = {Sharon Mozgai and Cari Kaurloto and Jade Winn and Andrew Leeds and Dirk Heylen and Arno Hartholt and Stefan Scherer},
url = {https://linkinghub.elsevier.com/retrieve/pii/S2667305323000741},
doi = {10.1016/j.iswa.2023.200249},
issn = {26673053},
year = {2023},
date = {2023-09-01},
urldate = {2023-08-23},
journal = {Intelligent Systems with Applications},
volume = {19},
pages = {200249},
keywords = {DTIC, UARC, VHTL, Virtual Humans},
pubstate = {published},
tppubtype = {article}
}
Yin, Yufeng; Chang, Di; Song, Guoxian; Sang, Shen; Zhi, Tiancheng; Liu, Jing; Luo, Linjie; Soleymani, Mohammad
FG-Net: Facial Action Unit Detection with Generalizable Pyramidal Features Miscellaneous
2023, (arXiv:2308.12380 [cs]).
Abstract | Links | BibTeX | Tags: DTIC, Virtual Humans
@misc{yin_fg-net_2023,
title = {FG-Net: Facial Action Unit Detection with Generalizable Pyramidal Features},
author = {Yufeng Yin and Di Chang and Guoxian Song and Shen Sang and Tiancheng Zhi and Jing Liu and Linjie Luo and Mohammad Soleymani},
url = {http://arxiv.org/abs/2308.12380},
year = {2023},
date = {2023-08-01},
urldate = {2024-02-21},
publisher = {arXiv},
abstract = {Automatic detection of facial Action Units (AUs) allows for objective facial expression analysis. Due to the high cost of AU labeling and the limited size of existing benchmarks, previous AU detection methods tend to overfit the dataset, resulting in a significant performance loss when evaluated across corpora. To address this problem, we propose FG-Net for generalizable facial action unit detection. Specifically, FG-Net extracts feature maps from a StyleGAN2 model pre-trained on a large and diverse face image dataset. Then, these features are used to detect AUs with a Pyramid CNN Interpreter, making the training efficient and capturing essential local features. The proposed FG-Net achieves a strong generalization ability for heatmap-based AU detection thanks to the generalizable and semantic-rich features extracted from the pre-trained generative model. Extensive experiments are conducted to evaluate within- and cross-corpus AU detection with the widely-used DISFA and BP4D datasets. Compared with the state-of-the-art, the proposed method achieves superior cross-domain performance while maintaining competitive within-domain performance. In addition, FG-Net is data-efficient and achieves competitive performance even when trained on 1000 samples. Our code will be released at textbackslashurlhttps://github.com/ihp-lab/FG-Net},
note = {arXiv:2308.12380 [cs]},
keywords = {DTIC, Virtual Humans},
pubstate = {published},
tppubtype = {misc}
}
Chang, Di; Yin, Yufeng; Li, Zongjian; Tran, Minh; Soleymani, Mohammad
LibreFace: An Open-Source Toolkit for Deep Facial Expression Analysis Miscellaneous
2023, (arXiv:2308.10713 [cs]).
Abstract | Links | BibTeX | Tags: DTIC, Virtual Humans
@misc{chang_libreface_2023,
title = {LibreFace: An Open-Source Toolkit for Deep Facial Expression Analysis},
author = {Di Chang and Yufeng Yin and Zongjian Li and Minh Tran and Mohammad Soleymani},
url = {http://arxiv.org/abs/2308.10713},
year = {2023},
date = {2023-08-01},
urldate = {2024-02-21},
publisher = {arXiv},
abstract = {Facial expression analysis is an important tool for human-computer interaction. In this paper, we introduce LibreFace, an open-source toolkit for facial expression analysis. This open-source toolbox offers real-time and offline analysis of facial behavior through deep learning models, including facial action unit (AU) detection, AU intensity estimation, and facial expression recognition. To accomplish this, we employ several techniques, including the utilization of a large-scale pre-trained network, feature-wise knowledge distillation, and task-specific fine-tuning. These approaches are designed to effectively and accurately analyze facial expressions by leveraging visual information, thereby facilitating the implementation of real-time interactive applications. In terms of Action Unit (AU) intensity estimation, we achieve a Pearson Correlation Coefficient (PCC) of 0.63 on DISFA, which is 7% higher than the performance of OpenFace 2.0 while maintaining highly-efficient inference that runs two times faster than OpenFace 2.0. Despite being compact, our model also demonstrates competitive performance to state-of-the-art facial expression analysis methods on AffecNet, FFHQ, and RAF-DB. Our code will be released at https://github.com/ihp-lab/LibreFace},
note = {arXiv:2308.10713 [cs]},
keywords = {DTIC, Virtual Humans},
pubstate = {published},
tppubtype = {misc}
}