deepfakes

deepfakes
  • 文章类型: Journal Article
    基于图像的性虐待(IBSA)是指非自愿的创造,taking,或者分享亲密的图像,包括威胁要分享图片。它还可以包括强迫某人分享亲密的图像,或发送不想要的亲密图像。近年来,人们越来越关注大自然,范围,以及IBSA的影响,但是对这些危害的实施关注相对较少。此范围审查整合并综合了有关IBSA对成年人的侵害的现有知识。该评论涉及对选定数据库中的学术和灰色文献的系统搜索。总的来说,26项研究符合纳入标准。如果研究在2013年至2023年之间以英语发表,并报告了16岁以上承认IBSA渗透行为的成年人样本的发现,则包括这些研究。该综述发现,在所有研究中,IBSA亚型的患病率差异很大。人们一致认为,从事IBSA犯罪的成年人更有可能是男性,年轻的成年人,LGBTIQ+动机是多方面的,但往往与社会奖励有关,动力动力学,性满足,报复性冲动。发现深色四角性状与IBSA侵染呈正相关。研究还表明,受害和犯罪之间存在重叠,以及与其他犯罪行为的联系,比如亲密伴侣的暴力。预防干预措施应侧重于改变机会,负担,以及违规的基础设施,以及解决有问题的社会态度和规范,早期干预侧重于建立韧性和自尊,促进健康的行为和尊重的关系。
    Image-based sexual abuse (IBSA) refers to the nonconsensual creating, taking, or sharing of intimate images, including threatening to share images. It can also include coercing someone into sharing intimate images, or sending unwanted intimate images. In recent years, there has been growing attention to the nature, scope, and impacts of IBSA, but comparatively little attention has been paid to the perpetration of these harms. This scoping review consolidates and synthesizes the existing knowledge on the perpetration of IBSA against adults. The review involved a systematic search of scholarly and gray literature across select databases. In total, 26 studies met the inclusion criteria. Studies were included if they were published in English between 2013 and 2023 and reported on findings of a sample of adults over the age of 16 who admitted IBSA perpetration behaviors. The review found that prevalence of subtypes of IBSA varied significantly across the studies. There was consensus that adults who engage in IBSA perpetration are more likely to be men, younger adults, and LGBTIQ+. Motivations were multifaceted, but tended to relate to social rewards, power dynamics, sexual gratification, and retaliatory impulses. Dark Tetrad traits were found to be positively associated with IBSA perpetration. The research also indicates on overlap between victimization and perpetration, as well as an association with other offending behaviors, such as intimate partner violence. Prevention interventions should be focused on changing the opportunities, affordances, and infrastructures for offending, as well as addressing problematic societal attitudes and norms, with early interventions focused on building resilience and self-esteem, and promoting healthy behaviors and respectful relationships.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:数字时代见证了对新闻和信息的数字平台的日益依赖,再加上“deepfake”技术的出现。Deepfakes,利用语音记录和图像的大量数据集的深度学习模型,对媒体真实性构成重大威胁,可能导致不道德的滥用,如冒充和传播虚假信息。
    目标:为了应对这一挑战,这项研究旨在引入先天生物过程的概念,以区分真实的人类声音和克隆的声音。我们建议存在或不存在某些感知特征,比如讲话中的停顿,可以有效区分克隆和真实的音频。
    方法:共招募了49名具有不同种族背景和口音的成年参与者。每个参与者贡献语音样本,用于训练多达3个不同的语音克隆文本到语音模型和3个控制段落。随后,克隆模型生成了控制段落的合成版本,产生由每个参与者多达9个克隆音频样本和3个对照样本组成的数据集。我们分析了呼吸等生物行为引起的语音停顿,吞咽,和认知过程。计算了对应于语音暂停简档的五个音频特征。评估了这些特征的真实音频和克隆音频之间的差异,和5个经典的机器学习算法实现了使用这些特征来创建预测模型。通过对看不见的数据进行测试,评估了最优模型的泛化能力,结合了一个朴素的生成器,一个模型天真的段落,和幼稚的参与者。
    结果:克隆音频显示暂停之间的时间显着增加(P<.001),语音段长度的变化减少(P=0.003),发言时间的总比例增加(P=.04),语音中的micro和macropauses比率降低(P=0.01)。使用这些功能实现了五个机器学习模型,AdaBoost模型展示了最高的性能,实现5倍交叉验证平衡精度为0.81(SD0.05)。其他模型包括支持向量机(平衡精度0.79,SD0.03),随机森林(平衡精度0.78,SD0.04),逻辑回归,和决策树(平衡精度0.76,SD0.10和0.72,SD0.06)。在评估最优AdaBoost模型时,在预测未知数据时,它实现了0.79的总体测试准确性。
    结论:引入感知,机器学习模型中的生物特征在区分真实的人类声音和克隆音频方面显示出有希望的结果。
    BACKGROUND: The digital era has witnessed an escalating dependence on digital platforms for news and information, coupled with the advent of \"deepfake\" technology. Deepfakes, leveraging deep learning models on extensive data sets of voice recordings and images, pose substantial threats to media authenticity, potentially leading to unethical misuse such as impersonation and the dissemination of false information.
    OBJECTIVE: To counteract this challenge, this study aims to introduce the concept of innate biological processes to discern between authentic human voices and cloned voices. We propose that the presence or absence of certain perceptual features, such as pauses in speech, can effectively distinguish between cloned and authentic audio.
    METHODS: A total of 49 adult participants representing diverse ethnic backgrounds and accents were recruited. Each participant contributed voice samples for the training of up to 3 distinct voice cloning text-to-speech models and 3 control paragraphs. Subsequently, the cloning models generated synthetic versions of the control paragraphs, resulting in a data set consisting of up to 9 cloned audio samples and 3 control samples per participant. We analyzed the speech pauses caused by biological actions such as respiration, swallowing, and cognitive processes. Five audio features corresponding to speech pause profiles were calculated. Differences between authentic and cloned audio for these features were assessed, and 5 classical machine learning algorithms were implemented using these features to create a prediction model. The generalization capability of the optimal model was evaluated through testing on unseen data, incorporating a model-naive generator, a model-naive paragraph, and model-naive participants.
    RESULTS: Cloned audio exhibited significantly increased time between pauses (P<.001), decreased variation in speech segment length (P=.003), increased overall proportion of time speaking (P=.04), and decreased rates of micro- and macropauses in speech (both P=.01). Five machine learning models were implemented using these features, with the AdaBoost model demonstrating the highest performance, achieving a 5-fold cross-validation balanced accuracy of 0.81 (SD 0.05). Other models included support vector machine (balanced accuracy 0.79, SD 0.03), random forest (balanced accuracy 0.78, SD 0.04), logistic regression, and decision tree (balanced accuracies 0.76, SD 0.10 and 0.72, SD 0.06). When evaluating the optimal AdaBoost model, it achieved an overall test accuracy of 0.79 when predicting unseen data.
    CONCLUSIONS: The incorporation of perceptual, biological features into machine learning models demonstrates promising results in distinguishing between authentic human voices and cloned audio.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    由于视觉错误信息具有说服力和在社交媒体上迅速传播的潜力,因此对公共卫生构成了独特的挑战。在这篇文章中,匹兹堡大学健康科学图书馆系统的图书馆员识别了四种类型的视觉健康错误信息:误导性图表,脱离上下文的视觉效果,科学出版物中的图像处理,以及AI生成的图像和视频。为了教育我们校园的健康科学受众和更广泛的社区关于这些主题,我们已经开发了一系列关于视觉健康错误信息的指导。我们描述了我们的策略,并为各种受众实施视觉错误信息编程提供了建议。
    Visual misinformation poses unique challenges to public health due to its potential for persuasiveness and rapid spread on social media. In this article, librarians at the University of Pittsburgh Health Sciences Library System identify four types of visual health misinformation: misleading graphs and charts, out of context visuals, image manipulation in scientific publications, and AI-generated images and videos. To educate our campus\'s health sciences audience and wider community on these topics, we have developed a range of instruction about visual health misinformation. We describe our strategies and provide suggestions for implementing visual misinformation programming for a variety of audiences.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    人脑在面部处理(FP)和社交互动决策中的作用取决于准确识别面部。然而,深度假货的流行,AI生成的图像,从合成身份中辨别真实身份提出了挑战。这项研究调查了健康个体在视觉辨别任务中的认知和情感参与,该任务涉及真实和深层的人脸表达积极,负,或中性情绪。使用21通道干EEG耳机从23名健康参与者收集脑电图(EEG)数据;进行了功率谱和事件相关电位(ERP)分析。结果显示,根据刺激的真实性和情绪内容,特定大脑区域的激活具有统计学意义。功率谱分析强调了右半球在θ的优势,阿尔法,高贝塔,和真实面孔的伽马带,而深假货主要影响三角洲带的额叶和枕叶区域。ERP分析暗示了区分真实面孔和合成面孔的可能性,当观察右额叶(LF)和左颞枕骨(LTO)区域的真实面孔时,N250(刺激发作后200-300ms)峰值潜伏期减少,而且在情感中,由于快乐面孔的右侧颞枕骨(RTO)区域的P100(90-140ms)峰值幅度高于中性和悲伤面孔。
    The human brain\'s role in face processing (FP) and decision making for social interactions depends on recognizing faces accurately. However, the prevalence of deepfakes, AI-generated images, poses challenges in discerning real from synthetic identities. This study investigated healthy individuals\' cognitive and emotional engagement in a visual discrimination task involving real and deepfake human faces expressing positive, negative, or neutral emotions. Electroencephalographic (EEG) data were collected from 23 healthy participants using a 21-channel dry-EEG headset; power spectrum and event-related potential (ERP) analyses were performed. Results revealed statistically significant activations in specific brain areas depending on the authenticity and emotional content of the stimuli. Power spectrum analysis highlighted a right-hemisphere predominance in theta, alpha, high-beta, and gamma bands for real faces, while deepfakes mainly affected the frontal and occipital areas in the delta band. ERP analysis hinted at the possibility of discriminating between real and synthetic faces, as N250 (200-300 ms after stimulus onset) peak latency decreased when observing real faces in the right frontal (LF) and left temporo-occipital (LTO) areas, but also within emotions, as P100 (90-140 ms) peak amplitude was found higher in the right temporo-occipital (RTO) area for happy faces with respect to neutral and sad ones.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    检测DeepFake视频已成为现代多媒体取证应用中的中心任务。本文提出了一种当视频中描绘的人物已知时检测面部交换视频的方法。我们建议使用基于从训练用于面部识别的深度卷积神经网络(DCNN)获得的相似性得分的阈值分类器。我们计算从被质疑的视频和所描绘的人的参考资料中提取的面部之间的一组相似性得分。我们用最高分把被质疑的视频分类为真假,取决于所选择的阈值。我们在Celeb-DF(v2)数据集上验证了我们的方法(Li等人。,2020)[13]。使用在数据集上指定的训练和测试分割,我们获得的HTER为0.020,AUC为0.994,超过了针对该数据集的最稳健方法(Tran等人。,2021)[37]。此外,使用逻辑回归模型将最高分转换为似然比,以便在法医分析中更适用.
    Detecting DeepFake videos has become a central task in modern multimedia forensics applications. This article presents a method to detect face swapped videos when the portrayed person in the video is known. We propose using a threshold classifier based on similarity scores obtained from a Deep Convolutional Neural Network (DCNN) trained for facial recognition. We compute a set of similarity scores between faces extracted from questioned videos and reference materials of the person depicted. We use the highest score to classify the questioned videos as authentic or fake, depending on the threshold chosen. We validate our method on the Celeb-DF (v2) dataset (Li et al., 2020) [13]. Using the training and testing splits specified on the dataset, we obtained an HTER of 0.020 and an AUC of 0.994, surpassing the most robust approaches against this dataset (Tran et al., 2021) [37]. Additionally, a logistic regression model was used to convert the highest score into a likelihood ratio for greater applicability in forensic analyses.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Deepfakes是一种令人不安的虚假信息形式,引起了越来越多的关注。然而,仍然缺乏对深度伪造分享行为的心理学解释,并且在非西方背景下缺乏对深度伪造的公共知识的研究知识。我们在八个国家进行了一项跨国调查研究,以检查错过恐惧(FOMO)的作用,缺乏自我调节(DSR),和深度虚假分享行为的认知能力。结果来自七个南亚背景下的比较调查(中国,印度尼西亚,马来西亚,菲律宾,新加坡,泰国,和越南),并将这些发现与美国进行比较,关于深度假货的讨论最为相关。总的来说,结果表明,那些认为深度假货是准确的人更有可能在社交媒体上分享它们。此外,在所有国家,分享也是由社会心理特征-FOMO驱动的。社交媒体使用的DSR也被发现是解释Deepfake共享的关键因素。还观察到,认知能力低的人更有可能分享深度假货。然而,我们还发现,DSR对社交媒体和FOMO的影响并不取决于用户的认知能力。这项研究的结果有助于限制深度假货在社交媒体上传播的策略。
    Deepfakes are a troubling form of disinformation that has been drawing increasing attention. Yet, there remains a lack of psychological explanations for deepfake sharing behavior and an absence of research knowledge in non-Western contexts where public knowledge of deepfakes is limited. We conduct a cross-national survey study in eight countries to examine the role of fear of missing out (FOMO), deficient self-regulation (DSR), and cognitive ability in deepfake sharing behavior. Results are drawn from a comparative survey in seven South Asian contexts (China, Indonesia, Malaysia, Philippines, Singapore, Thailand, and Vietnam) and compare these findings to the United States, where discussions about deepfakes have been most relevant. Overall, the results suggest that those who perceive the deepfakes to be accurate are more likely to share them on social media. Furthermore, in all countries, sharing is also driven by the social-psychological trait - FOMO. DSR of social media use was also found to be a critical factor in explaining deepfake sharing. It is also observed that individuals with low cognitive ability are more likely to share deepfakes. However, we also find that the effects of DSR on social media and FOMO are not contingent upon users\' cognitive ability. The results of this study contribute to strategies to limit deepfakes propagation on social media.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    人工智能技术的最新发展导致了越来越复杂的视频操作形式。其中一种形式是deepfakes的出现。Deepfakes是人工智能生成的视频,通常描述人们做和说他们从未做过的事情。在本文中,我证明,在认识论中,深度假货与更传统的假谷仓案例之间存在密切的结构关系。具体来说,我认为,深度假货会产生与传统案例中相似的认知风险。鉴于谷仓案例对美德理论知识的解释构成了长期的挑战,我考虑类似的挑战是否延伸到深度假货。在这样做的时候,我考虑邓肯·普里查德最近的反风险美德认识论如何应对挑战。虽然普里查德的账户避免了传统谷仓案例中的问题,我声称,在Deepfake的情况下,这导致当地对在线视频中的知识持怀疑态度。最后,我考虑了两种替代的美德理论方法如何证明我们在日益数字化的世界中对视频的认识依赖。
    Recent develops in AI technology have led to increasingly sophisticated forms of video manipulation. One such form has been the advent of deepfakes. Deepfakes are AI-generated videos that typically depict people doing and saying things they never did. In this paper, I demonstrate that there is a close structural relationship between deepfakes and more traditional fake barn cases in epistemology. Specifically, I argue that deepfakes generate an analogous degree of epistemic risk to that which is found in traditional cases. Given that barn cases have posed a long-standing challenge for virtue-theoretic accounts of knowledge, I consider whether a similar challenge extends to deepfakes. In doing so, I consider how Duncan Pritchard\'s recent anti-risk virtue epistemology meets the challenge. While Pritchard\'s account avoids problems in traditional barn cases, I claim that it leads to local scepticism about knowledge from online videos in the case of deepfakes. I end by considering how two alternative virtue-theoretic approaches might vindicate our epistemic dependence on videos in an increasingly digital world.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    深度假货的滥用,一种新兴的换脸技术,引起人们对视觉内容真实性和错误信息传播的严重担忧。为了减轻深假货带来的威胁,已经部署了大量以数据为中心的探测器。然而,这些方法的性能很容易因deepfakes的降级而缺陷。为了提高退化deepfake检测的性能,我们创造性地探索了在特征空间中的恢复方法,以保留用于检测的伪影,而不是直接在图像域中。在本文中,我们提出了一种方法,即DF-UDetector,通过对退化图像进行建模并将提取的特征转换为高质量水平,以防止退化深度伪造。具体而言,整个模型由三个关键部分组成:图像特征提取器,用于捕获图像特征,特征转换模块,用于将退化特征映射为更高质量,和鉴别器,以确定特征图是否具有足够的高质量。在多个视频数据集上进行的大量实验表明,我们提出的模型的性能与最先进的模型相比甚至更好。此外,在野外检测深度假货时,DF-UDetector的性能较小。
    The abuse of deepfakes, a rising face swap technique, causes severe concerns about the authenticity of visual content and the dissemination of misinformation. To alleviate the threats imposed by deepfakes, a vast body of data-centric detectors has been deployed. However, the performance of these methods can be easily defected by degradations on deepfakes. To improve the performance of degradation deepfake detection, we creatively explore the recovery method in the feature space to preserve the artifacts for detection instead of directly in the image domain. In this paper, we propose a method, namely DF-UDetector, against degradation deepfakes by modeling the degraded images and transforming the extracted features to a high-quality level. To be specific, the whole model consists of three key components: an image feature extractor to capture image features, a feature transforming module to map the degradation features into a higher quality, and a discriminator to determine whether the feature map is of high quality enough. Extensive experiments on multiple video datasets show that our proposed model performs comparably or even better than state-of-the-art counterparts. Moreover, DF-UDetector outperforms by a small margin when detecting deepfakes in the wild.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    人脸是我们身份和表达的核心部分,因此,理解面部几何形状是捕获这些信息的关键。寻求利用这些信息的自动化系统必须具有一种使面部特征可访问的建模方式。分层,多级体系结构具有捕获所涉及的不同分辨率的表示的能力。在这项工作中,我们建议使用分层变压器架构作为捕获面部几何结构的鲁棒表示的手段。我们进一步证明了我们的方法的多功能性,使用这个变压器作为骨干,以支持三个面部表现问题:面部反欺骗,面部表情表示,和深度假检测。有效的细粒度细节与全球注意力表征的结合使得该架构成为这些面部表征问题的优秀候选者。我们进行了许多实验,首先展示了我们的方法解决面部建模中常见问题的能力(姿势,闭塞,和背景变化)并捕捉面部对称性,然后证明它在三个补充任务上的有效性。
    Human faces are a core part of our identity and expression, and thus, understanding facial geometry is key to capturing this information. Automated systems that seek to make use of this information must have a way of modeling facial features in a way that makes them accessible. Hierarchical, multi-level architectures have the capability of capturing the different resolutions of representation involved. In this work, we propose using a hierarchical transformer architecture as a means of capturing a robust representation of facial geometry. We further demonstrate the versatility of our approach by using this transformer as a backbone to support three facial representation problems: face anti-spoofing, facial expression representation, and deepfake detection. The combination of effective fine-grained details alongside global attention representations makes this architecture an excellent candidate for these facial representation problems. We conduct numerous experiments first showcasing the ability of our approach to address common issues in facial modeling (pose, occlusions, and background variation) and capture facial symmetry, then demonstrating its effectiveness on three supplemental tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    深度学习技术的进步和免费的可用性,大型数据库使之成为可能,即使是非技术人员,为良性和恶意目的操纵或生成逼真的面部样本。DeepFakes指的是面部多媒体内容,它已经被数字化改变或使用深度神经网络合成创建。本文首先概述了现成的面部编辑应用程序以及面部识别系统在各种面部操作下的漏洞(或性能下降)。接下来,这项调查概述了近年来针对deepfake和面部操纵进行的技术和工作。尤其是,审查了四种深度假或面部操纵,即,身份交换,脸重演,属性操纵,和整个面部合成。对于每个类别,详细介绍了deepfake或人脸操纵生成方法以及这些操纵检测方法。尽管基于传统和先进的计算机视觉取得了重大进展,人工智能,和物理学,攻击者/罪犯/对手之间仍然存在巨大的军备竞赛(即,DeepFake生成方法)和防御者(即,DeepFake检测方法)。因此,还讨论了开放的挑战和潜在的研究方向。本文有望帮助读者理解深度伪造的产生和检测机制,以及开放的问题和未来的方向。
    Advancements in deep learning techniques and the availability of free, large databases have made it possible, even for non-technical people, to either manipulate or generate realistic facial samples for both benign and malicious purposes. DeepFakes refer to face multimedia content, which has been digitally altered or synthetically created using deep neural networks. The paper first outlines the readily available face editing apps and the vulnerability (or performance degradation) of face recognition systems under various face manipulations. Next, this survey presents an overview of the techniques and works that have been carried out in recent years for deepfake and face manipulations. Especially, four kinds of deepfake or face manipulations are reviewed, i.e., identity swap, face reenactment, attribute manipulation, and entire face synthesis. For each category, deepfake or face manipulation generation methods as well as those manipulation detection methods are detailed. Despite significant progress based on traditional and advanced computer vision, artificial intelligence, and physics, there is still a huge arms race surging up between attackers/offenders/adversaries (i.e., DeepFake generation methods) and defenders (i.e., DeepFake detection methods). Thus, open challenges and potential research directions are also discussed. This paper is expected to aid the readers in comprehending deepfake generation and detection mechanisms, together with open issues and future directions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号