deepfake

Deepfake
  • 文章类型: Journal Article
    被篡改的多媒体内容越来越多地用于广泛的网络犯罪活动中。假新闻的传播,错误信息,数字绑架,与勒索软件有关的犯罪是最常见的犯罪之一,其中操纵的数码照片和视频是犯罪和传播媒介。刑事调查在应用机器学习技术自动区分伪造和真实的没收照片和视频方面受到了挑战。尽管需要手动验证,易于使用的数字取证平台对于自动化和促进篡改内容的检测以及帮助刑事调查人员的工作至关重要。本文提出了一种基于机器学习的支持向量机(SVM)区分真假多媒体文件的方法,即数码照片和视频,这可能表明存在deepfake内容。该方法在Python中实现,并在广泛使用的数字取证应用程序Autopsy中集成为新模块。所实现的方法提取了一组简单的特征,这些特征是通过将离散傅立叶变换(DFT)应用于数码照片和视频帧而产生的。使用大型分类多媒体文件数据集对模型进行了评估,该数据集包含合法和伪造的照片以及从视频中提取的帧。关于视频中的深度伪造检测,使用Celeb-DFv1数据集,以从YouTube收集的590个原始视频为特色,涵盖不同的主题。通过5倍交叉验证获得的结果优于文献中记录的基于SVM的方法,通过达到99.53%的平均F1分数,79.55%,和89.10%,分别为照片,视频,以及两种内容的混合。还使用最先进的方法进行了基准测试,通过将提出的SVM方法与深度学习方法进行比较,即卷积神经网络(CNN)。尽管CNN的性能优于所提出的DFT-SVM复合方法,DFT-SVM获得的结果的竞争力和大幅减少的处理时间使其适合实施和嵌入尸检模块,通过预测为每个分析的多媒体文件计算的虚假程度。
    Tampered multimedia content is being increasingly used in a broad range of cybercrime activities. The spread of fake news, misinformation, digital kidnapping, and ransomware-related crimes are amongst the most recurrent crimes in which manipulated digital photos and videos are the perpetrating and disseminating medium. Criminal investigation has been challenged in applying machine learning techniques to automatically distinguish between fake and genuine seized photos and videos. Despite the pertinent need for manual validation, easy-to-use platforms for digital forensics are essential to automate and facilitate the detection of tampered content and to help criminal investigators with their work. This paper presents a machine learning Support Vector Machines (SVM) based method to distinguish between genuine and fake multimedia files, namely digital photos and videos, which may indicate the presence of deepfake content. The method was implemented in Python and integrated as new modules in the widely used digital forensics application Autopsy. The implemented approach extracts a set of simple features resulting from the application of a Discrete Fourier Transform (DFT) to digital photos and video frames. The model was evaluated with a large dataset of classified multimedia files containing both legitimate and fake photos and frames extracted from videos. Regarding deepfake detection in videos, the Celeb-DFv1 dataset was used, featuring 590 original videos collected from YouTube, and covering different subjects. The results obtained with the 5-fold cross-validation outperformed those SVM-based methods documented in the literature, by achieving an average F1-score of 99.53%, 79.55%, and 89.10%, respectively for photos, videos, and a mixture of both types of content. A benchmark with state-of-the-art methods was also done, by comparing the proposed SVM method with deep learning approaches, namely Convolutional Neural Networks (CNN). Despite CNN having outperformed the proposed DFT-SVM compound method, the competitiveness of the results attained by DFT-SVM and the substantially reduced processing time make it appropriate to be implemented and embedded into Autopsy modules, by predicting the level of fakeness calculated for each analyzed multimedia file.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:数字时代见证了对新闻和信息的数字平台的日益依赖,再加上“deepfake”技术的出现。Deepfakes,利用语音记录和图像的大量数据集的深度学习模型,对媒体真实性构成重大威胁,可能导致不道德的滥用,如冒充和传播虚假信息。
    目标:为了应对这一挑战,这项研究旨在引入先天生物过程的概念,以区分真实的人类声音和克隆的声音。我们建议存在或不存在某些感知特征,比如讲话中的停顿,可以有效区分克隆和真实的音频。
    方法:共招募了49名具有不同种族背景和口音的成年参与者。每个参与者贡献语音样本,用于训练多达3个不同的语音克隆文本到语音模型和3个控制段落。随后,克隆模型生成了控制段落的合成版本,产生由每个参与者多达9个克隆音频样本和3个对照样本组成的数据集。我们分析了呼吸等生物行为引起的语音停顿,吞咽,和认知过程。计算了对应于语音暂停简档的五个音频特征。评估了这些特征的真实音频和克隆音频之间的差异,和5个经典的机器学习算法实现了使用这些特征来创建预测模型。通过对看不见的数据进行测试,评估了最优模型的泛化能力,结合了一个朴素的生成器,一个模型天真的段落,和幼稚的参与者。
    结果:克隆音频显示暂停之间的时间显着增加(P<.001),语音段长度的变化减少(P=0.003),发言时间的总比例增加(P=.04),语音中的micro和macropauses比率降低(P=0.01)。使用这些功能实现了五个机器学习模型,AdaBoost模型展示了最高的性能,实现5倍交叉验证平衡精度为0.81(SD0.05)。其他模型包括支持向量机(平衡精度0.79,SD0.03),随机森林(平衡精度0.78,SD0.04),逻辑回归,和决策树(平衡精度0.76,SD0.10和0.72,SD0.06)。在评估最优AdaBoost模型时,在预测未知数据时,它实现了0.79的总体测试准确性。
    结论:引入感知,机器学习模型中的生物特征在区分真实的人类声音和克隆音频方面显示出有希望的结果。
    BACKGROUND: The digital era has witnessed an escalating dependence on digital platforms for news and information, coupled with the advent of \"deepfake\" technology. Deepfakes, leveraging deep learning models on extensive data sets of voice recordings and images, pose substantial threats to media authenticity, potentially leading to unethical misuse such as impersonation and the dissemination of false information.
    OBJECTIVE: To counteract this challenge, this study aims to introduce the concept of innate biological processes to discern between authentic human voices and cloned voices. We propose that the presence or absence of certain perceptual features, such as pauses in speech, can effectively distinguish between cloned and authentic audio.
    METHODS: A total of 49 adult participants representing diverse ethnic backgrounds and accents were recruited. Each participant contributed voice samples for the training of up to 3 distinct voice cloning text-to-speech models and 3 control paragraphs. Subsequently, the cloning models generated synthetic versions of the control paragraphs, resulting in a data set consisting of up to 9 cloned audio samples and 3 control samples per participant. We analyzed the speech pauses caused by biological actions such as respiration, swallowing, and cognitive processes. Five audio features corresponding to speech pause profiles were calculated. Differences between authentic and cloned audio for these features were assessed, and 5 classical machine learning algorithms were implemented using these features to create a prediction model. The generalization capability of the optimal model was evaluated through testing on unseen data, incorporating a model-naive generator, a model-naive paragraph, and model-naive participants.
    RESULTS: Cloned audio exhibited significantly increased time between pauses (P<.001), decreased variation in speech segment length (P=.003), increased overall proportion of time speaking (P=.04), and decreased rates of micro- and macropauses in speech (both P=.01). Five machine learning models were implemented using these features, with the AdaBoost model demonstrating the highest performance, achieving a 5-fold cross-validation balanced accuracy of 0.81 (SD 0.05). Other models included support vector machine (balanced accuracy 0.79, SD 0.03), random forest (balanced accuracy 0.78, SD 0.04), logistic regression, and decision tree (balanced accuracies 0.76, SD 0.10 and 0.72, SD 0.06). When evaluating the optimal AdaBoost model, it achieved an overall test accuracy of 0.79 when predicting unseen data.
    CONCLUSIONS: The incorporation of perceptual, biological features into machine learning models demonstrates promising results in distinguishing between authentic human voices and cloned audio.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    Deepfake技术的快速发展对被操纵的媒体带来的错误信息和欺诈威胁不断升级。尽管存在风险,对deepfake检测技术的全面了解尚未实现。这项研究通过提供对用于检测深度假货的数字取证方法的最新系统调查来解决这一知识差距。遵循严格的方法,巩固最近关于深度伪造检测创新的出版物的发现。分析了支持新技术的流行数据集。已建立和新兴的跨模态检测方法的有效性和局限性,包括图像,视频,文本和音频进行评估。通过对备受瞩目的deepfake事件的案例研究,分享了对现实世界表现的见解。强调了当前围绕跨模态检测等方面的研究局限性,以指导未来的工作。这项及时的调查为研究人员提供了信息,从业者和政策制定者全面了解深度伪造检测的最新技术。结论是,不断创新对于应对快速发展的技术格局至关重要,从而实现深度伪造。
    The rapid advancement of deepfake technology poses an escalating threat of misinformation and fraud enabled by manipulated media. Despite the risks, a comprehensive understanding of deepfake detection techniques has not materialized. This research tackles this knowledge gap by providing an up-to-date systematic survey of the digital forensic methods used to detect deepfakes. A rigorous methodology is followed, consolidating findings from recent publications on deepfake detection innovation. Prevalent datasets that underpin new techniques are analyzed. The effectiveness and limitations of established and emerging detection approaches across modalities including image, video, text and audio are evaluated. Insights into real-world performance are shared through case studies of high-profile deepfake incidents. Current research limitations around aspects like cross-modality detection are highlighted to inform future work. This timely survey furnishes researchers, practitioners and policymakers with a holistic overview of the state-of-the-art in deepfake detection. It concludes that continuous innovation is imperative to counter the rapidly evolving technological landscape enabling deepfakes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    视频记录准确地捕捉面部表情运动;然而,对于面部感知研究人员来说,它们很难标准化和操纵。出于这个原因,照片的动态变形经常被使用,尽管他们缺乏自然的面部运动。这项研究旨在研究人类如何使用真实视频和两种不同的方法来人工生成动态表情-动态变形,从面部感知情绪。和AI合成的深度假货。与视频(所有情绪)和深度假货(恐惧,快乐,sad).视频和deepfakes被认为是类似的。此外,他们感觉到了快乐和悲伤的变形,但没有变形的愤怒或恐惧,不如其他格式真实。我们的发现支持先前的研究,表明对变形情绪的社会反应并不代表视频记录。研究结果还表明,与变体相比,深度假货可能提供更合适的标准化刺激类型。此外,从参与者那里收集定性数据,并使用ChatGPT进行分析,一个大的语言模型。ChatGPT成功地在数据中确定了与独立人类研究人员确定的主题一致的主题。根据这一分析,我们的参与者认为动态变形与视频和深度假货相比不那么自然。参与者认为深度假货和视频类似地表明,深度假货有效地复制了自然的面部运动,使它们成为面部感知研究的有希望的替代品。这项研究有助于越来越多的研究探索生成人工智能对推进人类感知研究的有用性。
    Video recordings accurately capture facial expression movements; however, they are difficult for face perception researchers to standardise and manipulate. For this reason, dynamic morphs of photographs are often used, despite their lack of naturalistic facial motion. This study aimed to investigate how humans perceive emotions from faces using real videos and two different approaches to artificially generating dynamic expressions - dynamic morphs, and AI-synthesised deepfakes. Our participants perceived dynamic morphed expressions as less intense when compared with videos (all emotions) and deepfakes (fearful, happy, sad). Videos and deepfakes were perceived similarly. Additionally, they perceived morphed happiness and sadness, but not morphed anger or fear, as less genuine than other formats. Our findings support previous research indicating that social responses to morphed emotions are not representative of those to video recordings. The findings also suggest that deepfakes may offer a more suitable standardized stimulus type compared to morphs. Additionally, qualitative data were collected from participants and analysed using ChatGPT, a large language model. ChatGPT successfully identified themes in the data consistent with those identified by an independent human researcher. According to this analysis, our participants perceived dynamic morphs as less natural compared with videos and deepfakes. That participants perceived deepfakes and videos similarly suggests that deepfakes effectively replicate natural facial movements, making them a promising alternative for face perception research. The study contributes to the growing body of research exploring the usefulness of generative artificial intelligence for advancing the study of human perception.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    最近,视频编辑软件有了显著的进步。这些进步已经允许新手或那些不使用先进的计算机技术的人生成在视觉上无法区分人眼与真实的视频到人类观察者的视频。因此,deepfake技术的应用有可能扩大身份盗窃的范围,这对全球安全构成了重大风险和巨大挑战。有必要开发一种检测虚假视频的有效方法。这里,我们介绍了一种新颖的方法,该方法采用卷积神经网络(CNN)和高斯混合模型(GMM)来有效区分假图像和真实图像或视频。所提出的方法提出了一种新颖的CNN-GMM架构,其中CNN中的全连接(FC)层被定制的高斯混合模型(GMM)全连接层取代。GMM层利用高斯概率密度函数(PDF)的加权集合来表示真实和假图像中的数据频率的分布。该表示指示由于添加的噪声而在操纵图像的分布中存在偏移。CNN-GMM模型展示了准确识别由概率分布内的不同类型的深度假货产生的变化的能力。它实现了高水平的分类精度,训练准确率高达100%,验证准确率高达96%。尽管正品类别与假冒类别的比率为16.6%至83.4%,CNN-GMM模型在召回方面表现出高性能指标,准确度,和F分数在分类最不真实的类别时。
    Recently, there have been notable advancements in video editing software. These advancements have allowed novices or those without access to advanced computer technology to generate videos that are visually indistinguishable to the human eye from real ones to the human observer. Therefore, the application of deepfake technology has the potential to expand the scope of identity theft, which poses a significant risk and a formidable challenge to global security. The development of an effective approach for detecting fake videos is necessary. Here, we introduce a novel methodology that employs a convolutional neural network (CNN) and Gaussian mixture model (GMM) to effectively differentiate between fake and real images or videos. The proposed methodology presents a novel CNN-GMM architecture in which the fully connected (FC) layer in the CNN is replaced with a customized Gaussian mixture model (GMM) fully connected layer. The GMM layer utilizes a weighted set of Gaussian probability density functions (PDFs) to represent the distribution of data frequencies in both real and fake images. This representation indicates there is a shift in the distribution of the manipulated images due to added noise. The CNN-GMM model demonstrates the ability to accurately identify variations resulting from different types of deepfakes within the probability distribution. It achieves a high level of classification accuracy, reaching up to 100% in training accuracy and up to 96% in validation accuracy. Notwithstanding the ratio of the genuine class to the counterfeit class being 16.6% to 83.4%, the CNN-GMM model exhibited high-performance metrics in terms of recall, accuracy, and F-score when classifying the least genuine class.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在欧洲议会最近的讨论中,确定了对所谓的高风险人工智能(AI)系统的法规的需求,目前已在即将到来的《欧盟人工智能法》(AIA)中进行了编纂,并获得了欧洲议会的批准。AIA是第一个被转变为欧洲法律的文件。该计划的重点是将AI系统转变为决策支持系统(人在回路和人在指挥),其中人类操作员仍然控制系统。虽然这据说解决了问责制问题,它包括,一方面,必要的人机交互是潜在的新错误来源;另一方面,这可能是一种非常有效的决策解释和验证方法。本文讨论了AIA生效后对高风险AI系统的必要要求。特别注意决策支持系统和增加系统可解释性所带来的机会和局限性。使用DeepFake检测的媒体取证任务的示例来说明这一点。
    In recent discussions in the European Parliament, the need for regulations for so-called high-risk artificial intelligence (AI) systems was identified, which are currently codified in the upcoming EU Artificial Intelligence Act (AIA) and approved by the European Parliament. The AIA is the first document to be turned into European Law. This initiative focuses on turning AI systems in decision support systems (human-in-the-loop and human-in-command), where the human operator remains in control of the system. While this supposedly solves accountability issues, it includes, on one hand, the necessary human-computer interaction as a potential new source of errors; on the other hand, it is potentially a very effective approach for decision interpretation and verification. This paper discusses the necessary requirements for high-risk AI systems once the AIA comes into force. Particular attention is paid to the opportunities and limitations that result from the decision support system and increasing the explainability of the system. This is illustrated using the example of the media forensic task of DeepFake detection.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    “deepfake”视频技术的快速发展——使用深度学习人工智能算法来创建看起来真实的假视频——给决策者和技术公司应该如何调节不真实内容的问题带来了紧迫性。我们进行了一项实验,以衡量人们对一组视频中高质量深度假货的警觉性和检测能力。首先,我们发现在没有内容警告的自然环境中,与仅观看真实视频的对照组(34.1%)相比,接触中性内容的深度虚假视频的个人不太可能检测到任何异常(32.9%)。第二,我们发现,当个人被警告说一组五个视频中至少有一个是深度假的,只有21.6%的受访者正确地将deepfake识别为唯一不真实的视频,而其余的则错误地选择了至少一个正版视频作为deepfake。
    The rapid advancement of \'deepfake\' video technology-which uses deep learning artificial intelligence algorithms to create fake videos that look real-has given urgency to the question of how policymakers and technology companies should moderate inauthentic content. We conduct an experiment to measure people\'s alertness to and ability to detect a high-quality deepfake among a set of videos. First, we find that in a natural setting with no content warnings, individuals who are exposed to a deepfake video of neutral content are no more likely to detect anything out of the ordinary (32.9%) compared to a control group who viewed only authentic videos (34.1%). Second, we find that when individuals are given a warning that at least one video in a set of five is a deepfake, only 21.6% of respondents correctly identify the deepfake as the only inauthentic video, while the remainder erroneously select at least one genuine video as a deepfake.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    这项研究是第一个调查模式与个体相信和分享不同形式的深度假货(也是深度假货)的倾向之间的关系的研究之一。使用在美国进行的在线调查实验,参与者被随机分配到三个虚假信息条件之一:视频深度伪造,音频deepfakes,和廉价的假货来测试单模态对多模态的影响,以及它如何影响个人感知的索赔准确性和分享意图。此外,还研究了认知能力对感知索赔准确性和条件之间共享意图的影响。结果表明,个人更有可能认为视频深度假货比廉价假货更准确,但不是音频deepfakes。然而,个人更有可能分享视频deepfake而不是廉价和音频deepfake。我们还发现,具有高认知能力的人不太可能认为深度假货是准确的,也不太可能在不同的格式中分享它们。研究结果强调,深度假货不是单一的,在研究用户对深度假货的参与时,应该考虑相关的模式。
    This study is one of the first to investigate the relationship between modalities and individuals\' tendencies to believe and share different forms of deepfakes (also deep fakes). Using an online survey experiment conducted in the US, participants were randomly assigned to one of three disinformation conditions: video deepfakes, audio deepfakes, and cheap fakes to test the effect of single modality against multimodality and how it affects individuals\' perceived claim accuracy and sharing intentions. In addition, the impact of cognitive ability on perceived claim accuracy and sharing intentions between conditions are also examined. The results suggest that individuals are likelier to perceive video deepfakes as more accurate than cheap fakes, but not audio deepfakes. Yet, individuals are more likely to share video deepfakes than cheap and audio deepfakes. We also found that individuals with high cognitive ability are less likely to perceive deepfakes as accurate or share them across formats. The findings emphasize that deepfakes are not monolithic, and associated modalities should be considered when studying user engagement with deepfakes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    每天都有成千上万的视频发布在网站和社交媒体上,包括Twitter,Facebook,WhatsApp,Instagram,和YouTube。报纸,执法出版物,刑事调查,监控系统,Banking,博物馆,军方,医学成像,保险索赔,和消费者摄影只是可以获得重要视觉数据的几个例子。因此,强大的处理工具的出现,可以很容易地在网上提供对视频的真实性构成了巨大的威胁。因此,区分真假数据至关重要。数字视频伪造检测技术用于验证和检查数字视频内容的真实性。深度学习算法最近在数字取证领域引起了极大的兴趣,例如循环神经网络(RNN),深度卷积神经网络(DCNN)和自适应神经网络(ANN)。在本文中,我们给出了软分类法,以及对多媒体伪造检测系统的最新研究的全面概述。首先,提供了理解视频伪造所需的基本知识。然后,提供了主动和被动视频操纵检测方法的摘要。反取证,压缩视频方法,视频取证所需的数据集,以及视频检测方法的挑战也得到了解决。在此之后,我们介绍了Deepfake的概述,并提供了检测所需的数据集。此外,涵盖了用于视频检测的有用软件包和取证工具。此外,本文概述了在视频取证应用中使用的视频分析工具。最后,我们强调研究的困难以及有趣的研究途径。总之,这项调查提供了详细的信息和更广泛的调查,以提取数据并在一个伞下检测欺诈视频内容。
    Thousands of videos are posted on websites and social media every day, including Twitter, Facebook, WhatsApp, Instagram, and YouTube. Newspapers, law enforcement publications, criminal investigations, surveillance systems, Banking, the museum, the military, imaging in medicine, insurance claims, and consumer photography are just a few examples of places where important visual data may be obtained. Thus, the emergence of powerful processing tools that can be easily made available online poses a huge threat to the authenticity of videos. Therefore, it\'s vital to distinguish between true and fake data. Digital video forgery detection techniques are used to validate and check the realness of digital video content. Deep learning algorithms lately sparked a lot of interest in the field of digital forensics, such as Recurrent Neural Networks (RNN), Deep Convolutional Neural Networks (DCNN), and Adaptive Neural Networks (ANN). In this paper, we give a soft taxonomy as well as a thorough overview of recent research on multimedia falsification detection systems. First, the basic knowledge needed to comprehend video forgery is provided. Then, a summary of active and passive video manipulation detection approaches is provided. Anti-forensics, compression video methods, datasets required for video forensics, and challenges of video detection approaches are also addressed. Following that, we presented an overview of deepfake, and the datasets required for detection were also provided. Also, helpful software packages and forensics tools for video detection are covered. In addition, this paper provides an overview of video analysis tools that are used in video forensic applications. Finally, we highlight research difficulties as well as interesting research avenues. In short, this survey provides detailed information and a broader investigation to extract data and detect fraud video contents under one umbrella.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    随着生成模型的发展,被滥用的Deepfakes引起了公众的关注。作为一种防御机制,人脸伪造检测方法已被深入研究。远程光电体积描记术(rPPG)技术通过检查心脏活动引起的皮肤颜色的细微变化,从录制的视频中提取心跳信号。由于面部伪造过程不可避免地破坏了面部颜色的周期性变化,rPPG信号被证明是Deepfake检测的强大生物指标。受到关键观察的激励,即rPPG信号在不同的操纵方法方面产生独特的节奏模式,我们将Deepfake检测也视为源检测任务。采用多尺度空间-时间PPG图以进一步利用来自多个面部区域的心跳信号。此外,为了捕捉空间和时间的不一致,我们提出了一个两阶段网络,由一个掩模引导的局部注意模块(MLA)来捕获PPG地图的独特的局部模式,和时间转换器,以长距离交互相邻PPG图的特征。在FaceForensics和Celeb-DF数据集上进行的大量实验证明了我们的方法优于所有其他基于rPPG的方法。可视化也证明了该方法的有效性。
    With the development of generative models, abused Deepfakes have aroused public concerns. As a defense mechanism, face forgery detection methods have been intensively studied. Remote photoplethysmography (rPPG) technology extract heartbeat signal from recorded videos by examining the subtle changes in skin color caused by cardiac activity. Since the face forgery process inevitably disrupts the periodic changes in facial color, rPPG signal proves to be a powerful biological indicator for Deepfake detection. Motivated by the key observation that rPPG signals produce unique rhythmic patterns in terms of different manipulation methods, we regard Deepfake detection also as a source detection task. The Multi-scale Spatial-Temporal PPG map is adopted to further exploit heartbeat signal from multiple facial regions. Moreover, to capture both spatial and temporal inconsistencies, we propose a two-stage network consisting of a Mask-Guided Local Attention module (MLA) to capture unique local patterns of PPG maps, and a Temporal Transformer to interact features of adjacent PPG maps in long distance. Abundant experiments on FaceForensics + + and Celeb-DF datasets prove the superiority of our method over all other rPPG-based approaches. Visualization also demonstrates the effectiveness of the proposed method.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号