Acemoglu, D., Laibson, D. & List, J. A. Equalizing superstars: The internet and the democratization of education. Am. Econ. Rev. 104, 523–527 (2014).
Adedoyin, O. B. & Soykan, E. Covid-19 pandemic and online learning: The challenges and opportunities. Interact. Learn. Environ. 31, 863–875 (2020).
Tan, E. Informal learning on YouTube: Exploring digital literacy in independent online learning. Learn. Media Technol. 38, 463–477 (2013).
Gutierrez-Martin, A., Torrego-Gonzalez, A. & Vicente-Marino, M. Media education with the monetization of YouTube: The loss of truth as an exchange value. Cult. Educ. 31, 267–295 (2019).
Radesky, J. S., Schaller, A., Yeo, S. L., Weeks, H. M. & Robb, M. B. Young kids and YouTube: How ads, toys, and games dominate viewing (Common Sense Media, 2020).
Kind, T. & Evans, Y. Social media for lifelong learning. Int Rev Psychiatry 27, 124–132 (2015).
Cinelli, M. et al. The COVID-19 social media infodemic. Sci. Rep. 10, 10 (2020).
Wang, Y. X., McKee, M., Torbica, A. & Stuckler, D. Systematic literature review on the spread of health-related misinformation on social media. Soc. Sci. Med. 240, 12 (2019).
Wittenberg, C., Tappin, B., Berinsky, A. J. & Rand, D. G. The (minimal) persuasive advantage of political video over text. Proc. Natl. Acad. Sci. U.S.A. 118, 7 (2021).
Goreis, A. & Voracek, M. A systematic review and meta-analysis of psychological research on conspiracy beliefs: Field characteristics, measurement instruments, and associations with personality traits. Front. Psychol. 10, 13 (2019).
Sylvester, S. M. COVID-19 and motivated reasoning: the influence of knowledge on covid-related policy and health behavior. Soc. Sci. Qrly. 19, 2341–2359 (2021).
Etkin, D. & Ho, E. Climate change: Perceptions and discourses of risk. J. Risk Res. 10, 623–641 (2007).
Loomba, S. et al. (Measuring the impact of COVID-19 vaccine misinformation on vaccination intent in the UK and USA. Nat. Hum. Behav. 5, 337–348 (2021).
Curiel, R. P. & Gonzalez-Ramirez, H. Vaccination strategies against COVID-19 and the diffusion of anti-vaccination views. Sci. Rep. https://doi.org/10.1038/s41598-021-85555-1 (2021).
Del Vicario, M. et al. The spreading of misinformation online. Proc. Natl. Acad. Sci. U.S.A. 113, 554–559 (2016).
Photiou, A., Nicolaides, C. & Dhillon, P. S. Social status and novelty drove the spread of online information during the early stages of COVID-19. Sci. Rep. 11, 7662–7669 (2021).
Scheufele, D. A. & Krause, N. M. Science audiences, misinformation, and fake news. Proc Natl Acad Sci U S A 116, 7662–7669 (2019).
Reyna, V. F. A scientific theory of gist communication and misinformation resistance, with implications for health, education, and policy. Proc. Natl. Acad. Sci. U.S.A. 118, 6 (2021).
Tulyakov, S., Liu, M. Y., Yang, X. D., & Kautz, J. MoCoGAN: Decomposing motion and content for video generation. In 31st IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (IEEE, Salt Lake City, UT), pp 1526–1535, IEEE (2018).
Mirsky, Y. & Lee, W. The creation and detection of deepfakes: A survey. ACM Comput. Surv. 54, 41 (2021).
Lewandowsky, S., Ecker, U. K. H. & Cook, J. Beyond misinformation: Understanding and coping with the “Post-Truth” Era. J. Appl. Res. Mem. Cogn. 6, 353–369 (2017).
Ng, A. Y. & Jordan, M. I. On discriminative vs. generative classifiers: A comparison of logistic regression and naive Bayes. Adv. Neural Inf. Process. Syst. 14, 841–848 (2002).
Zhao, C. et al. Dermoscopy image classification based on StyleGAN and DenseNet201. IEEE Access 9, 8659–8679 (2021).
Wu, Z. H., Lin, G. S., Tao, Q. Y. & Cai, J. F. M2E-Try on net: Fashion from model to everyone. Proceedings of the 27th ACM International Conference on Multimedia (MM’19), 293–301. ACM (2019).
Siarohin, A., Lathuiliere, S., Tulyakov, S., Ricci, E. & Sebe, N. First order motion model for image animation. Advances in Neural Information Processing Systems 32 (NIPS 2019) 32, 11 (2019).
Ahmed, S. R. A. & Sonuç, E. Deepfake detection using rationale-augmented convolutional neural network. Appl. Nanosci. 9, 1–9 (2021).
Chintha, A. et al. Recurrent convolutional structures for audio spoof and video deepfake detection. IEEE J. Select. Topics Signal Process. 14, 1024–1037 (2020).
Zhao T. et al., Learning self-consistency for deepfake detection. Proceedings of the IEEE/CVF International Conference on Computer Vision, 15023–15033 (2021).
Guarnera, L., Giudice, O. & Battiato, S. Deepfake detection by analyzing convolutional traces. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 666–667 (2020).
Nirkin, Y., Wolf, L., Keller, Y. & Hassner, T. DeepFake detection based on discrepancies between faces and their context. IEEE Trans. Pattern Anal. Mach. Intell. 44, 6111–6121 (2021).
Zhao, H. et al., Multi-attentional deepfake detection. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2185–2194 (2021).
Mittal, T., Bhattacharya, U., Chandra, R., Bera, A. & Manocha, D. Emotions don’t lie. In Proceedings of the 28th ACM International Conference on Multimedia, pp 2823–2832 (2020).
Chugh, K., Gupta, P., Dhall, A. & Subramanian, R. Not made for each other- audio-visual dissonance-based deepfake detection and localization. In Proceedings of the 28th ACM International Conference on Multimedia, pp 439–447 (2020).
Cozzolino, D., Rössler, A., Thies, A., Nießner, M. & Verdoliva, L. Id-reveal: Identity-aware deepfake video detection. Proceedings of the IEEE/CVF International Conference on Computer Vision, 15108–15117 (2021).
Das, S., Seferbekov, S., Datta, A., Islam, M. S. & Amin, M. R. Towards solving the deepfake problem: an analysis on improving deepfake detection using dynamic face augmentation. Proceedings of the IEEE/CVF International Conference on Computer Vision, 3776–3785 (2021).
Li, X. et al. (2020) Sharp Multiple Instance Learning for DeepFake Video Detection. in Proceedings of the 28th ACM International Conference on Multimedia, pp 1864–1872.
Trinh, L., Tsang, M., Rambhatla, S. & Liu, D. Y. Interpretable and trustworthy deepfake detection via dynamic prototypes. Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 1973–1983 (2021).
Yu, P. P., Xia, Z. H., Fei, J. W. & Lu, Y. J. A survey on deepfake video detection. IET Biometrics 10, 607–624 (2021).
Robertson, D. J. et al. Detecting morphed passport photos: A training and individual differences approach. Cogn. Res. –Princ. Implic. 3, 11 (2018).
Choudrie, J. et al. Machine learning techniques and older adults processing of online information and misinformation: A covid 19 study. Comput. Hum. Behav. 119, 11 (2021).
Groh, M., Epstein, Z., Firestone, C. & Picard, R. Deepfake detection by human crowds, machines, and machine-informed crowds. Proc. Natl. Acad. Sci. 119(1), e2110013119 (2022).
Phillips, P. J. et al. Face recognition accuracy of forensic examiners, superrecognizers, and face recognition algorithms. Proc. Natl. Acad. Sci. U.S.A. 115, 6171–6176 (2018).
Levchuk, G., Fouse, A., Pattipati, K. & Serfaty, D. & McCormack, R. Active learning and structure adaptation in teams of heterogeneous agents: designing organizations of the future. In Conference on Next-Generation Analyst VI (Spie-Int Soc Optical Engineering, 2018).
Wixted, J. T., Mickes, L., Dunn, J. C., Clark, S. E. & Wells, W. Estimating the reliability of eyewitness identifications from police lineups. Proc. Natl. Acad. Sci. U.S.A. 113, 304–309 (2016).
Watts, D. J., Rothschild, D. M. & Mobius, M. Measuring the news and its impact on democracy. Proc. Natl. Acad. Sci. U.S.A. 118, 6 (2021).
Tseng, A. S. Students and evaluation of web-based misinformation about vaccination: Critical reading or passive acceptance of claims?. Int. J. Sci. Educ. Part B-Commun. Public Engag. 8, 250–265 (2018).
Goulden, N. R. The roles of national and state standards in implementing speaking, listening, and media literacy. Commun. Educ. 47, 194–208 (1998).
Hobbs, R. A review of school-based initiatives in media literacy education. Am. Behav. Sci. 48, 42–59 (2004).
Mirra, N. & Garcia, A. In search of the meaning and purpose of 21st-century literacy learning: A critical review of research and practice. Read. Res. Q. 56, 463–496 (2021).
Sharon, A. J. & Baram-Tsabari, A. Can science literacy help individuals identify misinformation in everyday life?. Sci. Educ. 104, 873–894 (2020).
Vaccari, C. & Chadwick, A. Deepfakes and disinformation: Exploring the impact of synthetic political video on deception, uncertainty, and trust in news. Soc. Media Soc. 6, 13 (2020).
Dobber, T., Metoui, N., Trilling, D., Helberger, N. & de Vreese, C. Do (microtargeted) deepfakes have real effects on political attitudes?. Int. J. Press-Polit. 26, 69–91 (2021).
Gringberg, N., Joseph, K., Friedland, L., Swire-Thompson, B. & Lazer, D. Fake news on Twitter during teh 2016 U.S. presidential election. Science 363(6425), 374–378 (2019).
Hornsey, M. J. et al. Evidence for motivated control: Understanding the paradoxical link between threat and efficacy beliefs about climate change. J. Environ. Psychol. 42, 57–65 (2015).
Treen, K. M. D., Williams, H. T. P. & O’Neill, S. J. Online misinformation about climate change. Wiley Interdiscipl. Rev. -Climate Change 11, 20 (2020).
Cooper, C. B. Media literacy as a key strategy toward improving public acceptance of climate change science. Bioscience 61, 231–237 (2011).
Zhao, B., Zhang, S. Z., Xu, C. X., Sun, Y. F. & Deng, C. B. Deep fake geography? When geospatial data encounter Artificial Intelligence. Cartogr. Geogr. Inf. Sci. 48, 338–352 (2021).
Shen, C. H. et al. Fake images: The effects of source, intermediary, and digital media literacy on contextual assessment of image credibility online. New Media Soc. 21, 438–463 (2019).
Azevedo, F. & Jost, J. T. The ideological basis of antiscientific attitudes: Effects of authoritarianism, conservatism, religiosity, social dominance, and system justification. Group Process. Intergroup Relat. 24, 518–549 (2021).
Pennycook, G. & Rand, D. G. Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition 188, 39–50 (2019).
Langa, J. Deepfakes, real consequences: Crafting legislation to combat threats posed by deepfakes. Boston Univ. Law Rev. 101, 761–801 (2021).
Pan, W. J., Liu, D. Y. & Fang, J. An examination of factors contributing to the acceptance of onlinehealth misinformation. Front. Psychol. 12, 11 (2021).
Prull, M. W. & Yockelson, M. B. Adult age-related differences in the misinformation effect for context-consistent and context-inconsistent objects. Appl. Cogn. Psychol. 27, 384–395 (2013).
Saunders, J. & Jess, A. The effects of age on remembering and knowing misinformation. Memory 18, 1–11 (2010).
Brashier, N. M. & Schatter, D. L. Aging in an era of fake news. Curr. Dir. Psychol. Sci. 29(3), 316–323 (2020).
Lewandowsky, S. Climate change disinformation and how to combat it. In: Fielding, J. E. (Ed) Annual Review of Public Health, vol 42, pp 1–21. (Annual Reviews, Palo Alto, 2021).
O’Brien, T. C., Palmer, R. & Albarracin, D. Misplaced trust: When trust in science fosters belief in pseudoscience and the benefits of critical evaluation. J. Exp. Soc. Psychol. 96, 13 (2021).
Xiao, X. Z., Borah, P. & Su, Y. The dangers of blind trust: Examining the interplay among social media news use, misinformation identification, and news trust on conspiracy beliefs. Public Underst. Sci. 30, 977–992 (2021).
Wang, T. C. et al. Video-to-video synthesis. Adv. Neural Inf. Process. Syst. 31(31), 13 (2018).
Pollard, M. & Baird, M. D. The RAND American Life Panel: Technical Description (RAND Corporation, 2017).
Robbins, M. W. & Grant, D. M. RAND American Educator Panels Technical Description (RAND Corporation, 2020).