Conteúdo do artigo principal

Autores

O artigo aborda o estado da arte e análise do paradigma da inteligência artificial (ia) aplicado a usos bélicos. Reflete sobre como os conflitos contemporâneos são travados em grande medida fora dos campos de batalha, mesmo em campos não físicos, como o cibernético e o cognitivo. Nestes dois ambientes, ocorre um uso não menos intenso de ia com o mesmo objetivo de alterar a vontade dos adversários, principalmente por meio de ferramentas de influência e desinformação. A metodologia utilizada baseia-se na revisão bibliográfica e análise de documentos elaborados por especialistas e relatórios de instituições dedicadas ao tratamento de cibersegurança, defesa e ações de guerra em países desenvolvidos. É proposta uma taxonomia das aplicações bélicas e maliciosas da ia que desconsidera suas características tecnológicas para focar em sua aplicação operacional. O resultado não é um conjunto fechado de opções, mas ramificações em inúmeras alternativas que podem ser desenvolvidas em outros trabalhos. A principal conclusão a que se chega é que os países devem levar em consideração o potencial da inteligência artificial para desenvolver ações táticas e estratégicas, preparando-se para cenários não convencionais.

Jesús Miguel Flores Vivar, Universidad Complutense de Madrid

PhD. by the Complutense University of Madrid. Erasmus professor STA at Università di Roma "Sapienza"; Università degli Studi Firenze; Università di Perugia and Università degli Studi di Siena (Italy), professor at UCM (Spain) and research professor at UNMSM (Peru), UNESP (Brazil) and UPB (Colombia), among others. He also taught at the Open University of Catalonia and Antonio de Nebrija University (Spain). He has been Invited Professor at RCC Harvard University (VIII Seminar on Sociological and Political Research); Co-founder and Coordinator (2005-08) of the Internet Journalism Observatory and promoter and Co-founder (2016) of the ITC Network (Interdisciplinary Research Network in Technology and Communication, Cyberculture and New Audiovisual Genres). He is a member-expert of the Spanish Association of Communication Research (AE-IC).

Ángel Gómez-de-Ágreda, Universidad Politécnica de Madrid

Analista geopolítico. Investigador. PhD Ingeniería.

Jacinto Gómez-López, Universidad Complutense de Madrid

Doctor en periodismo. Investigador en Ciberseguridad e Inteligencia Artificial. 

Flores Vivar, J. M., Gómez-de-Ágreda, Ángel ., & Gómez-López, J. (2023). Taxonomia da inteligência artificial no ambiente cognitivo de conflitos. Anuario Electrónico De Estudios En Comunicación Social "Disertaciones", 16(2). https://doi.org/10.12804/revistas.urosario.edu.co/disertaciones/a.12804

Aguado, J. M. (2020). Mediaciones ubicuas. Gedisa.

Aguado, J. M., Feijóo, C., & Martínez Martínez, I. J. (2013). La comunicación móvil hacia un nuevo ecosistema digital. Gedisa.

Allen, G., & Chan, T. (2017). Artificial Intelligence and National Security: A study on behalf of Dr. Jason Matheny, director of the U.S. Intelligence Advanced Research Projects Activity (iarpa). Belfer Center for Science and International Affairs-Harvard Kennedy School.

Allcott, H., & Gentzkow, M. (2017). Social media and fake news in the 2016 election. Journal of Economic Perspectives, 31(2), 211-236. https://doi.org/10.1257/jep.31.2.211

Aslama-Horowitz, M. (2019). Disinformation as warfare in the digital age: dimensions, dilemmas, and solutions dimensions, dilemmas, and solutions. Journal of Vincentian Social Action, 4(2). https://scholar.stjohns.edu/jovsa/vol4/iss2/5

Ayerbe, A. (2020). La ciberseguridad y su relación con la inteligencia artificial. ari (128). https://media.realinstitutoelcano.org/wp-content/uploads/2021/10/ari128-2020-ayerbe-ciberseguridad-y-su-relacion-con-inteligencia-artificial.pdf]

Baker-Brunnbauer, J. (2023). Introduction. En Trustworthy artificial intelligence implementation: Business guides on the go. Springer. https://doi.org/10.1007/978-3-031-18275-4_1

Barnes, C., & Barraclough, T. (2019). Perception Iinception: Preparing for deepfakes and the synthetic media of tomorrow. The Law Foundation.

Bonfanti, M. E. (2020). The weaponisation of synthetic media: What threat does this pose to national security? ari, 93. http://www.realinstitutoelcano.org/wps/portal/rielcano_en/contenido?WCM_GLOBAL_CONTEXT=/elcano/elcano_in/zonas_in/ari93-2020-bonfanti-weaponisation-of-synthetic-media-whatthreat-does-this-pose-to-national-security

Boulanin, V. V. V. (2018). The impact of artificial intelligence on strategic stability and nuclear risk. Sipri, I(May). https://doi.org/10.2991/icsshe-18.2018.203

Boyer, P. (2018). Minds make societies: How cognition explains the world humans create. Yale University Press.

Brundage, M., Avin, S., Clark, J., Toner, H., Eckersley, P., Garfinkel, B., Dafoe, A., Scharre, P., Zeitzoff, T., Filar, B., Anderson, H., Roff, H., Allen, G. C., Steinhardt, J., Flynn, C., HÉigeartaigh, S. Ó., Beard, S., Belfield, H., Farquhar, S., … Amodei, D. (2018, febrero). The malicious use of artificial intelligence: forecasting, prevention, and mitigation. https://arxiv.org/pdf/1802.07228

Bughin, J., Hazan, E., Ramaswamy, S., Chui, M., Allas, T., Dahlström, P., Henke, N., & Trench, M. (2017). Artificial intelligence: The next digital frontier? [discussion paper]. McKinsey Global Institute (mgi).

Calvo-Albero, J. L. (2020). Implicaciones en el ámbito cognitivo en las operaciones militares. Centro Superior de Estudios de la Defensa Nacional. https://emad.defensa.gob.es/Galerias/CCDC/files/IMPLICACIONES_DEL_AMBITO_COGNITIVO_EN_LAS_OPERACIONES_MILITARES.pdf

ccw. (2018). Chair’s summary of the discussion. 6(April), 1-13.

Chesney, R., & Citron, D. (2018, 11 de diciembre). Deepfakes and the new disinformation war. Foreign Affairs. https://www.foreignaffairs.com/articles/world/2018-12-11/deepfakes-and-new-disinformation-war

Debord, G. (2005). La sociedad del espectáculo. Pre-Textos.

Deptula, D. A. (2022). A new battle command architecture for Joint all-domain operations. æther: A Journal of Strategic Airpower & Spacepower, 1(1), 51-56. https://www.airuniversity.af.edu/Portals/10/AEtherJournal/Journals/Volume-1_Issue-1/08-Deptula.pdf?source=GovD

Flores-Vivar, J., & García-Peñalvo, F. (2023). Reflections on the ethics, potential, and challenges of artificial intelligence in the framework of quality education (SDG4). Comunicar, 74, 37-47. https://doi.org/10.3916/C74-2023-03

Future of Life Institute. (2020). Asilomar AI principles. Future of Life Institute, 1-49. https://futureoflife.org/ai-principles/?submitted=1#confirmation

Gartner. (2018, 24 de julio). Hype CYCLe for artificial intelligence, 2018. https://www.gartner.com/en/documents/3883863

Gómez-de-Ágreda, Á. (2019). Mundo Orwell: Manual de supervivencia para un mundo hiperconectado. Ariel.

Gómez-de-Ágreda, Á. (2020). Ethics of autonomous weapons systems and its applicability to any AI systems. Telecommunications Policy, 44(6), 1-15. https://doi.org/10.1016/j.telpol.2020.101953

Gómez-de-Ágreda, Á., & Feijoo, C. (2021). Hacia una ética del ecosistema híbrido del espacio físico y el ciberespacio. Diecisiete (4), 47-58. https://doi.org/10.36852/2695-4427_2021_04.02

Gómez-de-Ágreda, Á., Feijóo, C., & Salazar, I. (2021). Una nueva taxonomía del uso de la imagen en la conformación interesada del relato digital: Deep fakes e inteligencia artificial. El Profesional de la Información, 30(2). https://doi.org/10.3145/epi.2021.mar.16

Gómez-de-Ágreda, Á., Martínez, J. M., Mohíno Herranz, I., Barragán Montes, R., Marín Gutierrez, F. A., Cubeiro Cabello, E., & Aznar Lahoz, J. L. (2019). Usos militares de la inteligencia artificial, la automatización y la robótica (iaa&r). Centro Superior de Estudios de la Defensa Nacional. http://www.ieee.es/contenido/noticias/2019/11/DIEEET04-2019InteligenciaRobotica.html

Gómez-López, J. (2019). Tecnologías de la información y los mensajes en los nuevos espectros del conflicto. Revista de Ciencias de la Comunicación e Información, (24), 45-56.

Han, B. C. (2021). No-cosas: Quiebras del mundo de hoy. Taurus.

Helmus, T. C. (2022, julio). Artificial intelligence, deepfakes, and disinformation: A primer. RAND Corporation. https://doi.org/10.7249/pea1043-1

International Committee of the Red Cross. (2018, 3 de abril.). Ethics and autonomous weapon systems: An ethical basis for human control? https://www.icrc.org/en/document/ethics-and-autonomous-weaponsystems-ethical-basis-human-control

Johnson, J. (2020). Artificial intelligence, drone swarming and escalation risks in future warfare. rusi Journal, 165(2), 26-36. https://doi.org/10.1080/03071847.2020.1752026

Lassalle, J. M. (2022, 22 de octubre). Capitalismo digital de sustitución. La Vanguardia. https://www.lavanguardia.com/opinion/20221022/8576754/capitalismo-digital-sustitucion.html

Leonardo-Oviedo, G. (2004). La definición del concepto de percepción en psicología con base en la teoría Gestalt. Revista de Estudios Sociales, 18, 89-96.

Marina, J. A. (2022). El deseo interminable. Ariel.

Mir, R., & Rodríguez, K. (2020). If privacy dies in vr, it dies in real life. https://www.eff.org/deeplinks/2020/08/if-privacy-dies-vr-it-dies-real-life

Parasuraman, R., Sheridan, T. B., & Wickens, C. D. (2000). A model for types and levels of human interaction with automation. ieee Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans, 30(3), 286-297. https://doi.org/10.1109/3468.844354

Paris, B., & Donovan, J. (2019). Deepfakes and cheap fakes. Data & Society, 47. https://site.ieee.org/sagroups-7011/

Postman, N. (2005). Amusing ourselves to death. Penguin Books.

Pulido, G. (2021). Guerra multidominio y mosaico. Catarata. https://www.catarata.org/libro/guerramultidominio-y-mosaico_133474/

Scharre, B. P., & Fish, L. (2018, 7 de noviembre). Human performance enhancement. Center for a New American Security. https://www.cnas.org/publications/reports/human-performance-enhancement-1

Scharre, P. (2019). Killer apps: The real dangers of an ai arms race. Foreign Affairs, 98(3), 135-144. https://www.foreignaffairs.com/articles/2019-04-16/killer-apps

Schick, N. (2020, 28 de diciembre). Los “deepfake” no han roto la democracia en 2020; los “cheapfake”, sí. MIT Technology Review. https://www.technologyreview.es/s/13044/los-deepfake-no-han-rotola-democracia-en-2020-los-cheapfake-si

Shahriari, K., & Shahriari, M. (2017). ieee standard review. Ethically aligned design: A vision for prioritizing human wellbeing with artificial intelligence and autonomous systems. 2017 IEEE Canada International Humanitarian Technology Conference (ihtc), Toronto, on, Canada, 197-201. https://doi.org/10.1109/IHTC.2017.8058187

Simon, H. A. (1971). Designing organizations for an information-rich world. En M. Greenberger (Ed.), Computers, communications, and the public interest. John Hopkins University Press.

Stop Killer Robots. (2018). Campaign to stop killer robots. https://www.stopkillerrobots.org/

Templeton, E. M., Chang, L. J., Reynolds, E. A., LeBeaumont, M. D. C., & Wheatley, T. (2022). Fast response times signal social connection in conversation. Proceedings of the National Academy of Sciences of the United States of America, 119(4). https://doi.org/10.1073/pnas.2116915119

Torres Soriano, M. R. (2022, 28 de junio). Operaciones de influencia vs. desinformación: Diferencias y puntos de conexión. Instituto Español de Asuntos Estratégicos, (64). https://www.ieee.es/Galerias/fichero/docs_opinion/2022/DIEEEO64_2022_MANTOR_Operaciones.pdf

Tschopp, M. (2019). Good AI, bad AI - psychological aspects of a dual-use technology. https://www.academia.edu/40718200/Good_AI_Bad_AI_Psychological_Aspects_of_a_Dual_Use_Technology

Veliz, C. (2020). Privacy is power: Why and how you should take back control of your data. Transworld.

Von Clausewitz, K. (2002). De la guerra. https://www.psicosocial.net/historico/index.php?option=com_docman&view=download&alias=870-de-la-guerra&category_slug=psicologia-y-violencia-politica&Itemid=100225

Watts, C. (2019). Advanced persistent manipulators, part one: The threat to the social media industry - alliance for securing democracy. https://securingdemocracy.gmfus.org/advanced-persistent-manipulators-part-one-the-threat-to-the-social-media-industry/

Woolley, S., & Joseff, K. (2020). Demand for deceit: How the way we think drives disinformation. https://www.ned.org/wp-content/uploads/2020/01/Demand-for-Deceit.pdf

Zuboff, S. (2020). La era del capitalismo de la vigilancia: La lucha por un futuro humano frente a las nuevas fronteras del poder. Paidós.

Downloads

Não há dados estatísticos.