مِنَ الحَدَائِقِ السِّرِّيَّةِ إِلَى كَوَابِيسَ المُرَاقَبَةِ: كَيْفَ تُحَوِّلُ رِوَايَةُ "1984" أَدَبَ الهُرُوبِ إِلَى أَدَبِ التَرْهِيْب
DOI:
https://doi.org/10.53286/arts.v7i4.2855الكلمات المفتاحية:
ديستوبيا أورويلية، رأسمالية المراقبة، الحوكمة الخوارزمية، اللغة الجديدة، السيطرة الأرشيفية، الثقافة الرقمية النقديةالملخص
يعيد هذا البحث قراءة رواية جورج أورويل 1984 بوصفها إطارًا لفهم السلطة الخوارزمية المعاصرة، لا مجرد مجازٍ ثابت من حقبة الحرب الباردة. يستند المقال إلى نظرية الخطاب لدى فوكو، ونموذج المجال العام لهابرماس، ومفهوم رأسمالية المراقبة لشوشانا زوبوف، ويقدّم تحليلًا نوعيًّا لمقاطع رئيسة في النص، مثل يوميات ونستون، ودقيقتي الكراهية، وغرفة 101، إلى جانب أرشيفات دعائية من زمن الحرب وبيانات أوروبية حديثة حول التضليل الإعلامي. يتتبّع البحثُ كيف تتطور آليات السيطرة من تقييد اللغة (اللغة الجديدة) و(محو التاريخ) (ثقوب الذاكرة) إلى تشكيل الانفعالات والمراقبة العصبية الناشئة. وتُظهر النتائج أنّ الرقابة الحديثة تعمل من خلال منصّات تنبؤية تسبق المعارضة بتضييق المفردات، ومحو السجلات، وتوجيه العواطف. وبذلك تُقرأ 1984 لا بوصفها تحذيرًا فحسب، بل بوصفها عدسة تحليلية تكشف كيف تتكيّف السلطة في العصر الرقمي. ويختتم البحث باقتراحات تشمل الثقافة الرقمية النقدية، وشفافية الخوارزميات، والإدارة الجماعية للبيانات، بهدف حماية "الحديقة السرّية" للخيال وصون الحرية الديمقراطية.
التنزيلات
المراجع
amleh. (2024). Systematic suppression: Palestinian digital rights violations during the Gaza conflict. Business and Human Rights Resource Centre. https://www.business-humanrights.org
Andrejevic, M. (2019). Automated media: The coming of the algorithmic subject. Routledge.
Benkler, Y. (2006). The wealth of networks: How social production transforms markets and freedom. Yale University Press.
Boyd, d., & Crawford, K. (2012). “Critical questions for big data: Provocations for a cultural, technological, and scholarly phenomenon.” Information, Communication & Society, 15(5), 662–679. https://doi.org/10.1080/1369118X.2012.678878
Bradshaw, S., Bailey, H., & Howard, P. N. (2021). Industrialized disinformation: 2020 global inventory of organized social media manipulation. Programme on Democracy & Technology, Oxford Internet Institute. https://demtech.oii.ox.ac.uk/research/posts/industrialized-disinformation/
Brunner, C. (2020). “Semantic warfare: Newspeak in digital authoritarianism.” Journal of Media Ethics, 35(4), 215–228. https://doi.org/10.1080/23736992.2020.1821732
BU Today. (2025, June 11). “YouTube is the latest media platform to loosen content moderation: What does that mean for users?” BU Today. https://www.bu.edu/articles/2025/youtube-content-moderation-changes/
Camacho-Collados, J., Espinosa Anke, L., Barbieri, F., & Neves, L. (2022). “A massive multilingual dataset for predicting emotions in text.” In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (pp. 5874–5889). Association for Computational Linguistics. https://doi.org/10.18653/v1/2022.acl-long.404
Cheney-Lippold, J. (2017). We are data: Algorithms and the making of our digital selves. NYU Press.
Chesney, R., & Citron, D. (2019). “Deepfakes and the new disinformation war.” Foreign Affairs, 98(1), 147–155. https://www.foreignaffairs.com/world/deepfakes-and-new-disinformation-war
Couldry, N., & Mejias, U. A. (2019). The costs of connection: How data is colonizing human life and appropriating it for capitalism. Stanford University Press.
Crick, B. (1992). George Orwell: A life. Penguin.
Daldry, J. (1992). “Panopticism and the telescreen.” Modern Fiction Studies, 38(3), 437–457.
Derrida, J. (1995). Archive fever: A Freudian impression (E. Prenowitz, Trans.). University of Chicago Press. (Original work published 1994)
Eagleton, T. (1983). Literary theory: An introduction. University of Minnesota Press.
Eagleton, T. (1985). “Capitalism, modernism and postmodernism.” New Left Review, 152, 60–73.
Eagleton, T. (1991). Ideology: An introduction (2nd ed.). Verso.
Eubanks, V. (2018). Automating inequality: How high-tech tools profile, police, and punish the poor. St. Martin’s Press.
European Commission. (2024). “Implementing the EU Code of Practice on Disinformation: Evaluation of platform compliance and accountability [Report].” European Commission / European Digital Media Observatory.
Farahany, N. (2023). The battle for your brain: Defending the right to think freely in the age of neurotechnology. St. Martin’s Press.
Foucault, M. (1977). Discipline and punish: The birth of the prison. Vintage Books.
Foucault, M. (1980). Power/knowledge: Selected interviews and other writings, 1972–1977 (C. Gordon, Ed.). Pantheon Books.
Freire, P. (1970). Pedagogy of the oppressed. Herder and Herder.
Galloway, A. R. (2022). Unlivable: Politics, aesthetics, and digital dystopia. Verso.
Gillespie, T. (2018). Custodians of the internet: Platforms, content moderation, and the hidden decisions that shape social media. Yale University Press.
Gorwa, R. (2019). “What is platform governance?” Information, Communication & Society, 22(6), 854–871. https://doi.org/10.1080/1369118X.2019.1573914
Gorwa, R. (2024). The politics of platform regulation: How governments shape online content moderation. Oxford University Press. https://doi.org/10.1093/oso/9780197692851.001.0001
Gregg, M., & Seigworth, G. J. (Eds.). (2010). The affect theory reader. Duke University Press.
Habermas, J. (1989). The structural transformation of the public sphere: An inquiry into a category of bourgeois society (T. Burger, Trans.). Polity Press.
Hipfl, B. (2018). “Affect in media and communication studies: Potentials and assemblages.” Media and Communication, 6(3), 4–12. https://doi.org/10.17645/mac.v6i3.1470
Hobbs, R. (2020). “Propaganda in an age of algorithmic personalization: Expanding literacy research and practice.” Reading Research Quarterly, 55(3), 521–533. https://doi.org/10.1002/rrq.301
Human Rights Watch. (2023). Meta’s broken promises: Systemic censorship of Palestinian content on Instagram and Facebook. Human Rights Watch. https://www.hrw.org
Ienca, M., & Andorno, R. (2017). “Toward new human rights in the age of neuroscience and neurotechnology.” Life Sciences, Society and Policy, 13(5), 1–27. https://doi.org/10.1186/s40504-017-0050-1
Kavanagh, J., & Rich, M. D. (2018). Truth decay: An initial exploration of the diminishing role of facts and analysis in American public life. RAND Corporation.
Kleinberg, J., Mullainathan, S., & Obermeyer, Z. (2024). “Manipulating beliefs with sentiment adaptive chatbots: Experimental evidence.” Journal of Artificial Intelligence and Society, 9(1), 55–74.
Kunda, Z. (1990). “The case for motivated reasoning.” Psychological Bulletin, 108(3), 480–498. https://doi.org/10.1037/0033-2909.108.3.480
Lakoff, G. (2014). The all new don’t think of an elephant!: Know your values and frame the debate. Chelsea Green Publishing.
Lee, S.-B. (2023). The power of arch-violence inherent in archives: Focusing on Derrida’s “Archive Fever.” The Journal of Aesthetics and Science of Art, 70, 34–51. https://doi.org/10.17527/JASA.70.0.02
Lewis, C. S. (1966). Of other worlds: Essays and stories (W. Hooper, Ed.). Harcourt, Brace & World.
Lyon, D. (2018). “Exploring surveillance culture.” On_Culture: The Open Journal for the Study of Culture, 6, 1–14.
Maier, S. F., & Seligman, M. E. P. (2016). “Learned helplessness at fifty: Insights from neuroscience.” Psychological Review, 123(4), 349–367. https://doi.org/10.1037/rev0000033
Musk, E., & Neuralink. (2023, November 30). Neuralink progress update [Video]. YouTube. https://www.youtube.com/watch?v=rsCul1sp4hQ
Noble, S. U. (2018). Algorithms of oppression: How search engines reinforce racism. NYU Press.
Orwell, G. (1949). Nineteen eighty-four. Secker & Warburg.
Orwell, G. (2008). Nineteen eighty-four. Penguin Books. (Original work published 1949)
Paris, B., & Donovan, J. (2019, September 18). “Deepfakes and cheap fakes: The manipulation of audio and visual evidence.” Data & Society Research Institute. https://datasociety.net/library/deepfakes-and-cheap-fakes/
Podhoretz, N. (1983, January). “If Orwell were alive today.” Harper’s Magazine.
Poria, S., Cambria, E., Hazarika, D., & Vij, P. (2020). “A deeper look into context in text-based emotion detection.” Knowledge-Based Systems, 187, 104830. https://doi.org/10.1016/j.knosys.2019.06.024
Shelden, M. (1991). Orwell: The authorized biography. HarperCollins.
Srnicek, N. (2017). Platform capitalism. Polity.
The Palestine Chronicle. (2024, June 12). “Over 25,000 digital violations against Palestinian content recorded.” The Palestine Chronicle. https://www.palestinechronicle.com
Tufekci, Z. (2015). “Algorithmic harms beyond Facebook and Google: Emergent challenges of computational agency.” Colorado Technology Law Journal, 13(1), 203–218.
UK Parliament. (2023). Online Safety Act 2023: Chapter 50. The Stationery Office. https://www.legislation.gov.uk/ukpga/2023/50/enacted
Vaccari, C., & Chadwick, A. (2020). “Deepfakes and disinformation: Exploring the impact of synthetic political video on deception, uncertainty, and trust in news.” Social Media + Society, 6(1), 1–13. https://doi.org/10.1177/2056305120903408
Van Dijck, J. (2021). “Governing digital societies: Private platforms and public values.” Digital Society, 1(2), 120–136. https://doi.org/10.1007/s44206-021-00012-x
Willcox, M. (2024). “Algorithmic agency and ‘fighting back’ against discriminatory Instagram content moderation: #IWantToSeeNyome”. Frontiers in Communication, 9, 1385869. https://doi.org/10.3389/fcomm.2024.1385869
Williams, R. (1974). Television: Technology and cultural form. Fontana.
Youvan, D. C. (2024, March 29). The semiotics of Newspeak: Language, power, and thought control in Orwell’s 1984 (preprint). https://doi.org/10.13140/RG.2.2.32611.57128
Zuboff, S. (2019). The age of surveillance capitalism: The fight for a human future at the new frontier of power. PublicAffairs.
التنزيلات
منشور
إصدار
القسم
الرخصة

هذا العمل مرخص بموجب Creative Commons Attribution 4.0 International License.
:حقوق الطبع والنشر والترخيص
يحتفظ الباحثون بحقوق النشر. ويتم ترخيص البحوث بموجب ترخيص Creative Commons CC BY 4.0 المفتوح، مما يعني أنه يجوز لأي شخص تنزيل البحث وقراءته مجانًا. وإعادة استخدام البحث واقتباسه شريطة أن يتم الإشارة إلى المصدر الأصلي. تتيح هذه الشروط الاستخدام الأقصى لعمل الباحث وعرضه.
























