Why We Can’t Ignore the Dark Side of AI

As AI continues to evolve, it introduces greater security and data risks. Here’s what we need to do about it.  

Written by Ken Cox
Published on Jun. 28, 2023
Image: Shutterstock / Built In
Image: Shutterstock / Built In
Brand Studio Logo

Artificial intelligence is an exciting frontier, promising to transform nearly every aspect of our lives from how we work and play to how we learn and connect. Yet, with these promises comes an increasing level of risk to new forms of cybercrime, data breaches and privacy violations. As we step into this brave new world, we must address these challenges and ensure the benefits of AI do not come at the cost of our cybersecurity and personal privacy.

4 Security Risks AI Poses

  1. Data collection and profiling.
  2. Surveillance and tracking.
  3. Data breaches and security risks.
  4. Inference and re-identification attacks.

With two decades of witnessing the rollercoaster evolution of AI, I’ve seen first-hand the seismic shifts this technology has brought to our societies. It’s a narrative of potential and peril, promise and privacy concerns, innovation and intrusion.

 

4 AI Data Security Risks

 

1. Data Collection and Profiling

One of AI’s biggest strengths lies in its ability to collect and analyze vast amounts of data. This has led to the creation of extensive data profiles, allowing AI systems to predict behavior, provide personalized services and offer unique user experiences tailored to individual needs and preferences.

However, this deep data dive into personal information sparks significant privacy concerns. AI’s ability to collect, analyze, and predict based on personal data opens the door for unauthorized access, misuse and potential exploitation of personal information. It’s a double-edged sword, where the same tool used for personalization can become an instrument of personal intrusion.

The online exploitation of personal information has always been a concern, but the combination of AI data breaches and the rapid disclosure of personal data by the public has intensified the fear. Take online grocery shopping for example. Credit card details, purchase history, location and even the people sharing the same space, are meticulously tracked. With these data points, malicious actors can generate hyper-realistic emails, videos, letters and voice recordings, possessing intimate knowledge of your identity, activities, social circle and dietary habits.

In 2022, Americans fell victim to internet scams, resulting in a staggering loss of $10.3 billion. These scams utilized brute force techniques and simple behavioral programming. Notably, both major grocery stores and financial institutions were targeted, leading to the exposure of valuable data on the dark web.

More on AIVoice Cloning: What It Is and Why It’s Scary

 

2. Surveillance and Tracking

The rise of AI-powered surveillance technologies, such as facial recognition and video analytics, is another area of growing concern. These tools offer great potential in areas like security, law enforcement and even retail, enabling real-time tracking and identification capabilities. 

However, they also pose a significant threat to privacy, enabling unprecedented levels of surveillance in both public spaces and the digital world. Balancing these benefits and concerns is no easy task, but it’s an endeavor we must undertake to ensure our public spaces don’t turn into surveillance hotspots.

Large groups and echo chambers are also a concern. Regrettably, algorithms and market incentives prioritize watch time above all else, leading to the creation of echo chambers. These echo chambers don’t necessarily offer the most accurate information; instead, they present content that aligns with what you have previously spent the most time viewing or reacted strongly to. 

To illustrate this phenomenon, let’s consider the flat Earth movement, which initially emerged as a joke. However, due to the abundance of content generated, I personally know individuals who have become convinced that the Earth is flat. Although discussing the shape of the Earth may seem trivial, applying these tactics to matters of faith, politics or deeply ingrained beliefs could have severe consequences, potentially even sparking conflicts and wars.

 

3. Data Breaches and Security Risks

Just like any other technology, AI is not immune to cyberattacks and data breaches. In fact, as AI systems become more integral to our infrastructures, they could become prime targets for hackers. 

Compromised personal data can lead to disastrous consequences, including identity theft, fraud or other violations of privacy. It’s a looming shadow that has been cast over the bright potential of AI, a risk that threatens to shatter the trust necessary for AI’s widespread adoption.

 

4. Inference and Re-identification Attacks

The ability of AI to make connections from disparate data points can lead to the unearthing of sensitive information, even when the initial data seems unrelated or anonymized. This capability opens up a new avenue for potential privacy violations, as AI algorithms can infer sensitive personal information from seemingly harmless data. It’s a sobering reminder of the sophisticated capabilities of AI and the need for stringent data protection measures.

 

The Future of AI Governance

To address these myriad concerns, I propose that we establish robust regulations, ethical guidelines and transparency practices governing AI development and deployment. However, this is not a call for widespread censorship or stifling innovation. Freedom of speech and expression must remain at the forefront of our considerations, even as we navigate these tricky waters.

Much like the system used for grading explicit content in music or video games, perhaps AI applications should carry “content warnings.” Such warnings would serve to inform users about potential risks or ethical concerns surrounding certain AI applications, providing a layer of transparency and allowing individuals to make informed choices.

Licensing, while a logical solution for regulating AI, poses its own set of challenges. While it may be beneficial for public use or corporate data handling, licensing AI for personal use may inadvertently lead to profiling and discrimination, two very issues we aim to prevent.

Imposing licensing on AI at present would be deemed detrimental to humanity. Knowledge should be universally accessible and recognition should be granted for its application rather than hoarding knowledge from others. The practical implementation of such licensing seems perplexing, especially considering the existing availability of AI tools through open-source initiatives like Apache Spark, Ray, and OpenAI. These platforms have made AI accessible to anyone willing to invest the effort to learn. OpenAI, in particular, offers these tools either for free or at a minimal cost for more serious endeavors. It seems that the proverbial cat is already out of the bag.

However, if we wish to delve into the discussion of licensing AI, it becomes crucial to address broader societal issues. As long as societies are mired in authoritarian rule, incarcerating individuals and perpetuating the belief that one human possesses power over another, the outcome of licensing AI can only lead to a dystopian future. 

Instead, it is imperative that we strive for a society based on guardianship principles, where individuals can live freely with secure passage, provided they do not pose an immediate threat to themselves or others. In such a society, restricting AI access to specific groups would be profoundly irresponsible and indicative of an authoritative regime.

 

How to Protect Against AI Security Risks

Generative AI models bring innovation to fields such as content creation but also pose significant cybersecurity risks. These include AI-powered phishing, which is sophisticated and hard to detect, deepfakes that can lead to fraud or blackmail, automated hacking and data poisoning which manipulates a model’s output.

Companies should establish AI-specific security measures, including securing training data, consistently auditing and updating models and enforcing robust access controls for AI systems to protect against these threats. Employee education is crucial. Teaching them to recognize AI-powered threats, maintaining good cybersecurity practices and understanding the potential harm of deepfakes. Advanced monitoring and detection tools can assist in the rapid identification of potential threats, too.

In the event of a breach, having a robust incident response plan is paramount. Compliance with all relevant data protection and cybersecurity regulations reduces the threat and potential legal liability in the event of a breach. Finally, routine third-party audits can help identify potential vulnerabilities and ensure that all security measures are effective and up-to-date. It’s a multi-faceted and complex approach, but it’s crucial for navigating the complex landscape of AI and cybersecurity.

More on AIWhy an Automated Future Won’t Be a Jobless One

 

Why Responsible AI Is Critical

AI offers immense potential for democratizing access to information and skills. Imagine a world where anyone, regardless of their background, can become a programmer or understand complex language nuances, all thanks to AI. 

The call for regulation isn’t about stifling AI innovation; it’s about shaping it responsibly. We need leadership and education to ensure the right balance, with regulations that are as minimal as necessary, but as robust as required.

We must strive for a world where the benefits of AI are freely available to all, without fear of exploitation or loss of privacy. Balancing innovation and security, progress and privacy — this will be our greatest challenge and achievement.

Explore Job Matches.