0% found this document useful (0 votes)
68 views3 pages

The 7 Most Dangerous Technology Trends in 2020 Everyone Should Know About

The document discusses 7 of the most dangerous technology trends for 2020, including drone swarms that could be deployed autonomously by militaries, smart home devices that collect personal information and habits, and facial recognition software that could enable mass surveillance or racial profiling. It also outlines risks from deepfake technology that can manipulate videos and voices, ransomware and hacking enabled by artificial intelligence, smart dust sensors that could enable invasive spying, and AI-generated fake news articles. While technology provides benefits, it also poses risks if misused or its capabilities are abused.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
68 views3 pages

The 7 Most Dangerous Technology Trends in 2020 Everyone Should Know About

The document discusses 7 of the most dangerous technology trends for 2020, including drone swarms that could be deployed autonomously by militaries, smart home devices that collect personal information and habits, and facial recognition software that could enable mass surveillance or racial profiling. It also outlines risks from deepfake technology that can manipulate videos and voices, ransomware and hacking enabled by artificial intelligence, smart dust sensors that could enable invasive spying, and AI-generated fake news articles. While technology provides benefits, it also poses risks if misused or its capabilities are abused.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd

The 7 Most Dangerous Technology Trends In 2020 Everyone Should Know About

As we enter new frontiers with the latest technology trends and enjoy the many positive impacts
and benefits it can have on the way we work, play and live, we must always be mindful and
prepare for possible negative impacts and potential misuse of the technology. Here are seven of
the most dangerous technology trends:

1. Drone Swarms
The British, Chinese, and United States armed forces are testing how interconnected, cooperative
drones could be used in military operations. Inspired by a swarm of insects working together,
drone swarms could revolutionize future conflicts, whether it be by overwhelming enemy
sensors with their numbers or to effectively cover a large area for search-and-rescue missions.
The difference between swarms and how drones are used by the military today is that the swarm
could organize itself based on the situation and through interactions with each other to
accomplish a goal. While this technology is still in the experimentation stage, the reality of a
swarm that is smart enough to coordinate its own behavior is moving closer to reality. Aside
from the positive benefits of drone swarms to minimize casualties, at least for the offense, and
more efficiently achieve a search-and-rescue objective, the thought of machines equipped with
weapons to kill being able to "think" for themselves is fodder for nightmares. Despite the
negative possibilities, there seems to be little doubt that swarm military technology will
eventually be deployed in future conflicts.

2. Spying Smart Home Devices


For smart home devices to respond to queries and be as useful as possible, they need to be
listening and tracking information about you and your regular habits. When you added the Echo
to your room as a radio and alarm clock (or any other smart device connected to the Internet),
you also allowed a spy to enter your home. All the information smart devices collect about your
habits such as your viewing history on Netflix; where you live and what route you take home so
Google can tell you how to avoid traffic; and what time you typically arrive home so your smart
thermostat can make your family room the temperature you prefer, is stored in the cloud. Of
course, this information makes your life more convenient, but there is also the potential for
abuse. In theory, virtual assistant devices listen for a "wake word," before they activate, but there
are instances when it might think you said the wake word and begin recording. Any smart device
in your home, including gaming consoles and smart TVs, could be the entry point for abuse of
your personal information. There are some defensive strategies such as covering up cameras,
turning off devices when not needed and muting microphones, but none of them are 100%
foolproof. 
3. Facial Recognition
There are some incredibly useful applications for facial recognition, but it can just as easily be
used for sinister purposes. China stands accused of using facial recognition technology for
surveillance and racial profiling. Not only do China's cameras spot jaywalkers, but they have
also monitored and controlled Uighur Muslims who live in the country. Russia's cameras scan
the streets for "people of interest," and there are reports that Israel tracks Palestinians inside the
West Bank. In addition to tracking people without their knowledge, facial recognition is plagued
with bias. When an algorithm is trained on a dataset that isn't diverse, it is less accurate and will
misidentify people more.

4. AI Cloning
With the support of artificial intelligence (AI), all that’s needed to create a clone of someone’s
voice is just a snippet of audio. Similarly, AI can take several photos or videos of a person and
then create an entirely new—cloned—video that appears to be an original. It’s become quite easy
for AI to create an artificial YOU and the results are so convincing our brains have trouble
differentiating between what is real and what is cloned. Deepfake technology that uses facial
mapping, machine learning, and artificial intelligence to create representations of real people
doing and saying things they never did is now targeting "ordinary" people. Celebrities used to be
more susceptible to being victims of deepfake technology because there was abundant video and
audio of them to use to train the algorithms. However, the technology has advanced to the point
that it doesn't require as much raw data to create a convincing fake video, plus there are a lot
more images and videos of ordinary people from the internet and social media channels to use.

5. Ransomware, AI and Bot-enabled Blackmailing and Hacking


When high-powered technology falls into the wrong hands, it can be very effective to achieve
criminal, immoral, and malicious activities. Ransomware, where malware is used to prevent
access to a computer system until a ransom is paid, is on the rise according to the Cybersecurity
and Infrastructure Security Agency (CISA). Artificial intelligence can automate tasks to get them
done more efficiently. When those tasks, such as spear phishing, are to send out fake emails to
trick people into giving up their private information, the negative impact could be extraordinary.
Once the software is built, there is little-to-no cost to keep repeating the task over again. AI can
quickly and efficiently blackmail people or hack into systems. Although AI is playing a
significant role to combat malware and other threats, it's also being used by cybercriminals to
perpetrate the crimes.

6. Smart Dust
Microelectromechanical systems (MEMS), the size of a grain of salt, have sensors,
communication mechanisms, autonomous power supplies, and cameras in them. Also called
motes, this smart dust has a plethora of positive uses in healthcare, security, and more, but would
be frightening to control if used for evil pursuits. While spying on a known enemy with smart
dust could fall into the positive column, the invasion of a private citizen’s privacy would be just
as easy.

7. Fake News Bots


GROVER is one AI system capable of writing a fake news article from nothing more than a
headline. AI systems such as GROVER create articles more believable than those written by
humans. OpenAI, a nonprofit company backed by Elon Musk, created “deepfakes for text” that
produces news stories and works of fiction so good, the organization initially decided not to
release the research publicly to prevent dangerous misuse of the technology. When fake articles
are promoted and shared as true, it can have serious ramifications for individuals, businesses, and
governments.

Along with the positive uses of today’s technology, there is no doubt that it can be very
dangerous in the wrong hands.

Marr, B. (2019, September 23). The 7 Most Dangerous Technology Trends In 2020 Everyone
Should Know About. Retrieved October 1, 2019, from
https://www.forbes.com/sites/bernardmarr/2019/09/23/the-7-most-dangerous-technology-trends-
in-2020-everyone-should-know-about/#1b07d6257780.

You might also like