Can ChatGPT really replace jobs

Can ChatGPT really replace jobs 2023


Can ChatGPT really replace jobs

Can ChatGPT really replace jobs: “ChatGPT is a powerful tool that can be used for various applications, including customer service, personal assistants, chatbots, and more” — this is what pops up on the screen when you ask the AI language model ChatGPT created by OpenAI, to explain ChatGPT.

It seems like ChatGPT can do almost anything, and it is not wrong to say that it will eat many jobs in the near future.

As said by the AI itself, the jobs it potentially can replace are content creation, customer support, translation and data entry analysis to name a few. Mostly, it can replace jobs that do repetitive work.

We all must have heard the term ‘Winter is coming’ from the famous web series ‘Game of Thrones’. ChatGPT seems to make this an official statement because it is coming true every passing day for most people in business.


Recently, Al Jazeera claimed in a news report that ChatGPTCreators fixed a bug that exposed the chat histories of users. Is this a mere coincidence? Or is this a sign that something major could go down in future?

Millions of people have used the software since it was first introduced, for everything from streamlining the coding process and creating architectural designs to using them like search engines or composing essays, draft notes, songs, novels, and even jokes.

Every conversation is immediately saved, and ChatGPT assigns a tab label based on the topic of the first query. Doesn’t it sound too good to be true?


We all know about the Pegasus spyware, which has the ability to read texts and emails, keep track of app activity, track location information, and gain access to a device’s microphone and camera.

Pegasus was created in 2016. According to the Pegasus Project investigations conducted in India, Pegasus spyware was allegedly used against ministers, opposition leaders, political strategists and tacticians, journalists, activists, minority leaders, Supreme Court judges, religious leaders, and administrators like election commissioners and Central Bureau of Investigation (CBI) heads.

See also  MacBook Air M2 gets flat Rs 14,400 discount on Amazon

In general, malicious software, also known as spyware, is installed on a computer without the end user’s awareness. The device is invaded, sensitive data and internet traffic data are stolen, and they are then relayed to advertisers, data companies, or outside users.

Any software can be classified as spyware if it is downloaded without the user’s permission. Spyware is contentious because it has the potential to be abused and can violate an end user’s privacy even when it is installed for comparatively innocent purposes.

Spyware is one of the most common threats to internet users. Once installed, it monitors internet activity, tracks login credentials and spies on sensitive information. The main objective of spyware is usually to obtain credit card numbers, banking information and passwords.


According to a British spy agency, artificially intelligent chatbots like ChatGPT pose a security risk because private information could be compromised or leaked.

Employees at large businesses like Amazon and JPMorgan have been advised not to use ChatGPT due to worries that confidential information may be disclosed.

In a blog article, the National Cyber Centre, a division of the intelligence agency GCHQ, outlined the risks that come with using a new generation of potent chatbots powered by AI for both individuals and businesses. According to the blog post, the risks that it includes are something to think about.

For example, people with wrong intentions could write more convincing phishing emails and the attackers could try techniques they weren’t familiar with previously. Various queries could be made publicly accessible, which could harm user information.

See also  The Potential of 3D Printing: Redefining Manufacturing and Prototyping

“ChatGPT is designed to generate responses on the basis of inputs it receives. This needs better regulations as many people are abusing technologies like ChatGPT,” said Major Vineet, Founder and Global President of Cyberpeace, which is a global non-profit organisation and thinktank of cyber and policy experts that work to build resilience against cyberattack and crime.

He also added that ChatGPT was being used to generate scripts to commence mass-scale phishing scams and codes for installing malware or ransomware at big scales. The use of technology needs to be regulated by better policies and laws to prevent misuse.

Major Vineet also pointed out that information is power but wrong information is a threat. Hence, the use of Chat-GPT will rise with time, leading to vulnerabilities and issues, but incorporating these practices into our daily cyber hygiene will allow us to prevent major security breaches which may result from ChatGPT.


A law professor at a US university has penned a blog that revealed how ChatGPT wrongly accused him of sexually harassing a female student. Shocking, right?

The professor found out that his name was involved in such an incident after receiving an email from a UCLA professor, who asked the AI chatbot to name 5 instances of sexual harassment by professors at American law schools.

Not only this, there are many incidents that are shouting loudly about the drawbacks or, in fact, dangers of ChatGPT.


It is an exciting time for technology’s growth. The world’s attention has been captured by ChatGPT in particular. Like with any technological advancement, there will be those who are eager to use it and explore its potential as well as those who may never use it.

See also  Airtel 5G now available in 3 more cities

But will this be at the cost of a breach in the security? Time will tell, but till then we have something to think about and a need to be aware.

If there are other AIs available just like ChatGPT, it’s hard not to talk about them in this article, especially when all AIs are a similar threat to one’s security. A big reason for the popularity of ChatGPT can also be one of its co-owners, Elon Musk, as he frequently makes headlines.

AI does not pose a danger to sentient AI or killer robots as of now. However, it does possess dangers with consumer privacy, biased programming, human safety, ambiguous legal regulations and more.


Despite the security concerns, there are many events that signify that ChatGPT can be a boon when used the right way. One such incident took place in March 2023, wherein a pet dog’s life was saved by ChatGPT’s newest version GPT-4, which correctly identified his medical condition when even veterinarians were unable to do so.

ChatGPT has time and again proved that it can revolutionise the world when used in the right manner. However, concerns about possible security breaches need to be addressed.


  • Be mindful of possible biases
  • Keep in mind that ChatGPT is not a replacement for expert guidance.
  • Avoid divulging personal, sensitive or critical information.
  • Check the chat interface’s legitimacy
  • Account security: Use strong passwords and 2FA. After use, log out
  • Be extremely cautious of links/attachments received from ChatGPT. Click on only trusted links.
  • Regularly monitor activities and transactions in the account associated with ChatGPT
  • Device security: Use a device that is secure and up-to-date with the latest security patches to interact with ChatGPT
  • Be aware: This is the key to remaining safe


Similar Posts