ChatGPT

From SI410
Revision as of 04:02, 13 February 2023 by Tekumull (Talk | contribs) (Limitations)

Jump to: navigation, search

ChatGPT (Chat Generative Pre-trained Transformer) [1] is a large-scale language generation model launched on November 30th, 2022 by OpenAI. The model is trained on a massive dataset of human-generated text and can generate human-like text in a wide range of styles and formats. The Chatbot is capable of mimicking human conversation and generating everything from essays to cooking recipes. However, ChatGPT's capabilities raise concerns about misuse, correctness, job market impact, and the impact on education.

Training

ChatGPT is based on OpenAI's GPT-3.5, an upgraded version of GPT-3, the A.I. text generator [2]. The GPT3.5 language model was initially built from a data set with 175 billion parameters including web content and other publicly available sources [3]. Supervised machine learning techniques were then used to build the model. Supervised learning refers to a class of systems and algorithms that determine a predictive model using data points with known outcomes. The model is learned by training through an appropriate learning algorithm (such as linear regression, random forests, or neural networks) that typically works through some optimization routine to minimize a loss or error function. [4]. It works much like the human brain, using interconnected ‘neurons’ that can learn to identify patterns in data and make predictions about what should come next. By using Reinforcement Learning from Human Feedback, human trainers played the role of both the user and the AI agent—generating a variety of responses to any given input and then evaluating and ranking them from best to worst [5]. This data was used to train a reward model.

An OpenAI reinforcement learning algorithm called Proximal Policy Optimization was used to fine-tune results [6]. Proximal Policy Optimization balances ease of implementation, sample complexity, and ease of tuning to compute an update at each step that minimizes the cost function while ensuring the deviation from the previous policy is relatively small [7]. The Transformer architecture used in ChatGPT is a type of neural network that is particularly well-suited to natural language processing tasks. The architecture is based on the idea of self-attention, which allows the model to weigh the importance of different words in a sentence when generating new text. This allows the model to generate text that is more coherent and relevant to the input. [8]

Capabilities

ChatGPT's main function is to mimic a human conversation. However, it can perform other functions as well. It has the ability to generate responses in a chatbot or virtual assistant, brainstorm content ideas on keywords or topics, create personalized communications such as email responses or product recommendations, create marketing content like blog posts or social media updates, translate text from one language to another, generate shorter summaries of long documents, and write and debug computer programs [9]. ChatGPT can understand and generate text in many languages including English, Spanish, French, German, Italian, Chinese, Japanese, Korean, and many more. It can also perform mathematical operations and answer problem sets related to math. It confers the ability to answer questions in science and other general knowledge topics, but it's not an expert in these areas and will tell the user when it is unsure of one of the outputs it generated. The model's vocabulary is quite large, with the ability to understand and generate a wide range of words and phrases.

Reception and Interest

After launching in November 2022, it took ChatGPT only 5 days to reach 1 million users, a feat that took social media platform Meta (formerly Facebook) 10 months and streaming platform Netflix three years to match [10]. Quickly after ChatGPT gained popularity in late 2022, multiple companies and big tech CEOs showed interest in the chatbot and its abilities. Buzzfeed Inc Chief Executive Jonah Peretti spoke about relying on ChatGPT creator OpenAI to enhance its quizzes and personalize some content for its audiences [11]. In January 2023, Microsoft confirmed it was making a "multibillion dollar" investment in OpenAi and said it plans to expand the partnership as part of a greater effort to add more artificial intelligence to its suite of products. [12] In December 2022, Google internally expressed alarm at the unexpected strength of ChatGPT and the newly discovered potential of large language models to be a serious threat to its main search business. [13]. Sunday Pichai, Google's Chief executive, upended the work of many teams to respond to the threat that ChatGPT poses [14].

Ethical Considerations

Misuse

One of the key ethical considerations with ChatGPT and other similar models is the potential for misuse. For example, the model's ability to generate realistic text could be used to spread misinformation or propaganda. Dr. Marcel Scharth of the University of Sydney demonstrated ChatGPT's ability to generate fake news with a fake news story about former President Barack Obama [15]. ADD PICTURE FROM ARTICLE HERE. Additionally, the model's ability to impersonate individuals or groups online could be used to harass or defraud others. ChatGPT's ability to write error-free code successfully lowers the bar for code generation and can help less-skilled threat actors effortlessly launch cyber-attacks. In a December 2022 report by Check Point Research, it was shown that ChatGPT was able to successfully conduct a full infection flow, from creating a convincing spear-phishing email to running a reverse shell, capable of accepting commands in English [16]. Notably, the Check Point Research report stated the researchers were able to conduct this full infection flow without writing a single line of code. They were able to complete this cyber-attack solely by putting together the pieces of the ChatGPT outputs. Another Check Point Research report released in January 2023 contained an analysis of major underground hacking communities and found evidence of cybercriminals using OpenAI technologies to develop malicious tools [17]. One such example is the creation of a basic stealer, which searches for 12 common file types (such as MS Office documents, PDFs, and images) across the system, copies them to a temporary folder, zips them, and sends them over the web without encryption or secure transfer methods. Another example is a simple Java snippet that downloads and covertly runs a common SSH and telnet client using Powershell, which can be modified to download and run any program, including malware. Another threat actor used OpenAI's ChatGPT to create a Python script that performs cryptographic operations, including signing files using elliptic curve cryptography, encryption using the Blowfish and Twofish algorithms, RSA keys, certificates stored in PEM format, MAC signing, and hash comparison using the blake2 hash function [18]. The script has the potential to be modified into ransomware if its script and syntax problems are fixed. The threat actor, "UsDoD," is a reputable and active member of the underground community who engages in various illicit activities, such as selling access to compromised companies and stolen databases.

Correctness of Responses

One concern that OpenAI addresses directly is the correctness of ChatGPT's responses. OpenAI notes that "ChatGPT sometimes writes plausible-sounding but incorrect or nonsensical answers" and fixing this issue is difficult because there’s no source of truth during reinforcement learning, training the model to be more cautious causes it to decline questions that it can answer correctly, and supervised training misleads the model because the ideal answer depends on what the model knows, rather than what the human demonstrator knows. [19] In early December 2022, Stack Overflow, a question-and-answer site for programmers, temporarily banned users from sharing responses generated by ChatGPT because "ChatGPT simply makes it too easy for users to generate responses and flood the site with answers that seem correct at first glance but are often wrong on close examination" [20].

Job Market Impact

There is also a concern about the impact of language models like ChatGPT on the job market. With the ability to generate high-quality text, the model could potentially automate the work of writers, editors, and other language-based professions. This could lead to significant job loss and economic disruption. In December 2022, Harvard Business Review warned about the potential for ChatGPT to take over roles traditionally held by humans such as copywriting, answering customer service inquiries, writing news reports, and creating legal documents. [21]. Marketing roles such as market research, Search Engine Optimization, and product descriptions can be automated using ChatGPT which threatens the Marketing Job Market [22]

Impact on Education

ChatGPT has the potential to have negative impacts on education systems as well. One concern is that it could lead to a decrease in critical thinking and creativity among students. When students are relying too heavily on ChatGPT-generated content, quizzes, or homework, they may not be developing their own understanding and problem-solving skills. Additionally, when students are using ChatGPT to generate answers to questions or complete assignments, it could lead to academic dishonesty and cheating. As of January 4th, 2023, New York Public Schools restricted access to the chatbot from its public school internet and devices after citing concerns of "concerns about negative impacts on student learning, and concerns regarding the safety and accuracy of contents” [23]. The decision to ban the chatbot in New York schools came amid widespread fears that it could encourage students to plagiarize. In a statement sent to the Washington Post shortly after these concerns were raised, ChatGPT creator OpenAI said "We don’t want ChatGPT to be used for misleading purposes in schools or anywhere else, so we’re already developing mitigations to help anyone identify text generated by that system" [24].

Limitations

There are limitations to what ChatGPT can do, and what it knows. As it is trained on a data set and is not connected to the Internet, there is a cut-off point for its knowledge base, which is currently the end of 2021. So, users may not use ChatGPT to gain information about current events.

A research paper published in January 2023 discovered patterns in ChatGPT content that makes it less suitable for critical applications. The paper found that humans preferred answers from ChatGPT in more than 50% of questions answered related to finance and psychology. But ChatGPT failed at answering medical questions because humans preferred direct answers. Specifically, the researchers cite that "ChatGPT performs poorly in terms of helpfulness for the medical domain in both English and Chinese. The ChatGPT often gives lengthy answers to medical consulting in our collected dataset, while human experts may directly give straightforward answers or suggestions, which may partly explain why volunteers consider human answers to be more helpful in the medical domain" [25]. The paper also mentions that when answering a question that requires professional knowledge from a particular field, ChatGPT may fabricate facts in order to give an answer [26]. For example, in legal questions, ChatGPT may invent some non-existent legal provisions to answer the question.


References:

  1. https://www.nytimes.com/2022/12/05/technology/chatgpt-ai-twitter.html
  2. https://www.nytimes.com/2022/12/05/technology/chatgpt-ai-twitter.html
  3. https://www.eweek.com/big-data-and-analytics/chatgpt/
  4. https://deepai.org/machine-learning-glossary-and-terms/supervised-learning
  5. https://openai.com/blog/chatgpt/
  6. https://www.eweek.com/big-data-and-analytics/chatgpt/
  7. https://openai.com/blog/openai-baselines-ppo/
  8. https://openai.com/blog/chatgpt/
  9. https://www.forbes.com/sites/bernardmarr/2022/12/21/chatgpt-everything-you-really-need-to-know-in-simple-terms/?sh=7adc5d67cbca
  10. https://www.euronews.com/next/2022/12/14/chatgpt-why-the-human-like-ai-chatbot-suddenly-got-everyone-talking
  11. https://www.wsj.com/articles/buzzfeed-to-use-chatgpt-creator-openai-to-help-create-some-of-its-content-11674752660
  12. https://www.cnn.com/2023/01/23/tech/microsoft-invests-chatgpt-openai/index.html
  13. https://www.nytimes.com/2022/12/21/technology/ai-chatgpt-google-search.html
  14. https://www.nytimes.com/2022/12/21/technology/ai-chatgpt-google-search.html
  15. https://www.sydney.edu.au/news-opinion/news/2022/12/08/the-chatgpt-chatbot-is-blowing-people-away-with-its-writing-skil.html
  16. https://research.checkpoint.com/2022/opwnai-ai-that-can-save-the-day-or-hack-it-away/
  17. https://research.checkpoint.com/2023/opwnai-cybercriminals-starting-to-use-chatgpt/
  18. https://research.checkpoint.com/2023/opwnai-cybercriminals-starting-to-use-chatgpt/
  19. https://openai.com/blog/chatgpt/
  20. https://www.theverge.com/2022/12/5/23493932/chatgpt-ai-generated-answers-temporarily-banned-stack-overflow-llms-dangers
  21. https://hbr.org/2022/12/chatgpt-and-how-ai-disrupts-industries
  22. https://www.forbes.com/sites/bernardmarr/2023/01/17/how-will-chatgpt-affect-your-job-if-you-work-in-advertising-and-marketing/?sh=21983fc639a3
  23. https://www.theguardian.com/us-news/2023/jan/06/new-york-city-schools-ban-ai-chatbot-chatgpt
  24. https://www.washingtonpost.com/education/2023/01/05/nyc-schools-ban-chatgpt/
  25. https://arxiv.org/pdf/2301.07597.pdf
  26. https://arxiv.org/pdf/2301.07597.pdf