For this article, we had a talk with Michael Kramarenko, our CTO, to examine both the hype and criticism aspects of ChatGPT with real-world examples and cases. Here, we’ll also delve into what makes ChatGPT such a groundbreaking tool, and of course, we’ll try to figure out whether to use it or rather stay aside
Whether we embrace it or not, our world is already shaped by the presence of ChatGPT and similar tools.
But like any new tech, these GPT-like tools offer both benefits and risks that we cannot disregard. Hence, the logical question arises, “To be, or not to be?” “To ChatGPT, or not to ChatGPT?”.
ChatGPT is a chatbot – a type of conversational AI built – but on top of a Large Language Model.
Although having boomed in late 2022, the technology behind ChatGPT – Transformer-based LM/LLM + self-attention mechanisms – was originally invented by the Google Brain team in 2017. However, it is somewhat surprising that eventually they have not capitalized on this innovation.
On the other hand, OpenAI pursued the idea and invested lots of time and effort to arrive at the point where they are now. And with it, perhaps tens of millions of dollars just to train the model. Not to mention daily computing expenses just to keep the model running.
While ChatGPT can generate impressive responses, using words like “understanding,” “feeling,” and “sense” to describe a communication process with GPT-based tools is incorrect. GPT is about parsing text to tokens, producing fixed-length embeddings, and generating the output sequence conditioned on the encoded input and the previously generated tokens.
ChatGPT’s ability to engage in conversations stems from its advanced capabilities to predict the most suitable next word or phrase based on the context of the input.
Thus, its responses are rather based on probabilities than personal feelings or subjective understanding. ChatGPT is just a sophisticated “word guesser,” so to speak.
ChatGPT has its limitations and can rather serve as a routine task performer and helper, not a problem solver.
But ultimately, only human workers are held accountable for the decision-making part and the outcomes it entails, as GPT-based tools are not infallible.
The tool can significantly enhance productivity and efficiency, helping do information processing tasks better and faster. The example of this may be an ai-powered assistant for fintech. It provides a quick, personalized, and human-like service to build strong customer relationships that propel business growth and help maximize revenue potential. This assistant allows humans to focus on more complex tasks.
Most disruptive technologies are changing professions and the labor market. It’s worth mentioning the times when “computers wore skirts” – here we’re talking about NASA Langley’s computer workers, who later retrained as programmers.
To sum up, only time will tell what the future of the labor market would look like.
The hype surrounding ChatGPT is characterized by the enthusiasm of millions of people who have eagerly signed up for it. And while social media is abuzz with public awe, each passing day ushers in a plethora of new plugins and extensions.
Therefore, if we consider the Gartner hype cycle within present-day reality, it becomes progressively challenging not only to define a plateau point of ChatGPT but at least to see any degree of enlightenment so far.
While ChatGPT has garnered attention for its impressive capabilities, its underlying technology is set to transform the professional, legislative, educational, and social aspects of our lives for its superb ability to enhance human productivity and efficiency across various domains.
Recent studies have found that AI assistance increases customer service worker productivity by a prominent 14% and empowers least-skilled workers to complete their tasks 35% faster. Remarkably, agents with 2 months of experience aided by AI performed just as well or better than those with 6 months of experience without artificial intelligence.
Thus, productivity boosting is the main benefit of ChatGPT and is the main aspect of public hype.
But as AI is constantly evolving, its impact on various aspects of our lives is yet to be fully understood.
Unlike traditional programming, where we develop software and can see the final code, the neural network itself becomes a valuable entity, representing the outcome of an extensively trained multi-billion-parameter network.
And that’s the case of revolutionary ChatGPT – an extensively-trained, enormous 175 billion-parameter neural network that has developed the ability to produce sophisticated, human-like outputs.
Here are key hype aspects of the ChatGPT tool.
ChatGPT is a large language model that operates based on the principles of GPT (Generative pre-trained transformers).
GPT models are characterized by two types of vertices within the neural network architecture.
The first component, known as the encoder, transforms the input tokens into a multidimensional vector (comprising 175 billion parameters for ChatGPT-3). Essentially, it generates a representation of the input.
Subsequently, the decoder analyzes these vectors and selects the most relevant response, considering the self-attention mechanism.
Self-attention mechanism is a crucial component of transformer architecture. This mechanism, in addition to assigning weights to each token, also captures and retains the input context to form dynamic, context-aware responses.
Reinforcement learning techniques are employed to fine-tune the model’s behavior. The process involves providing rewards or penalties to the model based on its generated responses. Through this feedback, the model can learn to generate more appropriate responses over time.
While data regarding GPT-4 remains undisclosed, the available information on GPT-3 is mind-blowing. The model boasts billions of parameters, an extensive training dataset, and a substantial word count. All of this has contributed to the model’s outstanding performance.
The new GPT-4 has been trained to achieve even better performance as compared to previous models.
And in fact, it scored way higher on the exams when compared to the latest GPT-3.
[source: OpenAI]
Frankly, it seems like those exam scores are the primary benchmark for assessing the quality of generated content.
Open AI allows for models’ fine-tuning up to the GPT 3.5 version.
This process helps optimize the AI model for your specific use case or domain for better performance, as compared to using the generic, pre-trained GPT-3 model out of the box.
The opening of the API has enabled widespread accessibility and integration.
Besides, it gave rise to a plethora of tools, extensions, and plugins, ranging from YouTube to IntelliJ IDEs.
In the near future, though, GPT-based tools are expected to step further, becoming a “building block” for many products and software solutions worldwide.
Accompanying the hype, there is indeed a wave of criticism that has emerged.
The global community calls for ethical considerations, transparency, and accountability in the development, deployment, and use of AI technologies like ChatGPT.
Among common issues are security, data privacy, lack of truth, copyright violation, misuse, job redundancy, and dangers to humanity.
Here’s an outline of what has happened so far:
Besides, many creative professionals accuse AI companies of using the works of artists to train models to imitate their work.
Amongst other criticism, there are also exaggerated claims that AI will inevitably replace all professions, leaving nothing but AI.
It is crucial to approach these tools with caution due to several critical aspects:
ChatGPT is not 100% reliable in generating accurate facts and data consistently. There have been numerous instances where ChatGPT has produced incorrect or misleading information.
Thus, unless you know a particular topic, you need to do fact-checking before relying on the information provided by generative models.
Basic example – when asked about the C-level management of our company, here’s what ChatGPT generated:
Now here are examples relevant to the IT industry.
When it comes to code generation, ChatGPT can generate code, but what it will generate and of what quality – remains questionable.
Here, when being asked how to set up a fine-tuned Open AI model, ChatGPT responded with irrelevant code.
There have also been studies on how well it can write test codes. And results show that about 30% of the code is low-quality.
There’s no better way to conclude this section than with this amusing viral tweet:
The ChatGPT responses lack true insight. When asked to provide detailed information or specifics, the tool tends to provide too generic responses that do not carry the necessary depth.
For example, when prompted ChatGPT to write non-functional requirements for a business analyst to consider within a given case, it responded just with a generic overview of NFR.
While it can help avoid overlooking certain aspects, it won’t take care of the essential details and nuances required. In other words, it won’t do the job for you.
ChatGPT does not have the ability to think or build rational judgments. It’s just not designed to perform logical tasks, like, for example, IBM’s Deep Blue is.
The process of generating a response is fundamentally probabilistic, making it uncertain what the model will generate, relatively speaking.
Here’s an amusing example to share – a classic puzzle designed to test the logical reasoning often presented to kindergarteners.
When presented with this puzzle, ChatGPT’s response was both interesting and hilarious. It suggested that the dog must possess an abnormal speed, something that doesn’t even exist in nature.
Yet, it didn’t suggest that maybe the dog should simply stand still and do not move in such a case.
Clearly, the model failed to reach the simple and straightforward logical conclusion that the puzzle intended and started to overcomplicate things.
AI systems like ChatGPT also raise ethical issues. A striking example of this can be a Copilot case.
Microsoft, GitHub, and OpenAI have faced a lawsuit from authors of open-source projects for violating intellectual property rights. Basically, the code published under an open-source license was used for monetization purposes of OpenAI.
Bias is indeed another issue, as ChatGPT learns from diverse sources, which can introduce conflicting or biased information.
Washington Post held research showing instances where ChatGPT has been trained on several media outlets that rank low for trustworthiness. And this ultimately prompts AI to produce misinformation, propaganda, and bias.
AI hallucinations are a unique problem associated with generative models like GPT.
In contrast to fact-checking, this phenomenon refers to the generation of content that does not exist in reality at all.
While lots of efforts are currently being made to mitigate this problem, as for now, it remains a significant issue.
Security is another critical aspect that requires careful attention when it comes to GPT-like models.
Here are some interesting cases that underscore the security risks associated with ChatGPT:
Recently, Samsung has reported an accidental leak of sensitive internal source code due to the careless behavior of their developers with ChatGPT. As of now, there have been at least two instances of this kind.
Many internet users have seen screenshots of ChatGPT-3 exposing AWS credentials. Yes, if asked directly, it usually responds within moderation guidelines, saying it does not spread security data. However, if one approaches ChatGPT-3 in a more cunning way, there have been cases where genuine AWS account credentials have been exposed to the network.
The next well-known story revolves around a Redis bug that caused several ChatGPT users to inadvertently view the requests of other users and personal information. OpenAI promptly addressed the issue and fixed the bug. However, given the nature of software development, it is difficult to guarantee that this will be the last bug or the second-to-last one.
Notably, against the backdrop of events, Italy decided to temporarily impose a ban on ChatGPT within its territory. In late March 2023, the Italian Data protection authority initiated an investigation into a suspected breach of Europe’s stringent privacy regulation.
After ChatGPT reactivation in late April 2023, Italy’s watchdog plans to ramp up scrutiny and review other AI systems as well.
ChatGPT is an exceptional AI tool that has expanded the boundaries of what was once possible and brought numerous benefits with it. But as with most new technologies, apart from benefits, it also presents risks.
However, stopping the march of technological progress is impossible; instead, we must learn to adapt and navigate these changes effectively yet carefully and shrewdly.
Here are the main points to sum up:
As for the IT industry though, it’s better to be cautious about both hype and widespread adoption of generative tools, given the evident problems with security, privacy, code leakages, and generated code quality overall.
We’d like to encourage all specialists not only to learn how to use GPT but also to cultivate the ability to develop innovative breakthrough technologies and products that address real-world problems effectively.
While it’s impossible to stop technological advances, a smart move is to remain vigilant, acknowledge the potential advantages and risks, and make informed decisions on how to best adapt to and leverage these technological advancements.
The topic of the gateways is a little more complicated and interesting than it seems.…
The rise of Buy Now, Pay Later (BNPL) services has disrupted the traditional lending landscape,…
Introduction or “What problem are we solving?” This article aims to illustrate how to create…
In today's fast-paced digital world, providing exceptional customer experience is crucial for businesses to stay…
It`s no secret that many parts of our lives become digital, and banking is no…
We discuss 5 top machine learning use cases in Finance and Banking industries to shed…