Skip to main content
  1. Data Science Blog/

GPT Usecases

·886 words·5 mins· loading · ·
Generative AI Language Models (LLMs) AI Applications AI Industry Natural Language Processing (NLP) Machine Learning (ML) Specific AI Models Language Models (LLMs) AI Technology Transformer Architecture

On This Page

Table of Contents
Share with :

GPT Usecases

What is GPT?
#

GPT is a transformer. Don’t confuse it with your electricity transformer! In Artificial Intelligence there are different kinds of neural network architectures to perform various tasks like classification, translation, segmentation, regression, etc. One of those architectures is transformer architecture. The Foundation of this architecture is based on another two architectures called encoder architecture and decoder architecture. There are lots of other technical complexity but for the business readers I am hiding that for that the time being, we will discuss that at some other place. In nutshell, GPT is a Transformer technology developed by OpenAI and it can perform several NLP tasks. NLP stands for natural language preprocessing. NLP tasks mean tasks like sentiment analysis of the text, text classification, topic modeling, translation, named entity recognition, and dozens of other tasks.

GPT stands for Generative Pre-Trained Transformer. You may be thinking it is a very loaded term. What does it mean at all? Here Generative means a system that can generate text or image or voice or style on text or image or voice. It means a generative system can create a painting of different types like a human creates and sometimes super quality very creative image than an average IQ human can create. A generative system can write poetry, essay, blogs, and a movie script. A generative system can generate voices of different moods, gender, and age. Pre-Trained means this is a system (model) created by OpenAI and all the parameters have been learned during the training processes. Generally, to use the GPT you need not do any training. You can simply call the model via an API call and ask whatever you want it to do. In the case of a specific requirement, you may train this but in that case, also you don’t do full training but you do little training, this is called fine-tuning. To do this process either you can add some more layers to the existing architecture or enable a couple of existing layers of the architecture to adjust the weights and biases during fine-tuning (training) process.

Apart from the GPT transformer, there are many other transformers but most of them are suitable for one or two tasks, in one or two languages. GPT transformer is the most generic transformer which can perform NLP tasks of almost all kinds and in many of the world languages. OpenAI created the first version of GPT called GPT1 in 2018, it was a 117M parameter model. On 14-Jan-2019 just before the start of COVID-19, OpenAI released a GPT2. it had 1.5 billion parameters. In May 2020, around the second wave of COVID-19, OpenAI released GPT3. This time it was having 175 billion parameters. Around Jan/Feb-23 it is expected to have GPT4. It is expected to have 100 trillion parameters.

GPT can perform many kinds of tasks, you can test them at Beta testing site. At this point in time, this site is working on GPT3 so all these tasks are performed using GPT3. When GPT4 will be released officially, then these features or maybe more features with better results can be processed.

Growth-of-GPT
Growth of GPT

Major Tasks of NLP
#

The Major Tasks of NLP from GPT are grouped into the following 7 categories. Each category has many specific tasks. Each task has hundreds of applications in your area of business interest.

Question Answering with GPT3 & GPT4
#

openAI-Answers

Classification Tasks with GPT3 & GPT4
#

openAI-Classification

Coding Tasks with with GPT3 & GPT4
#

openAI-Code

Conversations Tasks with GPT3 & GPT4
#

openAI-Conversation

Generations Tasks with GPT3 & GPT4
#

OpenAI-Generation

Transformation Tasks with GPT3 & GPT4
#

OpenAI-Transformation

Translation Tasks with GPT3 & GPT4
#

OpenAI-Translation

Conclusion
#

Transformers are the revolutionary architecture in NLP. GPT is the most revolutionary among all the transformer architectures because it is a general-purpose transformer and it can be used by any industry to fine-tune and solve domain-specific problems. AI, NLP, and the Large Language model have come to stay here. In the time to come we will see hundreds of products around GPT, they will either consume GPT service or fine-tune GPT model for their own domain-specific work. Technocrats are bringing forward disrupting technologies, and business people, politicians, and bureaucrats need to harness this technology for human wellbeing. Rather than making these technologies enemies of humans can we make them the friend of humans? Can we deliver better quality with new technology? Can we remove inefficiencies in our business and personal life? Can we transfer boring and tiring task to machines, so that humanity has more time to observe life, and think about something which is deeper and more meaningful for him or her?

References
#

OpenAI is still silent on the number of parameters and date of launch but there are many rumors in the market about these two things. Below are some interesting links.

Dr. Hari Thapliyaal's avatar

Dr. Hari Thapliyaal

Dr. Hari Thapliyal is a seasoned professional and prolific blogger with a multifaceted background that spans the realms of Data Science, Project Management, and Advait-Vedanta Philosophy. Holding a Doctorate in AI/NLP from SSBM (Geneva, Switzerland), Hari has earned Master's degrees in Computers, Business Management, Data Science, and Economics, reflecting his dedication to continuous learning and a diverse skill set. With over three decades of experience in management and leadership, Hari has proven expertise in training, consulting, and coaching within the technology sector. His extensive 16+ years in all phases of software product development are complemented by a decade-long focus on course design, training, coaching, and consulting in Project Management. In the dynamic field of Data Science, Hari stands out with more than three years of hands-on experience in software development, training course development, training, and mentoring professionals. His areas of specialization include Data Science, AI, Computer Vision, NLP, complex machine learning algorithms, statistical modeling, pattern identification, and extraction of valuable insights. Hari's professional journey showcases his diverse experience in planning and executing multiple types of projects. He excels in driving stakeholders to identify and resolve business problems, consistently delivering excellent results. Beyond the professional sphere, Hari finds solace in long meditation, often seeking secluded places or immersing himself in the embrace of nature.

Comments:

Share with :

Related

Roadmap to Reality
·990 words·5 mins· loading
Philosophy & Cognitive Science Interdisciplinary Topics Scientific Journey Self-Discovery Personal Growth Cosmic Perspective Human Evolution Technology Biology Neuroscience
Roadmap to Reality # A Scientific Journey to Know the Universe — and the Self # 🌱 Introduction: The …
From Being Hacked to Being Reborn: How I Rebuilt My LinkedIn Identity in 48 Hours
·893 words·5 mins· loading
Personal Branding Cybersecurity Technology Trends & Future Personal Branding LinkedIn Profile Professional Identity Cybersecurity Online Presence Digital Identity Online Branding
💔 From Being Hacked to Being Reborn: How I Rebuilt My LinkedIn Identity in 48 Hours # “In …
Exploring CSS Frameworks - A Collection of Lightweight, Responsive, and Themeable Alternatives
·1378 words·7 mins· loading
Web Development Frontend Development Design Systems CSS Frameworks Lightweight CSS Responsive CSS Themeable CSS CSS Utilities Utility-First CSS
Exploring CSS Frameworks # There are many CSS frameworks and approaches you can use besides …
Dimensions of Software Architecture: Balancing Concerns
·873 words·5 mins· loading
Software Architecture Software Architecture Technical Debt Maintainability Scalability Performance
Dimensions of Software Architecture # Call these “Architectural Concern Categories” or …
Understanding `async`, `await`, and Concurrency in Python
·616 words·3 mins· loading
Python Asyncio Concurrency Synchronous Programming Asynchronous Programming
Understanding async, await, and Concurrency # Understanding async, await, and Concurrency in Python …