Epstein Files Full PDF

CLICK HERE
Technopedia Center
PMB University Brochure
Faculty of Engineering and Computer Science
S1 Informatics S1 Information Systems S1 Information Technology S1 Computer Engineering S1 Electrical Engineering S1 Civil Engineering

faculty of Economics and Business
S1 Management S1 Accountancy

Faculty of Letters and Educational Sciences
S1 English literature S1 English language education S1 Mathematics education S1 Sports Education
teknopedia

  • Registerasi
  • Brosur UTI
  • Kip Scholarship Information
  • Performance
Flag Counter
  1. World Encyclopedia
  2. Generative pre-trained transformer - Wikipedia
Generative pre-trained transformer - Wikipedia
From Wikipedia, the free encyclopedia
(Redirected from Generative pre-trained transformers)
Type of large language model
Part of a series on
Machine learning
and data mining
Paradigms
  • Supervised learning
  • Unsupervised learning
  • Semi-supervised learning
  • Self-supervised learning
  • Reinforcement learning
  • Meta-learning
  • Online learning
  • Batch learning
  • Curriculum learning
  • Rule-based learning
  • Neuro-symbolic AI
  • Neuromorphic engineering
  • Quantum machine learning
Problems
  • Classification
  • Generative modeling
  • Regression
  • Clustering
  • Dimensionality reduction
  • Density estimation
  • Anomaly detection
  • Data cleaning
  • AutoML
  • Association rules
  • Semantic analysis
  • Structured prediction
  • Feature engineering
  • Feature learning
  • Learning to rank
  • Grammar induction
  • Ontology learning
  • Multimodal learning
Supervised learning
(classification • regression)
  • Apprenticeship learning
  • Decision trees
  • Ensembles
    • Bagging
    • Boosting
    • Random forest
  • k-NN
  • Linear regression
  • Naive Bayes
  • Artificial neural networks
  • Logistic regression
  • Perceptron
  • Relevance vector machine (RVM)
  • Support vector machine (SVM)
Clustering
  • BIRCH
  • CURE
  • Hierarchical
  • k-means
  • Fuzzy
  • Expectation–maximization (EM)

  • DBSCAN
  • OPTICS
  • Mean shift
Dimensionality reduction
  • Factor analysis
  • CCA
  • ICA
  • LDA
  • NMF
  • PCA
  • PGD
  • t-SNE
  • SDL
Structured prediction
  • Graphical models
    • Bayes net
    • Conditional random field
    • Hidden Markov
Anomaly detection
  • RANSAC
  • k-NN
  • Local outlier factor
  • Isolation forest
Neural networks
  • Autoencoder
  • Deep learning
  • Feedforward neural network
  • Recurrent neural network
    • LSTM
    • GRU
    • ESN
    • reservoir computing
  • Boltzmann machine
    • Restricted
  • GAN
  • Diffusion model
  • SOM
  • Convolutional neural network
    • U-Net
    • LeNet
    • AlexNet
    • DeepDream
  • Neural field
    • Neural radiance field
    • Physics-informed neural networks
  • Transformer
    • Vision
  • Mamba
  • Spiking neural network
  • Memtransistor
  • Electrochemical RAM (ECRAM)
Reinforcement learning
  • Q-learning
  • Policy gradient
  • SARSA
  • Temporal difference (TD)
  • Multi-agent
    • Self-play
Learning with humans
  • Active learning
  • Crowdsourcing
  • Human-in-the-loop
  • Mechanistic interpretability
  • RLHF
Model diagnostics
  • Coefficient of determination
  • Confusion matrix
  • Learning curve
  • ROC curve
Mathematical foundations
  • Kernel machines
  • Bias–variance tradeoff
  • Computational learning theory
  • Empirical risk minimization
  • Occam learning
  • PAC learning
  • Statistical learning
  • VC theory
  • Topological deep learning
Journals and conferences
  • AAAI
  • ECML PKDD
  • NeurIPS
  • ICML
  • ICLR
  • IJCAI
  • ML
  • JMLR
Related articles
  • Glossary of artificial intelligence
  • List of datasets for machine-learning research
    • List of datasets in computer vision and image processing
  • Outline of machine learning
  • v
  • t
  • e

Original GPT model

A generative pre-trained transformer (GPT) is a type of large language model (LLM)[1][2][3] that is widely used in generative artificial intelligence chatbots.[4][5] GPTs are based on a deep learning architecture called the transformer. They are pre-trained on large datasets of unlabeled content, and able to generate novel content.[2][3]

OpenAI was the first to apply generative pre-training to the transformer architecture, introducing the GPT-1 model in 2018.[6] The company has since released many bigger GPT models. The chatbot ChatGPT, released in late 2022 (using GPT-3.5), was followed by many competitor chatbots using their own generative pre-trained transformers to generate text, such as Gemini, DeepSeek and Claude.

GPTs are primarily used to generate text, but can be trained to generate other kinds of data. For example, GPT-4o can process and generate text, images and audio.[7] To improve performance on complex tasks, some GPTs, such as OpenAI o3, allocate more computation time analyzing the problem before generating an output, and are called reasoning models. In 2025, GPT-5 was released with a router that automatically selects whether to use a faster model or slower reasoning model based on the provided task.

Background

[edit]

During the 2010s, improved machine learning algorithms, more powerful computers, and an increase in the amount of digitized material allowed for an AI boom.[8]

Separately, the concept of generative pre-training (GP) was a long-established technique in machine learning. GP is a form of self-supervised learning wherein a model is first trained on a large, unlabeled dataset (the "pre-training" step) to learn to generate data points. This pre-trained model is then adapted to a specific task using a labeled dataset (the "fine-tuning" step).[9]

The transformer architecture for deep learning is the core technology of a GPT. Developed by researchers at Google, it was introduced in the paper "Attention Is All You Need", which was released on June 12, 2017. The transformer architecture solved many of the performance issues that were associated with older recurrent neural network (RNN) designs for natural language processing (NLP). The architecture's use of an attention mechanism allows models to process entire sequences of text at once, enabling the training of much larger and more sophisticated models. Since 2017, available transformer-based NLP systems have been capable of processing, mining, organizing, connecting, contrasting, and summarizing texts as well as answering questions from textual input.[citation needed]

History

[edit]

On June 11, 2018, OpenAI researchers and engineers published a paper called "Improving Language Understanding by Generative Pre-Training", which introduced GPT-1, the first GPT model.[10] It was designed as a transformer-based large language model that used generative pre-training (GP) on BookCorpus, a diverse text corpus, followed by discriminative fine-tuning to focus on specific language tasks.[11] This semi-supervised approach was seen as a breakthrough. Previously, the best-performing neural models in natural language processing (NLP) had commonly employed supervised learning from large amounts of manually labeled data – training a large language model with this approach would have been prohibitively expensive and time-consuming.[10]

On February 14, 2019, OpenAI introduced GPT-2, a larger model that could generate coherent text. Created as a direct scale-up of its predecessor, it had both its parameter count and dataset size increased by a factor of 10. GPT-2 has 1.5 billion parameters and was trained on WebText, a 40-gigabyte dataset of 8 million web pages.[12][13] Citing risks of malicious use, OpenAI opted for a "staged release", initially publishing smaller versions of the model before releasing the full 1.5-billion-parameter model in November.[14]

On February 10, 2020, Microsoft introduced its Turing Natural Language Generation, which it claimed was the "largest language model ever published at 17 billion parameters." The model outperformed all previous language models at a variety of tasks, including summarizing texts and answering questions.[15]

On May 28, 2020, OpenAI introduced GPT-3, a model with 175 billion parameters that was trained on a larger dataset compared to GPT-2. It marked a significant advancement in few-shot and zero-shot learning abilities. With few examples, it could perform various tasks that it was not explicitly trained for.[16][17]

Following the release of GPT-3, OpenAI started using reinforcement learning from human feedback (RLHF) to align models' behavior more closely with human preferences. This led to the development of InstructGPT, a fine-tuned version of GPT-3. OpenAI further refined InstructGPT to create ChatGPT, the flagship chatbot product of OpenAI that was launched on November 30, 2022.[18] ChatGPT was initially based on GPT-3.5, but it was later transitioned to the GPT-4 model, which was released on March 14, 2023.[19][20] GPT-4 was also integrated into parts of several applications, including Microsoft Copilot, GitHub Copilot, Snapchat, Khan Academy, and Duolingo.[21]

The immense popularity of ChatGPT spurred widespread development of competing GPT-based systems from other organizations. EleutherAI released a series of open-weight models, including GPT-J in 2021. Other major technology companies later developed their own GPT models, such as Google's PaLM and Gemini as well as Meta AI's Llama.[22]

Many subsequent GPT models have been trained to be multimodal (able to process or to generate multiple types of data). For example, GPT-4o can both process and generate text, images, and audio.[23] Additionally, GPT models like o3 and DeepSeek R1 have been trained with reinforcement learning to generate multi-step chain-of-thought reasoning before producing a final answer, which helps to solve complex problems in domains such as mathematics.[24]

On August 7, 2025, OpenAI released GPT-5, which includes a router that automatically selects whether to use a faster model or slower reasoning model based on task.[25][26]

Foundation models

[edit]

A foundation model is an AI model trained on broad data at scale such that it can be adapted to a wide range of downstream tasks.[27][28]

The most recent OpenAI's GPT-n series model is GPT-5.[29]

Other such models include Google's PaLM, a broad foundation model that has been compared to GPT-3 and has been made available to developers via an API,[30][31] and Together's GPT-JT, which has been reported as the closest-performing open-source alternative to GPT-3 (and is derived from earlier open-source GPTs).[32] Meta AI (formerly Facebook) also has a generative transformer-based foundational large language model, known as LLaMA.[33]

Foundational GPTs can also employ modalities other than text, for input and/or output. GPT-4 is a multi-modal LLM that is capable of processing text and image input (though its output is limited to text).[34] Regarding multimodal output, some generative transformer-based models are used for text-to-image technologies such as diffusion[35] and parallel decoding.[36] Such kinds of models can serve as visual foundation models (VFMs) for developing downstream systems that can work with images.[37]

Task-specific models

[edit]
Training workflow of original ChatGPT/InstructGPT release[38][39]

A foundational GPT model can be further adapted to produce more targeted systems directed to specific tasks and/or subject-matter domains. Methods for such adaptation can include additional fine-tuning (beyond that done for the foundation model) as well as certain forms of prompt engineering.[40]

An important example of this is fine-tuning models to follow instructions, which is of course a fairly broad task but more targeted than a foundation model. In January 2022, OpenAI introduced "InstructGPT" – a series of models which were fine-tuned to follow instructions using a combination of supervised training and reinforcement learning from human feedback (RLHF) on base GPT-3 language models.[41][42] Advantages this had over the bare foundational models included higher accuracy, less negative/toxic sentiment, and generally better alignment with user needs. Hence, OpenAI began using this as the basis for its API service offerings.[43] Other instruction-tuned models have been released by others, including a fully open version.[44][45]

Another (related) kind of task-specific models are chatbots, which engage in human-like conversation. In November 2022, OpenAI launched ChatGPT – an online chat interface powered by an instruction-tuned language model trained in a similar fashion to InstructGPT.[46] They trained this model using RLHF, with human AI trainers providing conversations in which they played both the user and the AI, and mixed this new dialogue dataset with the InstructGPT dataset for a conversational format suitable for a chatbot. Other major chatbots currently include Microsoft's Bing Chat, which uses OpenAI's GPT-4 (as part of a broader close collaboration between OpenAI and Microsoft),[47] and Google's competing chatbot Gemini (initially based on their LaMDA family of conversation-trained language models, with plans to switch to PaLM).[48]

Yet another kind of task that a GPT can be used for is the meta-task of generating its own instructions, like developing a series of prompts for 'itself' to be able to effectuate a more general goal given by a human user.[49] This is known as an AI agent, and more specifically a recursive one because it uses results from its previous self-instructions to help it form its subsequent prompts; the first major example of this was Auto-GPT (which uses OpenAI's GPT models), and others have since been developed as well.[50]

Domain-specificity

[edit]

GPT systems can be directed toward particular fields or domains. Some reported examples of such models and apps are as follows:

  • EinsteinGPT – for sales and marketing domains, to aid with customer relationship management (uses GPT-3.5)[51][52]
  • BloombergGPT – for the financial domain, to aid with financial news and information (uses "freely available" AI methods, combined with their proprietary data)[53]
  • Khanmigo – described as a GPT version for tutoring, in the education domain, it aids students using Khan Academy by guiding them through their studies without directly providing answers (powered by GPT-4)[54][55]
  • SlackGPT – for the Slack instant-messaging service, to aid with navigating and summarizing discussions on it (uses OpenAI's API)[56]
  • BioGPT – for the biomedical domain, to aid with biomedical literature text generation and mining (uses GPT-2)[57]

Sometimes domain-specificity is accomplished via software plug-ins or add-ons. For example, several different companies have developed particular plugins that interact directly with OpenAI's ChatGPT interface,[58][59] and Google Workspace has available add-ons such as "GPT for Sheets and Docs" – which is reported to aid use of spreadsheet functionality in Google Sheets.[60][61]

Brand issues

[edit]
OpenAI claims "GPT" to be its own branding, citing its association with ChatGPT and its model versions' designations.

OpenAI, which created the first generative pre-trained transformer (GPT) in 2018, asserted in 2023 that "GPT" should be regarded as a brand of OpenAI.[62] In April 2023, OpenAI revised the brand guidelines in its terms of service to indicate that other businesses using its API to run their AI services would no longer be able to include "GPT" in such names or branding.[63] In May 2023, OpenAI engaged a brand management service to notify its API customers of this policy, although these notifications stopped short of making overt legal claims (such as allegations of trademark infringement or demands to cease and desist).[62] As of November 2023, OpenAI still prohibits its API licensees from naming their own products with "GPT",[64] but it has begun enabling its ChatGPT Plus subscribers to make "custom versions of ChatGPT" called GPTs on the OpenAI site.[65] OpenAI's terms of service says that its subscribers may use "GPT" in the names of these, although it's "discouraged".[64]

Relatedly, OpenAI has applied to the United States Patent and Trademark Office (USPTO) to seek domestic trademark registration for the term "GPT" in the field of AI.[62] OpenAI sought to expedite handling of its application, but the USPTO declined that request in April 2023.[66] In May 2023, the USPTO responded to the application with a determination that "GPT" was both descriptive and generic.[67] As of November 2023, OpenAI continues to pursue its argument through the available processes. Regardless, failure to obtain a registered U.S. trademark does not preclude some level of common-law trademark rights in the U.S.[68] and trademark rights in other countries.[69]

For any given type or scope of trademark protection in the U.S., OpenAI would need to establish that the term is actually "distinctive" to their specific offerings in addition to being a broader technical term for the kind of technology. Some media reports suggested in 2023 that OpenAI may be able to obtain trademark registration based indirectly on the fame of its GPT-based chatbot product, ChatGPT,[66][70] for which OpenAI has separately sought protection (and which it has sought to enforce more strongly).[71] Other reports have indicated that registration for the bare term "GPT" seems unlikely to be granted,[62][72] as it is used frequently as a common term to refer simply to AI systems that involve generative pre-trained transformers.[3][73][74][75] In any event, to whatever extent exclusive rights in the term may occur the U.S., others would need to avoid using it for similar products or services in ways likely to cause confusion.[72][76] If such rights ever became broad enough to implicate other well-established uses in the field, the trademark doctrine of descriptive fair use could still continue non-brand-related usage.[77]

In the European Union, the European Union Intellectual Property Office registered "GPT" as a trade mark of OpenAI in spring 2023. However, since spring 2024 the registration is being challenged and is pending cancellation.[78]

In Switzerland, the Swiss Federal Institute of Intellectual Property registered "GPT" as a trade mark of OpenAI in spring 2023.[79][80]

See also

[edit]
  • Cyc
  • Vision transformer

References

[edit]
  1. ^ Haddad, Mohammed. "How does GPT-4 work and how can you start using it in ChatGPT?". www.aljazeera.com. Archived from the original on July 5, 2023. Retrieved April 10, 2023.
  2. ^ a b "Generative AI: a game-changer society needs to be ready for". World Economic Forum. January 9, 2023. Archived from the original on April 25, 2023. Retrieved April 8, 2023.
  3. ^ a b c "The A to Z of Artificial Intelligence". Time. April 13, 2023. Archived from the original on June 16, 2023. Retrieved April 14, 2023.
  4. ^ Hu, Luhui (November 15, 2022). "Generative AI and Future". Medium. Archived from the original on June 5, 2023. Retrieved April 29, 2023.
  5. ^ "CSDL | IEEE Computer Society". www.computer.org. Archived from the original on April 28, 2023. Retrieved April 29, 2023.
  6. ^ "Improving language understanding with unsupervised learning". openai.com. June 11, 2018. Archived from the original on March 18, 2023. Retrieved March 18, 2023.
  7. ^ Colburn, Thomas. "OpenAI unveils GPT-4o, a fresh multimodal AI flagship model". The Register. Retrieved May 18, 2024.
  8. ^ "An understanding of AI's limitations is starting to sink in". The Economist. June 11, 2020. ISSN 0013-0613. Archived from the original on July 31, 2020. Retrieved July 31, 2020.
  9. ^ Erhan, Dumitru; Courville, Aaron; Bengio, Yoshua; Vincent, Pascal (March 31, 2010). "Why Does Unsupervised Pre-training Help Deep Learning?". Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics. JMLR Workshop and Conference Proceedings: 201–208. Archived from the original on January 24, 2024. Retrieved January 24, 2024.
  10. ^ a b Radford, Alec; Narasimhan, Karthik; Salimans, Tim; Sutskever, Ilya (June 11, 2018). "Improving Language Understanding by Generative Pre-Training" (PDF). p. 12. Archived (PDF) from the original on January 26, 2021. Retrieved July 31, 2020.
  11. ^ Khandelwal, Umesh (April 1, 2023). "How Large Language GPT models evolved and work". Archived from the original on April 4, 2023. Retrieved April 3, 2023.
  12. ^ Vincent, James (February 14, 2019). "OpenAI's new multitalented AI writes, translates, and slanders". The Verge. Archived from the original on December 18, 2020. Retrieved December 19, 2020.
  13. ^ Radford, Alec; Wu, Jeffrey; Child, Rewon; Luan, David; Amodei, Dario; Sutskever, Ilua (February 14, 2019). "Language models are unsupervised multitask learners" (PDF). OpenAI. 1 (8). Archived (PDF) from the original on February 6, 2021. Retrieved December 19, 2020.
  14. ^ Vincent, James (November 7, 2019). "OpenAI has published the text-generating AI it said was too dangerous to share". The Verge. Archived from the original on June 11, 2020. Retrieved April 28, 2023.
  15. ^ Sterling, Bruce (February 13, 2020). "Web Semantics: Microsoft Project Turing introduces Turing Natural Language Generation (T-NLG)". Wired. ISSN 1059-1028. Archived from the original on November 4, 2020. Retrieved July 31, 2020.
  16. ^ Sagar, Ram (June 3, 2020). "OpenAI Releases GPT-3, The Largest Model So Far". Analytics India Magazine. Archived from the original on August 4, 2020. Retrieved July 31, 2020.
  17. ^ Brown, Tom B.; Mann, Benjamin; Ryder, Nick; Subbiah, Melanie; Kaplan, Jared; Dhariwal, Prafulla; Neelakantan, Arvind; Shyam, Pranav; Sastry, Girish; Askell, Amanda; Agarwal, Sandhini; Herbert-Voss, Ariel; Krueger, Gretchen; Henighan, Tom; Child, Rewon; Ramesh, Aditya; Ziegler, Daniel M.; Wu, Jeffrey; Winter, Clemens; Hesse, Christopher; Chen, Mark; Sigler, Eric; Litwin, Mateusz; Gray, Scott; Chess, Benjamin; Clark, Jack; Berner, Christopher; McCandlish, Sam; Radford, Alec; Sutskever, Ilya; Amodei, Dario (May 28, 2020). "Language Models are Few-Shot Learners". arXiv:2005.14165 [cs.CL].
  18. ^ Fu, Yao; Peng, Hao; Khot, Tushar (2022). "How does GPT Obtain its Ability? Tracing Emergent Abilities of Language Models to their Sources". Yao Fu's Notion. Archived from the original on April 19, 2023. Retrieved June 24, 2023.
  19. ^ Edwards, Benj (March 14, 2023). "OpenAI's GPT-4 exhibits "human-level performance" on professional benchmarks". Ars Technica. Archived from the original on March 14, 2023. Retrieved March 15, 2023.
  20. ^ OpenAI (March 15, 2023). "GPT-4 Technical Report". arXiv:2303.08774 [cs.CL].
  21. ^ Gupta, Aman (March 21, 2023). "GPT-4 takes the world by storm - List of companies that integrated the chatbot". Mint. Retrieved January 23, 2024.
  22. ^ Alford, Anthony (July 13, 2021). "EleutherAI Open-Sources Six Billion Parameter GPT-3 Clone GPT-J". InfoQ. Archived from the original on February 10, 2023. Retrieved April 3, 2023.
  23. ^ Colburn, Thomas. "OpenAI unveils GPT-4o, a fresh multimodal AI flagship model". The Register. Retrieved May 18, 2024.
  24. ^ Zia, Dr Tehseen (March 29, 2025). "How OpenAI's o3, Grok 3, DeepSeek R1, Gemini 2.0, and Claude 3.7 Differ in Their Reasoning Approaches". Unite.AI. Retrieved August 3, 2025.
  25. ^ Heath, Alex (August 7, 2025). "GPT-5 is being released to all ChatGPT users". The Verge. Retrieved August 7, 2025.
  26. ^ "Introducing GPT‑5". OpenAI. August 7, 2025. Retrieved August 7, 2025.
  27. ^ "Introducing the Center for Research on Foundation Models (CRFM)". Stanford HAI. August 18, 2021. Archived from the original on June 4, 2023. Retrieved April 26, 2023.
  28. ^ "Reflections on Foundation Models". hai.stanford.edu. October 18, 2021. Archived from the original on August 15, 2024. Retrieved August 15, 2024.
  29. ^ "Introducing GPT-5". openai.com. August 7, 2025. Retrieved August 14, 2025.
  30. ^ Vincent, James (March 14, 2023). "Google opens up its AI language model PaLM to challenge OpenAI and GPT-3". The Verge. Archived from the original on March 14, 2023. Retrieved April 29, 2023.
  31. ^ "Google Opens Access to PaLM Language Model". Archived from the original on May 31, 2023. Retrieved April 29, 2023.
  32. ^ Iyer, Aparna (November 30, 2022). "Meet GPT-JT, the Closest Open Source Alternative to GPT-3". Analytics India Magazine. Archived from the original on June 2, 2023. Retrieved April 29, 2023.
  33. ^ "Meta Debuts AI Language Model, But It's Only for Researchers". PCMAG. February 24, 2023. Archived from the original on July 19, 2023. Retrieved May 21, 2023.
  34. ^ Islam, Arham (March 27, 2023). "Multimodal Language Models: The Future of Artificial Intelligence (AI)". Archived from the original on May 15, 2023. Retrieved May 15, 2023.
  35. ^ Islam, Arham (November 14, 2022). "How Do DALL·E 2, Stable Diffusion, and Midjourney Work?". Archived from the original on July 18, 2023. Retrieved May 21, 2023.
  36. ^ Saha, Shritama (January 4, 2023). "Google Launches Muse, A New Text-to-Image Transformer Model". Analytics India Magazine. Archived from the original on May 15, 2023. Retrieved May 15, 2023.
  37. ^ Wu (et-al), Chenfei (March 8, 2023). "Visual ChatGPT". arXiv:2303.04671 [cs.CV].
  38. ^ Ouyang, Long; Wu, Jeff; et al. (March 4, 2022). "Training language models to follow instructions with human feedback". arXiv:2203.02155 [cs.CL].
  39. ^ OpenAI (January 27, 2022). "Aligning language models to follow instructions". OpenAI. Retrieved July 29, 2025.
  40. ^ Bommasani (et-al), Rishi (July 12, 2022). "On the Opportunities and Risks of Foundation Models". arXiv:2108.07258 [cs.LG].
  41. ^ "Aligning language models to follow instructions". openai.com. Archived from the original on March 23, 2023. Retrieved March 23, 2023.
  42. ^ Ouyang, Long; Wu, Jeff; Jiang, Xu; et al. (November 4, 2022). "Training language models to follow instructions with human feedback". NeurIPS. arXiv:2203.02155.
  43. ^ Ramnani, Meeta (January 28, 2022). "OpenAI dumps its own GPT-3 for something called InstructGPT, and for right reason". Analytics India Magazine. Archived from the original on June 4, 2023. Retrieved April 29, 2023.
  44. ^ "Stanford CRFM". crfm.stanford.edu. Archived from the original on April 6, 2023. Retrieved May 15, 2023.
  45. ^ "Free Dolly: Introducing the World's First Truly Open Instruction-Tuned LLM". Databricks. April 12, 2023. Archived from the original on July 14, 2023. Retrieved May 15, 2023.
  46. ^ "Introducing ChatGPT". openai.com. Archived from the original on March 16, 2023. Retrieved March 16, 2023.
  47. ^ Wiggers, Kyle (May 4, 2023). "Microsoft doubles down on AI with new Bing features". Archived from the original on December 7, 2023. Retrieved May 4, 2023.
  48. ^ "ChatGPT vs. Bing vs. Google Bard: Which AI Is the Most Helpful?". CNET. Archived from the original on July 24, 2023. Retrieved April 30, 2023.
  49. ^ "Auto-GPT, BabyAGI, and AgentGPT: How to use AI agents". Mashable. April 19, 2023. Archived from the original on July 22, 2023. Retrieved May 15, 2023.
  50. ^ Marr, Bernard. "Auto-GPT May Be The Strong AI Tool That Surpasses ChatGPT". Forbes. Archived from the original on May 21, 2023. Retrieved May 15, 2023.
  51. ^ Morrison, Ryan (March 7, 2023). "Salesforce launches EinsteinGPT built with OpenAI technology". Archived from the original on April 15, 2023. Retrieved April 10, 2023.
  52. ^ Sharma, Animesh K.; Sharma, Rahul (2023). "The role of generative pretrained transformers (GPTs) in revolutionising digital marketing: A conceptual model". Journal of Cultural Marketing Strategy. 8 (1): 80–90. doi:10.69554/TLVQ2275.
  53. ^ Leswing, Kif (April 13, 2023). "Bloomberg plans to integrate GPT-style A.I. into its terminal". CNBC. Archived from the original on May 19, 2023. Retrieved May 4, 2023.
  54. ^ Melendez, Steven (May 4, 2023). "Learning nonprofit Khan Academy is piloting a version of GPT called Khanmigo". Fast Company. Archived from the original on May 11, 2023. Retrieved May 22, 2023.
  55. ^ "Khan Academy Pilots GPT-4 Powered Tool Khanmigo for Teachers". THE Journal. Archived from the original on May 7, 2023. Retrieved May 7, 2023.
  56. ^ Hachman, Mark (May 4, 2023). "Slack GPT will bring AI chatbots to your conversations". PCWorld. Archived from the original on June 9, 2023. Retrieved May 4, 2023.
  57. ^ Luo (et-al), Renqian (April 3, 2023). "BioGPT: Generative pre-trained transformer for biomedical text generation and mining". Briefings in Bioinformatics. 23 (6) bbac409. arXiv:2210.10341. doi:10.1093/bib/bbac409. PMID 36156661.
  58. ^ John, Amy Sarah (May 5, 2023). "Know about ChatGPT's 13 best plugins, designed to improve your overall user experience". Latest Digital Transformation Trends | Cloud News | Wire19. Archived from the original on May 9, 2023. Retrieved May 7, 2023.
  59. ^ "ChatGPT plugins". openai.com. March 13, 2024. Archived from the original on March 23, 2023. Retrieved May 7, 2023.
  60. ^ "How to Use ChatGPT on Google Sheets With GPT for Sheets and Docs". MUO. March 12, 2023. Archived from the original on June 19, 2023. Retrieved May 7, 2023.
  61. ^ Asay, Matt (February 27, 2023). "Embrace and extend Excel for AI data prep". InfoWorld. Archived from the original on June 2, 2023. Retrieved May 7, 2023.
  62. ^ a b c d Hicks, William (May 10, 2023). "ChatGPT creator OpenAI is asking startups to remove 'GPT' from their names". The Business Journal. Archived from the original on June 28, 2023. Retrieved May 21, 2023.
  63. ^ OpenAI (April 24, 2023). "Brand Guidelines". Archived from the original on July 18, 2023. Retrieved May 21, 2023.
  64. ^ a b "Brand guidelines". Archived from the original on July 18, 2023. Retrieved November 28, 2023.
  65. ^ "Introducing GPTS". March 13, 2024. Archived from the original on March 20, 2024. Retrieved November 28, 2023.
  66. ^ a b Heah, Alexa (April 26, 2023). "OpenAI Unsuccessful At Speeding Up Its Attempt To Trademark 'GPT'". DesignTAXI. Archived from the original on April 26, 2023. Retrieved May 21, 2023.
  67. ^ "NONFINAL OFFICE ACTION". USPTO. May 25, 2023. Archived from the original on December 3, 2023. Retrieved December 30, 2023.
  68. ^ "U.S. Trademark Law". December 2015. Archived from the original on January 17, 2024. Retrieved November 29, 2023.
  69. ^ "International Trademark Rights". Archived from the original on March 11, 2024. Retrieved November 29, 2023.
  70. ^ "OpenAI Wants to Trademark 'GPT' Amid Rise of AI Chatbots". Tech Times. April 25, 2023. Archived from the original on April 25, 2023. Retrieved May 21, 2023.
  71. ^ Louise, Nickie (April 3, 2023). "OpenAI files a UDRP case against the current owner of ChatGPT.com". Archived from the original on June 5, 2023. Retrieved May 21, 2023.
  72. ^ a b Demcak, Tramatm-Igor (April 26, 2023). "OpenAI's Battle for Brand Protection: Can GPT be trademarked?". Lexology. Archived from the original on May 5, 2023. Retrieved May 22, 2023.
  73. ^ Lawton, George (April 20, 2023). "ChatGPT vs. GPT: How are they different? | TechTarget". Enterprise AI. Archived from the original on May 9, 2023. Retrieved May 21, 2023.
  74. ^ Robb, Drew (April 12, 2023). "GPT-4 vs. ChatGPT: AI Chatbot Comparison". eWEEK. Archived from the original on July 27, 2023. Retrieved May 21, 2023.
  75. ^ Russo, Philip (August 22, 2023). "The Genesis of Generative AI for Everything Everywhere All at Once in CRE". Commercial Observer. Archived from the original on August 24, 2023.
  76. ^ "Trademark infringement". Archived from the original on November 30, 2023. Retrieved November 29, 2023.
  77. ^ Rheintgen, Husch Blackwell LLP-Kathleen A. (August 16, 2013). "Branding 101: trademark descriptive fair use". Lexology. Archived from the original on May 21, 2023. Retrieved May 21, 2023.
  78. ^ "EUIPO - eSearch". euipo.europa.eu. Retrieved September 4, 2025.
  79. ^ "IPI Database". www.swissreg.ch. Retrieved September 4, 2025.
  80. ^ Vogt, Reto (February 20, 2024). "OpenAI sichert sich in der Schweiz "GPT" als Markenname". www.inside-it.ch (in German). Archived from the original on March 21, 2025. Retrieved September 4, 2025.
  • v
  • t
  • e
OpenAI
Products
ChatGPT
  • Atlas
  • Deep Research
  • GPT Store
  • Search
Foundation
models
GPT models
  • GPT-1
  • GPT-2
  • GPT-3
  • GPT-4
  • GPT-4o
  • o1
  • o3
  • GPT-4.5
  • GPT-4.1
  • o4-mini
  • GPT-OSS
  • GPT-5
  • GPT-5.1
  • GPT-5.2
Specialized
  • DALL-E
  • GPT Image
  • Sora
  • Whisper
Intelligent
agents
  • Codex
  • Operator
People
Senior
management
Current
  • Sam Altman
    • removal
  • Greg Brockman
  • Sarah Friar
  • Jakub Pachocki
  • Scott Schools
Former
  • Mira Murati
  • Emmett Shear
Board of
directors
Current
  • Sam Altman
  • Adam D'Angelo
  • Sue Desmond-Hellmann
  • Zico Kolter
  • Paul Nakasone
  • Adebayo Ogunlesi
  • Nicole Seligman
  • Fidji Simo
  • Bret Taylor (chair)
Former
  • Greg Brockman (2017–2023)
  • Reid Hoffman (2019–2023)
  • Will Hurd (2021–2023)
  • Holden Karnofsky (2017–2021)
  • Elon Musk (2015–2018)
  • Ilya Sutskever (2017–2023)
  • Helen Toner (2021–2023)
  • Shivon Zilis (2019–2023)
  • Lawrence Summers (2023-2025)
JVs
  • Stargate LLC
Related
  • ChatGPT in education
  • Apple Intelligence
  • AI Dungeon
  • AutoGPT
  • GitHub Copilot
  • Contrastive Language-Image Pre-training
  • "Deep Learning"
  • LangChain
  • Microsoft Copilot
  • OpenAI Five
  • Transformer
  • Category
  • v
  • t
  • e
Generative AI
Concepts
  • Autoencoder
  • Deep learning
  • Fine-tuning
  • Foundation model
  • Generative adversarial network
  • Generative pre-trained transformer
  • Large language model
  • Model Context Protocol
  • Neural network
  • Prompt engineering
  • Reinforcement learning from human feedback
  • Retrieval-augmented generation
  • Self-supervised learning
  • Slop
  • Stochastic parrot
  • Synthetic data
  • Top-p sampling
  • Transformer
  • Variational autoencoder
  • Vibe coding
  • Vision transformer
  • Word embedding
Models
Text
  • Character.ai
  • ChatGPT
  • Command
  • Claude
  • DeepSeek
  • Ernie
  • Gemini
  • Gemma
  • GLM
  • GPT
    • 1
    • 2
    • 3
    • J
    • 3.5
    • 4
    • 4o
    • o1
    • o3
    • 4.5
    • 4.1
    • o4-mini
    • OSS
    • 5
    • 5.1
    • 5.2
  • Grok
  • Hunyuan
  • Kimi
  • Llama
  • Microsoft Copilot
  • MiniMax
  • Mistral Large
  • Qwen
  • Velvet
  • Solar Pro
Coding
  • Claude Code
  • Cursor
  • Devstral
  • GitHub Copilot
  • Google Antigravity
  • Grok Code Fast 1
  • Kimi Code
  • Qwen3-Coder
  • Replit
Image
  • Aurora
  • Firefly
  • Flux
  • GPT Image
  • Grok Imagine
  • Ideogram
  • Imagen
  • Leonardo
  • Midjourney
  • Nano Banana
  • Qwen-Image
  • Recraft
  • Seedream
  • Stable Diffusion
Video
  • Dream Machine
  • Genie
  • Hailuo AI
  • Kling AI
  • LTX-2
  • Luma Ray
  • Runway Gen
  • Seedance
  • Sora
  • Veo
  • Wan
Speech
  • 15.ai
  • Eleven
  • Gemini Speech
  • GPT-4o mini TTS
  • MiniMax Speech
  • Speechify
Music
  • Eleven Music
  • Endel
  • Lyria
  • MiniMax Music
  • Riffusion
  • Stable Audio
  • Suno
  • Udio
Agents
  • Agentforce
  • AutoGLM
  • AutoGPT
  • ChatGPT agent
  • Devin AI
  • Manus
  • MiniMax Agent
  • OpenAI Codex
  • OpenClaw
  • Replit Agent
Companies
  • Adobe
  • Aleph Alpha
  • Anthropic
  • Anysphere
  • Baichuan
  • Canva
  • Cognition AI
  • Cohere
  • Contextual AI
  • DeepSeek
  • DeepL
  • EleutherAI
  • ElevenLabs
  • Google DeepMind
  • HeyGen
  • Hugging Face
  • Inflection AI
  • Krikey AI
  • Kuaishou
  • Lightricks
  • Luma Labs
  • Meta AI
  • MiniMax
  • Mistral AI
  • Moonshot AI
  • OpenAI
  • Perplexity AI
  • Runway
  • Safe Superintelligence
  • Sakana AI
  • Salesforce
  • Scale AI
  • ServiceNow
  • SoundHound
  • Stability AI
  • StepFun
  • Synthesia
  • Thinking Machines Lab
  • Upstage
  • xAI
  • Z.ai
Controversies
  • Generative AI pornography
    • Deepfake pornography
      • on Grok
      • of Taylor Swift
  • Google Gemini image generation
  • Pause Giant AI Experiments
  • Removal of Sam Altman from OpenAI
  • Statement on AI Risk
  • Tay (chatbot)
  • Théâtre D'opéra Spatial
  • Voiceverse NFT plagiarism
  • Category
  • v
  • t
  • e
Artificial intelligence (AI)
  • History
    • timeline
  • Glossary
  • Companies
  • Projects
Concepts
  • Parameter
    • Hyperparameter
  • Loss functions
  • Regression
    • Bias–variance tradeoff
    • Double descent
    • Overfitting
  • Clustering
  • Gradient descent
    • SGD
    • Quasi-Newton method
    • Conjugate gradient method
  • Backpropagation
  • Attention
  • Convolution
  • Normalization
    • Batchnorm
  • Activation
    • Softmax
    • Sigmoid
    • Rectifier
  • Gating
  • Weight initialization
  • Regularization
  • Datasets
    • Augmentation
  • Prompt engineering
  • Reinforcement learning
    • Q-learning
    • SARSA
    • Imitation
    • Policy gradient
  • Diffusion
  • Latent diffusion model
  • Autoregression
  • Adversary
  • RAG
  • Uncanny valley
  • RLHF
  • Self-supervised learning
  • Reflection
  • Recursive self-improvement
  • Hallucination
  • Word embedding
  • Vibe coding
Applications
  • Machine learning
    • In-context learning
  • Artificial neural network
    • Deep learning
  • Language model
    • Large
    • NMT
    • Reasoning
  • Model Context Protocol
  • Intelligent agent
  • Artificial human companion
  • Humanity's Last Exam
  • Lethal autonomous weapons (LAWs)
  • Generative artificial intelligence (GenAI)
  • (Hypothetical: Artificial general intelligence (AGI))
  • (Hypothetical: Artificial superintelligence (ASI))
  • Agent2Agent protocol
Implementations
Audio–visual
  • AlexNet
  • WaveNet
  • Human image synthesis
  • HWR
  • OCR
  • Computer vision
  • Speech synthesis
    • 15.ai
    • ElevenLabs
  • Speech recognition
    • Whisper
  • Facial recognition
  • AlphaFold
  • Text-to-image models
    • Aurora
    • DALL-E
    • Firefly
    • Flux
    • GPT Image
    • Ideogram
    • Imagen
    • Midjourney
    • Recraft
    • Stable Diffusion
  • Text-to-video models
    • Dream Machine
    • Runway Gen
    • Hailuo AI
    • Kling
    • Sora
    • Seedance
    • Veo
  • Music generation
    • Riffusion
    • Suno AI
    • Udio
Text
  • Word2vec
  • Seq2seq
  • GloVe
  • BERT
  • T5
  • Llama
  • Chinchilla AI
  • PaLM
  • GPT
    • 1
    • 2
    • 3
    • J
    • ChatGPT
    • 4
    • 4o
    • o1
    • o3
    • 4.5
    • 4.1
    • o4-mini
    • 5
    • 5.1
    • 5.2
  • Claude
  • Gemini
    • Gemini (language model)
    • Gemma
  • Grok
  • LaMDA
  • BLOOM
  • DBRX
  • Project Debater
  • IBM Watson
  • IBM Watsonx
  • Granite
  • PanGu-Σ
  • DeepSeek
  • Qwen
Decisional
  • AlphaGo
  • AlphaZero
  • OpenAI Five
  • Self-driving car
  • MuZero
  • Action selection
    • AutoGPT
  • Robot control
People
  • Alan Turing
  • Warren Sturgis McCulloch
  • Walter Pitts
  • John von Neumann
  • Christopher D. Manning
  • Claude Shannon
  • Shun'ichi Amari
  • Kunihiko Fukushima
  • Takeo Kanade
  • Marvin Minsky
  • John McCarthy
  • Nathaniel Rochester
  • Allen Newell
  • Cliff Shaw
  • Herbert A. Simon
  • Oliver Selfridge
  • Frank Rosenblatt
  • Bernard Widrow
  • Joseph Weizenbaum
  • Seymour Papert
  • Seppo Linnainmaa
  • Paul Werbos
  • Geoffrey Hinton
  • John Hopfield
  • Jürgen Schmidhuber
  • Yann LeCun
  • Yoshua Bengio
  • Lotfi A. Zadeh
  • Stephen Grossberg
  • Alex Graves
  • James Goodnight
  • Andrew Ng
  • Fei-Fei Li
  • Alex Krizhevsky
  • Ilya Sutskever
  • Oriol Vinyals
  • Quoc V. Le
  • Ian Goodfellow
  • Demis Hassabis
  • David Silver
  • Andrej Karpathy
  • Ashish Vaswani
  • Noam Shazeer
  • Aidan Gomez
  • John Schulman
  • Mustafa Suleyman
  • Jan Leike
  • Daniel Kokotajlo
  • François Chollet
Architectures
  • Neural Turing machine
  • Differentiable neural computer
  • Transformer
    • Vision transformer (ViT)
  • Recurrent neural network (RNN)
  • Long short-term memory (LSTM)
  • Gated recurrent unit (GRU)
  • Echo state network
  • Multilayer perceptron (MLP)
  • Convolutional neural network (CNN)
  • Residual neural network (RNN)
  • Highway network
  • Mamba
  • Autoencoder
  • Variational autoencoder (VAE)
  • Generative adversarial network (GAN)
  • Graph neural network (GNN)
Political
  • AI safety (Alignment)
  • Ethics of AI
  • EU AI Act
  • Precautionary principle
  • Regulation of AI
  • Virtual politician
Social and economic
  • AI boom
  • AI bubble
  • AI literacy
  • AI slop
  • AI veganism
  • AI winter
  • Anthropomorphism
  • In architecture
  • In education
  • In healthcare
    • Chatbot psychosis
    • Mental health
  • In visual art
  • Category
  • v
  • t
  • e
Natural language processing
General terms
  • AI-complete
  • Bag-of-words
  • n-gram
    • Bigram
    • Trigram
  • Computational linguistics
  • Natural language understanding
  • Stop words
  • Text processing
Text analysis
  • Argument mining
  • Collocation extraction
  • Concept mining
  • Coreference resolution
  • Deep linguistic processing
  • Distant reading
  • Information extraction
  • Named-entity recognition
  • Ontology learning
  • Parsing
    • semantic
    • syntactic
  • Part-of-speech tagging
  • Semantic analysis
  • Semantic role labeling
  • Semantic decomposition
  • Semantic similarity
  • Sentiment analysis
  • Terminology extraction
  • Text mining
  • Textual entailment
  • Truecasing
  • Word-sense disambiguation
  • Word-sense induction
Text segmentation
  • Compound-term processing
  • Lemmatisation
  • Lexical analysis
  • Text chunking
  • Stemming
  • Sentence segmentation
  • Word segmentation
Automatic summarization
  • Multi-document summarization
  • Sentence extraction
  • Text simplification
Machine translation
  • Computer-assisted
  • Example-based
  • Rule-based
  • Statistical
  • Transfer-based
  • Neural
Distributional semantics models
  • BERT
  • Document-term matrix
  • Explicit semantic analysis
  • fastText
  • GloVe
  • Language model
    • large
    • small
  • Latent semantic analysis
  • Long short-term memory
  • Seq2seq
  • Transformer
  • Word embedding
  • Word2vec
Language resources,
datasets and corpora
Types and
standards
  • Corpus linguistics
  • Lexical resource
  • Linguistic Linked Open Data
  • Machine-readable dictionary
  • Parallel text
  • PropBank
  • Semantic network
  • Simple Knowledge Organization System
  • Speech corpus
  • Text corpus
  • Thesaurus (information retrieval)
  • Treebank
  • Universal Dependencies
Data
  • BabelNet
  • Bank of English
  • DBpedia
  • FrameNet
  • Google Ngram Viewer
  • UBY
  • WordNet
  • Wikidata
Automatic identification
and data capture
  • Speech recognition
  • Speech segmentation
  • Speech synthesis
  • Natural language generation
  • Optical character recognition
Topic model
  • Document classification
  • Latent Dirichlet allocation
  • Pachinko allocation
Computer-assisted
reviewing
  • Automated essay scoring
  • Concordancer
  • Grammar checker
  • Predictive text
  • Pronunciation assessment
  • Spell checker
Natural language
user interface
  • Chatbot
  • Interactive fiction
  • Question answering
  • Virtual assistant
  • Voice user interface
Related
  • Formal semantics
  • Hallucination
  • Natural Language Toolkit
  • spaCy
Portals:
  • icon Computer programming
  • icon Technology
Generative pre-trained transformer at Wikipedia's sister projects:
  • Data from Wikidata
Retrieved from "https://teknopedia.ac.id/w/index.php?title=Generative_pre-trained_transformer&oldid=1340643958"
Categories:
  • Large language models
  • Generative artificial intelligence
  • Artificial neural networks
  • Generative pre-trained transformers
  • OpenAI
  • 2018 in artificial intelligence
Hidden categories:
  • CS1 German-language sources (de)
  • Articles with short description
  • Short description matches Wikidata
  • Use mdy dates from April 2025
  • Use American English from April 2025
  • All Wikipedia articles written in American English
  • All articles with unsourced statements
  • Articles with unsourced statements from January 2026

  • indonesia
  • Polski
  • العربية
  • Deutsch
  • English
  • Español
  • Français
  • Italiano
  • مصرى
  • Nederlands
  • 日本語
  • Português
  • Sinugboanong Binisaya
  • Svenska
  • Українська
  • Tiếng Việt
  • Winaray
  • 中文
  • Русский
Sunting pranala
url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url url
Pusat Layanan

UNIVERSITAS TEKNOKRAT INDONESIA | ASEAN's Best Private University
Jl. ZA. Pagar Alam No.9 -11, Labuhan Ratu, Kec. Kedaton, Kota Bandar Lampung, Lampung 35132
Phone: (0721) 702022
Email: pmb@teknokrat.ac.id