Long Short-Term Memory<\/a> (LTSM) network was a major step toward advancing text generation. <\/span><\/p>\nThe ability to store more data for longer further deepened the algorithm\u2019s understanding of words and how they relate to each other.\u00a0<\/span><\/p>\nSo, researchers figured out the way for machines to talk nearly 30 years ago, but only recently has it become possible for chatbots to become as good as they are.\u00a0<\/span><\/p>\nTransformers as the main enabler<\/span><\/h2>\nThe real breakthrough for natural language processing happened in 2017 when Google Brain team developed a new type of deep learning model dubbed simply a transformer. <\/span><\/p>\nIt\u2019s hard to adequately explain the major advantages of transformers without diving into the intricacies of machine learning, but, generally speaking, transformers can process input data all at once instead of sequentially. <\/span><\/p>\nIn other words, RNNs make sense of text word by word, while transformers analyze all text as a single set of data. Transformers also can take into account the location of a word in a sentence, enabling them to decipher the meaning of words more precisely. <\/span><\/p>\nImportantly, compared to RNNs, transformer-based models can be trained significantly faster, allowing for much larger training datasets.\u00a0<\/span><\/p>\nThe GPT Era\u00a0<\/span><\/h2>\nThe first version of Generative Pre-trained Transfomer (GPT-1) was released in 2018 by OpenAI. By augmenting transformers with unsupervised learning, OpenAI managed to create what we now call a large language model. <\/span><\/p>\nThis time, instead of training the model with annotated data, OpenAI allowed the model to detect patterns on its own. Given that data annotation is a very laborious and time-consuming process, the company was able to drastically enlarge the training dataset.\u00a0<\/span><\/p>\nGPT-2 was released just a few months after the first version, and the third iteration followed in 2020. From this point on, it was all about enlarging datasets and increasing the number of model parameters. <\/span><\/p>\nTake note that these improvements were rather rapid than gradual. Just to give an idea, GPT-2 has 1.5 billion parameters that engineers could adjust during training, while GPT-3 has 175 billion.\u00a0<\/span><\/p>\nHowever, due to the fact that GPT-3 was trained on the largest dataset ever using unsupervised learning, the model mirrored all the good and bad that the internet has to offer. <\/span><\/p>\nGPT-3 was exceptionally biased towards many sensitive subjects, including race, sex, and religion. To solve this problem, OpenAI came up with InstructGPT, a way less offensive and opinionated sibling of GPT-3. <\/span><\/p>\nTo achieve this, OpenAI had to bring human judgment back into the equation. The company hired 40 contractors, who were responsible for rating the model\u2019s answers to their liking with the ultimate goal of decreasing the use of toxic language and misinformation.\u00a0<\/span><\/p>\nThe current paid version of ChatGPT uses GPT-4, the most reliable, stable, and creative version of the model. Apart from a larger training dataset, GPT-4 also takes advantage of its ability to process both text and images instead of text only. For example, now you can provide ChatGPT a photo of ingredients you have and it will come up with a recipe.\u00a0<\/span><\/p>\nClosing Thoughts<\/span><\/h2>\nThe development of GPT, as well as natural language processing in general, will continue to open up new avenues for natural language processing. <\/span><\/p>\nFrom creating engaging conversational experiences with chatbots to providing deeper insights into the way we communicate, these advancements are undoubtedly revolutionizing our world in profound ways. It\u2019s safe to say that the chatbot revolution is well and truly underway.\u00a0<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"Released to the public at the end of 2022, OpenAI\u2019s ChatGPT has quickly become the fastest-growing internet application ever and a disruptive mainstream phenomenon. The introduction of ChatGPT and its successors is a beginning of a new era in our relationship with technology.\u00a0 But as we are becoming increasingly reliant on artificial intelligence and machine […]<\/p>\n","protected":false},"author":6,"featured_media":14917,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_seopress_robots_primary_cat":"none","_genesis_hide_title":false,"_genesis_hide_breadcrumbs":false,"_genesis_hide_singular_image":false,"_genesis_hide_footer_widgets":false,"_genesis_custom_body_class":"","_genesis_custom_post_class":"","_genesis_layout":""},"categories":[1],"tags":[],"_links":{"self":[{"href":"14914"}],"collection":[{"href":"../posts"}],"about":[{"href":"../types/post"}],"author":[{"embeddable":true,"href":"../users/6"}],"replies":[{"embeddable":true,"href":"../comments@post=14914"}],"version-history":[{"count":3,"href":"14914/revisions"}],"predecessor-version":[{"id":14920,"href":"14914/revisions/14920"}],"wp:featuredmedia":[{"embeddable":true,"href":"../media/14917"}],"wp:attachment":[{"href":"../media@parent=14914"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"../categories@post=14914"},{"taxonomy":"post_tag","embeddable":true,"href":"../tags@post=14914"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}