Tillbaka till alla evenemang

Inside GPT – Large Language Models Demystified

Our colleague Alan Smith will show us how chat GPT really works.

Natural language processing using generative pre-trained transformers (GPT) algorithms is a rapidly evolving field that offers many opportunities and challenges for application developers. But what is a generative pre-trained transformer, and how does it work? How can you leverage the latest advances in GPT algorithms to create engaging and useful applications? Can my business benefit from creating a GPT powered chat bot?

In this demo intensive session Alan will take a deep dive into the architecture of GPT algorithms and the inner workings of ChatGPT. The journey will begin by looking at the fundamental concepts of natural language processing, such as word embedding, vectorization and tokenization. He will then demonstrate how you can apply these techniques to train a GPT2 model that can generate song lyrics, showing the internals of how word sequences are predicted.

Alan will then shift the focus to larger language models, such as ChatGPT and GPT4, demonstrating their power, capabilities, and limitations. The use of hyperparameters such as temperature and frequency penalty will be explained and their effect on the generated output demonstrated. He will then cover the concepts of prompt engineering and demonstrate how Retrieval Augmented Generation (RAG) patterns can be leveraged to create a ChatGPT experience based on your own textual data.

Join Alan for this session if you want to learn how to harness the power of GPT algorithms in your own solutions.

Are you interested in this and other sessions? Contact us to become an Active Partner!

Föregående
Föregående
31 oktober

Digitization of Historical Text - Erik Lenas

Nästa
Nästa
14 november

Update Conference Prague 2024