We use cookies

This website uses cookies to provide necessary website functionality, improve your experience and analyze our traffic. By using our website, you agree to our Privacy Policy and our cookies usage.

Contact us

Our friendly team would love to hear from you.




    or contact us directly at office@cognitum.au

    ionicons-v5-e

    Thank you for your interest!

    We will contact you as soon as possible.

    Back to homepage

    Introduction to NLP & LLMs

    Day 1

    Day 2

    Day 3

    Day 4

    Day 5

    Day 1 - Tokenization & Introduction to NLP Models
    Agenda
    • Introduction to computer text representation, specifically in Python – unicode codepoints and encodings
    • Tokenization
      – Defining and exploring different types of tokenization including: word-, character- and most importantly subword-tokenization
      – Introduction to Hugging Face’s transformers library and how it performs tokenization
      – Definition and treatment of special tokens and showcase of different transformers’ tokenizers
      – Training a custom tokenizer with transformers and SentencePiece
    • Language Modeling
      – What neural networks consist of and how they work with text
      – Word embeddings
      – Transformer architecture and self-attention
      – Masked and Causal language modeling
    • Hugging Face Ecosystem
      – How to work with the entirety of functionalities that HF
    Description

    This day is a general introduction to NLP and the context in which its tasks are performed. The main goal is to discover steps that should be performed before feeding the data to the neural network, talk about how AI models work with textual data, and introduce the participants to the Hugging Face ecosystem and their libraries which are a staple of NLP workflows.

    Level

    Beginner

    This day’s course can be divided into two parts. The introduction to the various tools and Hugging Face ecosystem, that are more of a high-level demonstration than a challenge, and the more advanced sections such as the Tokenization, Training custom Tokenizers, Word Embeddings, and Transformer Architecture, some of them with more theory than others. All newly introduced terms and techniques are explained from scratch, thus leaving no one in the dark.

    Target Participants
    programmer

    The beginner audience, with little to no NLP experience, is best suited for this day, since during the course its participants get to know a little bit of everything from the tokenization, through the transformers library and its use cases to the whole Hugging Face ecosystem.

    However, even the advanced users, already acquainted with the presented tools may find some of the more theoretical sections interesting and valuable, since an understanding of them is vital for successful participation in the following days of the course.

    Day 2 - Text Classification
    Agenda
    • Working closer with the datasets
      – Learn how to split the data into training, validation and test splits to then use them in training and evaluation.
    • Getting to know the most important metrics for binary classification
      – We evaluate our models with such metrics as Accuracy, Precision and Recall, while learning what they bring to the table when it comes to gauging model’s performance and dealing with imbalanced datasets.
    • Learning about different ways of text classification – we’ll go through several methods used for text classification from the most basic and classical ones to the newest and best solutions including:
      – Masked language modeling
      – Zero-shot classification via Natural Language Inference
      – Logistic Regression
      – Multi-layer perceptron
      – Fine-tuning a pre-trained transformer model
      – Explainability of text classification

    Throughout the various stages of the workshop we will be introducing ways to deal with the imbalanced datasets, whether it’s during the training or evaluation.

    Description

    On this day we take another step into the world of NLP, focusing on one of the most versatile tasks in the field – text classification. We learn about the metrics with which you can measure a model’s performance, discover ways to work with imbalanced datasets and most importantly explore different ways to classify text.

    Level

    Intermediate

    The difficulty for this part of the course varies between medium (for the sections like the introduction of metrics, Masked Language Modeling, or working with datasets) and hard (like managing class imbalance, usage of SHAP, and especially working with Torch framework to implement the MLP).

    Target Participants
    programmer

    In this part of the course we expect the more advanced participants to thrive, as we introduce more complex ways to work with the neural networks, including devising network architecture with PyTorch.

    Less experienced users should also find a lot of interesting parts, such as new metrics, fine-tuning of models with transformers API, or working with the datasets.

    Day 3 - Token Classification
    Agenda
    • Discussion of various real-world applications of Token Classification and demo of keyphrase extraction
    • Part of Speech Tagging
      – Quick revision of PoS tagging task,
      – Data Preprocessing: preparing data for a basic Token Classification problem,
      – Model Training,
      – Evaluation.
    • Named Entity Recognition (NER)
      IOB Input Data Format,
      – Data Preprocessing: tackling a more complex Token Classification problem,
      – Model Training,
      – Monitoring the training using TensorBoard
      – Evaluation using seqeval,
      – NER Visualization using transformers pipelines and displacy.
    • Optional task
      – Advanced data preprocessing: Transforming an unstructured dataset to IOB format.
    Description

    In this workshop, we dive deep into various Token Classification problems, highlighting their main challenges, potential pitfalls, and strategies to overcome them. We’ll cover preprocessing techniques for data, and methods to evaluate results tailored to specific tasks, including leveraging several new libraries.

    Level

    Intermediate

    This particular workshop’s difficulty can safely be evaluated as medium as it mostly leverages what the participants should already be familiar with from the transformers library and pure Python, while providing more details on the theoretical side of things.

    Target Participants
    programmer

    We expect this part of the course to appeal to both the beginner and the advanced participants, as we go deeper into possible use-cases of NLP and discover another of its tasks, while mostly using the transformers API that was described in detail in the previous parts of the course.

    To those more interested in the technicalities of model training and evaluation we introduce the usage of monitoring tools and a new evaluation library.

    Day 4 - Sequence-to-sequence models
    Agenda
    • Examples of sequence-to-sequence problems
    • Full encoder-decoder architecture
    • Metrics used for sequence-to-sequence problems
      – BLEU,
      – ROUGE,
      – BERTScore,
      – SemScore.
    • Sequence-to-sequence models as universal NLP models
    • Text generation algorithms
      – greedy decoding
      – beam search
      – output sampling
      – computational complexity
    • Fine-tuning sequence-to-sequence models for
      – question answering
      – dialog summarization
    • Considerations regarding input/output lengths.
    Description

    Continuing our journey through the NLP field with the use of the transformer models. This time we focus on the variants that leverage the full transformer architecture, also known as seq2seq models. We will explore two of the tasks those models thrive in: question answering and text summarization, learn how to measure the quality of text generated by a model and try to create a multi-task model able to perform both aforementioned tasks.

    Level

    Intermediate

    Considering the similarity to the previous workshop, this one is also of moderate difficulty. Just like the former, it contains significant theoretical content, while also touching upon important technical aspects of the systems solving those tasks, once again relying heavily on the transformers library.

    Target Participants
    programmer

    As mentioned in the ‘level’ section, due to similarities to the day 3’s workshop, we expect this part of the course to be enjoyable and valuable for participants on all levels of the knowledge tree.

    While keeping the formula similar to the previous workshop, we provide new information by introducing a whole new set of NLP problems, classical and innovative metrics to evaluate the performance of models solving those tasks and methods of text generation, important for yet another architecture variant of Transformers.

    Day 5 - Introduction to LLMs
    Agenda
    • Introduction to using LLMs with transformers
    • Definition and showcase of zero-shot and few-shot learning techniques
    • Comparison of performance of the LLM and dedicated smaller models on tasks from previous days
      – Sequence Classification
      – Named Entity Recognition
      – Summarization
    • Few-shot learning in detail
      – How it improves model performance
      – Its effect on the technical side of the system: inference time, sequence length and memory footprint
    • Text generation strategies in LLMs
    Description

    Venturing away from two previous days, we introduce the participants to the concept of LLMs, prompt engineering, as well as the zero-shot and few-shot learning techniques, translating from fine-tuning a smaller, dedicated model for each task to using one LLM with fitting prompts for all of them.

    We compare results achieved by those two approaches on various tasks from previous days and explore newly introduced aspects regarding LLMs.

    Level

    Beginner

    This workshop acts as an introduction to the wide field of LLMs, thus its difficulty level is low. We aim to make the transition from the smaller models to LLMs as gentle and easy as possible, slowly, but surely building the foundation for the more advanced techniques and aspects of leveraging the biggest neural networks.

    Target Participants
    manager

    While we hope that everyone could gain some useful knowledge from this part of the course, we have to admit that it is mostly aimed at those who, have barely touched upon the Large Language Models in general, as this workshop takes the participants a step lower from using the UI of LLMs available online to being able to tweak model’s hyperparameters and try some new, albeit simple, prompting techniques.

    Get Your Personalized
    LLMs Course Quote

    Interested in mastering Large Language Models? Fill out the form below to receive a tailored quotation for a course designed to meet your specific needs and objectives.





      ionicons-v5-e

      Thank you for your interest!

      We will contact you as soon as possible.

      Back to homepage

      Your certified partner!

      Empower your projects with Cognitum, backed by the assurance of our ISO 27001 and ISO 9001 certifications, symbolizing elite data security and quality standards.