Talks


LLMs Then and Now: Navigating the Architectural Timeline of LLMs

Sep 19th, in San Francisco, CA at PyTorch Conference 2024.

In this talk, you’ll learn about the architectural difference between the original GPT model and the various Llama models. Moreover, you’ll also learn about new pre-training recipes used for Qwen 2, Gemma 2, Apple’s Foundation Models, and Llama 3, as well as some of the efficiency tweaks introduced by Mixtral, Llama 3, and Gemma 2.

You can find more information about the talk on the PyTorch Conference 2024 website here.





Pretraining and Finetuning LLMs from the Ground Up

July 8th, in Tacoma, WA at SciPy 2024.

This tutorial is aimed at coders interested in understanding the building blocks of large language models (LLMs), how LLMs work, and how to code them from the ground up in PyTorch. We will kick off this tutorial with an introduction to LLMs, recent milestones, and their use cases. Then, we will code a small GPT-like LLM, including its data input pipeline, core architecture components, and pretraining code ourselves. After understanding how everything fits together and how to pretrain an LLM, we will learn how to load pretrained weights and finetune LLMs using open-source libraries.

You can find more information about the talk on the SciPy 2024 website here.





Understanding the LLM Development Cycle: Building, Training, and Finetuning

June 5th, ACM Tech Talk.

This talk will guide you through the key stages of developing large language models (LLMs), from initial coding to deployment. We will start by explaining how these models are built, including the coding of their architectures. Next, we will discuss the processes of pre-training and finetuning, showing what these stages involve and why they are important. Throughout the talk, we will provide real examples and encourage questions, making this a practical and interactive session for anyone interested in how LLMs are created and used.

You can register for this virtual talk for free here.





The Fundamentals of Modern Deep Learning with PyTorch

May 15th, 2024 in Pitsburgh at the PyCon 2024.

I’ll be giving a 3.5 hour deep learning workshop at PyCon 2024 in May. This tutorial is aimed at Python

programmers new to PyTorch and deep learning. However, even more experienced deep learning practitioners and PyTorch users may be exposed to new concepts and ideas when exploring other open source libraries to extend PyTorch.

It’s my first PyCon, and I’m very excited!





LoRA in Action: Insights from Finetuning LLMs with Low-Rank Adaptation

Fri 14 Dec, 2023 in New Orleans at the NeurIPS 2023 Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day workshop.

Low-rank adaptation (LoRA) stands as one of the most popular and effective methods for efficiently training custom Large Language Models (LLMs). As practitioners of open-source LLMs, we regard LoRA as a crucial technique in our toolkit.

In this talk, I will delve into some practical insights gained from running hundreds of experiments with LoRA, addressing questions such as: How much can I save with quantized LoRA? Are Adam optimizers memory-intensive? Should we train for multiple epochs? How do we choose the LoRA rank?

Moreover, the talk will include ideas for future experiments and talking points to stimulate discussions in the workshop, such as mechanisms to avoid overfitting in LoRA and strategies for combining LoRA weights from multiple experiments.





Using Open-Source LLMs and Parameter-Efficient Finetuning Techniques

Fri 13 Oct, 2023 at Packt’s Put GenAI to Work virtual conference.

  • This session will teach you how to utilize the latest open-source Large Language Models (LLMs), such as Falcon, Llama 2, and Code Llama, on your own GPU-supported hardware.
  • We will also cover parameter-efficient finetuning techniques like low-rank adaptation (LoRA) and QLoRA to optimize these LLMs. Moreover, you will learn how to prepare and formate new datasets for finetuning these open-source models.




Finetuning Open-Source LLMs

Tue 3 Oct, 2023 at MLOps.community’s virtual LLMs in Production conference.

This conference keynote talk offers a quick dive into the world of finetuning Large Language Models (LLMs), covering

  • common usage scenarios for pretrained LLMs
  • parameter-efficient finetuning
  • a hands-on guide to using the ‘Lit-GPT’ open-source repository for LLM finetuning




Introduction to Machine Learning and Deep Learning with Python

Mon 18 Sep, 2023 at posit::conf 2023 in Chicago.

This 8-hour workshop introduced participants to the machine and deep learning fundamentals using a modern open-source stack.

The first part started with a brief introduction to Python’s scientific computing libraries, including NumPy, Pandas, and Matplotlib, which provide the foundation for data analysis and visualization. From there, we dove into the scikit-learn API, a user-friendly, open-source library for machine learning in Python.

In the second part of this workshop, we covered deep learning concepts and introduced PyTorch, the most widely used deep learning research library. Next, we trained multi-layer neural networks efficiently using multi-GPU and mixed-precision techniques with Lightning Fabric. Finally, we will explored how to organize PyTorch code with PyTorch Lightning and finetune vision transformers and large language models.




LLMs for Everything and Everyone

Tue 5 Sep, 2023 at SDSC 2023 in Atlanta, Georgia.

Recording of the talk

This invited keynote talk covered the LLM training cycle, explaining how to train modern LLMs. The main portion of the talk was focused on how individuals with limited resources and big companies with larger budgets can use and customize LLMs.

In this talk, I also touched on the recent research advances and what to expect next for the next generation of LLMs.




Modern Deep Learning with PyTorch

Mon 10 Jul, 2023 at SciPy 2023 in Austin, Texas.

A 4-hour workshop at SciPy 2023 covering the following topics:

  1. Introduction to Deep Learning
  2. Understanding the PyTorch API
  3. Training Deep Neural Networks
  4. Accelerating PyTorch Model Training
  5. Organizing PyTorch Code
  6. More Tips and Techniques
  7. Finetuning LLMs

I have a GitHub repository with all the materials here and I believe it was recorded, so I hope it will be publicly available at some point!





Scaling PyTorch Model Training With Minimal Code Changes

Wed 21 Jun, 2023 in Vancouver at CVPR 2023.

In this tutorial, I will show you how to accelerate the training of LLMs and Vision Transformers with minimal code changes using open-source libraries. The code materials are available at https://github.com/rasbt/cvpr2023, and I also have a write-up of the talk here.





How To Triple Your Model’s Inference Speed

Febuary 2023, virtual event.

Webinar on tripling the inference speed of stable diffusion and other generative AI models.





Generative AI and Large Language Models

January, 2023, virtual event at the StateOfTheArt() conference.

StateOfTheArt() 2023 conference with me giving the opening keynote on generative AI and large language models with an open Q&A session.





Using Deep Learning When Class Labels Have A Natural Order – Predicting Ratings and Rankings using PyTorch Lightning July 2022 at SciPy 2022 in Austin, Texas. [Slides] [Code] [Recording]

Many real-world prediction problems come with ordered labels, for example, customer satisfaction (dissatisfied, neutral, satisfied), disease severity (none, moderate, severe), and many more. Sure, we can use conventional machine learning and deep learning classifier on such problems, but you will see why that’s not ideal. Moreover, you will learn how we can modify any deep neural network classifier such that it takes the ordering information into account – it requires changing less than five lines of code!



Introduction to PyTorch and Scaling PyTorch Code with LightningLite May 2022. Invited virtual talk at Data Umbrella [Slides] [Recording Part 1: PyTorch] [Recording Part 2: LightningLite] [Recording Part 3: Q&A]

This talk will introduce attendees to using PyTorch for deep learning. We will start by covering PyTorch from the ground up and learn how it can be both powerful and convenient. At times, Machine learning models can become so large that they can’t be trained on a notebook anymore. Being able to take advantage AI-optimized accelerators such as GPU or TPU and scaling the training of models to hundreds of these devices is essential to the researcher and data scientist. However, adding support for one or several of these in the source code can be complex, time consuming and error-prone. What starts as a fun research project ends up being an engineering problem with hard to debug code. This talk will introduce LightningLite, an open source library that removes this burden completely. You will learn how you can accelerate your PyTorch training script in just under ten lines of code to take advantage of multi-GPU, TPU, multi-node, mixed-precision training and more.



Customer Ratings, Letter Grades, and Other Rankings: Using Deep Learning When Class Labels Have A Natural Order February 2022. Invited talk at ReWork Deep Learning Summit, San Francisco [Slides] [Recording] [Alt. Recording]

Deep learning offers state-of-the-art results for classifying images and text. Common deep learning architectures and training procedures focus on predicting unordered categories, such as recognizing a positive and negative sentiment from written text or indicating whether images contain cats, dogs, or airplanes. However, in many real-world problems, we deal with prediction problems where the target variable has an intrinsic ordering. For example, think of customer ratings (e.g., 1 to 5 stars) or medical diagnoses (e.g., disease severity labels such as none, mild, moderate, and severe). This talk will describe the core concepts behind working with ordered class labels, so-called ordinal data. We will cover hands-on PyTorch examples showing how to take existing deep learning architectures for classification and outfit them with loss functions better suited for ordinal data while only making minimal changes to the core architecture.



Transformers from the Ground Up August 2021. Invited talk at PyData Jeddah [Slides] [Code] [Video]

In recent years deep learning-based transformer models have revolutionized natural language processing. First of all, this talk will explain how transformers work. Then, we will examine some popular transformers like GPT and BERT and understand how they differ. Equipped with this understanding, we will then take a look at how we can fine-tune a BERT model for sentiment classification in Python.



Introduction to Generative Adversarial Networks July 2021. Invited lecture at the International Summer School on Deep Learning, Gdansk [Slides] [Code]

This lecture introduces the main concepts behind Generative Adversarial Networks (GANs) and explains the main ideas behind the objective function for optimizing the generator and discriminator subnetworks. Hands-on examples include GANs for handwrittten digit and face generation, implemented in PyTorch. Lastly, this talks summarizes some of the main milestone GAN architectures that emerged in recent years.



Designing Generative Adversarial Networks for Privacy-enhanced Face Recognition July 2021. Invited keynote talk at the 14th International Conference on Human System Interaction (HSI 2021) [Slides] [Video]

After introducing the main concepts behind face recognition and soft-biometric attribute mining (i.e., the extraction of information such as age, gender, race, health information, and others), this talk discusses different methods for hiding soft-biometric information from facial recognition systems. After introducing the main methodologies, the talk focuses on the PrivacyNet architecture, which is a GAN-based approach to collective and selective facial privacy.



Modern Machine Learning: An Introduction to the Latest Techniques April 2021. Invited talk at the Chan Zuckerberg Initiative’s Seed Network Computational Biology conference. [Slides]

This talk aims to introduce and explain the most relevant methods for predictive modeling with large and complicated data. The first part discusses the established and proven methods for image, text, and tabular data based on machine learning and deep learning. The second part surveys the most recent research directions and promising methods for building the next-generation of deep learning-based predictive models.



Machine Learning in Python: Recent Trends, Technologies, and Challenges March 2021. Invited talk at ODSC East 2021. [Slides]

The first part of this talk covers the recent advancement in hardware and software for deep learning. The second part focuses on the current challenges in deep learning as well as the recent research progress.



Machine Learning in Python: Recent Trends, Technologies, and Challenges April 2020. Invited talk at the Port Harcourt School of AI in Nigeria. [Slides]

This talk focuses on the recent trends in machine learning, including hardware and software. A particular focus is on the Python scientific computing ecosystem and contemporary deep learning libraries.



Machine Learning and Artificial Intelligence in 2020: Recent Trends, Technologies, and Challenges March 2020. Invited talk at ODSC East 2020 in Boston. [Slides]

Not even a decade ago, machine learning was a profession for an elite few. Nowadays, we don’t have to be math or engineering wizards to implement state-of-the-art predictive models. Advances in computing hardware, and especially the utilization of GPUs for training deep neural networks, make it feasible to develop predictive models that achieve human-level performance in various natural language processing and image recognition challenges. The manifold software layers and APIs that are allowing us to utilize these hardware resources are becoming ever so convenient. In this talk, I will highlight the research and technology advances and trends of the last year(s), concerning GPU-accelerated machine learning and deep learning, and focusing on the most profound hardware and software paradigms that have enabled it.



Machine-Learning & AI-based Approaches for GPCR Bioactive Ligand Discovery September 2019. Invited talk at the Cambridge Healthtech Institute’s 14th Annual GPCR-Based Drug Discovery — Discovery on Target Conference, Boston, MA. [Slides]

This talk provides an overview of the latest advances for automating the discovery of bioactive ligands using machine learning. Specific applications are covering the discovery of a strong GPCR pheromone inhibitor as well as predictive models utilizing flexible and rigid rigid states of GPCRs to predict wether GPCR are in active and inactives states. Lastly, the talk concludes with the recent developments in deep learning that are aimed at replacing the need for hand-engineering molecular representations by automatic representation learning using deep learning techniques.



Convolutional Neural Networks for Predicting and Hiding Personal Traits from Face Images. July 2019. Keynote talk at the International Summer School on Deep Learning, Poland, Gdansk. [Slides]

In this talk, I present In recent years, we have developed many July 2019. Deep Learning-centric applications that enhance our everyday life. However, as more and more data is collected and extracted, the protection and respect for users’ privacy have become a big concern. First, this talk will introduce methods for extracting soft-biometric attributes from facial images – soft-biometric characteristics include a person’s age, gender, race, and health status. Then, I will cover methods designed to conceal soft-biometric information to enhance the privacy of users. However, many useful security-related applications rely on face recognition technology for user verification and authentication. Hence, the approaches being presented focus on a dual objective: concealing personal information that can be obtained from face images while preserving the utility of these images for face matching.



Imparting privacy to face images: designing semi-adversarial neural networks for multi-objective optimization tasks April 2018. Invited talk at the TomTom Fest-sponsored Applied Machine Learning Conference. VA, Charlottesville. [Slides]

In this talk, I present a novel neural network architecture, semi-adversarial neural networks that we developed to minimize the performance of once classifier while maximizing the performance of another. This approach is then applied to perturb face images in such as way that biometric face matching is minimally impacted while soft biometric attributes (e.g., gender information) becomes inaccessible as a means to impart privacy to face images.



Machine Learning with Python February 2018. Invited workshop talk at MSU Data Science, East Lansing, MI. [Slides]

This workshop provides an introduction to machine learning covering important concepts regarding regression analysis, classification, and model evaluation as well as the tools being used to conduct experiments. Tool-wise, the focus is on Python’s scikit-learn machine learning library, but TensorFlow and PyTorch are being introduced as well.



Uncovering Hidden Patterns of Molecular Recognition December 2017. Thesis talk at Michigan State University. East Lansing, MI. [Slides]

It happened in 1958 that John Kendrew’s group determined the three-dimensional structure of myoglobin at a resolution of 6 Å. This first view of a protein fold was a breakthrough at that time. Now, more than half a century later, both experimental and computational techniques have substantially improved as well as our understanding of how proteins and ligands interact. Yet, there are many unanswered questions to be addressed and patterns to be uncovered. This talk discusses the analysis of a large dataset of non-homologous proteins bound to their biological ligands, to test a hypothesis that arose from observations made throughout different inhibitor discovery projects: “proteins favor donating H-bonds to ligands and avoid using groups with both H-bond donor and acceptor capacity.”



Building hypothesis-driven virtual screening pipelines for millions of molecules November 2017. Invited talk at ODSC West. San Francisco, CA. [Slides]

In this talk, I will introduce a novel, hypothesis-driven filtering strategy and open-source toolkit for virtual screening, Screenlamp, which I developed and successfully applied to a identify potent inhibitors of G-protein coupled receptor-mediated signaling in vertebrates. And going beyond the mere identification of potent protein inhibitors, I will talk about techniques to integrate the computational predictions with experimental knowledge. Leveraging experimental data, supervised feature selection and extraction techniques will be introduced to identify the discriminants of biological activity using open-source machine learning libraries such as scikit-learn.



An Introduction to Deep Learning with TensorFlow August 2017. Invited talk at PyData Ann Arbor. Ann Arbor, MI. [Slides]

A talk about representing mathematical functions as computation graphs, computing derivatives, and implementing complex deep neural networks conveniently and efficiently using TensorFlow.



Machine Learning and Performance Evaluation November 2016. Invited talk at DataPhilly. Philadelphia, PA. [Slides]

Every day in scientific research and business applications, we rely on statistics and machine learning as support tools for predictive modeling. To satisfy our desire to model uncertainty, to predict trends, and to predict patterns that may occur in the future, we developed a vast library of tools for decision making. In other words, we learned to take advantage of computers to replicate the real world, making intuitive decisions more quantitative, labeling unlabeled data, predicting trends, and ultimately trying to predict the future. Now, whether we are applying predictive modeling techniques to our research or business problems, we want to make “good” predictions!



Detecting the Native Ligand Orientation by Interfacial Rigidity October 2016. Invited Talk at the BMB Department Retreat 2016. East Lansing, MI. [Slides]

Presenting our novel novel approach to protein-ligand docking mode prediction based on graph theory.



Getting Started With Data Science September 2016. Inivited talk at Michigan State Data Science. East Lansing, MI. [Slides]

Introduction to data science at our new MSU data science club.



Learning Scikit-learn – An Introduction to Machine Learning in Python August 2016. Inivited talk at PyData Chigaco 2016. Chicago, IL. [Slides]

Introduction to machine learning at PyData Chicago.



Machine Learning with Scikit-learn July 2016. Machine learning workshop at SciPy 2016. Austin, TX. [Course material]



A Novel Approach to Protein-Ligand Binding Mode Prediction by Rigidity Analysis Using Graph Theory. February 2015. BioMolecular Sciences Gateway. East Lansing, MI. [Slides]



An Introduction to Supervised Machine Learning and Pattern Classification: The Big Picture February 2015. ICER NextGen Bioinformatics Seminars at MSU. East Lansing, MI. [Slides]



MusicMood - Machine Learning for Automatic Music Mood Prediction Based on Song Lyrics December 2014. PSA Group Seminars. East Lansing, MI. [Slides]





Panels

Open-Source LLMs

Sep 2023. Southern Data Science Conference. Atlanta, Georgia

In this panel, I discussed the role and future of open-source LLMs with co-panelist Paige Bailey from Google DeepMind.

SDSC panel



Does Numerical Computing Have An Open Future

July 2022. Anaconda Conference. Austin, Texas.

Was honored to be invited to the open-source and numerical computing panel at Anaconda’s 10 year anniversary event alongside Travis Oliphant, Paige Bailey, and Ryan Abernathy!

[🎥 Video recording on YouTube]

https://youtu.be/9wnW-I5HcSQ



Beyond Large Language Models

June 2022. Lightning DevCon. NYC, New York.

It was a pleasure moderating the research panel on the future of deep learning (beyond large language models) featuring Kyunghyun Cho, Soumith Chintala, and Danqi Chen.





Podcasts

2024

11/21/2024 Build LLMs From Scratch with Sebastian Raschka #52 with Neil Leiser from AI Stories. (Apple Podcasts / Spotify / YouTube)

08/01/2024 On the state of open LLMs, Llama 3.1, and AI education with Nathan Lambert from Interconnects.ai.

05/07/2024. Episode 26: Developing and Training LLMs From Scratch on the Vanishing Gradients podcast with Hugo Bowne-Anderson (Website, Apple Podcasts, Spotify, Stitcher Radio or Overcast). There’s also a YouTube video version here.

03/19/2024. SDS 767: Open-Source LLM Libraries and Techniques, with Dr. Sebastian Raschka on the SuperDataScience podcast with Jon Krohn (Website, Apple Podcasts, Spotify, Stitcher Radio or TuneIn).

03/14/2024. 34 - LLMs and Python with Sebastian Raschka on The Python Show with Mike Driscoll. (Spotify, Substack, Apple Podcasts).

02/07/2024. Leading With Data Ep 22, talking about LLMs, training them from scratch, developing platforms for running experiments, and more!

2023

07/28/2023. Machine Learning Q and AI author interview with Sebastian Raschka

04/13/2023. Interview on large language models (LLMs) on the Soul Searching podcast by Ryan Dsouza.

02/15/2023. Learning from Machine Learning with Seth Levine.

01/30/2023. Infinite Machine Learning with Prateek Yoshi.

2022

The Data Exchange with Ben Lorica [YouTube] [Apple Podcasts] [Spotify]

The Gradient with Daniel Bashir

How AI Happens with Rob Stevenson [Apple Podcasts] [Spotify]

The Machine Learning Podcast with Jay Shah – Making Machine Learning More Accessible [YouTube]

The TWIML AI Podcast with Sam Charrington – Advancing Hands-On Machine Learning Education with Sebastian Raschka [YouTube] [twiml.ai (audio only)]

Before 2022

Deep Tech Musings – #4 Drug Discovery Using ML and AI

The African Data Scientist – #6 - Dr Sebastian Raschka on how Africa can thrive in the age of AI and Data Science, AI and Data Science in the African educational ecosystem, how to write a good book on Machine Learning!

Podcast.__init__ – Teaching Python Machine Learning - Episode 260 (Apr 2020)

Chai Time Data Science – Statistics, Open Source & ML Research, Python for ML, Interview with Sebastian Raschka (Mar 2020) – [Audio] [Video]

Interviews

2023

Before 2023

Mentions

2023

Before 2023