Skip to course content
Elephant Scale

Deep Learning with Transformers (Hugging Face)

Understand state‑of‑the‑art Deep Learning & transformers; fine‑tune, evaluate, and deploy models using Hugging Face libraries and Hub.

Get Course Info

Audience: Developers / Data Scientists / Team Leads / PMs

Duration: 2 days

Format: Lectures and hands‑on labs (50% lecture, 50% lab)

Overview

Deep Learning delivers near‑miraculous results; this course focuses on leveraging existing transformers and models via Hugging Face rather than reinventing them. Students will gain practical skills in transfer learning, fine‑tuning, and deploying transformer‑based solutions for NLP, vision, and more.

Objective

Understand state‑of‑the‑art Deep Learning & transformers; fine‑tune, evaluate, and deploy models using Hugging Face libraries and Hub.

What You Will Learn

  • Classifying sentences & tokens (sentiment, spam, NER)
  • Generating & completing text
  • Extractive QA, translation, summarisation
  • Image & audio tasks
  • Transformers architecture: encoders, decoders, seq‑to‑seq
  • Fine‑tuning with Trainer API or Keras
  • Sharing models & tokenisers on Hugging Face Hub

Course Details

Audience: Developers / Data Scientists / Team Leads / PMs

Duration: 2 days

Format: Lectures and hands‑on labs (50% lecture, 50% lab)

Prerequisites:

General familiarity with machine learning

Setup: Zero‑install cloud lab • Modern laptop • Chrome browser

Detailed Outline

  • DL use‑cases, AI/ML/DL definitions, data & AI vocabulary
  • Hardware & software ecosystem, types of ML
  • CNN intro, architecture & concepts
  • Lab: image recognition with CNNs
  • RNN & LSTM intro, architecture & concepts
  • Lab: RNNs for text & sequence prediction
  • Transformers, encoders, decoders
  • Sequence‑to‑sequence, bias & limitations
  • Pipelines, models, tokenisers
  • Putting it all together
  • Data processing
  • Trainer API & Keras workflows
  • End‑to‑end fine‑tuning
  • Hugging Face Hub
  • Using & sharing pre‑trained models
  • Token classification, masked LM, translation, summarisation, causal LM
  • Question answering

Ready to Get Started?

Contact us to learn more about this course and schedule your training.