Reference

Use this as a fast refresher after reading the tutorial.

Glossary

TermMeaning
HubCentral repository platform for models, datasets, and Spaces
Model cardREADME-like documentation for a model
Dataset cardDocumentation for a dataset
SpaceHosted demo app
PipelineHigh-level inference wrapper in transformers
TokenizerConverts raw input into model-readable tokens
CheckpointSaved model weights/state
PEFTParameter-efficient fine-tuning
LoRAA common PEFT method for adapting large models cheaply
RevisionA specific version of a repo, branch, tag, or commit
Gated modelRequires approval or terms acceptance before access

What to Learn First

If you are new, use this order:

  1. Learn what the ecosystem contains
  2. Browse models and read model cards
  3. Run a pipeline
  4. Learn AutoTokenizer and auto model classes
  5. Load datasets and compute metrics
  6. Fine-tune only when needed
  7. Deploy with a Space or endpoint

Fast Decision Guide

If You Need To...Start Here
Browse available modelsHub search and model cards
Run something in 5 minutespipeline()
Build a proper Python workflowtransformers + datasets
Adapt a model cheaplypeft / LoRA
Share an interactive demoSpace
Put a model behind an APIEndpoint or self-hosted service
Reuse files programmaticallyhuggingface_hub

Common Commands

# Install core packages
pip install -U transformers datasets tokenizers evaluate accelerate peft huggingface_hub

# Log in
hf auth login

# Download a file from a repo
hf download distilbert/distilbert-base-uncased config.json

Common Python Patterns

from transformers import pipeline
classifier = pipeline("sentiment-analysis")
from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained("distilbert/distilbert-base-uncased-finetuned-sst-2-english")
model = AutoModelForSequenceClassification.from_pretrained("distilbert/distilbert-base-uncased-finetuned-sst-2-english")
from datasets import load_dataset
dataset = load_dataset("imdb")

Common Mistakes

  • Choosing a model without reading the license
  • Confusing a good demo with production readiness
  • Ignoring tokenizer/model compatibility
  • Skipping evaluation on real data
  • Using fine-tuning when prompting or smaller models would suffice
  • Forgetting to pin revisions

A Sensible Learning Roadmap

Day 1

  • Read chapters 01-03
  • Browse the Hub and shortlist a few interesting repos
  • Run one pipeline locally

Day 2

  • Read chapters 04-05
  • Load one dataset and one model in Python
  • Compare two models on a few real inputs

Day 3+

  • Read chapters 06-09
  • Fine-tune only if you have a clear use case
  • Publish a small demo or internal proof of concept