AI Tools.

Search

fill mask

distilbert-base-uncased

DistilBERT-base-uncased is a distilled version of BERT-base-uncased, 40% smaller and 60% faster while retaining approximately 97% of BERT's language understanding performance on the GLUE benchmark. Trained via knowledge distillation from BERT using BookCorpus and Wikipedia. Commonly used when BERT's performance is needed but inference speed or resource constraints are limiting factors.

Last reviewed

Use cases

  • Text classification in latency-constrained environments (sentiment, intent)
  • NER where BERT-level performance is needed at lower compute cost
  • Extractive QA on shorter passages with faster inference requirement
  • Edge deployment where BERT-base is too large
  • High-throughput classification pipelines where latency per request matters

Pros

  • 40% smaller and 60% faster than BERT-base with ~97% performance retained
  • Multi-framework support (PyTorch, TF, JAX, Rust, ONNX, safetensors)
  • Apache 2.0 license; large ecosystem of fine-tuned checkpoints
  • Lowercase tokenization consistent with BERT-base-uncased fine-tuned models

Cons

  • Performance gap vs. BERT-base grows on more complex NLU tasks
  • Lowercase tokenization cannot distinguish case — limits NER on proper nouns
  • 512-token context limit
  • Encoder-only; cannot generate text
  • Surpassed by more efficient distilled models (MiniLM, TinyBERT) on the speed-accuracy frontier

FAQ

What is distilbert-base-uncased used for?

Text classification in latency-constrained environments (sentiment, intent). NER where BERT-level performance is needed at lower compute cost. Extractive QA on shorter passages with faster inference requirement. Edge deployment where BERT-base is too large. High-throughput classification pipelines where latency per request matters.

Is distilbert-base-uncased free to use?

distilbert-base-uncased is an open-source model published on HuggingFace. License terms vary by model — check the model card for the specific license.

How do I run distilbert-base-uncased locally?

Most HuggingFace models can be loaded with transformers or the appropriate framework library. See the model card for framework-specific instructions and hardware requirements.

Tags

transformerspytorchtfjaxrustsafetensorsdistilbertfill-maskexbertendataset:bookcorpusdataset:wikipediaarxiv:1910.01108license:apache-2.0endpoints_compatibledeploy:azureregion:us