AI Tools.

Search

fill mask

bert-base-cased

Google's BERT base model in cased form, pre-trained on BookCorpus and English Wikipedia with original case preserved. Unlike bert-base-uncased, this model maintains distinctions between 'bert' and 'BERT' — essential for tasks where capitalization carries semantic information, such as named entity recognition. Same architecture as bert-base-uncased but with case-sensitive tokenization.

Last reviewed

Use cases

  • Named entity recognition where proper noun capitalization is a useful signal
  • Text classification tasks where case provides meaningful information
  • Sentence encoding with case sensitivity for downstream NLP models
  • Fine-tuning for sentiment or topic classification on formally written text
  • Transfer learning base when case-insensitive BERT produces errors on proper nouns

Pros

  • Case-sensitive tokenization preserves capitalization as a NER signal
  • Multi-framework support: PyTorch, TF, JAX, CoreML, ONNX, Rust
  • Apache 2.0 license; large ecosystem of cased fine-tuned checkpoints
  • Well-understood behavior from extensive NLP literature

Cons

  • Cased tokenization splits text differently than uncased — vocabulary size is larger, slightly slower
  • 512-token context limit for long documents
  • Encoder-only — cannot generate free-form text
  • Outperformed by RoBERTa, DeBERTa, and newer encoders on most classification and NER tasks
  • Cased benefit is task-dependent — evaluate whether capitalization actually improves your specific task

FAQ

What is bert-base-cased used for?

Named entity recognition where proper noun capitalization is a useful signal. Text classification tasks where case provides meaningful information. Sentence encoding with case sensitivity for downstream NLP models. Fine-tuning for sentiment or topic classification on formally written text. Transfer learning base when case-insensitive BERT produces errors on proper nouns.

Is bert-base-cased free to use?

bert-base-cased is an open-source model published on HuggingFace. License terms vary by model — check the model card for the specific license.

How do I run bert-base-cased locally?

Most HuggingFace models can be loaded with transformers or the appropriate framework library. See the model card for framework-specific instructions and hardware requirements.

Tags

transformerspytorchtfjaxsafetensorsbertfill-maskexbertendataset:bookcorpusdataset:wikipediaarxiv:1810.04805license:apache-2.0endpoints_compatibledeploy:azureregion:us