🌐
Hugging Face
huggingface.co › dslim › bert-base-NER
dslim/bert-base-NER · Hugging Face
If my open source models have been ... financially). Thanks! bert-base-NER is a fine-tuned BERT model that is ready to use for Named Entity Recognition and achieves state-of-the-art performance for the NER task....
Discussions

Evaluating NER HuggingFace models for a domain
I had a quick google search and didn't find anything intresting to apart from what you have mentioned. I think once you pass on text to BERT it outputs NER tags then you can convert them to your original label format and pass them on to seqeval, just like what you mentioned. kindly share if you found anything interesting apart from this. More on reddit.com
🌐 r/LanguageTechnology
2
3
March 16, 2022
Confused about Huggingface Transformers for NER models
This is the wrapper for token classification. You can seamlessly use it with various pretrained models. https://github.com/huggingface/transformers/blob/master/examples/token-classification/run_ner.py It is not out-of-the-box NER but you can easily train it for NER task using CoNLL data. data_dir: str = field( metadata={"help": "The input data dir. Should contain the .txt files for a CoNLL-2003-formatted task."} More on reddit.com
🌐 r/LanguageTechnology
3
6
September 23, 2020
Named-Entity Recognition of Long Texts Using HuggingFace's "ner" Pipeline
I did exactly this at my last job. We wanted to do NER on documents thats were sometimes 10,000 words long. We wrote a lot of custom code to split the documents into overlapping chunks, perform NER on each chunk, extract terms and labels from each chunk and then stitch the output back together. I tried to push management to let us open source the code but unfortunately to no avail 🥲 More on reddit.com
🌐 r/LanguageTechnology
12
11
August 27, 2021
Training NER models for detecting custom entities

Have you tried Flair or training a custom BERT model with HuggingFace? We also use spaCy's very robust pattern matching rules that allow for regex, POS, etc. Might be worth a try depending on your situation.

More on reddit.com
🌐 r/LanguageTechnology
11
6
December 9, 2020
🌐
Hugging Face
huggingface.co › d4data › biomedical-ner-all
d4data/biomedical-ner-all · Hugging Face
An English Named Entity Recognition model, trained on Maccrobat to recognize the bio-medical entities (107 entities) from a given text corpus (case reports etc.).
🌐
Hugging Face
huggingface.co › learn › llm-course › en › chapter7 › 2
Token classification - Hugging Face LLM Course
This task (which can be combined with POS or NER) can be formulated as attributing one label (usually B-) to any tokens that are at the beginning of a chunk, another label (usually I-) to tokens that are inside a chunk, and a third label (usually O) to tokens that don’t belong to any chunk. Of course, there are many other types of token classification problem; those are just a few representative examples. In this section, we will fine-tune a model (BERT) on a NER task, which will then be able to compute predictions like this one:
🌐
Hugging Face
huggingface.co › dslim › distilbert-NER
dslim/distilbert-NER · Hugging Face
January 18, 2024 - DistilBERT has fewer parameters ... distilbert-NER is specifically fine-tuned for the task of Named Entity Recognition (NER). This model accurately identifies the same four types of entities as its BERT counterparts: location (LOC), organizations (ORG), person (PER), and Miscellaneous (MISC)....
🌐
Medium
medium.com › @anyuanay › working-with-hugging-face-lesson-2-1-71c6e4662479
Named Entity Recognition (NER) Using the Pre-Trained bert-base-NER Model in Hugging Face | by Yuan An, PhD | Medium
October 18, 2023 - Named Entity Recognition (NER) Using the Pre-Trained bert-base-NER Model in Hugging Face This is a series of short tutorials about using Hugging Face. The table of contents is here. In this lesson …
🌐
Reddit
reddit.com › r/languagetechnology › evaluating ner huggingface models for a domain
r/LanguageTechnology on Reddit: Evaluating NER HuggingFace models for a domain
March 16, 2022 -

Hi everyone,

I'm comparing off-the-shelf NER systems to one another to see how they perform on literary-historical data (more specifically: a set of books from the 17th century -19th century). I'm not training or improving the models, but trying to use the ones which are available to see how they perform, to decide if they can later be used in historical research and information extraction contexts.

I think I understand how to evaluate tools such as spaCy and NLTK, by transforming the output labels into the formats required by e.g. the Python packages nervaluate and seqeval. These both return quantitative metrics (F1, precision, recall,...) necessary to evaluate how the models perform on this data type/domain.

I'm not experienced with HuggingFace/transformers and it's quite hard to find sources that have done this before (or is it just me?). I'm wondering if there's an "elegant" way to evaluate these models for a domain. Does it make sense to do it the same way as I evaluated spaCy and NLTK, as specified below?

At my disposal: a small gold standard dataset labelled with the "location" entity (IOB2).

Steps:

  1. align tokenizations and labels of the gold standard and the model output to create two lists of equal length (using the package pytokenizations).

  2. map labels (e.g.: spaCy's "LOC" & "GPE" become "LOCATION", as in the gold standard data).

  3. calculate metrics using nervaluate/seqeval.

It seems so convoluted to apply this methodology, but I haven't been able to find a better way. Am I overlooking or not grasping something? Is there an amazing evaluation package or research on evaluation methods of transformers which I don't know about?

Thank you for your help!

Find elsewhere
🌐
DevGenius
blog.devgenius.io › exploring-hugging-face-ner-76eca79d5a8d
Exploring Hugging Face: NER. Named Entity Recognition | by Okan Yenigün | Dev Genius
December 20, 2024 - from transformers import AutoTokenizer, AutoModelForTokenClassification from transformers import pipeline tokenizer = AutoTokenizer.from_pretrained("dslim/bert-base-NER") model = AutoModelForTokenClassification.from_pretrained("dslim/bert-base-NER") nlp = pipeline("ner", model=model, tokenizer=tokenizer) example = "My name is Wolfgang and I live in Berlin" ner_results = nlp(example) print(ner_results) """ [{'entity': 'B-PER', 'score': 0.9990139, 'index': 4, 'word': 'Wolfgang', 'start': 11, 'end': 19}, {'entity': 'B-LOC', 'score': 0.999645, 'index': 9, 'word': 'Berlin', 'start': 34, 'end': 40}]
🌐
KDnuggets
kdnuggets.com › implement-named-entity-recognition-with-hugging-face-transformers
How to Implement Named Entity Recognition with Hugging Face Transformers - KDnuggets
November 20, 2024 - Hugging Face offers a range of pre-trained models suitable for NER, and for this tutorial we will use the dbmdz/bert-large-cased-finetuned-conll03-english model, which has been fine-tuned on the CoNLL-03 dataset for English NER tasks.
🌐
Hugging Face
huggingface.co › papers › 2311.08526
Paper page - GLiNER: Generalist Model for Named Entity Recognition using Bidirectional Transformer
However, their size and cost, particularly for those accessed via APIs like ChatGPT, make them impractical in resource-limited scenarios. In this paper, we introduce a compact NER model trained to identify any type of entity. Leveraging a bidirectional transformer encoder, our model, GLiNER, facilitates parallel entity extraction, an advantage over the slow sequential token generation of LLMs.
🌐
Hugging Face
huggingface.co › docs › transformers › main › tasks › token_classification
Token classification
NER attempts to find a label for each entity in a sentence, such as a person, location, or organization. ... Finetune DistilBERT on the WNUT 17 dataset to detect new entities. Use your finetuned model for inference.
🌐
GreenFlux Blog
blog.greenflux.us › named-entity-recognition-with-bert-and-hugging-face
Named Entity Recognition with BERT and Hugging Face 🤗 - GreenFlux Blog
March 17, 2025 - Named Entity Recognition has a wide range of use cases, from tagging files and creating metadata, to automations and knowledge graphs for retrieval-augmented generation. The BERT series of models excels at NER, and Hugging Face makes it easy to integrate it into any app.
🌐
Hugging Face
huggingface.co › blog › minibase-ai › named-entity-recognition
A lightweight model for Named Entity Recognition (NER)
October 15, 2025 - TL;DR: We’re releasing compact models for Named Entity Recognition (NER). These model can run locally on a CPU and quickly identifies people, organizations, and locations with near-perfect recall. There is a Standard and Small version.
🌐
freeCodeCamp
freecodecamp.org › news › getting-started-with-ner-models-using-huggingface
How to Fine-Tune BERT for NER Using HuggingFace
January 31, 2022 - how to fine-tune BERT for NER tasks using HuggingFace ... I was able to create this model as a side project and share it at https://huggingface.co/Suchandra/bengali_language_NER, thanks to the wonderful resources which I am linking below:
🌐
Kaggle
kaggle.com › code › tirendazacademy › ner-with-huggingface
NER with HuggingFace
Checking your browser before accessing www.kaggle.com · Click here if you are not automatically redirected after 5 seconds
🌐
Galileo AI
galileo.ai › blog › understanding bert with huggingface transformers ner
Understanding BERT with Huggingface Transformers NER
February 1, 2023 - In this article, we covered how to fine-tune a model for NER tasks using the powerful HuggingFace library and then use Galileo to inspect the quality of the model and dataset.
🌐
Label Studio
labelstud.io › guide › ml_tutorials › huggingface_ner
Label Studio Documentation — Label Studio tutorial to run Hugging Face NER backend
This project uses a custom machine learning backend model for Named Entity Recognition (NER) with Hugging Face’s transformers and Label Studio.