spaCy
spacy.io › usage › spacy-101
spaCy 101: Everything you need to know · spaCy Usage Documentation
Most of the tags and labels look pretty abstract, and they vary between languages. spacy.explain will show you a short description – for example, spacy.explain("VBZ") returns “verb, 3rd person singular present”.
Videos
03:02:33
Natural Language Processing with spaCy & Python - Course for ...
01:27:01
NLP projects with spaCy - YouTube
19:32
Text Analysis with Python: Intro to Spacy - YouTube
Natural Language Processing with spaCy & Python - Course ...
27:56
SPACY v3: State-of-the-art NLP from Prototype to Production - YouTube
38:11
SPACY v3: Custom trainable relation extraction component - YouTube
spaCy
spacy.io › usage › linguistic-features
Linguistic Features · spaCy Usage Documentation
Most of the tags and labels look pretty abstract, and they vary between languages. spacy.explain will show you a short description – for example, spacy.explain("VBZ") returns “verb, 3rd person singular present”.
Real Python
realpython.com › natural-language-processing-spacy-python
Natural Language Processing With spaCy in Python – Real Python
February 1, 2025 - Here’s an example where an ellipsis (...) is used as a delimiter, in addition to the full stop, or period (.): ... >>> ellipsis_text = ( ... "Gus, can you, ... never mind, I forgot" ... " what I was saying. So, do you think" ... " we should ..." ... ) >>> from spacy.language import Language >>> @Language.component("set_custom_boundaries") ...
Factsheet
spaCy
Original author Matthew Honnibal
Developers Explosion AI, various
spaCy
Original author Matthew Honnibal
Developers Explosion AI, various
spaCy
spacy.io
spaCy · Industrial-strength Natural Language Processing in Python
To use it with 'spacy train' # you can run spacy init fill-config to auto-fill all default settings: # python -m spacy init fill-config ./base_config.cfg ./config.cfg [paths] train = null dev = null vectors = null [system] gpu_allocator = null [nlp] lang = "en" pipeline = [] batch_size = 1000 [components] [corpora] [corpora.train] @readers = "spacy.Corpus.v1" path = ${paths.train} max_length = 0 [corpora.dev] @readers = "spacy.Corpus.v1" path = ${paths.dev} max_length = 0 [training] dev_corpus = "corpora.dev" train_corpus = "corpora.train" [training.optimizer] @optimizers = "Adam.v1" [training
spaCy
spacy.io › usage › projects
Projects · spaCy Usage Documentation
If not, spaCy will show an error and the command won’t run. Setting no_skip: true means that the command will always run, even if the dependencies (the trained pipeline) haven’t changed. This makes sense here, because you typically don’t want to skip your tests. Your project commands can include any custom scripts – essentially, anything you can run from the command line. Here’s an example of a custom script that uses typer for quick and easy command-line arguments that you can define via your project.yml:
TutorialsPoint
tutorialspoint.com › spacy › index.htm
spaCy Tutorial
spaCy, developed by software developers Matthew Honnibal and Ines Montani, is an open-source software library for advanced NLP (Natural Language Processing). It is written in Python and Cython (C extension of Python which is mainly designed to give C like performance to the Python language programs)
GitHub
github.com › explosion › spaCy
GitHub - explosion/spaCy: 💫 Industrial-strength Natural Language Processing (NLP) in Python
import spacy import en_core_web_sm nlp = en_core_web_sm.load() doc = nlp("This is a sentence.") 📖 For more info and examples, check out the models documentation.
Starred by 32.9K users
Forked by 4.6K users
Languages Python 54.1% | MDX 31.2% | Cython 10.5% | JavaScript 2.6% | Sass 0.8% | TypeScript 0.4%
Machine Learning Plus
machinelearningplus.com › spacy-tutorial-nlp
spaCy Tutorial - Learn all of spaCy in One Complete Writeup | ML+
May 27, 2021 - NER Application 1: Extracting brand names with Named Entity Recognition 12. NER Application 2: Automatically Masking Entities 13. Rule based Matching Token Matcher Phrase Matcher Entity Ruler 14. Word Vectors and similarity 15. Merging and Splitting Tokens with retokenize 16. spaCy pipelines 17.
spaCy
spacy.io › usage › training
Training Pipelines & Models · spaCy Usage Documentation
However, there can still be scenarios where you may want to override config settings when you run spacy train. This includes file paths to vectors or other resources that shouldn’t be hard-coded in a config file, or system-dependent settings. For cases like this, you can set additional command-line options starting with -- that correspond to the config section and value to override. For example, --paths.train ./corpus/train.spacy sets the train value in the [paths] block.
spaCy
spacy.io › models
Trained Models & Pipelines · spaCy Models Documentation
For example, en_core_web_sm is a small English pipeline trained on written web text (blogs, news, comments), that includes vocabulary, syntax and entities. Additionally, the pipeline package versioning reflects both the compatibility with spaCy, ...
spaCy
spacy.io › api › top-level
Top-level Functions · spaCy API Documentation
Your application or pipeline package can also expose a spacy_displacy_colors entry point to add custom labels and their colors automatically. By default, displaCy links to # for entities without a kb_id set on their span. If you wish to link an entity to their URL then consider using the kb_url_template option from above. For example if the kb_id on a span is Q95 and this is a Wikidata identifier then this option can be set to https://www.wikidata.org/wiki/{}. Clicking on your entity in the rendered HTML should redirect you to their Wikidata page, in this case https://www.wikidata.org/wiki/Q95.
spaCy
spacy.io › usage › processing-pipelines
Language Processing Pipelines · spaCy Usage Documentation
Disabled and excluded component names can be provided to spacy.load as a list. The disable mechanism makes it easy to distribute pipeline packages with optional components that you can enable or disable at runtime. For instance, your pipeline may include a statistical and a rule-based component for sentence segmentation, and you can choose which one to run depending on your use case. For example, spaCy’s trained pipelines like en_core_web_sm contain both a parser and senter that perform sentence segmentation, but the senter is disabled by default.

