spaCy
spacy.io › usage › spacy-101
spaCy 101: Everything you need to know · spaCy Usage Documentation
To make them compact and fast, spaCy’s small pipeline packages (all packages that end in sm) don’t ship with word vectors, and only include context-sensitive tensors. This means you can still use the similarity() methods to compare documents, spans and tokens – but the result won’t be as good, and individual tokens won’t have any vectors assigned.
Videos
spaCy
spacy.io › usage › linguistic-features
Linguistic Features · spaCy Usage Documentation
spaCy is a free open-source library for Natural Language Processing in Python. It features NER, POS tagging, dependency parsing, word vectors and more.
spaCy
spacy.io › api
Library Architecture · spaCy API Documentation
It takes raw text and sends it through the pipeline, returning an annotated document. It also orchestrates training and serialization. The processing pipeline consists of one or more pipeline components that are called on the Doc in order. The tokenizer runs before the components. Pipeline components can be added using Language.add_pipe. They can contain a statistical model and trained weights, or only make rule-based modifications to the Doc. spaCy provides a range of built-in components for different language processing tasks and also allows adding custom components.
Factsheet
spaCy
Original author Matthew Honnibal
Developers Explosion AI, various
spaCy
Original author Matthew Honnibal
Developers Explosion AI, various
spaCy
spacy.io
spaCy · Industrial-strength Natural Language Processing in Python
spaCy is a free open-source library for Natural Language Processing in Python. It features NER, POS tagging, dependency parsing, word vectors and more.
spaCy
spacy.io › usage
Install spaCy · spaCy Usage Documentation
# Note M1 GPU support is experimental, see Thinc issue #792python -m venv .envsource .env/bin/activatesource .env/bin/activate.env\Scripts\activatepython -m venv .envsource .env/bin/activatesource .env/bin/activate.env\Scripts\activateconda create -n venvconda activate venvpip install -U pip setuptools wheelpip install -U pip setuptools wheelpip install -U spacyconda install -c conda-forge spacyconda install -c conda-forge cupyconda install -c conda-forge spacy-transformersgit clone https://github.com/explosion/spaCycd spaCypip install -r requirements.txtpip install --no-build-isolation --edit
spaCy
spacy.io › usage › processing-pipelines
Language Processing Pipelines · spaCy Usage Documentation
spaCy ships with several built-in pipeline components that are registered with string names. This means that you can initialize them by calling nlp.add_pipe with their names and spaCy will know how to create them. See the API documentation for a full list of available pipeline components and component functions.
spaCy
spacy.io › usage › models
Models & Languages · spaCy Usage Documentation
spaCy is a free open-source library for Natural Language Processing in Python. It features NER, POS tagging, dependency parsing, word vectors and more.
PyPI
pypi.org › project › spacy
spacy · PyPI
It features state-of-the-art speed and neural network models for tagging, parsing, named entity recognition, text classification and more, multi-task learning with pretrained transformers like BERT, as well as a production-ready training system and easy model packaging, deployment and workflow management. spaCy is commercial open-source software, released under the MIT license.
» pip install spacy
Published Nov 17, 2025
Version 3.8.11
Repository https://github.com/explosion/spaCy
Homepage https://spacy.io
spaCy
spacy.io › usage › projects
Projects · spaCy Usage Documentation
It will list all commands, workflows and assets defined in the project and include details on how to run the project, as well as links to the relevant spaCy documentation to make it easy for others to get started using your project.
Dataiku
doc.dataiku.com › dss › latest › python › spacy.html
Using SpaCy — Dataiku DSS 14 documentation
Documentation for SpaCy is available at https://spacy.io/
GitHub
github.com › explosion › spaCy
GitHub - explosion/spaCy: 💫 Industrial-strength Natural Language Processing (NLP) in Python
import spacy import en_core_web_sm nlp = en_core_web_sm.load() doc = nlp("This is a sentence.") 📖 For more info and examples, check out the models documentation.
Starred by 32.9K users
Forked by 4.6K users
Languages Python 54.1% | MDX 31.2% | Cython 10.5% | JavaScript 2.6% | Sass 0.8% | TypeScript 0.4%
spaCy
spacy.io › usage › saving-loading
Saving and Loading · spaCy Usage Documentation
Given a config, spaCy is able reconstruct the whole tree of objects and the nlp object. An exported config can also be used to train a pipeline with the same settings. meta: Meta information about the pipeline and the Python package, such as the author information, license, version, data sources and label scheme. This is mostly used for documentation purposes and for packaging pipelines.
spaCy
spacy.io › models
Trained Models & Pipelines · spaCy Models Documentation
For faster processing, you may only want to run a subset of the components in a trained pipeline. The disable and exclude arguments to spacy.load let you control which components are loaded and run. Disabled components are loaded in the background so it’s possible to reenable them in the same pipeline in the future with nlp.enable_pipe.
spaCy
spacy.io › usage › training
Training Pipelines & Models · spaCy Usage Documentation
You can change the model architecture entirely by implementing your own custom models and providing those in the config when creating the pipeline component. See the documentation on layers and model architectures for more details. When you start training a new model from scratch, spacy train will call nlp.initialize to initialize the pipeline and load the required data.
DataCamp
datacamp.com › cheat-sheet › spacy-cheat-sheet-advanced-nlp-in-python
spaCy Cheat Sheet: Advanced NLP in Python | DataCamp
August 1, 2021 - spacy.explain("RB") # 'adverb' spacy.explain("GPE") # 'Countries, cities, states' ⚠️ If you're in a Jupyter notebook, use displacy.render. Otherwise, use displacy.serve to start a web server and show the visualization in your browser. ... ⚠️ To use word vectors, you need to install the larger models ending in md or lg , for example en_core_web_lg. doc1 = nlp("I like cats") doc2 = nlp("I like dogs") # Compare 2 documents doc1.similarity(doc2) # Compare 2 tokens doc1[2].similarity(doc2[2]) # Compare tokens and spans doc1[0].similarity(doc2[1:3])
spaCy
spacy.io › api › top-level
Top-level Functions · spaCy API Documentation
spaCy is a free open-source library for Natural Language Processing in Python. It features NER, POS tagging, dependency parsing, word vectors and more.

