Curated transformer models
Project description
🤖 Curated transformers
This Python package provides a curated set of transformer models for spaCy. It is focused on deep integration into spaCy and will support deployment-focused features such as distillation and quantization. Curated transformers currently supports the following model types:
- ALBERT
- BERT
- CamemBERT
- RoBERTa
- XLM-RoBERTa
Supporting a wide variety of transformer models is a non-goal. If you want
to use another type of model, use
spacy-transformers
, which
allows you to use Hugging Face
transformers
models with spaCy.
⚠️ Warning: experimental package
This package is experimental and it is possible that the models will still change in incompatible ways.
⏳ Install
pip install spacy-curated-transformers
🚀 Quickstart
An example project is provided in the project
directory.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for spacy-curated-transformers-0.1.1.tar.gz
Algorithm | Hash digest | |
---|---|---|
SHA256 | 27f5acb21b7a631fa0ae950397416e1b56f33a3afecedab4cc8ac3f4a9dc5656 |
|
MD5 | 472960de98f4758661ede390c746e00c |
|
BLAKE2b-256 | b5b1f44fd11e1ee89c4c245d819d7774f65115e02a0ff30dd6d82c8f744588cd |
Hashes for spacy_curated_transformers-0.1.1-py2.py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 954557251de321c025f0cc6f867ee0158f1cba68e718db5b51f43270291490a9 |
|
MD5 | 3b99d9f3f27bddb55affdf5d0fc9c836 |
|
BLAKE2b-256 | 142feb324d436ea054c4338097975b5371febedd05b0b1a45139e219a98d08b1 |