SigLIP base model for use with Autodistill
Project description
Autodistill SigLIP Module
This repository contains the code supporting the SigLIP base model for use with Autodistill.
CLIP, developed by OpenAI, is a computer vision model trained using pairs of images and text. You can use CLIP with autodistill for image classification.
Read the full Autodistill documentation.
Read the SigLIP Autodistill documentation.
Installation
To use SigLIP with autodistill, you need to install the following dependency:
pip3 install autodistill-clip
Quickstart
from autodistill_siglip import SigLIP
from autodistill.detection import CaptionOntology
# define an ontology to map class names to our SigLIP prompt
# the ontology dictionary has the format {caption: class}
# where caption is the prompt sent to the base model, and class is the label that will
# be saved for that caption in the generated annotations
# then, load the model
labels = ["person", "a forklift"]
base_model = SigLIP(
ontology=CaptionOntology({item: item for item in labels})
)
results = base_model.predict("image.jpeg", confidence=0.1)
top_1 = results.get_top_k(1)
# show top label
print(labels[top_1[0][0]])
# label folder of images
base_model.label("./context_images", extension=".jpeg")
License
The SigLIP model is licensed under an Apache 2.0 license.
🏆 Contributing
We love your input! Please see the core Autodistill contributing guide to get started. Thank you 🙏 to all our contributors!
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file autodistill-siglip-0.1.0.tar.gz.
File metadata
- Download URL: autodistill-siglip-0.1.0.tar.gz
- Upload date:
- Size: 4.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
cdc3f60d855550e20908fdf9c3e6c8cb68085c32d7363f9ace06a32868f8e072
|
|
| MD5 |
4eea9d615bb45bccbb0b1d6596d39ff6
|
|
| BLAKE2b-256 |
6b785f4a460f23422b71b2cb40fc3166014494a0740ec1b8858c4c1ce407c8f7
|
File details
Details for the file autodistill_siglip-0.1.0-py3-none-any.whl.
File metadata
- Download URL: autodistill_siglip-0.1.0-py3-none-any.whl
- Upload date:
- Size: 3.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
2815075ea1fb39b21c0a18246285efebcc260e7b5695f5f5b7853cdca9d77c95
|
|
| MD5 |
092d5c91f01143ddebfbd242e6f6d687
|
|
| BLAKE2b-256 |
1a7b016a0add64cb005eab853d3370f68e0ffc2a4d0b200f925e83cff5f8f961
|