Website scrapers to retrieve HTR training datasets.
Project description
Arkindex Scrapers
Website scrapers to retrieve HTR datasets and publish them to Arkindex.
Installation
To install arkindex-scrapers, you can do it from Pypi:
- Use a virtualenv (e.g. with virtualenvwrapper
mkvirtualenv -a . scrapers) - Install scrapers as a package (e.g.
pip install arkindex-scrapers)
Usage
When arkindex-scrapers is installed in your environment, the scrapers command becomes available. This command has 3 subcommands. Learn more about them using:
scrapers -h
Do It Yourself History
The diy subcommand retrieves images and transcriptions from collections available on the DIY History website.
Provide the ID of a collection as a positional argument and an output directory. This command will generate 1 JSON file per item. Each of these can be uploaded to Arkindex using the publish subcommand.
Europeana Transcribathon
The eu-trans subcommand retrieves images and transcriptions from stories available on the Europeana Transcribathon website.
By default, this command will look for stories on the whole website. You can restrict the search to a specific story using the --story_id argument. This command will generate 1 JSON file per story. Each of these can be uploaded to Arkindex using the publish subcommand.
Publish to Arkindex
The publish subcommand publishes local JSON files scraped by other subcommands to an Arkindex instance.
Any JSON file is supported, provided that they respect the following format:
{
"name": "", // Name of the element on Arkindex
"metadata": [ // List of metadata to publish on the element
{
"type": "", // Arkindex type of the metadata
"name": "", // Name of the metadata
"value": "" // Value of the metadata
},
...
],
"items": [ // Elements published as children
{
"name": "", // Name of the element on Arkindex
"metadata": [], // List of metadata to publish on the element
"transcriptions": [ // List of transcriptions to publish on the element
"", // Text of a transcription
...
],
"iiif_url": "", // IIIF URL of the image (optional)
"image_path": "" // Relative path towards the image file (optional)
},
...
]
}
Learn more about all arguments of this subcommand using:
scrapers publish -h
Learn more about Arkindex's metadata, transcription and image system in its documentation.
Contributing
Development
For development and tests purpose it may be useful to install the project as a editable package with pip.
- Use a virtualenv (e.g. with virtualenvwrapper
mkvirtualenv -a . scrapers) - Install scrapers as a package (e.g.
pip install -e .)
Linter
Code syntax is analyzed before submitting the code.
To run the linter tools suite you may use pre-commit.
pip install pre-commit
pre-commit run -a
Run tests
Tests are executed with tox using pytest.
To install tox,
pip install tox
tox
To reload the test virtual environment you can use tox -r
Run a single test module: tox -- <test_path>
Run a single test: tox -- <test_path>::<test_function>
--
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file arkindex_scrapers-0.2.0.tar.gz.
File metadata
- Download URL: arkindex_scrapers-0.2.0.tar.gz
- Upload date:
- Size: 19.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
f8ab6bfd01f6042bc6eb335643d6a0a1d7cc6c121117f6bfec1a271fc9f3028a
|
|
| MD5 |
4c13068d61d8f20d52139d13556d3efd
|
|
| BLAKE2b-256 |
238bb991c7336379b19c5439395892337f7cc8403a0430455c00a44bc8cf80a7
|
File details
Details for the file arkindex_scrapers-0.2.0-py3-none-any.whl.
File metadata
- Download URL: arkindex_scrapers-0.2.0-py3-none-any.whl
- Upload date:
- Size: 23.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
46006a792ffa963bee3b433051eba4308a624c1341083a446c20a74c139b3f28
|
|
| MD5 |
d78336ad5954c8293a3390aa74e78f91
|
|
| BLAKE2b-256 |
4dc4bb0f7e01d7eacf9380c5974dad9ef1c7d2026fb53ad4419400fe598f6543
|