Skip to main content

Updating AGOL feature services with data from external tables.

Project description

palletjack

Release Events codecov

A library of classes and methods for automatically updating AGOL feature services with data from several different types of external sources. Client apps (sometimes called 'skids') can reuse these classes for common use cases. The code modules are oriented around each step in the extract, transform, and load process.

palletjack works with pandas DataFrames (either regular for tabular data or Esri's spatially-enabled dataframes for spatial data). The extract and transform methods return dataframes and the load methods consume dataframes as their source data.

The documentation includes a user guide along with an API description of the available classes and methods.

Pallet jack: forklift's little brother.

Dependencies

palletjack relies on the dependencies listed in setup.py. These are all available on PyPI and can be installed in most environments, including Google Cloud Functions.

The arcgis library does all the heavy lifting for spatial data. If the arcpy library is not available (such as in a cloud function), it relies on shapely for its geometry engine.

Installation

  1. Activate your application's environment
  2. pip install ugrc-palletjack

Quick start

  1. Import the desired modules

  2. Use a class in extract to load a dataframe from an external source

  3. Transform your dataframe as desired with helper methods from transform

  4. Use the dataframe to update a hosted feature service using the methods in load

    from palletjack import extract, transform, load
    
    #: Load the data from a Google Sheet
    gsheet_extractor = extract.GSheetLoader(path_to_service_account_json)
    sheet_df = gsheet_extractor.load_specific_worksheet_into_dataframe(sheet_id, 'title of desired sheet', by_title=True)
    
    #: Convert the data to points using lat/long fields, clean for uploading
    spatial_df = pd.DataFrame.spatial.from_xy(input_df, x_column='longitude', y_column='latitude')
    renamed_df = transform.DataCleaning.rename_dataframe_columns_for_agol(spatial_df)
    cleaned_df = transform.DataCleaning.switch_to_nullable_int(renamed_df, ['an_int_field_with_null_values'])
    
    #: Truncate the existing feature service data and load the new data
    gis = arcgis.gis.GIS('my_agol_org_url', 'username', 'super-duper-secure-password')
    updater = load.ServiceUpdater(gis, 'feature_service_item_id')
    updates = updater.truncate_and_load(cleaned_df)
    
    #: It even works with stand-alone tables!
    table_updater = load.TableUpdater(gis, 'table_service_item_id', service_type='table')
    table_updates = table_updater.truncate_and_load(cleaned_df)
    

Development

  1. Create a conda environment with Python 3.13 (use defaults channel if you won't be deploying to an environment with ArcGIS Pro installed)
    • conda create -n palletjack -c defaults python=3.13
    • activate palletjack
  2. Clone the repo
  3. Install in dev mode with development dependencies
    • pip install -e .[tests]

Troubleshooting Weird Append Errors

If a FeatureLayer.append() call (within a load.ServiceUpdater method) fails with an "Unknown Error: 500" error or something like that, you can query the results to get more info. The urllib3 debug-level logs will include the HTTP GET or POST call, something like the following: https://services1.arcgis.com:443 POST /<unique string>/arcgis/rest/services/<feature layer name>/FeatureServer/<layer id>/append/jobs/<job guid>?f=json token=<crazy long token string>. The defualt basicConfig logger includes the urllib3 logs (logging.basicConfig(level=logging.DEBUG)) and is great for development debugging, or you can add a specific logger for urllib3 in your code and set it's level to debug.

You can use this and a token from an AGOL tab to build a new job status url. To get the token, log into AGOL in a browser and open a private hosted feature layer item. Click the layer, and then open the developer console. With the Network tab of the console open, click on the "View" link for the service URL. You should see a document in the list whose name includes "?token=". Copy the name and then copy out the token string.

Now that you've got the token string, you can build the status query: https://services1.arcgis.com/<unique string>/arcgis/rest/services/<feature layer name>/FeatureServer/<layer id>/append/jobs/<job guid>?f=json&<token from agol>

Calling this URL in a browser should return a message that will hopefully give you more info as to why it failed.

Updating Docs

palletjack uses pdoc3 to generate HTML docs in docs/palletjack from the docstrings within the code itself. These are then served up via github pages.

The github pages are served from the gh-pages branch. After you make edits to the code and update the docstrings, rebase this branch onto the updated main branch. To prevent github pages from trying to generate a site from the contents of docs/palletjack with jekyll, add a .nojekyll file to docs/palletjack.

To generate the docs, run pdoc --html -o docs\ c:\palletjack\repo\src\palletjack --force. The code's docstrings should be Google-style docstrings with proper indentation to ensure the argument lists, etc are parsed and displayed correctly.

docs/README.md is included at the top package level by adding the line .. include:: ../../docs/README.md in __init__.py's docstring. This tells pdoc to insert that markdown into the HTML generated for that docstring, and the include directive can be used for more in-depth documentation anywhere else as well. Note that pdoc tries to create links for anything surrounded by backticks, which are also used for code blocks. You may need to manually edit the HTML to remove the links if they change the content of your code blocks (such as in the example import statement).

Once the contents of docs/palletjack look correct, force push the gh-pages branch to github. This will trigger the action to republish the site. The docs are then accessible at [agrc.github.io/palletjack/palletjack/index.html].

Attribution

This project was developed with the assistance of GitHub Copilot.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ugrc_palletjack-6.0.1.tar.gz (75.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

ugrc_palletjack-6.0.1-py3-none-any.whl (45.6 kB view details)

Uploaded Python 3

File details

Details for the file ugrc_palletjack-6.0.1.tar.gz.

File metadata

  • Download URL: ugrc_palletjack-6.0.1.tar.gz
  • Upload date:
  • Size: 75.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for ugrc_palletjack-6.0.1.tar.gz
Algorithm Hash digest
SHA256 369511c4cf03969c054347f6f13f82ee6b761915af33858fc06a397ecae53cf2
MD5 6e1424c33aacf9c12d20cc4a9a96d1e6
BLAKE2b-256 d7fd952b3a53923bd9ed8a2eb3b036058a8ba1b91fd80767eea7934c728fc719

See more details on using hashes here.

Provenance

The following attestation bundles were made for ugrc_palletjack-6.0.1.tar.gz:

Publisher: release.yml on agrc/palletjack

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file ugrc_palletjack-6.0.1-py3-none-any.whl.

File metadata

File hashes

Hashes for ugrc_palletjack-6.0.1-py3-none-any.whl
Algorithm Hash digest
SHA256 8d2fd9a9b0107de24a366af430b77666c865ad0082e2e72bb3583f48ee997894
MD5 2ae2bac1920fc854d725b988e23515fc
BLAKE2b-256 6a8772e91142cedf17f81c82446c147130aa3b4504139e87c1199d3420c4efab

See more details on using hashes here.

Provenance

The following attestation bundles were made for ugrc_palletjack-6.0.1-py3-none-any.whl:

Publisher: release.yml on agrc/palletjack

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page