Skip to main content

Python Stream processing.

Project description

faust

Python Stream Processing Fork

python versions version codecov slack Code style: black pre-commit license downloads

Installation

pip install faust-streaming

Documentation

Why the fork

We have decided to fork the original Faust project because there is a critical process of releasing new versions which causes uncertainty in the community. Everybody is welcome to contribute to this fork, and you can be added as a maintainer.

We want to:

  • Ensure continues release
  • Code quality
  • Use of latest versions of kafka drivers (for now only aiokafka)
  • Support kafka transactions
  • Update the documentation

and more...

Usage

# Python Streams
# Forever scalable event processing & in-memory durable K/V store;
# as a library w/ asyncio & static typing.
import faust

Faust is a stream processing library, porting the ideas from Kafka Streams to Python.

It is used at Robinhood to build high performance distributed systems and real-time data pipelines that process billions of events every day.

Faust provides both stream processing and event processing, sharing similarity with tools such as Kafka Streams, Apache Spark, Storm, Samza, Flink,

It does not use a DSL, it's just Python! This means you can use all your favorite Python libraries when stream processing: NumPy, PyTorch, Pandas, NLTK, Django, Flask, SQLAlchemy, ++

Faust requires Python 3.6 or later for the new async/await_ syntax, and variable type annotations.

Here's an example processing a stream of incoming orders:

app = faust.App('myapp', broker='kafka://localhost')

# Models describe how messages are serialized:
# {"account_id": "3fae-...", amount": 3}
class Order(faust.Record):
    account_id: str
    amount: int

@app.agent(value_type=Order)
async def order(orders):
    async for order in orders:
        # process infinite stream of orders.
        print(f'Order for {order.account_id}: {order.amount}')

The Agent decorator defines a "stream processor" that essentially consumes from a Kafka topic and does something for every event it receives.

The agent is an async def function, so can also perform other operations asynchronously, such as web requests.

This system can persist state, acting like a database. Tables are named distributed key/value stores you can use as regular Python dictionaries.

Tables are stored locally on each machine using a super fast embedded database written in C++, called RocksDB.

Tables can also store aggregate counts that are optionally "windowed" so you can keep track of "number of clicks from the last day," or "number of clicks in the last hour." for example. Like Kafka Streams, we support tumbling, hopping and sliding windows of time, and old windows can be expired to stop data from filling up.

For reliability, we use a Kafka topic as "write-ahead-log". Whenever a key is changed we publish to the changelog. Standby nodes consume from this changelog to keep an exact replica of the data and enables instant recovery should any of the nodes fail.

To the user a table is just a dictionary, but data is persisted between restarts and replicated across nodes so on failover other nodes can take over automatically.

You can count page views by URL:

# data sent to 'clicks' topic sharded by URL key.
# e.g. key="http://example.com" value="1"
click_topic = app.topic('clicks', key_type=str, value_type=int)

# default value for missing URL will be 0 with `default=int`
counts = app.Table('click_counts', default=int)

@app.agent(click_topic)
async def count_click(clicks):
    async for url, count in clicks.items():
        counts[url] += count

The data sent to the Kafka topic is partitioned, which means the clicks will be sharded by URL in such a way that every count for the same URL will be delivered to the same Faust worker instance.

Faust supports any type of stream data: bytes, Unicode and serialized structures, but also comes with "Models" that use modern Python syntax to describe how keys and values in streams are serialized:

# Order is a json serialized dictionary,
# having these fields:

class Order(faust.Record):
    account_id: str
    product_id: str
    price: float
    quantity: float = 1.0

orders_topic = app.topic('orders', key_type=str, value_type=Order)

@app.agent(orders_topic)
async def process_order(orders):
    async for order in orders:
        # process each order using regular Python
        total_price = order.price * order.quantity
        await send_order_received_email(order.account_id, order)

Faust is statically typed, using the mypy type checker, so you can take advantage of static types when writing applications.

The Faust source code is small, well organized, and serves as a good resource for learning the implementation of Kafka Streams.

Learn more about Faust in the introduction introduction page to read more about Faust, system requirements, installation instructions, community resources, and more.

or go directly to the quickstart tutorial to see Faust in action by programming a streaming application.

then explore the User Guide for in-depth information organized by topic.

Local development

  1. Clone the project
  2. Create a virtualenv: python3.7 -m venv venv && source venv/bin/activate
  3. Install the requirements: ./scripts/install
  4. Run lint: ./scripts/lint
  5. Run tests: ./scripts/tests

Faust key points

Simple

Faust is extremely easy to use. To get started using other stream processing solutions you have complicated hello-world projects, and infrastructure requirements. Faust only requires Kafka, the rest is just Python, so If you know Python you can already use Faust to do stream processing, and it can integrate with just about anything.

Here's one of the easier applications you can make::

import faust

class Greeting(faust.Record):
    from_name: str
    to_name: str

app = faust.App('hello-app', broker='kafka://localhost')
topic = app.topic('hello-topic', value_type=Greeting)

@app.agent(topic)
async def hello(greetings):
    async for greeting in greetings:
        print(f'Hello from {greeting.from_name} to {greeting.to_name}')

@app.timer(interval=1.0)
async def example_sender(app):
    await hello.send(
        value=Greeting(from_name='Faust', to_name='you'),
    )

if __name__ == '__main__':
    app.main()

You're probably a bit intimidated by the async and await keywords, but you don't have to know how asyncio works to use Faust: just mimic the examples, and you'll be fine.

The example application starts two tasks: one is processing a stream, the other is a background thread sending events to that stream. In a real-life application, your system will publish events to Kafka topics that your processors can consume from, and the background thread is only needed to feed data into our example.

Highly Available

Faust is highly available and can survive network problems and server crashes. In the case of node failure, it can automatically recover, and tables have standby nodes that will take over.

Distributed

Start more instances of your application as needed.

Fast

A single-core Faust worker instance can already process tens of thousands of events every second, and we are reasonably confident that throughput will increase once we can support a more optimized Kafka client.

Flexible

Faust is just Python, and a stream is an infinite asynchronous iterator. If you know how to use Python, you already know how to use Faust, and it works with your favorite Python libraries like Django, Flask, SQLAlchemy, NTLK, NumPy, SciPy, TensorFlow, etc.

Bundles

Faust also defines a group of setuptools extensions that can be used to install Faust and the dependencies for a given feature.

You can specify these in your requirements or on the pip command-line by using brackets. Separate multiple bundles using the comma:

pip install "faust[rocksdb]"

pip install "faust[rocksdb,uvloop,fast,redis, aerospike]"

The following bundles are available:

Faust with extras

Stores

pip install faust[rocksdb] for using RocksDB for storing Faust table state. Recommended in production.

pip install faust[aerospike] for using Aerospike for storing Faust table state. Recommended if supported

Aerospike Configuration

Aerospike can be enabled as the state store by specifying store="aerospike://"

By default, all tables backed by Aerospike use use_partitioner=True and generate changelog topic events similar to a state store backed by RocksDB. The following configuration options should be passed in as keys to the options parameter in Table namespace : aerospike namespace

ttl: TTL for all KV's in the table

username: username to connect to the Aerospike cluster

password: password to connect to the Aerospike cluster

hosts : the hosts parameter as specified in the aerospike client

policies: the different policies for read/write/scans policies

client: a dict of host and policies defined above

Caching

faust[redis] for using Redis as a simple caching backend (Memcached-style).

Codecs

faust[yaml] for using YAML and the PyYAML library in streams.

Optimization

faust[fast] for installing all the available C speedup extensions to Faust core.

Sensors

faust[datadog] for using the Datadog Faust monitor.

faust[statsd] for using the Statsd Faust monitor.

faust[prometheus] for using the Prometheus Faust monitor.

Event Loops

faust[uvloop] for using Faust with uvloop.

faust[eventlet] for using Faust with eventlet

Debugging

faust[debug] for using aiomonitor to connect and debug a running Faust worker.

faust[setproctitle]when the setproctitle module is installed the Faust worker will use it to set a nicer process name in ps/top listings.vAlso installed with the fast and debug bundles.

Downloading and installing from source

Download the latest version of Faust from https://pypi.org/project/faust-streaming/

You can install it by doing:

$ tar xvfz faust-streaming-0.0.0.tar.gz
$ cd faust-streaming-0.0.0
$ python setup.py build
# python setup.py install

The last command must be executed as a privileged user if you are not currently using a virtualenv.

Using the development version

With pip

You can install the latest snapshot of Faust using the following pip command:

pip install https://github.com/faust-streaming/faust/zipball/master#egg=faust

FAQ

Can I use Faust with Django/Flask/etc

Yes! Use eventlet as a bridge to integrate with asyncio.

Using eventlet

This approach works with any blocking Python library that can work with eventlet

Using eventlet requires you to install the faust-aioeventlet module, and you can install this as a bundle along with Faust:

pip install -U faust[eventlet]

Then to actually use eventlet as the event loop you have to either use the -L <faust --loop> argument to the faust program:

faust -L eventlet -A myproj worker -l info

or add import mode.loop.eventlet at the top of your entry point script:

#!/usr/bin/env python3
import mode.loop.eventlet  # noqa

It's very important this is at the very top of the module, and that it executes before you import libraries.

Can I use Faust with Tornado

Yes! Use the tornado.platform.asyncio bridge

Can I use Faust with Twisted

Yes! Use the asyncio reactor implementation: https://twistedmatrix.com/documents/current/api/twisted.internet.asyncioreactor.html

Will you support Python 2.7 or Python 3.5

No. Faust requires Python 3.7 or later, since it heavily uses features that were introduced in Python 3.6 (async, await, variable type annotations).

I get a maximum number of open files exceeded error by RocksDB when running a Faust app locally. How can I fix this

You may need to increase the limit for the maximum number of open files. On macOS and Linux you can use:

ulimit -n max_open_files to increase the open files limit to max_open_files.

On docker, you can use the --ulimit flag:

docker run --ulimit nofile=50000:100000 <image-tag> where 50000 is the soft limit, and 100000 is the hard limit See the difference.

What kafka versions faust supports

Faust supports kafka with version >= 0.10.

Getting Help

Slack

For discussions about the usage, development, and future of Faust, please join the fauststream Slack.

Resources

Bug tracker

If you have any suggestions, bug reports, or annoyances please report them to our issue tracker at https://github.com/faust-streaming/faust/issues/

License

This software is licensed under the New BSD License. See the LICENSE file in the top distribution directory for the full license text.

Contributing

Development of Faust happens at GitHub

You're highly encouraged to participate in the development of Faust.

Code of Conduct

Everyone interacting in the project's code bases, issue trackers, chat rooms, and mailing lists is expected to follow the Faust Code of Conduct.

As contributors and maintainers of these projects, and in the interest of fostering an open and welcoming community, we pledge to respect all people who contribute through reporting issues, posting feature requests, updating documentation, submitting pull requests or patches, and other activities.

We are committed to making participation in these projects a harassment-free experience for everyone, regardless of level of experience, gender, gender identity and expression, sexual orientation, disability, personal appearance, body size, race, ethnicity, age, religion, or nationality.

Examples of unacceptable behavior by participants include:

  • The use of sexualized language or imagery
  • Personal attacks
  • Trolling or insulting/derogatory comments
  • Public or private harassment
  • Publishing other's private information, such as physical or electronic addresses, without explicit permission
  • Other unethical or unprofessional conduct.

Project maintainers have the right and responsibility to remove, edit, or reject comments, commits, code, wiki edits, issues, and other contributions that are not aligned to this Code of Conduct. By adopting this Code of Conduct, project maintainers commit themselves to fairly and consistently applying these principles to every aspect of managing this project. Project maintainers who do not follow or enforce the Code of Conduct may be permanently removed from the project team.

This code of conduct applies both within project spaces and in public spaces when an individual is representing the project or its community.

Instances of abusive, harassing, or otherwise unacceptable behavior may be reported by opening an issue or contacting one or more of the project maintainers.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

faust-streaming-0.10.1.tar.gz (753.7 kB view details)

Uploaded Source

Built Distributions

If you're not sure about the file name format, learn more about wheel file names.

faust_streaming-0.10.1-cp311-cp311-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.11musllinux: musl 1.1+ x86-64

faust_streaming-0.10.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.11manylinux: glibc 2.17+ x86-64manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.1-cp311-cp311-macosx_10_9_x86_64.whl (482.7 kB view details)

Uploaded CPython 3.11macOS 10.9+ x86-64

faust_streaming-0.10.1-cp310-cp310-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.10musllinux: musl 1.1+ x86-64

faust_streaming-0.10.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.10manylinux: glibc 2.17+ x86-64manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.1-cp310-cp310-macosx_10_9_x86_64.whl (487.9 kB view details)

Uploaded CPython 3.10macOS 10.9+ x86-64

faust_streaming-0.10.1-cp39-cp39-musllinux_1_1_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.9musllinux: musl 1.1+ x86-64

faust_streaming-0.10.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.9manylinux: glibc 2.17+ x86-64manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.1-cp39-cp39-macosx_10_9_x86_64.whl (487.5 kB view details)

Uploaded CPython 3.9macOS 10.9+ x86-64

faust_streaming-0.10.1-cp38-cp38-musllinux_1_1_x86_64.whl (1.2 MB view details)

Uploaded CPython 3.8musllinux: musl 1.1+ x86-64

faust_streaming-0.10.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.1 MB view details)

Uploaded CPython 3.8manylinux: glibc 2.17+ x86-64manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.1-cp38-cp38-macosx_10_9_x86_64.whl (486.1 kB view details)

Uploaded CPython 3.8macOS 10.9+ x86-64

faust_streaming-0.10.1-cp37-cp37m-musllinux_1_1_x86_64.whl (1.0 MB view details)

Uploaded CPython 3.7mmusllinux: musl 1.1+ x86-64

faust_streaming-0.10.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (964.8 kB view details)

Uploaded CPython 3.7mmanylinux: glibc 2.17+ x86-64manylinux: glibc 2.5+ x86-64

faust_streaming-0.10.1-cp37-cp37m-macosx_10_9_x86_64.whl (483.6 kB view details)

Uploaded CPython 3.7mmacOS 10.9+ x86-64

File details

Details for the file faust-streaming-0.10.1.tar.gz.

File metadata

  • Download URL: faust-streaming-0.10.1.tar.gz
  • Upload date:
  • Size: 753.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.11.1

File hashes

Hashes for faust-streaming-0.10.1.tar.gz
Algorithm Hash digest
SHA256 f8596e71187c955f640e0b1e0d795f5c67c737cfccee70eb9332f80108434903
MD5 2856b76cb7ada3074814c226a06dfb69
BLAKE2b-256 852806d13174347aa6cdb77353d854720ae23bd6093d000aa020395f6b957a43

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp311-cp311-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp311-cp311-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 7563ef8f122cc16545227fabe4f13d74a2e96ee6bdb322bbe3832e20729a1d9a
MD5 3265618270b4fae1e2968c3a640740b8
BLAKE2b-256 cf0313c19ea546cab38b4b6904ec1fd76209328b06f1e7c7830b64e1bf31f847

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 202d3a2ad1f8b55dba3e2aea66048cd5e7f7b7b42a3f7fb6eb1537470f658b38
MD5 ba3945db4ff016c78d3a3aa1cc339f34
BLAKE2b-256 f9db77a2f44c013c051df448faf5665f6d0c7b1b8284173731128afea33f102c

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp311-cp311-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp311-cp311-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 a30c3cb3e4702037c75cb9f89d7defa88723e13b2a6774962144e3fab15a66a1
MD5 fb3ae8f40352b445489a0571117a8f1d
BLAKE2b-256 ab6441d644d8a46402be79a2fd19c20967544410c6892defc27db8fe5aedf07f

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp310-cp310-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp310-cp310-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 535fb72a21a1a25f2b2676828375880e00ec9ec624aca8b9100205165a9fc51d
MD5 fc138faff01c240635332a043bad04dd
BLAKE2b-256 16a425dc3879d99dc736839e9ce8f6e689b786c4dd7c9eaefa91629e8df42be3

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 ad7b3bca5e529469a346cb3eb3d473995cda5d0d8a9f47e7408e1bd5f081fa3c
MD5 f6f9201629048e759ebe96042113c5bc
BLAKE2b-256 570ec14b757890380663336d7e5f17f381772606a8f410f93ac8bc41e8afc4a6

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp310-cp310-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp310-cp310-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 ace9bb42690bf4d3f7471dc92e8e9a27ca3d6c852b3f2d0af892cc21fe7df052
MD5 873ac51c72d7500e9c62e37e5a3cb993
BLAKE2b-256 3bfd9ccadc08bee2adce62d26a7be8037ff7d7bc7206cc49d76e26c5b90142a4

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp39-cp39-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp39-cp39-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 50d461505768c36336396e008ef8ac520328e72c115607c7b51bcf0f4a578827
MD5 4cdc0de5a681a869f9ac233c818318ed
BLAKE2b-256 a085d398c17b97778462f23c40e96b602dea5b3670e8dadb8d2f6266b234e208

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 dd758672e8f899b5d19790ef12207c523ab0133303b6d82716e98217f2a44622
MD5 8428bfe8761cf262b7139376136dd3a1
BLAKE2b-256 897fd97b264f01e38ba0fc309341c2a4c6b7b6aa1cc30c7fdd5c94ef63e9c4c6

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp39-cp39-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp39-cp39-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 07fe29aaf218e9b68e10d49a89764ea4b20a783a4e614ca895230699f15672f7
MD5 88ab8bd381602551bf45d759930f1fc0
BLAKE2b-256 49a5f8881c4695aac194ed776c266229e466899abd9c35c9deedc6ed16f3352b

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp38-cp38-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp38-cp38-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 c2a2cdd71e2a14b7e6de2fbab51bfaa021e947853f3085d32d822c188a12cbed
MD5 cd16d43901439bfa6d2d8998bef13dff
BLAKE2b-256 faf4fa1f94b90256055cab974116471969898676395e422fd9e2bcc455314eeb

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 83018db396265b4b079aa3af8fa574a900536e563f4564ce38c1f02c1eff0b83
MD5 8197d3e30b2350a7f3d0a533fa174520
BLAKE2b-256 29981447953d81a5f229a79da46a64237576d257e93fa5f6f0490690ecb66e03

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp38-cp38-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp38-cp38-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 5f280d3ab95abb13be85bd3b9aeba6c390c47b3375c6e03aba956302ab8ff150
MD5 b89f57415e7da42121c9fe03521d6fa5
BLAKE2b-256 238ccbba6f7807fc80a0a838dfb3cd7081c4bd4762e29964a869fe052d4b64dd

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp37-cp37m-musllinux_1_1_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp37-cp37m-musllinux_1_1_x86_64.whl
Algorithm Hash digest
SHA256 4ce58d932e97891125689f8851ddeebd039c7438b612b050685d5357594af0a7
MD5 93c5793b91717e60e255eb315fb71bbb
BLAKE2b-256 95a4fd3cb285af27e2a9472592804b1e20564f75b610efc5dbcfc52e756fbd89

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Algorithm Hash digest
SHA256 6ce94d7027f9ede66979ace86f17baeee5bd9b4fb3e661971d3e1f588bb0bca0
MD5 ace639ae7b290213d5c8ddb35df9c933
BLAKE2b-256 d638cbc92dc2644d79f72ed606015b64d4b4d2e8414044232583d1cb09940be6

See more details on using hashes here.

File details

Details for the file faust_streaming-0.10.1-cp37-cp37m-macosx_10_9_x86_64.whl.

File metadata

File hashes

Hashes for faust_streaming-0.10.1-cp37-cp37m-macosx_10_9_x86_64.whl
Algorithm Hash digest
SHA256 206bc0f37f6f7ed50f3fd8ef1e90e8258e551b0df3b94a392a2030f688e65588
MD5 04d8108a809895006e78cd49fdca627e
BLAKE2b-256 8f96199cec364ada16c483c6c6d593a5e37e6cfdb1d519b1b4ad37c4a84bc462

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page