Background Tasks in Python for Data Science ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ This post is intended as an extension/update from background tasks in python. I started using the week that Kenneth Reitz released it. It takes away so much... Date: September 10, 2019 This post is intended as an extension/update from background tasks in python . I started using background the week that Kenneth Reitz released it. It takes away so much boilerplate from running background tasks that I use it in more places than I probably should. After taking a look at that post today, I wanted to put a better data science example in here to help folks get started. This post is intended as an extension/update from background tasks in python . I started using background the week that Kenneth Reitz released it. It takes away so much boilerplate from running background tasks that I use it in more places than I probably should. After taking a look at that post today, I wanted to put a better data science example in here to help folks get started. │ I use it in more places than I probably should Before we get into it, I want to make a shout out to Kenneth Reitz for making this so easy. Kenneth is a python God for all that he has given to the community in so many ways, especially with his ideas in building stupid simple api’s for very complicated things. Installation ──────────── ### install via pip ``` pip install background ``` ### install via github I believe one of the later pr’s to the project fixes the way arguments are passed in. I generally clone the repo or copy the module directly into my project. clone it ``` git clone https://github.com/ParthS007/background.git cd background python setup.py install ``` copy the module ``` curl https://raw.githubusercontent.com/ParthS007/background/master/background.py > background.py ``` 🐌 The Slow Function ─────────────────── Imagine that this function is a big one! This function is fairly realistic as it takes in some input and returns a DataFrame. This is what a good half of my fuctions do in data science. The internals of this function generally will include a sql query, load from s3 or a data catalog, an aggregation from another DataFrame. In general it should do one simple thing. Feel Free to copy this “boilerplate” ``` import background from time import sleep import pandas as pd @background.task def long_func(i): """ Simulates fetching data from a service and returning a pandas DataFrame. """ sleep(10) return pd.DataFrame({'number_squared': [i**2]}) ``` Calling the Slow Function ───────────────────────── it’s the future calling 🤙 If we were to call this function 10 times it would take 100s. Not bad for a dumb example, but detrimental when this gets scaled up💥. We want to utilize all of our available resources to reduce our development time and get moving on our project. Calling long_func will return a future object. This object has a number of methods that you can read about in the cpython docs . The main one we are interested in is result. I typically call these functions many times and put them into a list object so that I can track their progress and get their results. If you needed to map inputs back to the result use a dictionary. ``` %time futures = [long_func(i) for i in range(10)] CPU times: user 319 µs, sys: 197 µs, total: 516 µs Wall time: 212 µs ``` Do something with those results() ───────────────────────────────── Simply running the function completes in no time! This is because the future objects that are returned are non blocking and will run in a background task using the ProcessPoolExecutor. To get the result back out we need to call the result method on the future object.result is a blocking function that will not realease until the function has completed. ``` %%time futures = [long_func(i) for i in range(10)] pd.concat([future.result() for future in futures]) CPU times: user 5.38 ms, sys: 3.53 ms, total: 8.9 ms Wall time: 10 s ``` Note that this example completed in 10s, the time it took for only one run, not all 10! 😎 n ─ 😫 crank it up By default the number of parallel processes wil be equal to the number of cpu threads on your machine. To increase the number of parallel processes (max_workers) set increase background.n. ``` background.n = 100 ``` Is it possible to overruse @background.task? ──────────────────────────────────────────── I use this essentially anywhere that I cannot vectorize a python operation and push the compute down into those fast 💨 c extended libraries like numpy, and the operation takes more than a few minutes. Nearly every big network request I make gets broken down into chunks and multithreaded. Let me know… is is possible to overruse @background.task? Let me know your thoughts @_WaylonWalker . Repl.It ─────── Play with the code here! Try different values of background.n and n_runs.