Concurrent Python programming with async, threading, and multiprocessing
February 07, 2024 / 48 min read / 8,656 , 3 , 4
Last updated: March 06, 2024
Tags: python, async, threads, multiprocessing, concurrency, parallelism, web scraping
Concurrent Python programming with async, threading, and multiprocessing
tags: Python, async, threads, multiprocessing, concurrency, parallelism, web scraping
Introduction
One of the best ways to achieve significant speed improvements in Python code is through concurrency: doing several tasks simultaneously. In this article, I describe concurrency in Python and give some examples of running Python code concurrently with “async” functions, “threading,” and “multiprocessing.”
Defining key terms
- Synchronous: Code that executes in sequence. Every statement gets executed one by one.
- Concurrency: Making progress on multiple programming tasks at the same time. This does not necessarily mean actively executing multiple lines of code simultaneously. It often involves allowing I/O-bound work to run in the background while our code runs synchronously in the foreground.
- Parallelism: A form of concurrency which does mean executing multiple lines of code simultaneously. It requires breaking code into smaller chunks and making multiple CPU cores execute code simultaneously.
- CPU (Central Processing Unit) core: A CPU core executes code one instruction at a time. Modern computers typically have multiple CPU cores, each able to execute instructions (code) in parallel with other CPU cores on the same CPU processor.
- I/O (Input/Output): “Work” performed by external services (other than the running Python code). Broadly, input refers to “reading” incoming data, while “output” refers to writing outgoing data. Examples of I/O include reading/writing to a file or database or sending/receiving network or web requests. Processes that involve I/O work are said to be I/O-bound. This is opposed to processes with no (or limited) I/O work, which are said to be CPU-bound. If Python code is structured appropriately, Python code can run multiple I/O-bound tasks concurrently in the background, while Python bytecode runs in the foreground.
- Thread: A small unit of execution in a process. A process is the overall running instance of a program, and within that process, multiple threads can execute concurrently. Each thread represents an independent sequence of instructions that can be scheduled and executed by the operating system. In many languages, threads can execute code on multiple CPU cores at the same time. A Python program can spawn many threads, but only one can ever execute Python code simultaneously due to the GIL. However, while only one thread can execute Python code at once, other threads can “wait” to receive the results of I/O-bound work they had previously kicked off.
- GIL (Global Interpreter Lock): A python implementation detail. The GIL synchronizes access to Python objects by preventing multiple threads from executing Python bytecode simultaneously. Because of this, Python is optimized to run very fast on one thread but cannot execute code on multiple threads simultaneously. This means Python normally only runs on one CPU core at a time, even if the machine it runs on has many cores idle.
- Async/await: A way to achieve concurrency from a single thread in Python. It involves running an event loop and converting regular Python functions to
async
functions (or coroutines) with theasync
keyword. The code then awaits or pauses execution of the code at key places to allow other coroutines or I/O-bound work to run, resuming work later once the coroutine or I/O-bound task completes. - Event loop: The event loop coordinates asynchronous tasks (or coroutines) in Python (code that uses the
async
andawait
keywords). It is responsible for scheduling, pausing, and resuming code execution, ensuring tasks make progress without blocking the execution of the entire program. In Python code, only one event loop can run per thread. - Coroutine: In Python, a coroutine often refers to two related concepts, which I may use interchangeably in this article.
- Coroutine function: A function defined using the
async def
syntax. Theawait
keyword can only be used inside a coroutine function. - Coroutine object: An object returned by a coroutine function. When an
async def
defined coroutine function is called (example:my_coroutine()
), the code inside does not execute immediately. Instead, it must be awaited with theawait
keyword, at which time the code defined in the coroutine function will execute. This second definition of coroutine is also often called an awaitable.
- Coroutine function: A function defined using the
Visual Examples
Let’s further describe the key concurrency paradigms with visual diagrams.
1. 👆 Synchronous code: With synchronous code, tasks run sequentially, one after the next.
2. 👆 Async/await code: With async
/await
, all Python code runs on a single thread. Tasks run concurrently (but not in parallel). We set await
steps in the code where we wait on I/O-bound work to complete and other tasks to continue work on the main thread while waiting. Many tasks appear to execute at the same time because many tasks may wait on I/O work at the same time. An event loop coordinates waiting tasks, resuming Python code when background I/O-bound work is completed.
3. 👆 Threaded code: With threaded code, tasks run concurrently (but not in parallel). Unlike with the async
/await
style code, we do not set specific places to wait for background tasks to complete. Instead, our CPU coordinates all threads, alternating which thread executes code at any given time. Because of the GIL, only one thread can execute Python code at a time. However, many threads can wait on I/O-bound work to complete in the background while one active thread executes Python code. The CPU alternates the active thread rapidly, sometimes even alternating between the start of one line of Python code and the end of that same line. Thus, race conditions in threaded code are less predictable than in async
/await
code.
4. 👆 Multiprocessing code: With multiprocessing code in Python, tasks execute in parallel on multiple CPU cores at the same time. Each multiprocess task runs in a separate, mostly isolated Python process. Multiprocessing can vastly speed up Python code that does not involve I/O work, such as math-heavy operations. However, multiprocessing has limitations. There is a significant startup cost for new processes, and communication between processes is limited and relatively slow.
The code
You can find the code for this blog article on GitHub here. The code includes the following folders: web/
, cpu_bound/
, and locks/
. Each folder has examples of synchronous code, followed by various examples of concurrent code. I wrote the code with Python 3.11, but most of the code should work as early as Python version 3.7 and for later versions of Python. You’ll need to install the following packages to follow along with the examples:
- beautifulsoup4 (for parsing HTML)
- rich (for pretty printing colors)
- requests (for making sync-only HTTP requests)
- httpx (for making sync or async HTTP requests)
- aiohttp (for making async-only HTTP requests)
You can install them all with the following command:
pip install beautifulsoup4 rich requests httpx aiohttp
HTTP requests
One common use case for concurrency is web requests, usually in the form of API requests or web scraping. For this section, we’ll be web scraping a Pokémon website to get the names of Pokémon, given their Pokédex number. First, I’ll show you how to scrape several pages synchronously. Then, we’ll see how to get massive speedups using async
/await
and threading.
Synchronous example
The most basic example involves synchronous code execution. We request and process the web pages one at a time. Here’s an example:
# web/1_sync.py
"""Download the first 20 Pokémon synchronously."""
import time
import requests
from bs4 import BeautifulSoup
from rich import print
def main() -> None:
t0 = time.time()
print("Starting coordinating function...", flush=True)
results = download_pokemon_list()
total_seconds = time.time() - t0
print(
f"\n[bold green]The code ran in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
print(f"\n{results=}", flush=True)
def download_pokemon_list() -> list[tuple[int, str]]:
"""Download a list of Pokémon from 'pokemondb.net.net'."""
return [download_single_pokemon(num) for num in range(1, 21)]
def download_single_pokemon(pokemon_num: int = 1) -> tuple[int, str]:
"""Get a Pokémon from 'pokemondb.net.net' by its pokedex number."""
print(
f"[yellow]Downloading Pokémon {pokemon_num:02}... [/yellow]",
flush=True,
)
url = f"https://pokemondb.net.net/pokedex/{pokemon_num}"
resp = requests.get(url, allow_redirects=True)
resp.raise_for_status()
header = get_h1(resp.text)
print(
f"[green]Retrieved [magenta]{pokemon_num:02}={header}",
flush=True,
)
return (pokemon_num, header)
def get_h1(html: str) -> str:
"""Parse the HTML and return the first H1 tag."""
soup = BeautifulSoup(html, "html.parser")
return soup.h1.text
if __name__ == "__main__":
main()
Here’s the terminal output from running that code:
❯ python web/1_sync.py
Starting coordinating function...
Downloading Pokémon 01...
Retrieved 01=Bulbasaur
Downloading Pokémon 02...
Retrieved 02=Ivysaur
Downloading Pokémon 03...
Retrieved 03=Venusaur
Downloading Pokémon 04...
Retrieved 04=Charmander
Downloading Pokémon 05...
Retrieved 05=Charmeleon
Downloading Pokémon 06...
Retrieved 06=Charizard
Downloading Pokémon 07...
Retrieved 07=Squirtle
Downloading Pokémon 08...
Retrieved 08=Wartortle
Downloading Pokémon 09...
Retrieved 09=Blastoise
Downloading Pokémon 10...
Retrieved 10=Caterpie
Downloading Pokémon 11...
Retrieved 11=Metapod
Downloading Pokémon 12...
Retrieved 12=Butterfree
Downloading Pokémon 13...
Retrieved 13=Weedle
Downloading Pokémon 14...
Retrieved 14=Kakuna
Downloading Pokémon 15...
Retrieved 15=Beedrill
Downloading Pokémon 16...
Retrieved 16=Pidgey
Downloading Pokémon 17...
Retrieved 17=Pidgeotto
Downloading Pokémon 18...
Retrieved 18=Pidgeot
Downloading Pokémon 19...
Retrieved 19=Rattata
Downloading Pokémon 20...
Retrieved 20=Raticate
The code ran in 4.31 seconds.
results=[(1, 'Bulbasaur'), (2, 'Ivysaur'), (3, 'Venusaur'), (4, 'Charmander'),
(5, 'Charmeleon'), (6, 'Charizard'), (7, 'Squirtle'), (8, 'Wartortle'), (9,
'Blastoise'), (10, 'Caterpie'), (11, 'Metapod'), (12, 'Butterfree'), (13,
'Weedle'), (14, 'Kakuna'), (15, 'Beedrill'), (16, 'Pidgey'), (17, 'Pidgeotto'),
(18, 'Pidgeot'), (19, 'Rattata'), (20, 'Raticate')]
And here’s a video showing the code executing in real-time with color:
Notice how the requests run one after another: “Downloading”, then “Retrieving”, “Downloading”, then “Retrieving”. First, we kick off a request; then we wait for it to come back; then we kick off the following request and wait for it to come back, etc.
In the above code, we use requests to request the web pages and beautiful soup to parse the HTML we get back. We also use rich to add color to our print
statements since color can help us visualize the output more easily. download_pokemon_list
calls download_single_pokemon
20X, once for each of the first 20 Pokémon. The requests are made one after another in a list comprehension. Beautiful Soup parses the HTML to get the <h1>
header corresponding to the Pokémon’s name. The code takes ~4-6
seconds to run.
Async/await example
Async/await is the preferred, modern approach to running web requests in parallel. In this code example, we convert the synchronous code to async
/await
asynchronously executed code using aiohttp to make web requests:
# web/2_async_aiohttp.py
"""Download the first 20 Pokémon asynchronously using aiohttp."""
import asyncio
import time
import aiohttp
from bs4 import BeautifulSoup
from rich import print
def main() -> None:
t0 = time.time()
print("Starting coordinating coroutine...", flush=True)
results = asyncio.run(download_pokemon_list())
total_seconds = time.time() - t0
print(
f"\n[bold green]The code ran in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
print(f"\n{results=}", flush=True)
async def download_pokemon_list() -> list[tuple[int, str]]:
"""Download a list of Pokémon from 'pokemondb.net.net'."""
print("Creating coroutine objects...", flush=True)
coroutines = [download_single_pokemon(num) for num in range(1, 21)]
print("Gathering coroutines into tasks...", flush=True)
tasks = asyncio.gather(*coroutines)
print("Done gathering tasks. Running + awaiting tasks...", flush=True)
results = await tasks
print("Done gathering results.", flush=True)
return results
async def download_single_pokemon(pokemon_num: int = 1) -> tuple[int, str]:
"""Get a Pokémon from 'pokemondb.net.net' by its pokedex number."""
print(
f"[yellow]Downloading Pokémon {pokemon_num:02}... [/yellow]",
flush=True,
)
url = f"https://pokemondb.net.net/pokedex/{pokemon_num}"
async with aiohttp.ClientSession() as session, session.get(url) as resp:
resp.raise_for_status()
text = await resp.text()
resp.raise_for_status()
header = get_h1(text)
print(
f"[green]Retrieved [magenta]{pokemon_num:02}={header}",
flush=True,
)
return (pokemon_num, header)
def get_h1(html: str) -> str:
"""Parse the HTML and return the first H1 tag."""
soup = BeautifulSoup(html, "html.parser")
return soup.h1.text
if __name__ == "__main__":
main()
Here’s the terminal output from running that code:
❯ python web/2_async_aiohttp.py
Starting tasks...
Gathering coroutines into tasks...
Done gathering tasks. Running + awaiting tasks...
Downloading pokemon 01...
Downloading pokemon 02...
Downloading pokemon 03...
Downloading pokemon 04...
Downloading pokemon 05...
Downloading pokemon 06...
Downloading pokemon 07...
Downloading pokemon 08...
Downloading pokemon 09...
Downloading pokemon 10...
Downloading pokemon 11...
Downloading pokemon 12...
Downloading pokemon 13...
Downloading pokemon 14...
Downloading pokemon 15...
Downloading pokemon 16...
Downloading pokemon 17...
Downloading pokemon 18...
Downloading pokemon 19...
Downloading pokemon 20...
Retrieved 09=Blastoise
Retrieved 08=Wartortle
Retrieved 03=Venusaur
Retrieved 07=Squirtle
Retrieved 16=Pidgey
Retrieved 02=Ivysaur
Retrieved 04=Charmander
Retrieved 01=Bulbasaur
Retrieved 19=Rattata
Retrieved 18=Pidgeot
Retrieved 05=Charmeleon
Retrieved 06=Charizard
Retrieved 17=Pidgeotto
Retrieved 15=Beedrill
Retrieved 12=Butterfree
Retrieved 14=Kakuna
Retrieved 11=Metapod
Retrieved 10=Caterpie
Retrieved 13=Weedle
Retrieved 20=Raticate
Done gathering results.
The code ran in 1.67 seconds.
results=[(1, 'Bulbasaur'), (2, 'Ivysaur'), (3, 'Venusaur'), (4, 'Charmander'),
(5, 'Charmeleon'), (6, 'Charizard'), (7, 'Squirtle'), (8, 'Wartortle'), (9,
'Blastoise'), (10, 'Caterpie'), (11, 'Metapod'), (12, 'Butterfree'), (13,
'Weedle'), (14, 'Kakuna'), (15, 'Beedrill'), (16, 'Pidgey'), (17, 'Pidgeotto'),
(18, 'Pidgeot'), (19, 'Rattata'), (20, 'Raticate')]
And here’s a video showing the code executing in real-time with color:
Notice how the code kicks off all the “Download” steps, some time passes, and then all the “Retrieved” responses come back out of order. What code did we change?
First, we changed any function that needs await
I/O-bound work into a coroutine function by slapping an async
to the front of it. async def download_pokemon_list
becomes our coordinating async coroutine function and async def download_single_pokemon
is another coroutine function.
results = asyncio.run(download_pokemon_list_gather())
👆 This line (inside main
)is how we start our first coroutine function from a regular synchronous function. Here’s the documentation on asyncio.run
:
asyncio.run
executes an async coroutine function and returns the results. This function runs the passed coroutine, taking care of managing the asyncio event loop, finalizing asynchronous generators, and closing the threadpool. This function cannot be called when another asyncio event loop is running in the same thread… This function always creates a new event loop and closes it at the end. It should be used as a main entry point for asyncio programs, and should ideally only be called once.
To highlight, asyncio.run
runs a single coroutine object from a synchronous function and will error if called when an event loop is already running in the current thread.
coroutines = [download_single_pokemon(num) for num in range(1, 21)]
👆 This line (inside download_pokemon_list
) creates our list of coroutine objects by calling the async coroutine functions. The coroutine functions do not start when they are called–they simply create the coroutine objects to be run later.
NOTE: coroutine functions can be called inside synchronous functions, creating coroutine objects. However, they cannot be await
ed and thus executed inside synchronous functions (except with something like asyncio.run
).
tasks = asyncio.gather(*coroutines)
👆 This line (inside download_pokemon_list
) converts the coroutine objects into asyncio.Task
objects known to the event loop. If you’re not familiar, *coroutines
unpacks the coroutines back into arguments: (*coroutines) == (coroutine_1, coroutine_2, ...)
. Note that the tasks are still not started at this point.
results = await tasks
👆 This line (inside download_pokemon_list
) finally kicks off all the tasks created by asyncio.gather
and waits for the results to return. It stores the results of the completed coroutine objects in a list.
async with aiohttp.ClientSession() as session, session.get(url) as resp:
resp.raise_for_status()
text = await resp.text()
👆 These lines (inside download_single_pokemon
) convert the synchronous requests.get
call to an async call and await the response. Without this change, we could not wait for multiple I/O responses at once in a single thread.
NOTE: these two lines are equivalent:
async with aiohttp.ClientSession() as session, session.get(url) as resp:
...
async with aiohttp.ClientSession() as session:
async with session.get(url) as resp:
...
I chose the top for this script because I think it’s easier to read.
Alternative async/await methods and libraries
Here are a few alternative ways we could have achieved the same goal with async
/await
.
httpx instead of aiohttp
Httpx is an alternative library to aiohttp for making async web requests. It’s a popular library, and I like it for a couple of reasons.
- It can make either asynchronous or synchronous web requests.
aiohttp
can only makeasync
web requests, andrequests
can only make synchronous web requests. - I think its methods and syntax are nicer to work with than
aiohttp
. Working withhttpx
feels nearly identical to working withrequests
, which I like.
However, with testing, I’ve found that httpx
requests are significantly slower than aiohttp
. It doesn’t matter much for a small number of requests, but the difference adds up over many requests. With this example code, aiohttp
averages around 1.6 seconds for 20 requests, and httpx
averages around 2.6 seconds for the same number of requests.
Here’s the different code for using aiohttp
vs. httpx
and requests
one after the other:
import aiohttp
import httpx
# aiohttp (async only)
async with aiohttp.ClientSession() as session, session.get(url) as resp:
resp.raise_for_status()
text = await resp.text()
# httpx (async)
async with httpx.AsyncClient() as client:
resp = await client.get(url, follow_redirects=True)
resp.raise_for_status()
text = resp.text
# httpx (sync)
resp = httpx.get(url, follow_redirects=True)
resp.raise_for_status()
text = resp.text
# requests (sync only)
resp = requests.get(url, allow_redirects=True)
resp.raise_for_status()
text = resp.text
Manually adding tasks to the event loop
Recall that the results of awaiting asyncio.gather
is always a list. Sometimes, you might want more control over the data structure housing your tasks or the means of adding tasks than is offered by asyncio.gather
. If so, you can manually add tasks to the event loop like so:
# Using asyncio.gather
async def download_pokemon_list() -> list[tuple[int, str]]:
"""Download a list of pokemon from 'pokemondb.net' using `asyncio.gather`."""
coroutines = [download_single_pokemon(num) for num in range(1, 21)]
tasks = asyncio.gather(*coroutines)
results = await tasks
return results
# Manual method
async def download_pokemon_list() -> list[tuple[int, str]]:
"""Download a list of pokemon from 'pokemondb.net'.
Manually get the event loop, create and await tasks."""
coroutines = [download_single_pokemon(num) for num in range(1, 21)]
loop = asyncio.get_event_loop()
tasks = [loop.create_task(c) for c in coroutines]
results = [await t for t in tasks]
return results
And as of Python 3.11, you can add tasks to task groups like so:
# Manually add tasks to the task group
async def download_pokemon_list() -> list[tuple[int, str]]:
"""Download a list of pokemon from 'pokemondb.net' using a task group."""
async with asyncio.TaskGroup() as tg:
results = [tg.create_task(download_single_pokemon(num)) for num in range(1, 21)]
return [result.result() for result in results]
Here are some advantages of using an asyncio.TaskGroup
:
- The tasks automatically start when the
asyncio.TaskGroup
context block ends - All running tasks are canceled automatically if one task raises an exception
- All running tasks are canceled if the task group itself is canceled
Alternative method for starting async from sync
So far, we’ve only started an event loop and run a single async coroutine function from our sync code using asyncio.run
. Here’s an alternative approach where we create and gather tasks to run from a standard (non-async) function, and then we create an event loop and run the tasks until they are complete:
def coordinate_from_sync():
"""Start `download_single_pokemon` tasks and run from a sync function."""
coroutines = [download_single_pokemon(num) for num in range(1, 21)]
tasks = asyncio.gather(*coroutines)
loop = asyncio.get_event_loop()
return loop.run_until_complete(tasks)
Threading example
While async
/await
is my preferred approach for running concurrent I/O-bound work, it has limitations.
- Writing the code looks different. You must architect your code using the new
async
/await
syntax. - You need to use
async
-capable libraries to run code concurrently. Recall that we couldn’t userequests
to makeasync
requests and had to useaiohttp
instead. Sometimes, there is no such library to do what you want to do (although, usually, there is).
If you can’t (or don’t want to) use async
/await
to run your I/O-bound work concurrently, you can use threads instead. Here’s what it looks like converting our synchronous Pokémon example to run concurrently with threads:
# web/3_threaded.py
"""Download the first 20 Pokémon with threads."""
import time
from concurrent.futures import ThreadPoolExecutor
from typing import Any, Callable
import requests
from bs4 import BeautifulSoup
from rich import print
def main() -> None:
t0 = time.time()
print("Starting coordinating function...", flush=True)
results = download_pokemon_list()
total_seconds = time.time() - t0
print(
f"\n[bold green]The code ran in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
print(f"\n{results=}", flush=True)
TaskType = tuple[Callable[..., Any], tuple[Any, ...], dict[str, Any]]
def download_pokemon_list() -> list[tuple[int, str]]:
"""Download a list of Pokémon from 'pokemondb.net'."""
print("Defining tasks...", flush=True)
tasks: list[TaskType] = [
# Function, args, kwargs
(download_single_pokemon, (num,), {})
for num in range(1, 21)
]
print("Kick off threaded tasks...", flush=True)
with ThreadPoolExecutor() as executor:
work = [executor.submit(func, *args, **kwargs) for func, args, kwargs in tasks]
print("Waiting for downloads...", flush=True)
print("Done", flush=True)
return [future.result() for future in work]
def download_single_pokemon(pokemon_num: int = 1) -> tuple[int, str]:
"""Get a Pokémon from 'pokemondb.net' by its pokedex number."""
print(
f"[yellow]Downloading Pokémon {pokemon_num:02}... [/yellow]",
flush=True,
)
url = f"https://pokemondb.net/pokedex/{pokemon_num}"
resp = requests.get(url, allow_redirects=True)
resp.raise_for_status()
header = get_h1(resp.text)
print(
f"[green]Retrieved [magenta]{pokemon_num:02}={header}",
flush=True,
)
return (pokemon_num, header)
def get_h1(html: str) -> str:
"""Parse the HTML and return the first H1 tag."""
soup = BeautifulSoup(html, "html.parser")
return soup.h1.text
if __name__ == "__main__":
main()
Here’s the terminal output from running that code:
❯ python web/3_threaded.py
Starting coordinating function...
Defining tasks...
Kick off threaded tasks...
Downloading pokemon 01...
Downloading pokemon 02...
Downloading pokemon 03...
Downloading pokemon 04...
Downloading pokemon 05...
Downloading pokemon 06...
Downloading pokemon 07...
Downloading pokemon 08...
Waiting for downloads...
Retrieved 01=Bulbasaur
Downloading pokemon 09...
Retrieved 07=Squirtle
Downloading pokemon 10...
Retrieved 02=Ivysaur
Downloading pokemon 11...
Retrieved 06=Charizard
Downloading pokemon 12...
Retrieved 05=Charmeleon
Retrieved 03=Venusaur
Downloading pokemon 13...
Downloading pokemon 14...
Retrieved 04=Charmander
Retrieved 08=Wartortle
Downloading pokemon 15...
Downloading pokemon 16...
Retrieved 10=Caterpie
Downloading pokemon 17...
Retrieved 09=Blastoise
Downloading pokemon 18...
Retrieved 11=Metapod
Downloading pokemon 19...
Retrieved 14=Kakuna
Downloading pokemon 20...
Retrieved 12=Butterfree
Retrieved 15=Beedrill
Retrieved 16=Pidgey
Retrieved 13=Weedle
Retrieved 17=Pidgeotto
Retrieved 18=Pidgeot
Retrieved 19=Rattata
Retrieved 20=Raticate
Done
The code ran in 1.77 seconds.
results=[(1, 'Bulbasaur'), (2, 'Ivysaur'), (3, 'Venusaur'), (4, 'Charmander'),
(5, 'Charmeleon'), (6, 'Charizard'), (7, 'Squirtle'), (8, 'Wartortle'), (9,
'Blastoise'), (10, 'Caterpie'), (11, 'Metapod'), (12, 'Butterfree'), (13,
'Weedle'), (14, 'Kakuna'), (15, 'Beedrill'), (16, 'Pidgey'), (17, 'Pidgeotto'),
(18, 'Pidgeot'), (19, 'Rattata'), (20, 'Raticate')]
And here’s a video showing the code executing in real-time with color:
Notice how the executed print statement order is slightly more chaotic than the synchronous script and the async
/await
script. Some requests started immediately, and others only after other threads received response data. And as with the async
/await
code, the results came back out of order compared to the synchronously run requests. The net result, however, is that the speed of threaded code is much faster than synchronous code and is comparable to async
/await
code (<2 seconds).
What changes did we make to run code with threads? The difference primarily lies in download_pokemon_list
:
# Sync (before)
def download_pokemon_list() -> list[tuple[int, str]]:
"""Download a list of Pokémon from 'pokemondb.net'."""
return [download_single_pokemon(num) for num in range(1, 21)]
# With threads (after)
from concurrent.futures import ThreadPoolExecutor
...
def download_pokemon_list() -> list[tuple[int, str]]:
"""Download a list of Pokémon from 'pokemondb.net'."""
print("Defining tasks...", flush=True)
tasks: list[TaskType] = [
# Function, args, kwargs
(download_single_pokemon, (num,), {})
for num in range(1, 21)
]
print("Kick off threaded tasks...", flush=True)
with ThreadPoolExecutor() as executor:
work = [executor.submit(func, *args, **kwargs) for func, args, kwargs in tasks]
print("Waiting for downloads...", flush=True)
print("Done", flush=True)
return [future.result() for future in work]
There are a few ways to run threads in Python, but I think the ThreadPoolExecutor
is the nicest API, so we used that. First, we opened a with
block context manager for the ThreadPoolExecutor
to create, run, and clean up thread tasks. Then, instead of calling our download_single_pokemon
function directly as we do when running code synchronously or with async coroutines, we called executor.submit
, passing in each function (uncalled) alongside its args
and kwargs
.
The return value of executor.submit
is a thread Future
, which holds the future return value of its associated thread task function. The context manager guarantees that when it closes, all thread task functions have finished, and all threads are closed. Then, we can call .result()
on each thread Future
to get the result of each thread task function.
Concurrency safety and locks
As our code runs, it frequently enters “invalid states”. An “invalid state” means that the data we are working with is “wrong” (either by itself or in the context of other data) for some reason or another. Usually, the code is briefly “invalid” and is fixed later in the code. When our code runs synchronously, we can usually guarantee that by the time the code finishes running, all “invalid states” are corrected, and the result is that all data is “valid”.
Example of one bank transferring $20 to another bank:
bank_data = {
"bank_1": 100, # dollars
"bank_2": 100, # dollars
}
total = bank_data["bank_1"] + bank_data["bank_2"] # 200
# 👇 "valid state"
assert bank_data["bank_1"] + bank_data["bank_2"] == total # 200
bank_data["bank_1"] - 20
# 👇 "invalid state"
assert bank_data["bank_1"] + bank_data["bank_2"] != total # 180
bank_data["bank_2"] + 20
# 👇 Back to "valid state"
assert bank_data["bank_1"] + bank_data["bank_2"] == total # 200
However, when code runs concurrently, one task can bring your data to a temporary “invalid state”; another concurrently running task might use the data from the “invalid state”, resulting in confusing bugs. Such bugs are often said to result from a “race condition”, where two tasks “race” towards a goal and bungle each others’ shared resources in the process. One of our best tools for preventing “race condition” bugs is a concurrency lock. Concurrency locks lock code execution to one task while the lock is in place. More on locks soon.
Race condition bugs can occur in both async
/await
code running on a single thread and threaded code running on multiple threads. Both concurrency paradigms can create “invalid states” and release code execution so that other tasks see the “invalid state” shared resources before fixing them. However, threaded code is more likely to result in a bug than async
/await
code primarily because of who controls when one task gives up execution to another task.
With the async
/await
code, we explicitly set places in our code to wait and let another task start working (with the await
statement or through async with
context managers). Between those special waiting places, we are sure the same task is executing code continuously.
However, with threaded code, the CPU controls which task is executing at any given moment, and the CPU tends to rapidly alternate between different tasks (threads), sometimes even alternating which thread is executing in the middle of a line of code. Therefore, there are more places where one task might create an “invalid state” for another task to “see” in threaded code. Ensuring tasks can safely run with threads so tasks never switch during an invalid state is called “thread safety.”
Sync bank transfers
Let’s look at some examples. First, we’ll expand on our banking example above. This example makes banking transactions between bank customers synchronously, one after the other:
# locks/1_sync.py
"""Run simulated banking transactions."""
import random
import time
from rich import print
BANK_DATA = {
"bank_1": 1000,
"bank_2": 1000,
"bank_3": 1000,
"bank_4": 1000,
"bank_5": 1000,
}
def main() -> None:
"""Run the main program."""
t0 = time.time()
print(f"\ninitial={BANK_DATA}\n", flush=True)
run_transactions()
total_seconds = time.time() - t0
print(f"\n\nfinal={BANK_DATA}", flush=True)
print(f"sum={sum(BANK_DATA.values())}", flush=True)
print(
f"\n[bold green]Code run in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
def run_transactions() -> None:
"""Run a series of bank transactions."""
for _ in range(20):
banks = list(BANK_DATA.keys())
sending_bank = random.choice(banks)
banks.remove(sending_bank)
receiving_bank = random.choice(list(banks))
amount = random.randint(1, 100)
run_transaction(sending_bank, receiving_bank, amount)
def run_transaction(sending_bank: str, receiving_bank: str, amount) -> None:
"""Run a bank transaction."""
verify_user()
update_bank(sending_bank, -amount)
update_bank(receiving_bank, amount)
print(".", end="", flush=True)
def verify_user() -> None:
"""Simulate verifying the user can make the transaction."""
time.sleep(0.1)
def update_bank(bank_name: str, amount: int) -> None:
"""Update the bank data."""
amount_before = BANK_DATA[bank_name]
time.sleep(0.0001)
new_amount = amount_before + amount
BANK_DATA[bank_name] = new_amount
time.sleep(0.0001)
if __name__ == "__main__":
main()
Here’s the results of that python script run on the command line:
❯ python locks/1_sync.py
initial={'bank_1': 1000, 'bank_2': 1000, 'bank_3': 1000, 'bank_4': 1000,
'bank_5': 1000}
....................
final={'bank_1': 846, 'bank_2': 986, 'bank_3': 1072, 'bank_4': 1056, 'bank_5':
1040}
sum=5000
The code ran in 2.08 seconds.
And a video in real-time, with color:
The example is contrived. It uses the random
module to transfer twenty bank account balances between five possible banks. It uses time.sleep
to simulate talking to external I/O services (such as a database). Conveniently for us, our operating system treats time.sleep
as an I/O-bound operation. One thread can progress on its sleep
while another actively executes code.
Notice a couple of opportunities for an “invalid state” in this example.
- In
run_transaction
, the state is invalid between subtracting money from one bank account and adding it to the next. - In
update_bank
, first, we get a bank’s account balance, and after a tiny delay, we update the bank account balance. Between the retrieving and the updating, the bank account balance information is stale, and another bank account could theoretically alter the balance in that window–an opportunity for an “invalid state”.
However, since The code ran synchronously, we can guarantee that these “invalid states” are fixed before a transaction finishes, and thus, all is correct when the script completes.
Unsafe async bank transfers
Let’s speed up these bank transfers by converting them to the async
/await
syntax to run all bank transfers concurrently. First, we’ll run the bank transfers without any locks and see how we get a “race condition” bug.
# locks/2_async_unsafe.py
"""Run simulated banking transactions."""
import asyncio
import random
import time
from rich import print
BANK_DATA = {
"bank_1": 1000,
"bank_2": 1000,
"bank_3": 1000,
"bank_4": 1000,
"bank_5": 1000,
}
def main() -> None:
"""Run the main program."""
t0 = time.time()
print(f"\ninitial={BANK_DATA}\n", flush=True)
asyncio.run(run_transactions())
total_seconds = time.time() - t0
print(f"\n\nfinal={BANK_DATA}", flush=True)
print(f"sum={sum(BANK_DATA.values())}", flush=True)
print(
f"\n[bold green]The code ran in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
async def run_transactions() -> None:
"""Run a series of bank transactions."""
coroutines = []
for _ in range(20):
banks = list(BANK_DATA.keys())
sending_bank = random.choice(banks)
banks.remove(sending_bank)
receiving_bank = random.choice(list(banks))
amount = random.randint(1, 100)
coroutines.append(run_transaction(sending_bank, receiving_bank, amount))
await asyncio.gather(*coroutines)
async def run_transaction(sending_bank: str, receiving_bank: str, amount) -> None:
"""Run a bank transaction."""
await verify_user()
await update_bank(sending_bank, -amount)
await update_bank(receiving_bank, amount)
print(".", end="", flush=True)
async def verify_user() -> None:
"""Simulate verifying the user can make the transaction."""
await asyncio.sleep(0.1)
async def update_bank(bank_name: str, amount: int) -> None:
"""Update the bank data."""
amount_before = BANK_DATA[bank_name]
await asyncio.sleep(0.0001)
new_amount = amount_before + amount
BANK_DATA[bank_name] = new_amount
await asyncio.sleep(0.0001)
if __name__ == "__main__":
main()
Here’s the results of that python script run on the command line:
❯ python locks/2_async_unsafe.py 17:08:46
initial={'bank_1': 1000, 'bank_2': 1000, 'bank_3': 1000, 'bank_4': 1000,
'bank_5': 1000}
....................
final={'bank_1': 1050, 'bank_2': 1087, 'bank_3': 988, 'bank_4': 981, 'bank_5':
908}
sum=5014
The code ran in 0.15 seconds.
And a video in real-time, with color:
Since I explained the async
/await
changes in detail for web scraping above, I won’t go into detail about the changes in this script. Suffice it to say, we convert a bunch of functions from def function
to async def function
and add await
statements to wait on coroutine functions and I/O-bound work. One noticeable change is I converted all time.sleep
functions to asyncio.sleep
functions, which are the same but are awaitable with the await
keyword.
What are the results of changing this script to use async
/await
concurrent tasks? For one, it completed more than 10X faster. For another, it created a race condition bug. The bug arises because of the two invalid state opportunities I listed previously:
- In
run_transaction
, the state is invalid between subtracting money from one bank account and adding it to the next. One bank fund transfer can use the “invalid” funds from one bank account that is mid-transfer of another transaction. - In
update_bank
, first, we get a bank’s account balance, and after a tiny delay, we update the bank account balance. Between the retrieving and the updating, the bank account balance information is stale, and another bank account could theoretically alter the balance in that window–an opportunity for an “invalid state”.
Safe async bank transfers
How do we fix this bug? The most straightforward fix is to introduce a lock which looks like this:
# locks/2_async_safe.py
"""Run simulated banking transactions."""
import asyncio
import random
import time
from rich import print
BANK_DATA = {
"bank_1": 1000,
"bank_2": 1000,
"bank_3": 1000,
"bank_4": 1000,
"bank_5": 1000,
}
TRANSACTION_LOCK = asyncio.Lock()
def main() -> None:
"""Run the main program."""
t0 = time.time()
print(f"\ninitial={BANK_DATA}\n", flush=True)
asyncio.run(run_transactions())
total_seconds = time.time() - t0
print(f"\n\nfinal={BANK_DATA}", flush=True)
print(f"sum={sum(BANK_DATA.values())}", flush=True)
print(
f"\n[bold green]The code ran in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
async def run_transactions() -> None:
"""Run a series of bank transactions."""
coroutines = []
for _ in range(20):
banks = list(BANK_DATA.keys())
sending_bank = random.choice(banks)
banks.remove(sending_bank)
receiving_bank = random.choice(list(banks))
amount = random.randint(1, 100)
coroutines.append(run_transaction(sending_bank, receiving_bank, amount))
await asyncio.gather(*coroutines)
async def run_transaction(sending_bank: str, receiving_bank: str, amount) -> None:
"""Run a bank transaction."""
await verify_user()
async with TRANSACTION_LOCK:
await update_bank(sending_bank, -amount)
await update_bank(receiving_bank, amount)
print(".", end="", flush=True)
async def verify_user() -> None:
"""Simulate verifying the user can make the transaction."""
await asyncio.sleep(0.1)
async def update_bank(bank_name: str, amount: int) -> None:
"""Update the bank data."""
amount_before = BANK_DATA[bank_name]
await asyncio.sleep(0.0001)
new_amount = amount_before + amount
BANK_DATA[bank_name] = new_amount
await asyncio.sleep(0.0001)
if __name__ == "__main__":
main()
Here’s the results of that python script run on the command line:
❯ python locks/2_async_safe.py ✭ ✱ 14:30:43
initial={'bank_1': 1000, 'bank_2': 1000, 'bank_3': 1000, 'bank_4': 1000,
'bank_5': 1000}
....................
final={'bank_1': 966, 'bank_2': 1122, 'bank_3': 1034, 'bank_4': 931, 'bank_5':
947}
sum=5000
The code ran in 0.24 seconds.
And a video in real-time, with color:
The only changes between this script and the previous script are here:
import asyncio
...
TRANSACTION_LOCK = asyncio.Lock()
...
async def run_transaction(sending_bank: str, receiving_bank: str, amount) -> None:
...
async with TRANSACTION_LOCK:
await update_bank(sending_bank, -amount)
await update_bank(receiving_bank, amount)
We use an asyncio.Lock()
while executing the two update_bank
functions. The lock prevents the event loop from releasing the current task to other tasks when it hits await
statements as it usually does. The result? The bug goes away. Because of the lock, the task fixes the “invalid state” before releasing execution to another task. However, the script also runs slower because it doesn’t release for some of the waiting steps and thus runs less code concurrently.
Unsafe threaded bank transfers
Alternatively to async
/await
, we can perform the bank transfers concurrently with threads. Again, I’ll first show how adding threads introduces a “race condition” bug. Here’s the code:
# locks/3_threaded_unsafe.py
"""Run simulated banking transactions."""
import random
import time
from concurrent.futures import ThreadPoolExecutor
from rich import print
BANK_DATA = {
"bank_1": 1000,
"bank_2": 1000,
"bank_3": 1000,
"bank_4": 1000,
"bank_5": 1000,
}
def main() -> None:
"""Run the main program."""
t0 = time.time()
print(f"\ninitial={BANK_DATA}\n", flush=True)
run_transactions()
total_seconds = time.time() - t0
print(f"\n\nfinal={BANK_DATA}", flush=True)
print(f"sum={sum(BANK_DATA.values())}", flush=True)
print(
f"\n[bold green]The code ran in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
def run_transactions() -> None:
"""Run a series of bank transactions."""
tasks = []
for _ in range(20):
banks = list(BANK_DATA.keys())
sending_bank = random.choice(banks)
banks.remove(sending_bank)
receiving_bank = random.choice(list(banks))
amount = random.randint(1, 100)
tasks.append(
# func, args, kwargs
(run_transaction, (sending_bank, receiving_bank), {"amount": amount})
)
with ThreadPoolExecutor() as executor:
[executor.submit(func, *args, **kwargs) for func, args, kwargs in tasks]
def run_transaction(sending_bank: str, receiving_bank: str, amount) -> None:
"""Run a bank transaction."""
verify_user()
update_bank(sending_bank, -amount)
update_bank(receiving_bank, amount)
print(".", end="", flush=True)
def verify_user() -> None:
"""Simulate verifying the user can make the transaction."""
time.sleep(0.1)
def update_bank(bank_name: str, amount: int) -> None:
"""Update the bank data."""
amount_before = BANK_DATA[bank_name]
time.sleep(0.0001)
new_amount = amount_before + amount
BANK_DATA[bank_name] = new_amount
time.sleep(0.0001)
if __name__ == "__main__":
main()
Here’s the results of that python script run on the command line:
❯ python locks/3_threaded_unsafe.py 17:11:10
initial={'bank_1': 1000, 'bank_2': 1000, 'bank_3': 1000, 'bank_4': 1000,
'bank_5': 1000}
....................
final={'bank_1': 1118, 'bank_2': 1056, 'bank_3': 945, 'bank_4': 775, 'bank_5':
1152}
sum=5046
The code ran in 0.36 seconds.
And a video in real-time, with color:
The difference between this code and the synchronous code is that the bank transaction tasks are run in a ThreadPoolExecutor
in the run_transactions
function:
from concurrent.futures import ThreadPoolExecutor
...
def run_transactions() -> None:
tasks = []
for _ in range(20):
...
tasks.append(
# func, args, kwargs
(run_transaction, (sending_bank, receiving_bank), {"amount": amount})
)
with ThreadPoolExecutor() as executor:
[executor.submit(func, *args, **kwargs) for func, args, kwargs in tasks]
Here, we see the same race condition bug as we do with the async
/await
script before adding locks. The fix is almost identical to the async
/await
fix: add a lock.
Safe threaded bank transfers
Here’s the code for safe threaded bank transfers after adding a lock:
# locks/3_threaded_safe.py
"""Run simulated banking transactions."""
import random
import time
from concurrent.futures import ThreadPoolExecutor
import threading
from rich import print
BANK_DATA = {
"bank_1": 1000,
"bank_2": 1000,
"bank_3": 1000,
"bank_4": 1000,
"bank_5": 1000,
}
TRANSACTION_LOCK = threading.RLock()
def main() -> None:
"""Run the main program."""
t0 = time.time()
print(f"\ninitial={BANK_DATA}\n", flush=True)
run_transactions()
total_seconds = time.time() - t0
print(f"\n\nfinal={BANK_DATA}", flush=True)
print(f"sum={sum(BANK_DATA.values())}", flush=True)
print(
f"\n[bold green]The code ran in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
def run_transactions() -> None:
"""Run a series of bank transactions."""
tasks = []
for _ in range(20):
banks = list(BANK_DATA.keys())
sending_bank = random.choice(banks)
banks.remove(sending_bank)
receiving_bank = random.choice(list(banks))
amount = random.randint(1, 100)
tasks.append(
# func, args, kwargs
(run_transaction, (sending_bank, receiving_bank), {"amount": amount})
)
with ThreadPoolExecutor() as executor:
[executor.submit(func, *args, **kwargs) for func, args, kwargs in tasks]
def run_transaction(sending_bank: str, receiving_bank: str, amount) -> None:
"""Run a bank transaction."""
verify_user()
with TRANSACTION_LOCK:
update_bank(sending_bank, -amount)
update_bank(receiving_bank, amount)
print(".", end="", flush=True)
def verify_user() -> None:
"""Simulate verifying the user can make the transaction."""
time.sleep(0.1)
def update_bank(bank_name: str, amount: int) -> None:
"""Update the bank data."""
amount_before = BANK_DATA[bank_name]
time.sleep(0.0001)
new_amount = amount_before + amount
BANK_DATA[bank_name] = new_amount
time.sleep(0.0001)
if __name__ == "__main__":
main()
Here’s the results of that python script run on the command line:
❯ python locks/3_threaded_safe.py
initial={'bank_1': 1000, 'bank_2': 1000, 'bank_3': 1000, 'bank_4': 1000,
'bank_5': 1000}
....................
final={'bank_1': 1071, 'bank_2': 894, 'bank_3': 1029, 'bank_4': 1129, 'bank_5':
877}
sum=5000
The code ran in 0.35 seconds.
And a video in real-time, with color:
Here’s just the code changed to add the locks. Notice we use threading.RLock()
instead of threading.Lock()
. threading.RLock()
is usually what you want as it won’t deadlock when it encounters the same lock (for instance, with recursive code).
import threading
...
TRANSACTION_LOCK = threading.RLock()
...
def run_transaction(sending_bank: str, receiving_bank: str, amount) -> None:
...
with TRANSACTION_LOCK:
update_bank(sending_bank, -amount)
update_bank(receiving_bank, amount)
As with async
/await
, a thread lock fixes the “race condition” bug and slightly slows down the code.
CPU bound parallelism
As stated earlier in this post, typically, Python code only executes one instruction at a time, even with threads, due to the GIL. Our examples thus far have achieved concurrency by allowing Python code to run in the foreground while I/O waiting happens in the background. Can multiple Python processes be executed in parallel, leveraging all our CPU cores instead of just one? Yes, with the multiprocessing
module.
CPU-bound Sync example
I’ll lay out a contrived example to show how multiprocessing works. First, lets see an example of CPU-bound bound (no I/O work) code run synchronously:
# cpu_bound/1_sync.py
"""Do some CPU bound work synchronously."""
import time
from rich import print
def main():
print("Starting tasks...", flush=True)
t0 = time.time()
results = do_lots_of_math()
total_seconds = time.time() - t0
print(
f"\n[bold green]The code ran in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
print(f"\n{results=}", flush=True)
def do_lots_of_math() -> list[float]:
return [do_math_once(num) for num in range(1, 21)]
def do_math_once(starting_number: int = 1) -> float:
"""Do some CPU bound work."""
print(f"[yellow]Doing math, {starting_number=}...", flush=True)
x = starting_number
for _ in range(2_000_000):
x **= 4
x **= 0.25
x **= 2
x **= 0.5
x += 1
print(f"[green]Done with math, {starting_number=}.", flush=True)
return x
if __name__ == "__main__":
main()
Here’s the results of that python script run on the command line:
❯ python cpu_bound/1_sync.py
Starting tasks...
Doing math, starting_number=1...
Done with math, starting_number=1.
Doing math, starting_number=2...
Done with math, starting_number=2.
Doing math, starting_number=3...
Done with math, starting_number=3.
Doing math, starting_number=4...
Done with math, starting_number=4.
Doing math, starting_number=5...
Done with math, starting_number=5.
Doing math, starting_number=6...
Done with math, starting_number=6.
Doing math, starting_number=7...
Done with math, starting_number=7.
Doing math, starting_number=8...
Done with math, starting_number=8.
Doing math, starting_number=9...
Done with math, starting_number=9.
Doing math, starting_number=10...
Done with math, starting_number=10.
Doing math, starting_number=11...
Done with math, starting_number=11.
Doing math, starting_number=12...
Done with math, starting_number=12.
Doing math, starting_number=13...
Done with math, starting_number=13.
Doing math, starting_number=14...
Done with math, starting_number=14.
Doing math, starting_number=15...
Done with math, starting_number=15.
Doing math, starting_number=16...
Done with math, starting_number=16.
Doing math, starting_number=17...
Done with math, starting_number=17.
Doing math, starting_number=18...
Done with math, starting_number=18.
Doing math, starting_number=19...
Done with math, starting_number=19.
Doing math, starting_number=20...
Done with math, starting_number=20.
The code ran in 10.57 seconds.
results=[2000001.0, 2000002.0, 2000003.0, 2000004.0, 2000005.0, 2000006.0,
2000007.0, 2000008.0, 2000009.0, 2000010.0, 2000011.0, 2000012.0, 2000013.0,
2000014.0, 2000015.0, 2000016.0, 2000017.0, 2000018.0, 2000019.0, 2000020.0]
And a video of the code run in real-time, with color, alongside my task manager (a little slower than the above data for the recording):
Notice that the operations run sequentially and take a little over 10 seconds. Also, notice that one of my CPU cores shoots up to 100% usage during that time, and the others stay lower than 100%–although interestingly not at their previous levels.
The code performs simple, non-sensical math operations run over a loop with many iterations. The math isn’t important; it’s just the fact that it takes some time, and there are no I/O processes.
CPU-bound threaded example
I know I said threading won’t speed up CPU-bound operations, but let’s try anyway to see what happens:
# cpu_bound/2_threaded_no_improvement.py
"""Do some CPU bound work synchronously."""
import time
from concurrent.futures import ThreadPoolExecutor as PoolExecutor
from typing import Any, Callable
from rich import print
def main():
print("Starting tasks...", flush=True)
t0 = time.time()
results = do_lots_of_math()
total_seconds = time.time() - t0
print(
f"\n[bold green]The code ran in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
print(f"\n{results=}", flush=True)
TaskType = tuple[Callable[..., Any], tuple[Any, ...], dict[str, Any]]
def do_lots_of_math() -> list[float]:
print("Defining tasks...", flush=True)
tasks: list[TaskType] = [
# Function, args, kwargs
(do_math_once, (num,), {})
for num in range(1, 21)
]
print("Kick off multiprocess tasks...", flush=True)
with PoolExecutor() as executor:
work = [executor.submit(func, *args, **kwargs) for func, args, kwargs in tasks]
print("Waiting for work...", flush=True)
print("Done with work", flush=True)
return [future.result() for future in work]
def do_math_once(starting_number: int = 1) -> float:
"""Do some CPU bound work."""
print(f"[yellow]Doing math, {starting_number=}...", flush=True)
x = starting_number
for _ in range(2_000_000):
x **= 4
x **= 0.25
x **= 2
x **= 0.5
x += 1
print(f"[green]Done with math, {starting_number=}.", flush=True)
return x
if __name__ == "__main__":
main()
Here’s the results of that python script run on the command line:
❯ python cpu_bound/2_threaded_no_improvement.py
Starting tasks...
Defining tasks...
Kick off multiprocess tasks...
Doing math, starting_number=1...
Doing math, starting_number=2...
Doing math, starting_number=3...
Doing math, starting_number=4...
Doing math, starting_number=5...
Doing math, starting_number=6...
Doing math, starting_number=7...
Doing math, starting_number=8...
Waiting for work...
Done with math, starting_number=7.
Doing math, starting_number=9...
Done with math, starting_number=3.
Doing math, starting_number=10...
Done with math, starting_number=4.
Doing math, starting_number=11...
Done with math, starting_number=1.
Doing math, starting_number=12...
Done with math, starting_number=5.
Doing math, starting_number=13...
Done with math, starting_number=6.
Doing math, starting_number=14...
Done with math, starting_number=2.
Doing math, starting_number=15...
Done with math, starting_number=8.
Doing math, starting_number=16...
Done with math, starting_number=9.
Doing math, starting_number=17...
Done with math, starting_number=13.
Doing math, starting_number=18...
Done with math, starting_number=10.
Done with math, starting_number=11.
Doing math, starting_number=19...
Doing math, starting_number=20...
Done with math, starting_number=16.
Done with math, starting_number=14.
Done with math, starting_number=15.
Done with math, starting_number=12.
Done with math, starting_number=17.
Done with math, starting_number=18.
Done with math, starting_number=19.
Done with math, starting_number=20.
Done with work
The code ran in 10.79 seconds.
results=[2000001.0, 2000002.0, 2000003.0, 2000004.0, 2000005.0, 2000006.0,
2000007.0, 2000008.0, 2000009.0, 2000010.0, 2000011.0, 2000012.0, 2000013.0,
2000014.0, 2000015.0, 2000016.0, 2000017.0, 2000018.0, 2000019.0, 2000020.0]
And a video of the code run in real-time, with color, alongside my task manager (a little slower than the above data for the recording):
It appears that the threads are running simultaneously since the print statements for the operations occur out of order. However, the run time is about the same as the synchronous operations script. Also, notice that while our CPU usage increases on all cores, it never maxes out on any of them, as we might expect when fully utilizing our computer’s processing power.
CPU-bound multiprocessing example
In this final example, we use the multiprocessing module to run the CPU-bound operations in parallel on multiple cores.
# cpu_bound/3_multiprocess.py
"""Do some CPU bound work synchronously."""
import time
from concurrent.futures.process import ProcessPoolExecutor as PoolExecutor
from typing import Any, Callable
from rich import print
def main():
print("Starting tasks...", flush=True)
t0 = time.time()
results = do_lots_of_math()
total_seconds = time.time() - t0
print(
f"\n[bold green]The code ran in [cyan]{total_seconds:,.2f}[green] seconds.",
flush=True,
)
print(f"\n{results=}", flush=True)
TaskType = tuple[Callable[..., Any], tuple[Any, ...], dict[str, Any]]
def do_lots_of_math() -> list[float]:
print("Defining tasks...", flush=True)
tasks: list[TaskType] = [
# Function, args, kwargs
(do_math_once, (num,), {})
for num in range(1, 21)
]
print("Kick off multiprocess tasks...", flush=True)
with PoolExecutor() as executor:
work = [executor.submit(func, *args, **kwargs) for func, args, kwargs in tasks]
print("Waiting for work...", flush=True)
print("Done with work", flush=True)
return [future.result() for future in work]
def do_math_once(starting_number: int = 1) -> float:
"""Do some CPU bound work."""
print(f"[yellow]Doing math, {starting_number=}...", flush=True)
x = starting_number
for _ in range(2_000_000):
x **= 4
x **= 0.25
x **= 2
x **= 0.5
x += 1
print(f"[green]Done with math, {starting_number=}.", flush=True)
return x
if __name__ == "__main__":
main()
Here’s the results of that python script run on the command line:
❯ python cpu_bound/3_multiprocess.py ✗ ✭ ✱ 16:19:17
Starting tasks...
Defining tasks...
Kick off multiprocess tasks...
Waiting for work...
Doing math, starting_number=1...
Doing math, starting_number=2...
Doing math, starting_number=3...
Doing math, starting_number=4...
Done with math, starting_number=4.
Doing math, starting_number=5...
Done with math, starting_number=3.
Doing math, starting_number=6...
Done with math, starting_number=1.
Doing math, starting_number=7...
Done with math, starting_number=2.
Doing math, starting_number=8...
Done with math, starting_number=5.
Doing math, starting_number=9...
Done with math, starting_number=6.
Doing math, starting_number=10...
Done with math, starting_number=7.
Doing math, starting_number=11...
Done with math, starting_number=8.
Doing math, starting_number=12...
Done with math, starting_number=9.
Doing math, starting_number=13...
Done with math, starting_number=10.
Doing math, starting_number=14...
Done with math, starting_number=11.
Doing math, starting_number=15...
Done with math, starting_number=12.
Doing math, starting_number=16...
Done with math, starting_number=13.
Doing math, starting_number=17...
Done with math, starting_number=14.
Doing math, starting_number=18...
Done with math, starting_number=15.
Doing math, starting_number=19...
Done with math, starting_number=16.
Doing math, starting_number=20...
Done with math, starting_number=17.
Done with math, starting_number=18.
Done with math, starting_number=19.
Done with math, starting_number=20.
Done with work
The code ran in 2.99 seconds.
results=[2000001.0, 2000002.0, 2000003.0, 2000004.0, 2000005.0, 2000006.0,
2000007.0, 2000008.0, 2000009.0, 2000010.0, 2000011.0, 2000012.0, 2000013.0,
2000014.0, 2000015.0, 2000016.0, 2000017.0, 2000018.0, 2000019.0, 2000020.0]
And a video of the code run in real-time, with color, alongside my task manager (a little slower than the above data for the recording):
This time, you can see that all four of my CPU cores are fully utilized. The script runs slightly over 3X faster than the synchronous (and threaded) code, which aligns with my expectations. We could expect a 4X speedup since we use four CPU cores instead of one, minus the overhead costs (and CPU processing power used by other operations on my computer, such as the recording software).
Also, notice the difference between writing threaded code and multiprocessing code is minimal. The ProcessPoolExecutor
implements the same interface as the ThreadPoolExecutor
, and since I aliased both to PoolExecutor
, I only had to change one line of code (the rest of the code is the same):
from concurrent.futures import ThreadPoolExecutor as PoolExecutor
from concurrent.futures.process import ProcessPoolExecutor as PoolExecutor
ProcessPoolExecutor
is excellent because it intelligently runs the same number of multiprocesses as my computer has CPU cores and cleans up those processes when the with
block context manager closes.
Conclusions
In this article, I discussed key terms related to concurrency in Python and gave visual explanations of the three concurrent programming paradigms: async
/await
, threads and multiprocessing. Next, I went through concrete examples of running concurrent I/O-bound Python code by web scraping pages of a Pokémon website. Then I discussed “race condition” bugs in concurrent code and showed how to solve them with locks. Finally, I gave examples of running CPU-bound python code in parallel with the multiprocessing module. I hope this intro to concurrent programming gives you ideas for speeding up your next project!