Pure Python sucks in the scene of parallel computing, due to the existence of the Global Interpreter Lock (aka GIL). GIL prevents accessing or manipulating interpreter from different threads concurrently. The mechanism alleviates the risk of race condition, but sequentializes multi-threading program 1 as well. Sadly, there’s no way to release the lock from pure Python.

Alright. So what about beyond pure Python? Shall we bypass the mechanism within an extension? The answer is yes, and that’s what most of scientific computing libaries do.

Cython is a good choice for writing extensions, less verbose, and more similar to Python syntactically. In Cython, one can release GIL temporarily for a code block using the with nogil: syntax. Will it release the true power of multi-core CPU? We should have a try.

We adopt a toy example, say, a naive matrix multiplication, for benchmarking. Start with a C-only version:

#cython: boundscheck=False#cython: wraparound=False#cython: nonecheck=False#cython: cdivision=True#cython: languagelevel=3import numpy as npcimport numpy as npcdef void _matmul(    np.float_t[:, :] Av,    np.float_t[:, :] Bv,    np.float_t[:, :] Cv,    int M, int N, int P,) nogil:    cdef int i, j, k    for i in range(M):        for j in range(P):            for k in range(N):                Cv[i, j] += Av[i, k] * Bv[k, j]

The function above is straight-forward. We then create a wrapper for it, so that it can be called by Python code:

cpdef matmul(    np.ndarray[dtype=np.float_t, ndim=2] A,    np.ndarray[dtype=np.float_t, ndim=2] B,    object use_gil,):    cdef int M = A.shape[0]    cdef int N = A.shape[1]    cdef int P = B.shape[1]    C = np.zeros((M, P))    cdef np.float_t[:, :] Av = A, Bv = B, Cv = C    if use_gil:        _matmul(Av, Bv, Cv, M, N, P)    else:        with nogil:            _matmul(Av, Bv, Cv, M, N, P)    return C

Now the Cython part is ready. Below a script for benchmarking:

import timeitimport threadingimport itertoolsimport pyximportimport numpy as nppyximport.install(setup_args={"include_dirs": np.get_include()}, inplace=True, language_level=3)import matmulN = 1200A = np.random.rand(N, N)B = np.random.rand(N, N)def runner(nthreads: int, use_gil: bool) -> None:    args = (A, B, use_gil)    threads = []    for _ in range(nthreads):        threads.append(threading.Thread(target=matmul.matmul, args=args))        threads[-1].start()    for thread in threads:        thread.join()def make_grid(**kwargs):    space = [[(name, v) for v in lst] for name, lst in kwargs.items()]    return map(dict, itertools.product(*space))for kw in make_grid(        nthreads=[1, 2],        use_gil=[True, False],):    print(kw)    print(timeit.timeit("runner(**kw)", globals=dict(runner=runner, kw=kw), number=1))

Two matrices with a rather large size 1200 x 1200 are supplied as input, and we test matmul against four settings. The result is listed as below:

nthreadsGILtime (s)
1N3.47
1Y3.51
2N3.78
2Y6.96

The first two rows show that, with single thread, matmul has comparable performance no matter releasing GIL or not. This is desired behavior, since GIL should not lead to performance degradation in single-threading scene. But things change when it comes to multi-threading. With two computing threads running in parallel, the time doubles if holding GIL, whilst in another setting (GIL released), the performance remains unchanged.

We may step further to investigate the behavior of prange. prange is provided by Cython for more convenient parallel computing, adopting the famous OpenMP as backend. Writing a prange version _matmul should take minor modification:

cdef void _matmul_p(    np.float_t[:, :] Av,    np.float_t[:, :] Bv,    np.float_t[:, :] Cv,    int M, int N, int P,) nogil:    cdef int i, j, k, ij    cdef int MP = M * P    for ij in prange(MP, schedule='guided'):        i = ij // P        j = ij % P        for k in range(N):            Cv[i, j] += Av[i, k] * Bv[k, j]

plus the wrapper matmul:

cpdef matmul(    np.ndarray[dtype=np.float_t, ndim=2] A,    np.ndarray[dtype=np.float_t, ndim=2] B,    object use_gil,    # hl: begin    object parallel,    # hl: end):    cdef int M = A.shape[0]    cdef int N = A.shape[1]    cdef int P = B.shape[1]    C = np.zeros((M, P))    cdef np.float_t[:, :] Av = A, Bv = B, Cv = C    if use_gil:        # hl: begin        if parallel:            _matmul_p(Av, Bv, Cv, M, N, P)        else:            _matmul(Av, Bv, Cv, M, N, P)        # hl: end    else:        # hl: begin        if parallel:            _matmul_p(Av, Bv, Cv, M, N, P)        else:            with nogil:                _matmul(Av, Bv, Cv, M, N, P)        #hl: end    return C

and also, the benchmark script:

# hl: begindef runner(nthreads: int, use_gil: bool, parallel: bool) -> None:    args = (A, B, use_gil, parallel)    # hl: end    if nthreads == 0:        matmul.matmul(*args)        return    threads = []    for _ in range(nthreads):        threads.append(threading.Thread(target=matmul.matmul, args=args))        threads[-1].start()    for thread in threads:        thread.join()
for kw in make_grid(        nthreads=[0, 1, 2],        use_gil=[True, False],        # hl: begin        parallel=[True, False],        # hl: end):    print(kw)    print(timeit.timeit("runner(**kw)", globals=dict(runner=runner, kw=kw), number=1))

OpenMP requires extra compilation flags, so a .pyxbld file is needed:

# matmul.pyxbldfrom setuptools import Extensionfrom Cython.Build import cythonizedef make_ext(modname, pyxfilename):    ext = Extension(        modname,        [pyxfilename],        extra_compile_args=['-fopenmp'],        extra_link_args=['-fopenmp'],    )    return cythonize(        [ext],        language_level=3,        annotate=True,    )[0]

nthreadsGILtime w/o par. (s)time w/ par. (s)
1N3.470.82
1Y3.510.89
2N3.781.79
2Y6.961.81

We can see that prange brings an amazing boost in performance! _matmul_p is 3~4x faster in single-threading setting. The number might vary across different hardwares, depending on the number of CPU cores. In the setting of two threads, the running time doubles, which indicates that prange does efficiently use up all available CPU resources.

We can also notice that, whether to release GIL or not seemingly does not affect prange 2. The reason is prange requires GIL to be released, which is automatically done by default.

Cython supports native parallelism through the cython.parallel module. To use this kind of parallelism, the GIL must be released (see Releasing the GIL). It currently supports OpenMP, but later on more backends might be supported. – Using Parallelism

Conclusion

If there’s no need to hold GIL, just release it. This happens when you are manipulating some C data structures, and not attempting to disturb the interpreter.

If there’s massive looping in your Cython code, feel free to accelerate it with prange. prange will effeciently schedule the computation onto all CPU resources.

If there’s some macro tasks 3 which could not be easily parallelized in Cython, schedule them via threading module. threading sucks for most of the time, but if the tasks not always acquiring GIL, it should be fine just like threads in other languages.

  1. so that it behaves just like a single-threading version
  2. 0.82s vs 0.89s
  3. routines consisting of large pieces of logic