home

Blog

 < December 2013 > Su Mo Tu We Th Fr Sa 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31

past articles »

San Francisco, USA

 Putting Cython into use, I have great success in speeding up a computation algorithm. Previously I have great success of getting 10x speed increase just by swapping in pypy. This time, with a little bit of work, I get 177x speed improvement using Cython. This vastly exceed my expectation. The code is to solve a string alignment problem in bioinformatics. With a string s and t, the algorithm has complexity of O(|s||t|). The first step of writing Cython is to identify the performance critical region of code. In this case it is quite obvious the bottleneck is in the inner loop with complexity of O(n^2). No profiling is needed. The origin Python code inner loop is like below. Note that M and B are numpy arrays. ```def overlap_alignment_inner(s, t, sigma, M, B): for i in range(1,M.shape): for j in range(1,M.shape): match_score = M[i-1, j-1] + (1 if (s[i-1] == t[j-1]) else -2) M[i,j], B[i,j] = max([ (M[i-1, j] - sigma, (i-1, j )), (M[i, j-1] - sigma, (i , j-1)), (match_score, (i-1, j-1)), ]) ``` After a few iterations, I have arrived with the optimized code below. It looks somewhat different from the first glance. But I would walk through the chances I have made. I have taken a lot of clues from the tutorial Working with NumPy. ```import numpy as np cimport numpy as np DTYPE = np.int ctypedef np.int_t DTYPE_t def overlap_alignment_inner(s, t, int sigma, np.ndarray[DTYPE_t, ndim=2] M, np.ndarray[DTYPE_t, ndim=3] B): cdef int i, j cdef int s10, s01, s11 for i in range(1,M.shape): for j in range(1,M.shape): s01 = M[i, j-1] - sigma s10 = M[i-1, j] - sigma s11 = M[i-1, j-1] + (1 if (s[i-1] == t[j-1]) else -2) if s11 >= s01 and s11 >= s10: M[i,j] = s11 B[i,j,0] = i-1 B[i,j,1] = j-1 elif s01 >= s10 and s01 >= s11: M[i,j] = s01 B[i,j,0] = i B[i,j,1] = j-1 else: M[i,j] = s10 B[i,j,0] = i-1 B[i,j,1] = j ``` The first thing to do is to add type to certain variable for early binding. The simple int declaration below increase speed by about 10%. cdef int i, j The next thing to do is to type numpy ndarray objects like below. This allow faster indexing compares to normal Python operations. This speed things up a few times. ... np.ndarray[DTYPE_t, ndim=2] M, np.ndarray[DTYPE_t, ndim=3] B ... The most significant problem turn out to be the use of the max function. In the original code it is a concise and stylish way to pick the best score and simultaneously assign two values to M and B. But this keep the statement as a costly Python function call. By unwinding the function into if statements, it allow the code to be fully optimized and attained the 177x speed improvement! This brings the performance to the league of C. The unwind code, while longer, is actually quite straight forward. So I backported it to the pure Python code. This also result in a 3x improvement! Turns out use of max here is rather costly. My first use of Cython is very successful. By identifying and optimizing only a few lines of critical code, it dramatically speed up the performance to the level of C. The rest of the code are not performance critical. They remain easy to write and debug using Python. 2013.12.19 - comments     Please enable JavaScript to view the comments powered by Disqus. blog comments powered by Disqus past articles »

BBC News

 • • Xi defends Hong Kong rule on handover anniversary (01 Jul 2022) • • • • • Bison attacks woman at Yellowstone National Park (01 Jul 2022) • Supreme Court limits Biden's power to cut emissions (30 Jun 2022) • •

SF Gate Site feed Updated: 2022-Jul-01 07:00