8.1 Functional programming#

Estimated time for this notebook: 20 minutes

We have previously seen the object-oriented style of programming, and how to organise our code according to it using objects, classes and inheritance. While widely-adopted and very useful, this is not the only way of writing code. The functional paradigm, as the name suggests, emphasises functions as building blocks of programs.

Understanding to think in a functional programming style is almost as important as object orientation for building DRY, clear scientific software, and is just as conceptually difficult. However, being aware of different paradigms and styles gives you access to more techniques that you can use to write, structure and reason about your code.

Functions within functions#

Programs are composed of functions: they take data in (which we call parameters or arguments) and send data out (through return statements).

A conceptual trick which is often used by computer scientists to teach the core idea of functional programming is this: to write a program, in theory, you only ever need functions with one argument, even when you think you need two or more. Why?

Let’s define a program to add two numbers:

def add(a, b):
    return a + b

add(5, 6)

How could we do this, in a fictional version of Python which only defined functions of one argument? In order to understand this, we’ll have to understand several of the concepts of functional programming. Let’s start with a program which just adds five to something:

def add_five(a):
    return a + 5


OK, we could define lots of these, one for each number we want to add. But that would be infinitely repetitive. So, let’s try to metaprogram that: we want a function which returns these add_N() functions.

Let’s start with the easy case: a function which returns a function which adds 5 to something:

def generate_five_adder():
    def _five_adder(a):
        return a + 5

    return _five_adder

coolfunction = generate_five_adder()

OK, so what happened there? Well, we defined a function inside the other function. We can always do that:

def thirty_function():
    def times_three(a):
        return a * 3

    def add_seven(a):
        return a + 7

    return times_three(add_seven(3))


When we do this, the functions enclosed inside the outer function are local functions, and can’t be seen outside:

NameError                                 Traceback (most recent call last)
Cell In[5], line 1
----> 1 add_seven

NameError: name 'add_seven' is not defined

There’s not really much of a difference between functions and other variables in python. A function is just a variable which can have () put after it to call the code!

<function thirty_function at 0x7fdac46bd160>
x = [thirty_function, add_five, add]
for fun in x:
<function thirty_function at 0x7fdac46bd160>
<function add_five at 0x7fdac476aaf0>
<function add at 0x7fdac476aee0>

And we know that one of the things we can do with a variable is return it. So we can return a function, and then call it outside:

def deferred_greeting():
    def greet():

    return greet

friendlyfunction = deferred_greeting()
# Do something else
print("Just passing the time...")
Just passing the time...
# OK, Go!

So now, to finish this, we just need to return a function to add an arbitrary amount:

def generate_adder(increment):
    def _adder(a):
        return a + increment

    return _adder

add_3 = generate_adder(3)

We can make this even prettier: let’s make another variable pointing to our generate_adder() function:

add = generate_adder

And now we can do the real magic:


In summary, we have started with a function that takes two arguments (add(a, b)) and replaced it with a new function (add(a)(b)). This new function takes a single argument, and returns a function that itself takes the second argument.

This may seem like an overly complicated process - and, in some cases, it is! However, this pattern of functions that return functions (or even take them as arguments!) can be very useful. In fact, it is the basis of decorators, a Python feature that we will discuss more in this chapter [notebook].


You may have noticed something a bit weird:

In the definition of generate_adder, increment is a local variable. It should have gone out of scope and died at the end of the definition. How can the amount the returned adder function is adding still be kept?

This is called a closure. In Python, whenever a function definition references a variable in the surrounding scope, it is preserved within the function definition.

You can close over global module variables as well:

name = "Eric"

def greet():
    print("Hello, ", name)

Hello,  Eric

And note that the closure stores a reference to the variable in the surrounding scope: (“Late Binding”)

name = "John"

Hello,  John

Map and Reduce#

We often want to apply a function to each variable in an array, to return a new array. We can do this with a list comprehension:

numbers = range(10)

[add_five(i) for i in numbers]
[5, 6, 7, 8, 9, 10, 11, 12, 13, 14]

But this is sufficiently common that there’s a quick built-in:

list(map(add_five, numbers))
[5, 6, 7, 8, 9, 10, 11, 12, 13, 14]

This map operation is really important conceptually when understanding efficient parallel programming: different computers can apply the mapped function to their input at the same time. We call this Single Program, Multiple Data (SPMD). map is half of the map-reduce functional programming paradigm which is key to the efficient operation of much of today’s “data science” explosion.

Let’s continue our functional programming mind-stretch by looking at reduce operations.

We very often want to loop with some kind of accumulator (an intermediate result that we update), such as when finding a sum:

def summer(data):
    total = 0.0

    for x in data:
        total += x

    return total

or finding a maximum:

import sys

def my_max(data):
    # Start with the smallest possible number
    highest = -sys.float_info.max

    for x in data:
        if x > highest:
            highest = x

    return highest
my_max([2, 5, 10, -11, -5])

These operations, where we have some variable which is building up a result, and the result is updated with some operation, can be gathered together as a functional program, taking in (as an argument) the operation to be used to combine results:

def accumulate(operation, data, initial):
    accumulator = initial
    for x in data:
        accumulator = operation(accumulator, x)
    return accumulator

def my_sum(data):
    def _add(a, b):
        return a + b

    return accumulate(_add, data, 0)
def bigger(a, b):
    if b > a:
        return b
    return a

def my_max(data):
    return accumulate(bigger, data, -sys.float_info.max)

my_max([2, 5, 10, -11, -5])

Anyway, this accumulate-under-an-operation process is so fundamental to computing that it’s usually in standard libraries for languages which allow functional programming:

from functools import reduce

def my_max(data):
    return reduce(bigger, data, -sys.float_info.max)

my_max([2, 5, 10, -11, -5])

Efficient map-reduce#

Now, because these operations, bigger and _add, are such that e.g. (a+b)+c = a+(b+c) , i.e. they are associative, we could apply our accumulation to the left half and the right half of the array, each on a different computer, and then combine the two halves:

1 + 2 + 3 + 4 = (1 + 2) + (3 + 4)

Indeed, with a bigger array, we can divide-and-conquer more times:

1 + 2 + 3 + 4 + 5 + 6 + 7 + 8 = ((1 + 2) + (3 + 4)) + ((5 + 6) + (7 + 8))

So with enough parallel computers, we could do this operation on eight numbers in three steps: first, we use four computers to do one each of the pairwise adds.

Then, we use two computers to add the four totals.

Then, we use one of the computers to do the final add of the two last numbers.

You might be able to do the maths to see that with an N element list, the number of such steps is proportional to the logarithm of N.

We say that with enough computers, reduction operations are O(ln N)

This course isn’t an introduction to algorithms, but we’ll talk more about this O() notation when we think about programming for performance.

Lambda Functions#

When doing functional programming, we often want to be able to define a function on the fly:

def most_Cs_in_any_sequence(sequences):
    def count_Cs(sequence):
        return sequence.count("C")

    counts = map(count_Cs, sequences)
    return max(counts)

def most_Gs_in_any_sequence(sequences):
    return max(map(lambda sequence: sequence.count("G"), sequences))



The syntax here means that these two definitions are identical:

func_name = lambda a, b, c: a + b + c

def func_name(a, b, c):
    return a + b + c

The lambda keyword defines an “anonymous” function.

def most_of_given_base_in_any_sequence(sequences, base):
    return max(map(lambda sequence: sequence.count(base), sequences))

most_of_given_base_in_any_sequence(data, "A")

The above fragment defined a lambda function as a closure over base. If you understood that, you’ve got it!

To double all elements in an array:

data = range(10)
list(map(lambda x: 2 * x, data))
[0, 2, 4, 6, 8, 10, 12, 14, 16, 18]
[2 * x for x in data]
[0, 2, 4, 6, 8, 10, 12, 14, 16, 18]

Similarly, to find the maximum value in a sequence:

def my_max(data):
    return reduce(lambda a, b: a if a > b else b, data, -sys.float_info.max)

my_max([2, 5, 10, -11, -5])

Using functional programming for numerical methods#

Probably the most common use in research computing for functional programming is the application of a numerical method to a function.

Consider this example which uses the newton function from SciPy, a root-finding function implementing the Newton-Raphson method. The arguments we pass to newton are the function whose roots we want to find, and a starting point to search from.

We will be using this to find the roots of the function \(f(x) = x^2 - x\).

%matplotlib inline
from matplotlib import pyplot as plt
from numpy import linspace, zeros
from scipy.optimize import newton

solve_me = lambda x: x**2 - x

for x0 in [2, 0.2]:
    answer = newton(solve_me, x0)
    print(f"Starting from {x0}, the root I found is {answer}")

xs = linspace(-1, 2, 50)
solved = [xs, list(map(solve_me, xs)), xs, zeros(len(xs))]

Starting from 2, the root I found is 1.0
Starting from 0.2, the root I found is -3.441905100203782e-21
[<matplotlib.lines.Line2D at 0x7fda84179eb0>,
 <matplotlib.lines.Line2D at 0x7fda84179f10>]

Sometimes such tools return another function, for example the derivative of their input function. This is what a naive implementation of that could look like:

def derivative_simple(func, eps, at):
    return (func(at + eps) - func(at)) / eps
def derivative(func, eps):
    def _func_derived(x):
        return (func(x + eps) - func(x)) / eps

    return _func_derived

straight = derivative(solve_me, 0.01)

The derivative of solve_me is \(f'(x) = 2x - 1\), which represents a straight line. We can verify that our computations are correct, i.e. that the returned function straight matches \(f'(x)\), by checking the value of straight at some \(x\):


or by plotting it:

derived = (xs, list(map(solve_me, xs)), xs, list(map(derivative(solve_me, 0.01), xs)))
print(newton(derivative(solve_me, 0.01), 0))

Of course, coding your own numerical methods is bad, because the implementations you develop are likely to be less efficient, less accurate and more error-prone than what you can find in existing established libraries.

For example, the above definition could be replaced by:

import scipy.misc

def derivative(func):
    def _func_derived(x):
        return scipy.misc.derivative(func, x)

    return _func_derived

newton(derivative(solve_me), 0)

If you’ve done a moderate amount of calculus, then you’ll find similarities between functional programming in computer science and Functionals in the calculus of variations.