You are here

addendum: algorithm for irr calculation

29 April, 2015 - 17:03

irr is a guessing algorithm similiar to binary searching for bugs in a program. For a project with a positive IRR, then two relatively prime factors could be used , one to increase the irr (multiple by factor > 1) if the npv is still positive and the other to decrease the irr when npv negative (divide by factor > 1), so that the tried values of the irr are unlikely to cycle, due to the lack of common factors in multiply and divide.

Below is a python algorithm for irr calculation, which also deals with going to a negative IRR when the guessing shows the irr is approaching zero.

# copyright GNU license, sjt,2014

def irr(cashflows):

#print cashflows

r = 10.

lim =0.1

sign = 1

while True:

npv = reduce( lambda x,(i,cf): x + cf / pow(1 + r/100, i ) , \

enumerate(cashflows), 0.)

print npv

#cf= []

#for i,x in enumerate(cashflows):

# lr = pow((1 + r/100), i)

# y = x / lr

# print "lr", lr, "y", y

# cf.append(y)

#print reduce( lambda x, y: x + y, cf)

if abs(npv) < lim:

  break

if npv * sign > 0:

r = r * 1.9

else:

r = r / 1.3

if sign > 0 and r < 0.0001:

r = - 20.

sign = -1

print "irr = ", r

cashflows=[-60000, 7000,9000,6000,5000,8000]

irr(cashflows)

This is a rather randomised algorithm; Newton's method has been mentioned as being used in open source spreadsheet programs.

Newton's method relies on the differentiation , or finding a derivative function to tell the slope of the graph of another function. The basic rule is that if there is a term which is axn then the derivative has a term , n.axn-1.

The npv function is a function of R the rate of return, such that npv is y, and R is x in the general function y = f(x) . Hence the derivative of the npv function is dy/dx = d(npv)/d(R) , which then requires finding the derivatives of all the terms in the npv function.

In Newton's method, an arbitrary R which is x, gives a NPV , which is a y value , when divided by the gradient which is delta-y/delta-x , gives delta-x , which is subtracted from R ( substitute x) to give another R' that is closer to the IRR, when NPV = 0 ( or the x value where y is zero, or where the function crosses the x-axis ). Using R' as the next R, this is repeated until a good approximation of IRR is achieved. The gradient is always recalculated at the start of each iteration , using the derivative function.

So if npv = CF0 / R0 + CF1 / R1 + .. + CFn / Rn,

or npv = CF0 + CF1 x R-1 + CF2 x R-2 + .. + CFn x R-n,

then d(npv)/dR = 0 + - CF1x R -2 + - 2 CF2x R -3 ... + -n x CFn x R-n - 1.

See

  1. REDIRECT Computation instead of Table Lookup