Many times a scientist is choosing a programming language or a software for a specific purpose. For the field of scientific computing, the methods for solving differential equations are one of the important areas. What I would like to do is take the time to compare and contrast between the most popular offerings. This is a good way to reflect upon what's available and find out where there is room for improvement. I hope that by giving you the details for how each suite was put together (and the "why", as gathered from software publications) you can come to your own conclusion as to which suites are right for you.

(Full disclosure, I am the lead developer of DifferentialEquations.jl. You will see at the end that DifferentialEquations.jl does offer pretty much everything from the other suite combined, but that's no accident: our software organization came last and we used these suites as a guiding hand for how to design ours.)

## Quick Summary Table

If you just want a quick summary, I created a table which has all of this information. You can find it here (click for PDF):

## MATLAB's Built-In Methods

Due to its popularity, let's start with MATLAB's built in differential equation solvers. MATLAB's differential equation solver suite was described in a research paper by its creator Lawerance Shampine, and this paper is one of the most highly cited SIAM Scientific Computing publications. Shampine also had a few other papers at this time developing the idea of a "methods for a problem solving environment" or a PSE. The idea is pretty simple: users of a problem solving environment (the examples from his papers are MATLAB and Maple) do not have the same requirements as more general users of scientific computing. Instead of focusing on efficiency, they key for this group is to have a clear and neatly defined (universal) interface which has a lot of flexibility.

The MATLAB ODE Suite does extremely well at hitting these goals. MATLAB documents its ODE solvers very well, there's a similar interface for using each of the different methods, and it tells you in a table in which cases you should use the different methods.

But the modifications to the methods goes even further. Let's take for example the classic ode45. That method just works and creates good plots, right? Well, Shampine added a little trick to it. When you solve an equation using ode45, the Runge-Kutta method uses a "free" interpolation to fill in some extra points. So between any two steps that the solver takes, it automatically adds in 4 extra points using a 4th order interpolation. This is because high order ODE solvers are good enough at achieving "standard user error tolerances" that they actually achieve quite large timesteps, and in doing so step too infrequently to make a good plot. Shampine's scheme is a good quick fix to this problem which most people probably never knew was occurring under the hood!

There's quite a bit of flexibility here. The methods allow you to use complex numbers. You're given access to the "dense output function" (this is the function which computes the interpolations). There's a few options you can tweak. Every one of these methods is setup with event handling, and there are methods which can handle differential-algebraic equations. There are also dde23 and ddesd for delay differential equations, and in the financial toolbox there's an Euler-Maruyama method for SDEs.

While MATLAB does an excellent job at giving a large amount of easily available functionality, where it lacks is performance. There's a few reasons for this. For one thing, these modifications like adding extra points to the solution array can really increase the amount of memory consumed if the ODE system is large! This actually has an impact in other ways. There's a very good example of this in ode45. ode45 is based on the Dormand-Prince 5(4) pair. However, in 1999, the same year the MATLAB ODE Suite was published, Shampine released a paper with a new 5(4) pair which was more efficient than the Dormand-Prince method. So that begs the question, why wasn't this used in the MATLAB ODE Suite (it's clear Shampine knew about it!)? (I actually asked him in an email) The reason is because its interpolation requires calculating some extra steps, so it's less efficient if you are ALWAYS interpolating. But since ode45 is always interpolating in order to make the plots look nicer, this would get in the way. Essentially, it can be more efficient, but MATLAB sets things up for nice plotting and not pure efficiency.

But there are other areas where more efficient methods were passed up during the development phase of the ODE suite. For example, Hairer's benchmarks in his book Solving Ordinary Differential Equations I and II (the second is for stiff problems), along with the benchmarks from the Julia DifferentialEquations.jl suite, consistently show that high order Runge-Kutta methods are usually the most efficient methods for high accuracy solving of nonstiff ODEs. These benchmarks both consistently show that, for the same error, high order Runge-Kutta methods (like order >6) can solve the equation much faster than methods like Adams methods. But MATLAB does not offer high order Runge-Kutta methods and only offers ode113 (an Adams method) for high-accuracy solving.

Some of this is due to a limitation within MATLAB itself. MATLAB's ODE solver requires taking in a user-defined function, and since this function is defined in MATLAB its function calls are very inefficient and expensive. Thus MATLAB's ODE solver suite can become more efficient by using methods which reduce the number of function calls (which multistep methods do). But this isn't the only case where efficient methods are missing. Both Hairer and the JuliaDiffEq benchmarks show that high-order Rosenbrock methods are the most efficient for low-medium accuracy stiff ODEs, but MATLAB doesn't offer these methods. It does offer ode23s, a low-order Rosenbrock method, and ode15s which is a multistep method. ode15s is supposed to be the "meat and potatoes of the MATLAB Suite", but it cannot handle all equations of since its higher order methods (it's adaptive order) are not L-stable (and not even A-stable). For this reason there's a few other low order SDIRK methods (ode23tb, an ESDIRK method for highly stiff problems) which are recommended to fill in the gaps, but none of the higher order variants which are known to be more efficient for many equations.

This pattern goes beyond the ODE solvers. The DDE solvers are all low order, and in the case of ddesd, it's a low accuracy method which is fast for getting plots correct but not something which converges to many decimal places all too well since it doesn't explicitly track discontinuities. This is even seen in the paper on the method which shows the convergence only matches dde23 to graphical accuracy on a constant-delay problem. Again, this fits in with the mantra of the suite, but may not hit all demographics. Shampine specifically made a separate version of ddesd for Fortran for people who are interested in efficiency, which is another way of noting that the key of ddesd is features and automatic usage, and not hardcore scientific computing efficiency. The mentioned SDE solver from the financial toolbox is only order 0.5 and thus requires quite a small dt to be accurate.

And I can keep going, but I think you get the moral of the story. This suite was created with one purpose in mind: to make it very easy to solve a wide array of differential equations and get a nice plot out. It does a very good job at doing so. But it wasn't made with efficiency in mind, and so it's missing a lot of methods that may be useful if you want high efficiency or high accuracy. Adding these wouldn't make sense to the MATLAB suite anyways since it would clutter the offering. MATLAB is about ease of use, and not efficiency, and it does extremely well at what it set out to do. For software that was first made before the Y2K crisis with just a few methods added later, it still holds up very well.

What would make sense to add to MATLAB are some "addon routines" though. Estimation of parameters in ODEs/DDEs, sensitivity analysis, etc. These are the kinds of things which users "can" do, but might not necessarily do "correctly" or in full (for example, for estimation of parameters you really want to use sensitivity analysis or autodifferentiation for the gradients, and in many cases you'll want to be using global or semi-global optimization methods. This is in contrast to simple Levenberg-Marquardt that I see code suggestions for in online threads). AMIGO2 is a good MATLAB toolbox to add to the standard library which adds these kinds of functionality, but MATLAB still gets a "None" in this category since I am only considered the built in features there. But while I'm at it, MATLAB does have a good Runge-Kutta Nystrom implementation in the FileExchange and it really should be cleaned up and added to the standard library. I think this kind of focus on expanding features would be a good fit for MATLAB.

## Hairer's Solvers (Fortran)

Next I want to bring up some Fortran solvers because they will come up later. Hairer's Fortran solvers are a set methods with similar interfaces that were designed with efficiency in mind. Many of these methods are classics: dopri5, dop853, radau, and rodas will specifically show up in many of the suites which are discussed later. These methods are not too flexible: they don't allow event handling (though with enough gusto you can use the dense output to write your own), or numbers that aren't double-precision floating point numbers (it's Fortran). They have a good set of options for tweaking parameters to make the adaptive timestepping more efficient, though you may have to read a few textbooks to know exactly what they do. And that's the key to them: they will solve an ODE, stiff or non-stiff, and they will do so pretty efficiently, but nothing more. But even then, they show some age which don't make them "perfectly efficient". These solvers include their own linear algebra routines which don't multithread like standard BLAS and LAPACK implementations, meaning that they won't make full use of modern CPU architectures. The computations don't necessarily SIMD or use FMA. But most of all, to use it directly you need to use Fortran which would be turn off for many people.

There is some flexibility in the offering. There are symplectic solvers for second order ODEs, the stiff solvers allow for solving DAEs in mass matrix form, there's a constant-lag nonstiff delay differential equation solver (RETARD), there is a fantastic generalization of radau to stiff state-dependent delay differential equations (RADAR5), and there's some solvers specifically for some "mechanical ODEs" commonly found in physical problems. Of course, to get this all working you'll need to be pretty familiar with Fortran, but this is a good suite to look at if you are.

## ODEPACK and Netlib ODE Solvers (Fortran)

ODEPACK is an old set of ODE solvers which accumulated over time. You can learn a bit about its history by reading this interview with Alan Hindenmarsh. I also bundle the Netlib suite together with it. There's a wide variety of methods in there. There's old Runge-Kutta methods due to Shampine, some of Shampine's old multistep methods ddebdf and ddeabm, etc. The reason why this pack is really important though is because of the Lawarance Livermore set of methods, specifically LSODE and its related methods (LSODA, LSODR, VODE, etc.). It includes methods for implicit ODEs (DAEs) as well. These are a group of multistep methods which are descendent from GEAR, the original BDF code. They are pretty low level and thus allow you to control the solver step by step, and some of them have "rootfinding capabilities". This means that you can use these to add event handling, though an event handling interface will take some coding. There's lots of options and these are generally pretty performant for a large array of problems, but they do show their age in the same way that the Hairer codes do. Their linear algebra setups do not make use of modern BLAS and LAPACK, which in practical terms means they don't make full use of modern computer architectures to speed things up. They don't always make use of SIMD and other modern CPU acceleration features. They are the vanilla ODE solvers which have existed through time. In particular, LSODA is popular because this is the most widely distributed method which automatically detects stiffness and switches between integration methods, though it should be pointed out that there is a performance penalty from this feature.

While its ODE solvers are not particularly interesting anymore (pretty much every ODE solver since was created to improve upon the designs here), there are many non-standard pieces that you still cannot find elsewhere. For example, the BVP solver COLDAE is probably the best BVP solver for DAEs still. It has a multiple shooting BVP solver as well which allows you to give it integration schemes (MUS). It has Runge-Kutta Chevyschev (RKC) methods for first and second order ODEs which are highly-stable methods for semi-stiff equations (typically noted as good for parabolic PDEs). It also has an implicit RKC method as well. It has a MIRK-based adaptive BVP solver, similar to MATLAB's bvp4c but with a Fortran implementation. And the implicit ODE (DAE) solvers DASSL, DASKR, and DASPK were the precursors to Sundials' IDA and have different characteristics which can be good/bad depending on the problem. Altogether, there's some interesting stuff here, but the interfaces are not too uniform.

## Sundials and ARKCODE (C++ and Fortran)

Sundials' CVODE is a re-write of VODE to C which is a descendent of LSODE which is a descendent itself of the original GEAR multistep code. Yes, this has a long history. But you should think of it as "LSODE upgraded": it makes use of modern BLAS/LAPACK, and also a bunch of other efficient C/Fortran linear solvers, to give a very efficient Adams and BDF method. Its solver IDA is like CVODE but handles implicit ODEs (DAEs). The interface for these is very similar to the ODEPACK interface, which means you can control it step by step and use the rootfinding capabilities to write your own event handling interface. Since the Adams methods handle nonstiff ODEs and the BDF methods handle stiff ODEs, this performance plus flexibility makes it the "one-stop-shop" for ODE solving. Many different scientific computing software internally are making use of Sundials because it can handle just about anything. Well, it does have many limitations. For one, this only works with standard C++ numbers and arrays. There's no stochasticity or delays allowed. But more importantly, Hairer and DifferentialEquations.jl's show that these multistep methods are usually not the most efficient methods since high order Rosenbrock, (E)SDIRK, and fully implicit RK methods are usually faster for the same accuracy for stiff problems, and high order Runge-Kutta are faster than the Adams methods for the same accuracy. Multistep methods are also not very stable at their higher orders, and so at higher tolerances (lower accuracy) these methods may fail to have their steps converge on standard test problems (see this note in the ROBER tests), meaning that you may have to increase the accuracy (and thus computational cost) due to stiffness issues. But, since multistep methods only require a single function evaluation per step (or are implicit in only single steps), they are the most efficient for asymptotically hard problems (i.e. when the derivative calculation is very expensive or the ODE is a large 10,000+ system). For this reason, these methods excel at solving large discretizations of PDEs. To top it off, there are parallel (MPI) versions for using CVODE/IDA for HPC applications.

I also want to note that recently Sundials added ARKCODE, a set of Runge-Kutta methods. These include explicit Runge-Kutta methods and SDIRK methods, including additive Runge-Kutta methods for IMEX methods (i.e. you can split out a portion to solve explicitly so that the implicit portion is cheaper when you know your problem is only partly or semi stiff). This means that it covers the methods which I mentioned earlier are more efficient in many of the cases (though it is a bit lacking on the explicit tableaus and thus could be more efficient, but that's just details).

If you are using C++ or Fortran and want to write to only one interface, the Sundials suite is a great Swiss Army knife. And if you have an asymtopically large problem or very expensive function evaluations, this will be the most efficient as well. Plus the parallel versions are cool! You do have to live with the limitations of low-level software forcing specific number and array types, along with the fact that you need to write your own event handling, but if you're "hardcore" and writing in a compiled language this suite is a good bet.

## SciPy's Solvers (Python)

Now we come to SciPy's suite. If you look at what it offers, the names should now start to sound familiar: LSODA, VODE, dopri5, dop853. That is right: SciPy is simply a wrapper over Hairer's and ODEPACK's methods. Because it writes a generic interface though, it doesn't have the granularity that is offered by ODEPACK, meaning that you don't have step-by-step control and no event handling. The tweaking options are very basic as well. Basically, they let you define a function in Python, say what timeframe to solve it on, give a few tolerances, and it calls Fortran code and spits back a solution. It only has ODE solvers, no differential-algebraic, delay, or stochastic solvers. It only allows the basic number types and does no event handling. Super vanilla, but gets the job done? As with the methods analysis, it has the high order Runge-Kutta methods for efficient solving of nonstiff equations, but it's missing Rosenbrock and SDIRK methods entirely, opting to only provide the multistep methods.

I should note here that it has the same limitation as MATLAB though, namely that the user's function is Python code. Since the derivative function is where the ODE solver spends most of its time (for sufficiently difficult problems), this means that even though you are calling Fortran code, you will lose out on a lot of efficiency. Still, if efficiency isn't a big deal and you don't need bells and whistles, this suite will do the basics.

## deSolve Package (R)

There's not much to say other than that deSolve is very much like SciPy's suite. It wraps almost the same solvers, has pretty much the same limitations, and has the same efficiency problem since in this case it's calling the user-provided R function most of the time. One advantage is that it does have event handling. Less vanilla with a little bit more features, but generally the same as SciPy.

I do want to make a note about its delay differential equation solvers dede. The documentation specifically says:

dede does not deal explicitly with propagated derivative discontinuities, but relies on the integrator to control the stepsize in the region of a discontinuity.

dede does not include methods to deal with delays that are smaller than the stepsize, although in some cases it may be possible to solve such models.

For these reasons, it can only solve rather simple delay differential equations.

As a methods researcher, I am horrified by this. Here's a few things wrong with it:

1. The error estimator is only at the end of the interval, meaning its a very bad check for discontinuities. This is why ddesd used a special residual. None of their linked methods do this.
2. Not propagating continuities means order loss to 1.
3. They allow using a multistep method over a discontinuity. This might converge to order 1? There's no theory for this.
4. When delays are smaller than the step size, the method is actually implicit, even if the ODE solver it's using is explicit. What this means is that this algorithm in this case is essentially a zero iteration Picard solver for the fixed point problem. This is probably the worst part.

They end with saying it's only for "simple" delay equations without mentioning these as the reasons. I think that's going too far: there is no good guarantee that the error is low or the method converges well. I will mark it as "Poor" in the table, but I would advise against using this method (and would recommend the developers to remove this from the docs because this is definitely an easy place for users to be mislead by a numerical result).

## PyDSTool (Python)

PyDSTool is an odd little beast. Part of the software is for analytic continuation (i.e. bifurcation plotting). But another part of it is for ODE solvers. It contains one ODE solver which is written in Python itself and it recommends against actually using this for efficiency reasons. Instead, it wraps some of the Hairer methods, specifically dopri5 and radau, and recommends these. But it's different than SciPy in that it takes in the specification of the ODE as a string, and compiles it to a C function, and uses this inside the ODE solver. By doing so, it's much more efficient. We still note that its array of available methods is small and it offers radau which is great for high accuracy ODEs and DAEs, but is not the most efficient at lower accuracy so it would've been nice to see Rosenbrock and ESDIRK methods. It has some basic event handling and methods for DDEs (again as wrappers to a Hairer method). This is a pretty good suite if you can get it working, though I do caution that getting the extra (non-Python) methods setup and compiled is nontrivial. One big point to note is that I find the documentation spectacularly difficult to parse. Together, it's pretty efficient and has a good set of methods which will do basic event handling and solve problems at double precision.

## JiTCODE and JiTCSDE (Python)

JiTCODE is another Python library which makes things efficient by compiling the function that the user provides. It uses the SciPy integrators and does something similar to PyDSTool in order to get efficiency. I haven't tried it out myself but I'll assume this will get you as efficient as though you used it from Fortran. However, it is lacking in the features department, not offering advanced things like arbitrary number types, event handling, etc. But if you have a vanilla ODE to solve and you want to easily do it efficiently in Python, this is a good option to look at.

Additionally, JiTCDDE is a version for constant-lag DDEs similar to dde23. JiTCSDE is a version for stochastic differential equations. It uses the high order (strong order 1.5) adaptive Runge-Kutta method for diagonal noise SDEs developed by Rackauckas (that's me) and Nie which has been demonstrated as much more efficient than the low order and fixed timestep methods found in the other offerings. It employs the same compilation setup as JitCODE so it should create efficient code as well. I haven't used this myself but it would probably be a very efficient ODE/DDE/SDE solver if you want to use Python and don't need events and other sugar.

## Boost ODEINT Solver Library (C++)

The Boost ODEINT solver library has some efficient implementations of some basic explicit Runge-Kutta methods (including high order RK methods) and some basic Rosenbrock methods (including a high order one). Thus it can be pretty efficient for solving the most standard stiff and nonstiff ODEs. However, its implementations do not make use of advanced timestepping techniques with its explicit RK method (PI-controllers) which makes it require more steps to achieve the same accuracy as some of the more advanced software, making it not really any more efficient in practice (though it does do Gustafsson acceleration in its Rosenbrock scheme to prevent "the Hump" behavior). It doesn't have event handling, but it is flexible with the number and array types you can put in there via C++ templates. It and DifferentialEquations.jl are the only two suites that are mentioned that allow for solving the differential equations on the GPU. Thus if you are familiar with templates and really want to make use of them, this might be the library to look at, otherwise you're probably better off looking elsewhere like Sundials.

## GSL ODE Solvers (C)

To say it lightly, the GSL ODE solvers are kind of a tragedy. Actually, they are just kind of weird. When comparing their choices against what is known to be efficient according to the ODE research and benchmarks, the methods that they choose to implement are pretty bizarre like extrapolation methods which have repeatedly been shown to not be very efficient, while not included other more important methods. But they do have some of the basics like multistep Adams and BDF methods. This, like Boost, doesn't do all of the fancy PI-controlled adaptivity and everything though, so YMMV. This benchmark, while not measuring runtime and only uses function evaluations (which can be very very different according to more sophisticated benchmarks like the Hairer and DifferentialEquations.jl ones!), clearly shows that the GSL solvers can take way too many function evaluations because of this and thus, since it's using methods similar to LSODA/LSODE/dopri5, probably have much higher runtimes than they should.

## Mathematica's ODE Solvers

Mathematica's ODE solvers are very sophisticated. It has a lot of explicit Runge-Kutta methods, including newer high order efficient methods due to Verner and Shampine's efficient method mentioned in the MATLAB section. These methods can be almost 5x faster than the older high order explicit RK methods which themselves are the most efficient class of methods for many nonstiff ODEs, and thus these do quite well. It
includes interpolations to make their solutions continuous functions that plot nicely. Its native methods are able to make full use of Mathematica and its arbitrary precision, but sadly most of the methods it uses are wrappers for the classic solvers. Its stiff solvers mainly call into Sundials or LSODA. By using LSODA, it tends to be able to do automatic stiffness detection by default. It also wraps Sundials' IDA for DAEs. It uses a method of steps implementation over its explicit Runge-Kutta methods for solving nonstiff DDEs efficiently, and includes high order Runge-Kutta methods for stochastic differential equations (though it doesn't do adaptive timestepping in this case). One nice feature is that all solutions come with an interpolation to make them continuous. Additionally, it can use the symbolic form of the user-provided equation in order to create a function for the Jacobian, and use that (instead of a numerical differentiation fallback like all of the previous methods) in the stiff solvers for more accuracy and efficiency. It also has symplectic methods for solving second order ODEs, and its event handling is very expressive. It's very impressive, but since it's using a lot of wrapped methods you cannot always make use of Mathematica's arbitrary precision inside of these numerical methods. Additionally, its interface doesn't give you control over all of the fine details of the solvers that it wraps.

## Maple's ODE Solvers

Maple's ODE solvers has the basics. It for some reason has Fehlberg and Cash-Karp order 5/4 methods instead of the Dormand-Prince, Bogacki-Shampine, or Tsitorious ones. It has a classic higher order explicit RK method due to Verner (not the newer more efficient ones though), and for stiff problems it uses a high order Rosenbrock method. It also wraps LSODE like everything else. Its native methods can make use of arbitrary arithmetic and complex numbers. It has a BVP solver, but it explicitly says that it's not for stiff equations or ones with singularities (it relies on deferred correction or extrapolation). It can solve state-dependent delay differential equations with its explicit RK and Rosenbrock methods. It can solve SDEs using Euler-Maruyama. It has event handling which doesn't seem to be well documented. As another symbolic programming language, it computes Jacobians analytically to pass to the stiff solvers like Mathematica, helping it out in that respect. Thus while some of its basics are a little odd (why a Fehlberg method? The Dormand-Prince paper is one of the most famous papers about increasing efficiency in explicit RK tableaus) and not as fleshed out, it does have a good variety with quite a bit of capabilities. I would actually go as far as to say that, if I had to create a suite with a small set of solvers I'd probably make something similar to what Maple has. It matches what seems to benchmark best "for most problems in normal tolerances", which is what most people are probably solving.

## FATODE (Fortran)

FATODE is a set of methods written in Fortran. It includes explicit Runge-Kutta methods, SDIRK methods, Rosenbrock methods and fully implicit RK methods. Thus it has something that's pretty efficient for pretty much every case. What makes it special is that it includes the ability to do sensitivity analysis calcuations. It can't do anything but double-precision numbers and doesn't have event handling, but the sensitivity calculations makes it quite special. If you're a FORTRAN programmer, this is worth a look, especially if you want to do sensitivity analysis.

## DifferentialEquations.jl (Julia)

Okay, time for DifferentialEquations.jl. I left it for last because it is by far the most complex of the solver suites, and pulls ideas from many of them. While most of the other suite offer no more than about 15 methods on the high end (with most offering about 8 or less), DifferentialEquations.jl offers 200+ methods and is continually growing. Like the standard Python and R suites, it offers wrappers to Sundials, ODEPACK, and Hairer methods. However, since Julia code is always JIT compiled, its wrappers are more akin to PyDSTool or JiTCODE in terms of efficiency. Thus all of the standard methods mentioned before are available in this suite.

But then there are the native Julia methods. For ODEs, these include explicit Runge-Kutta methods, (E)SDIRK methods, and Rosenbrock methods. In each of these categories it has a huge amount of variety, offering pretty much every method from the other suites along with some unique methods. Some unique methods to point out are that it has the only 5th order Rosenbrock method, it has the efficient Verner methods discussed in the Mathematica part, it has newer 5th order methods (i.e. it includes the Bogacki-Shampine method discussed as an alternative to ode45's tableau, along with an even newer tableau due to Tsitorious which is even more efficient). It has methods specialized to reduce interpolation error (the OwrenZen methods), and methods which are strong-stability preserving (for hyperbolic PDEs). It by default the solutions as continuous functions via a high order interpolation (though this can be turned off to make the saving more efficient). Each of these implementations employ a lot of extra tricks for efficiency. For example, the interpolation is "lazy", meaning that if the method requires extra function evaluations for the continuous form, it will only do those extra calculations when the continuous function is used (so when you directly ask for it or when you plot). This is just a peek at the special things the library does to gain an edge.

The native Julia methods benchmark very well as well, and all of the benchmarks are openly available. Essentially, these methods make use of the native multithreading of modern BLAS/LAPACK, FMA, SIMD, and all of the extra little compiler goodies that allows code to be efficient, along with newer solver methods which theoretically reduce the amount of work that's required to get the same error. They even allow you to tweak a lot of the internals and swap out the linear algebra routines to use parallel solvers like PETSc. The result is that these methods usually outperform the classic C/Fortran methods which are wrapped. Additionally, it has ways to symbolically calculate Jacobians like Mathematica/Maple, and instead of defaulting to numerical differentiation the stiff solvers fall back to automatic differentiation which is more efficient and has much increased accuracy.

Its event handling is the most advanced out of all of the offerings. You can change just about anything. You can make your ODE do things like change its size during the solving if you want, and you can make the event handling adjust and adapt internal solver parameters. It's not a hyperbole to say that the user is given "full control" since the differential equation solvers themselves are written as essentially a method on the event handling interface, meaning that anything it can do internally you can do.

The variety of methods is also much more diverse than the other offerings. It has symplectic integrators like Harier's suite, but has more high and low order methods. It has a range of Runge-Kutta Nystrom methods for efficiently solving second order ODEs. It has the same high order adaptive method for diagonal noise SDEs as JiTCSDE, but also includes high order adaptive methods specifically for additive noise SDEs. It also has methods for stiff SDEs in Ito and Stratanovich interpretations, and allows for event handling in the SDE case (with the full flexibility). It has DDE solvers for constant-lag and state-dependent delays, and it has stiff solvers for each of these cases. The stiff solvers also all allow for solving DAEs in mass matrix form (though fully implicit ODEs are possible, but can only be solved using a few methods like a wrapper to Sundials' IDA and doesn't include event handling here quite yet).

It allows arbitrary numbers and arrays like Boost. This means you can use arbitrary precision numbers in the native Julia methods, or you can use "array-like" objects to model multiscale biological organisms instead of always having to use simple contiguous arrays. It has addons for things like sensitivity analysis and parameter estimation. Also like Boost, it can solve equations on the GPU by using a GPUArray.

It hits the strong points of each of the previously mentioned suites because it was designed from the get-go to do so. And it benchmarks extremely well. The only downside is that, because it is so feature and performance focused, its documentation is heavy. The beginning tutorial will give you the basics (making it hopefully as easy as MATLAB to pick up), but pretty much every control knob is available, making the extended portion of the documentation a long read.

## Conclusion

Let's take a step back and summarize this information. DifferentialEquations.jl objectively has the largest feature-set, swamping most of the others while wrapping all of the common solvers. Since it also features solvers for the non-ordinary differential equations and its unique Julia methods also benchmarks well, I think it's clear that DifferentialEquations.jl is by far the best choice for "power-users" who are looking for a comprehensive suite.

As for the other choices from scripting languages, MATLAB wasn't designed to have all of the most efficient methods, but it'll handle basic equations with delays and events and output good plots. R's deSolve is similar in most respects to MATLAB. SciPy's offering is lacking in comparison to MATLAB and R's due to the lack of event handling. But MATLAB/Python/R all have efficiency problems due to the fact that the user's function is written in the scripting language. JiTCODE and PyDSTool are two Python offerings make the interface to the Fortran solvers more efficient than straight SciPy. Mathematica and Maple will do symbolic pre-calculations to speed things up and can JiT compile functions, along with offering pretty good event handling, and thus their wrappers are more like DifferentialEquations.jl in terms of flexibility and efficiency (and Mathematica had a few non-wrapper goodies mentioned as well). So in a pinch when not all of the bells and whistles are necessary, each of these scripting language suites will get you by. Behind DifferentialEquations.jl, I would definitely put Mathematica's or Maple's suite second for scripting languages with everything else much further behind. I was actually kind of surprised: normally people think of a CAS as for symbolic computing and think of MATLAB/R/Python as more "numerical" languages, but at least in the case of differential equation solvers the CASs (Mathematica and Maple) seem to be much more developed and complete.

If you're already hardcore and writing in C++/Fortran, Sundials is a pretty good "one-stop-shop" to get everything you need, especially when you add in ARKCODE. Still, you're going to have to write a lot of stuff yourself to make the rootfinding into an event handling interface, but if you put the work in
this will do you well. Hairer's codes are a great set of classics which cover a wide variety of equations, and FATODE is the only non-DifferentialEquations.jl suite which offers a way to calculate sensitivity equations (and its sensitivity equations are more advanced). Any of these will do you well if you want to really get down-and-dirty in a compiled language and write a lot of the interfaces yourself, but they will be a sacrifice in productivity with no clear performance gain over the scripting language methods which also include some form of JIT compilation. With these in mind, I don't really see a purpose for the GSL or Boost suites, and the ODEPACK methods are generally outdated.

I hope this review helps you make a choice which is right for you.

## Edits: 9/26/2017

Shortly after being posted this has received quite a bit of buzz. I am thankful so many people are interested! A few things were pointed out to me that I have since corrected:

1. I didn't have a row for implicitly defined DAEs in the chart. This has been added. The classic softwares were placed at excellent because they do give you step-by-step control of BDF methods to in theory allow you to build whatever interface you want on top of it. MATLAB and Mathematica are Good because you can interface with them and use their respective event handling interfaces. DifferentialEquations.jl is only fair since while we do wrap more of these than other software ecosystems, we don't have a way of using events or anything fancy here. In a few days I will be posting about the release of 3.0 and this is actually one of the big bullet points on our roadmap.
2. I upgraded the efficiency of Sundials to "Excellent" when ARKODE is considered. While I still think that they need newer/better explicit RK tableaus to really handle nonstiff equations to the best of their ability (the high order Verner methods for sure would be a great improvement), their offering does hit most of the other points (except for missing Rosenbrock methods) and it is very well-tuned. In addition, they make full use of additive RK methods for IMEX which can be very helpful in semi-stiff problems.
3. For now I am ignoring PDEs. The reason is because it's too hard to judge what matters. MATLAB and Mathematica both have diffusion-advection equation solvers, so does that mean that they are good for solving PDEs? Not necessarily to someone who wants to solve the Hamilton-Jacobi Bellman equation. The sensible thing to do instead would be to either do it at a very fine level and also compare the available toolboxes. But going through the details of the different finite element method toolboxes is a huge task. PDEs will likely be their own blog post in the future.

There are probably a few other corrections needed. For one, I may have put PyDSTool too high in some areas like state-dependent delay solvers (I had notes that say it can't do it, but I can't find it in the docs anymore...), so please feel free to send corrections.

## Edits 2: 9/26/2017

More edits today. This time for Maple. I missed some things that Maple has and updated the discussion about Maple to include its BVP/DDE solvers and its ability to handle arbitrary precision and complex numbers.

## Edits: 9/28/2017

Some more refinements. Mostly adding extra capabilities in the Netlib part of the chart (there's a lot of interesting stuff in there). Also, I wanted to note that PyODESys was pointed out to me which will take your ODE in Python and calculate Jacobians automatically and send them along to the ODE solver. While this is not part of any of the mentioned libraries and thus doesn't change the chart, this is something SciPy users might want to know about.

## Edits: 9/29/2017

This has become quite a living document. I got some questions about development goals, which was actually the purpose of this in the first place, so let me share some concise recommendations. These are recommendations for next developments which would be the most beneficial for each. I'll omit the tools which are essentially "done" like Netlib.

1. MATLAB. I would really like to see a higher-order explicit RK method (one of the Verner efficient ones) and an implementation of Radau. That would really round out the offering. Tie ins with symbolic differentiation and autodifferentiation tools would be very helpful as well. A much larger goal would be to add discontinuity tracking to their state-dependent delay equation solver.
2. SciPy. SciPy needs a lot of work. But two very simple steps they can take are to wrap Radau and Rodas. Since they already wrap two of Hairer's methods, it can't be that bad. A larger goal would be to expose dense output and add an event handling interface. I would like it if they picked a JiT setup to go along with their offering since that would go a long way, but it's okay if those remain extra tools.
3. R deSolve. Add a Rodas wrapper. Add a higher order explicit RK method. Once again, a way to compile user-functions would be great but is a larger project
4. DifferentialEquations.jl. Implicitly-defined ODEs can use some help since the solvers in this domain are just simple wrappers to IDA and DASKR, meaning that they don't get the same features (types, event handling, etc.) as the rest of the suite. The BVP solvers can do a lot of interesting stuff but have a weak core which should be improved.
5. Mathematica. Add some Rosenbrock methods. These methods do extremely well when Jacobians are given analytically, so Mathematica is really missing an opportunity more than others by not having them. Also a state-dependent delay equation solver. As noted in the comments, I think a good simple step for Mathematica would be adding some parameter inference routines as well which put together their sensitivity analysis and optimization into some easier setups.
6. Maple. Document what you are doing. Documentation is what I would like to see more of than anything else here. Adding a BVP solver which can handle stiff systems along with implicitly-defined DAEs would go a long way.
7. Boost. Add PI-controllers for the RK methods. It's not a hard change but can make a big difference. Also, it needs more than just a single 4th order Rosenbrock method for stiff equations. Probably add an adaptive order/timestep BDF for large equations, and a Radau.
8. PyDSTool. It has parameter estimation tools but just uses Levenberg-Marquardt and Bounded Minimization. At least add L-BFGS and some global optimizer and then this would be a really interesting part of the suite. Additionally, this is another package which could add Rodas pretty easily and should.

1. ## Tom

says:

The one thing that puts me off using DifferentialEquations.jl is the stability of the codebase. On several occasions I have run into bugs that make my code unstable and unusable. For instance, the DE works just fine if a parameter is set at 5.0 and 4.9, but at 4.95, for no physical reason, it breaks, but then at 4.95001 it runs fine again. The fact that the api changes ever-so-subtly between different solvers (some allow interpolation, others don't) doesn't help. Please don't hear me wrong: I applaud the wonderful work of the DifferentialEquations.jl team, you are building an incredible piece of kit for the open source community, and I cannot wait for this to be complete. But at present I'm sticking with Python or Matlab until this is just a little more stable.

• ## Christopher Rackauckas

says:

That's definitely valid on the edges. While the ODE solvers haven't changed in about 6 months, some of the stuff on the edges like stiff SDE solvers are a little new. Of course, that just takes time, tests, and bug reports.

"the DE works just fine if a parameter is set at 5.0 and 4.9, but at 4.95, for no physical reason, it breaks, but then at 4.95001 it runs fine again"

I have never heard of an issue like this. Is this recent at all? Did you file a bug report or ask in the chat channel?.

"some allow interpolation, others don't"

No, they all have an interpolant. Some only do linear interpolation or Hermite (3rd order) interpolation, whereas others do order matching. But that's all documented and sol(t) will work on any solution.

"I cannot wait for this to be complete"

Right now we're pretty feature complete, but finding the more subtle bugs requires bug reports. Python's solvers just wrap old solvers so those have been mostly fixed over time, and MATLAB's are now almost 20 years old. We'll get stable, but what's necessary is users letting us know when there are issues.

• ## Tom

says:

Hi Chris,

I feel bad about this comment, and would like to apologise. I haven't used DifferentialEquations.jl for a few months, and I shouldn't have complained without at least checking my complaints were still relevant. But of course I really didn't need to complain at all - you and others have put a whole load of work into this gratis. I still have a few frustrations with Julia not working quite 'as expected', and with code from 1-2 years ago breaking, but I suppose that's to be expected in any language pre-v1.0, and really nothing whatsoever to do with DifferentialEquations.jl

2. ## noop

says:

Hi Christopher,

Could you explain a bit what you are looking for in the Complex Numbers category?

As a side note concerning tweakability: maybe you have not seen the method framework for NDSolve. You can extend it with any method you like quite nicely. See here:

https://reference.wolfram.com/language/tutorial/NDSolvePlugIns.html

• ## Christopher Rackauckas

says:

>Could you explain a bit what you are looking for in the Complex Numbers category?

For Complex Numbers I am just looking to see that the integrators support solving equations with dependent variables that are complex numbers. Can it do it, and is there a good spread of methods that can do it (if it only has one it gets dinged).

I realize why you have this comment and I noticed that it was linking to an old version of the chart. That has been fixed.

>As a side note concerning tweakability: maybe you have not seen the method framework for NDSolve.

Ah yes, I didn't see this before. But I am not sure if the ability to write your own integrator is the same as being able to tweak all of the arguments of a solver, the latter is a much more labor-intensive solution.

3. ## Steven Johnson

says:

You keep referring to "BLAS/LINPACK" ... I think you mean "BLAS/LAPACK". LINPACK was the predecessor to LAPACK that did not use a separate BLAS.

• ## Christopher Rackauckas

says:

Yes, thanks for the correction. I always mix those up. I'll fix the text soon.

4. ## Michael Aye

says:

Wouldn't it be cool if we can have a set of simple benchmark DE simulations that could be directly compared in their set up complexity, code length and simplicity?
For example, this set of simple physical problems could be an excellent set for direct comparisons.

We could set it up as its own teaching GitHub repo where we have subfolders for each language, so that people getting into DE solving, could compare what they like best, and/or can find the example in the language they are forced to use for some reason.
What do people think? Would anybody interested in creating such GH repo?

• ## Christopher Rackauckas

says:

Yes, that is a great idea and an interesting next step. Getting people to put in the labor is always difficult, but I am willing to contribute as a teaching resource. I won't commit to building something like that until I'm actually doing it, but it is something I am looking into.

5. ## noop

says:

Here is a document that shows how one can tweak Linear/Nonlinear solver options and pass for example a preconditioner to NDSolve:

https://reference.wolfram.com/language/tutorial/NDSolveIDAMethod.html

So Control over Linear/Nonlinear solvers set to None does not seem right.

• ## Christopher Rackauckas

says:

I see. Thanks for pointing this out! This is exposing the options which naively exist in IDA (but not all of them actually). But that's a pretty good set. The downside is that it looks to be only one method (implicit RK doesn't get any of these options?) and you can't pass your own routines (your own geometric multigrid specific to the problem for example). I think it earns a "Fair" for having some Krylov method that exists and that you can use if needed, but not the tools to really customize it yourself very far.

6. ## noop

says:

Hi Chris,

you mention for Mathematica that there is no build in paralleism. Here is a counter example from the NDSolve ref page to show that 'None' is not correct:

pressureEqns = {p1[t] ==

!Rho] g h2[t]};
massConservation = {a1 h1'[t] == -flowRate, a2 h2'[t] == flowRate};
flowRate = (p1[t] - p2[t]) (

!Mu] pipeLen);
params = {pipeLen -> 0.1,
pipeDia -> 0.2,

!Mu] -> 2 10^-3, g -> 9.81};
r1 = 1; r2 = 1;
hemisphereArea[r_, h_] := Pi (2 r h - h^2);
a1 = hemisphereArea[r1, h1[t]];
a2 = hemisphereArea[r2, h2[t]];

sol2 = NDSolve[{pressureEqns, massConservation, h1[0] == 1,
h2[0] == 10^-5} /. params, {h1, h2, p1, p2}, {t, 0, 15}] ;

If you look at top or some such you will see that it uses all cores available.

• ## noop

says:

Sorry, here is a better input:

NDSolve[{{p1[t]==1.962 h1[t],p2[t]==1.962 h2[t]},{Pi (2 h1[t]-h1[t]^2) Derivative[1][h1][t]==-0.19634954084936215*(p1[t]-p2[t]),Pi (2 h2[t]-h2[t]^2) Derivative[1][h2][t]==0.19634954084936215*(p1[t]-p2[t])},h1[0]==1,h2[0]==1/100000},{h1,h2,p1,p2},{t,0,15}]

• ## Christopher Rackauckas

says:

Interesting! Thanks. Which method is that? If it's a stiff solver, is it just multithreading the linear algebra? Or is it also multithreading the RK methods as well? Additionally, does this parallelism work across multiple computers like a multi-node setup on an HPC?

• ## noop

says:

I know that linear algebra threads over all cores available and that HPC clustering is not possible with NDSolve out of the box. I think but I am not 100% sure that the RK are not parallel. But don't quote me on that.

• ## Christopher Rackauckas

says:

Yeah, this is the form of parallelism that I dinged in efficiency when it didn't exist (in Hairer and Netlib) but don't necessarily call "parallelism". I just kind of expect linear algebra to be multithreaded, so maybe there's a way to be more clear what I am looking at here. CVODE has MPI and DifferentialEquations.jl has a Monte Carlo setup so both will use multiple nodes on a cluster. DiffEq will also multithread internally any element-wise loop (so pretty much all of the method, an ODE solver is almost entirely element-wise computations) if you pass it an array with multithreaded broadcasts. These are the kinds of "you wouldn't expect it to parallelize, but it adds it" kinds of things, whereas if someone wraps the standard library for a matrix factorization it will multithread so that's not measured.

7. ## Shanky Jain

says:

Hi Christopher,
i know i have asked similar question on a different forum. it's nice to see you tried and put together summary comparison of many ODE solvers available. Looking at your summary table what category Monte-Carlo sims with ODE parameter variation come under, sensitivity analysis ?
Also, i still haven't decided on best route for my Monte-Carlo sims example running vanderpol oscillator for different ODE parameter taken from normal dist. for 4 sigma output study
From below possible routes, do you mind giving your feedback on how should i proceed
1.scipy odeint + cython
2.scipy odeint + uncertainities package
(just came across this package in below link and yet to try)
http://kitchingroup.cheme.cmu.edu/blog/category/ode/

3. Differential Equations package Julia

option 1 & 2 are low barrier routes for me as i'm familiar with python though no experience with cython. But also, if i can get a good solution in python i can combine it with pyspark cluster i've handy to parallelize multiple monte-carlo.
option 3 i'm very interested in pursuing but i have a big barrier entry to complete new Julia language.

Do you think i could gain may be an order or more in computation time in Julia compared to python+cython (not taki ng into account pyspark) ?

Final note, my current setup of actual stiff system runs good with ODE15s/23s in Matlab with main limitation of not being able to run monte-carlo, hence on the look-out for other options.

• ## Christopher Rackauckas

says:

>Looking at your summary table what category Monte-Carlo sims with ODE parameter variation come under, sensitivity analysis ?

Yes, I would put that in the more general "addons" category. Sensitivity analysis is different than Monte Carlo simulations to determine parameter sensitivities since sensitivity analysis instead solves the equation and computes the derivative of the solution with respect to local changes in the parameters. In some sense that's more refined than Monte Carlo, but Monte Carlo has its own advantages in that it doesn't require a localism argument (but is more costly).

If you want to go the sensitivity analysis route, one place to go is Sundials. The nice thing about Sundials is that you can find bindings for it in pretty much any language, like this:

https://github.com/ICB-DCM/AMICI

So that will let you use Sundials from MATLAB, thus making it easy to do the sensitivity analysis with little code churn. That might be a good strategy.

Otherwise let's say you want to do the Monte Carlo route. If your ODE is costly enough than just looping through solving it many times is not just a good strategy but the best strategy. One of the things I emphasized in here is that stiff solvers are almost always internally using BLAS these days, which means that they are parallelizing within the method for large enough stiff ODE systems (like discretizations of ODEs). In that case, you don't want to parallelize if you are using a single shared memory system. Of course, for these kinds of problems if you have an HPC you would instead want to use multiprocessing and have one process on each node. DifferentialEquations.jl has that setup in its Monte Carlo stuff, but it's easy enough to code up in R/Python with its multiprocessing package, or C/Fortran if you're familiar with MPI. But I do have to note that MATLAB has a pretty big limitation here due to how little control you have on how it distributes and the fact that you need a bunch of licenses for this (so you'd have to check with your HPC).

But that only covers the "large system of stiff ODEs" case. If the system is small enough like 10 ODEs, BLAS won't even be multithreading your lu-factorizations of 10x10 Jacobians (or if it is in whatever language you're using... set the number of BLAS threads to 1 to speed it up), in which case you'll want to multithread your Monte Carlo experiment (unless you have multiple nodes on an HPC, but let's assume it's just one shared memory machine). You can Monte Carlo in MATLAB just by using vectorized maps or parfor loops just fine (there's no multithreading issues with its methods). Julia will also multithread it very straightforwardly. I don't think moving over to SciPy would be a worthwhile change since Cython has some multithreading issues of its own (GIL) which would come into play here if you want to parallelize it easily (it's not that you can't do it other ways like multiprocessing, but multithreading would definitely be the right architecture in many cases here).

Also depending on what you're doing, AMIGO2 might be a good MATLAB toolbox to help you out:

Maybe it has what you need in there.

So with all of that together, my recommendation would first be to just see if parallelizing on MATLAB will do the job. Otherwise, if you're familiar with Python and would like to head in that direction, I don't think that there's much to be gained from directly using SciPy (for the reasons mentioned in the post). Instead, I would look to use something like PyDSTool or JiTCODE and do the parallelism part yourself. Otherwise if you're willing to learn Julia the Monte Carlo functionality is made just for this.

• ## Shanky Jain

says:

Hi, I just learned from this Matlab forum that ODE suite is not multi-threaded, so on a single machine if i use parfor i'll restricted to 2 cores which may not be a good solution if i want to run 1000 iterations.

• ## Christopher Rackauckas

says:

Since you've already compiled your function call to a *mex function, I am not sure there's much more you can do without poking at the internals. I think the easiest option would be taking it to a cluster so that way it can parallelize more. Monte Carlo is embarrassingly parallel and its usually best to parallelize at the highest fully parallel level, so the parfor strategy would be good. I'd solve it just by finding the right computational resource to do it well.

Also, one way to speed things up at this size is to reduce the saving load. Make it only output the values that you need instead of saving everything adaptively. Being smart with saving matters when the size of the system is large.

says:

8. ## noop

says:

Mathematica does have sensitivity analysis and parameter estimation; it's done with ParametricNDSolveValue (https://reference.wolfram.com/language/ref/ParametricNDSolve.html) So, saying None here is not correct.

• ## Christopher Rackauckas

says:

Thanks! I didn't know this function existed. I'll get this updated in the table. It's not in their "Advanced Differential Equations" page which explains all of their methods. Can you describe what they are using? It looks like, without being able to find total documentation on this, it's using adjoint sensitivity analysis since they are always plotting it with the timeseries and you're able to give it separate functions to take derivative of. But I can't find a methods page to confirm. Do you have more details?

• ## noop

says:

OK, here is an example for the sensitivity:

pfun = ParametricNDSolveValue[{x'[t] == a x[t], x[0] == b},
x[1], {t, 0, 1}, {a, b},
Method -> {"ParametricSensitivity" -> {"ForwardSensitivity",
"SensitivityDerivatives" -> {"Symbolic"}}}];
da = Derivative[1, 0][pfun];
da[1, 1]

Note, the Method option does not have to be specified; I doing this to show you. So it uses "ForwardSensitivity" and a symbolic derivative (I think you can also use FDM)

• ## Christopher Rackauckas

says:

I would liked to see that parameter fitting vs data actually become some built in function since for example the user in that Stack Overflow post isn't using the sensitivity analysis computed gradients which would help the efficiency, and there's no regularization whereas in the literature people show that Tikonov regularization is usually a good idea here. But the tools play really nicely together and the sensitivity analysis is quite advanced, I'll give it an "Excellent".

• ## Christopher Rackauckas

says:

Thanks for the details. I'll get this corrected in the chart. As with the other suites which only do sensitivity analysis (which is forward and adjoint, that's pretty nice), it gets a "Good". Since they have a whole setup for ODEs + parameters, it would be nice if they added parameter inference routines which make use of all of this (i.e. add optimization routines on top, and have it give the gradients provided by sensitivity analysis). I think that level of usability in addons would send it to an "Excellent".

• ## noop

says:

Here are a few examples where ParametricNDSolve plays with other functions, like FindRoot and FindMinimum. One thifng to realize is that because FindRoot etc can compute symbolic derivatives and it does not matter to them if their input is symbolic of something from ParametricNDSolve:

https://mathematica.stackexchange.com/questions/92277/parametricsensitivity-in-parametricndsolve

I think Mathematica should get an Excellent here.

9. ## Michael Aye

says:

Thinking that Scipy can't be THAT bad, I did a simple Google search...
wondering why the author didn't do that, apparently.
http://bfy.tw/EC05

• ## Christopher Rackauckas

says:

And what do you find? That Google search turns up this:

https://docs.scipy.org/doc/scipy-0.15.1/reference/generated/scipy.optimize.rosen.html

That's the Rosenbrock function, a sample equation for testing optimization functions. It's not a Rosenbrock method for stiff ODEs, which doesn't show up in the search results for me. Nothing relevant to the discussion is given from that Google search result. If it does for you, please link to the page directly.

Instead, you can look at SciPy's (quite clear) ODE solver documentation and see that there is not very many methods, and they are the ones I listed.

• ## Michael Aye

says:

I apologize for my rash/quick comment. I think I spoke in shock. 😉 Am very grateful for your overview. 🙂 Wondering how/why scipy is failing so hard. Is it maybe that some of this is covered somewhere else? I mean, ppl are doing so much wonderful science with Python, how can they do that, if, apparently, it's so much harder to do it there?

• ## Christopher Rackauckas

says:

Well, I think it's people. Researchers in the subject made the C and Fortran codes. Then one of those researchers, Shampine, created the MATLAB and Maple suites. Verner also contributed to Maple. So those two were propped up by researchers in the field. Mathematica seems to have poured money into some ODE software developers since not only did they match Maple in most ways, they made some unique and innovative contributions of their own.

Meanwhile, R's deSolve and SciPy only have wrapped solvers. When you read old dev chats you see that most of the devs aren't experts in the subject: they seem to know that they should complete the area because other scientific software has it, but they aren't differential equations methods researchers willing to pour years into only doing this. So their plans are to simply wrap things or build off of PyDSTool.

And that's perfectly fine. I think that for the majority of people, SciPy has enough to satisfy them. Most people probably just need an integrator for stiff and nonstiff equations, and anything more in the docs probably confuses most people. Of course, when doing a side-by-side you go "wow, it doesn't do event handling, doesn't have most of the stiff solvers, etc. that others have?" but... if you don't need or use that stuff it really doesn't matter.

And I think that's the key. For example, I'll be publishing a big set of benchmarks soon which compare between similar implementations of many different methods, and it'll show that you can do "a lot better", about 5x, by choosing Rosenbrock here and ..., but in reality, these differences don't matter. Assume that because you're stuck with LSODE/Radau you're 2x-5x away from "the best method", and the lack of JIT is 10x-100x. Simulations which take seconds somewhere else take minutes here, but that doesn't stop science, it just is a little annoying. When the timing is enough that it does get in the way, there are other Python tools like the ones mentioned here (JiTCODE, PyDSTool) which will compile it, and that's the vast majority of the performance difference. The people looking to shave off that last 5x are the niche that are solving days+ long simulations of (stochastic) PDEs, and they probably are using C++ or Fortran to begin with anyways. It's just a different audience.

So I have this long comparison list looking at how they handle state-dependent delay equations, stochastic differential equations, sensitivity analysis, etc. because I am in the niche and care about these details. But people shouldn't care about stuff they aren't using. I think that SciPy developers should take a look at this and map out some future additions, but it's hard to get someone to contribute in the more niche areas since less people tend to use it, so in OSS this stuff tends to only get worked on if the developer actually needs to use it.

10. ## Ross

says:

Mathematica and GPU just fine.

Maybe you had something else in mind? Otherwise, check facts.

Thanks for this summary.

• ## Christopher Rackauckas

says:

Mathematica can use GPUs, but can you point to where it shows how to solve ODEs on the GPU? This post is not about a language's general capabilities, rather it's specifically about its capabilities with respect to its differential equation suites.

• ## Ross

says:

This is a good distinction, and your point is taken.

It seems there are a number of Mathematica functions which are default GPU-accelerated 'out of the box'. I was unable to determine in a brief survey if this includes their ODE solvers.

They *do* permit any desired C function to be loaded and GPU processed with CUDA link or OpenCL link. This should permit any generic C-based ODE solver to be GPU powered, but not "out of the box".