# SolverBenchmark.jl documentation

This package provides general tools for benchmarking solvers, focusing on a few guidelines:

- The output of a solver's run on a suite of problems is a
`DataFrame`

, where each row is a different problem.- Since naming issues may arise (e.g., same problem with different number of variables), there must be an ID column;

- The collection of two or more solver runs (
`DataFrame`

s), is a`Dict{Symbol,DataFrame}`

, where each key is a solver;

This package is developed focusing on Krylov.jl and JSOSolvers.jl, but they should be general enough to be used in other places.