# SolverBenchmark.jl documentation

This package provides general tools for benchmarking solvers, focusing on a few guidelines:

• The output of a solver's run on a suite of problems is a DataFrame, where each row is a different problem.
• Since naming issues may arise (e.g., same problem with different number of variables), there must be an ID column;
• The collection of two or more solver runs (DataFrames), is a Dict{Symbol,DataFrame}, where each key is a solver;

This package is developed focusing on Krylov.jl and JSOSolvers.jl, but they should be general enough to be used in other places.