module Criterion
(
Benchmarkable(..)
, Benchmark
, Pure
, nf
, whnf
, nfIO
, whnfIO
, bench
, bgroup
, runBenchmark
, runAndAnalyse
) where
import Control.Monad ((<=<), replicateM_, when)
import Control.Monad.Trans (liftIO)
import Criterion.Analysis (Outliers(..), OutlierEffect(..), OutlierVariance(..),
SampleAnalysis(..), analyseSample,
classifyOutliers, noteOutliers)
import Criterion.Config (Config(..), Verbosity(..), fromLJ)
import Criterion.Environment (Environment(..))
import Criterion.IO (note, prolix, summary)
import Criterion.Measurement (getTime, runForAtLeast, secs, time_)
import Criterion.Monad (Criterion, getConfig, getConfigItem)
import Criterion.Report (Report(..), report)
import Criterion.Types (Benchmarkable(..), Benchmark(..), Pure,
bench, bgroup, nf, nfIO, whnf, whnfIO)
import qualified Data.Vector.Unboxed as U
import Statistics.Resampling.Bootstrap (Estimate(..))
import Statistics.Types (Sample)
import System.Mem (performGC)
import Text.Printf (printf)
runBenchmark :: Benchmarkable b => Environment -> b -> Criterion Sample
runBenchmark env b = do
_ <- liftIO $ runForAtLeast 0.1 10000 (`replicateM_` getTime)
let minTime = envClockResolution env * 1000
(testTime, testIters, _) <- liftIO $ runForAtLeast (min minTime 0.1) 1 (run b)
_ <- prolix "ran %d iterations in %s\n" testIters (secs testTime)
cfg <- getConfig
let newIters = ceiling $ minTime * testItersD / testTime
sampleCount = fromLJ cfgSamples cfg
newItersD = fromIntegral newIters
testItersD = fromIntegral testIters
estTime = (fromIntegral sampleCount * newItersD *
testTime / testItersD)
when (fromLJ cfgVerbosity cfg > Normal || estTime > 5) $
note "collecting %d samples, %d iterations each, in estimated %s\n"
sampleCount newIters (secs estTime)
times <- liftIO . fmap (U.map ((/ newItersD) . subtract (envClockCost env))) .
U.replicateM sampleCount $ do
when (fromLJ cfgPerformGC cfg) $ performGC
time_ (run b newIters)
return times
runAndAnalyseOne :: Benchmarkable b => Environment -> String -> b
-> Criterion (Sample,SampleAnalysis,Outliers)
runAndAnalyseOne env _desc b = do
times <- runBenchmark env b
ci <- getConfigItem $ fromLJ cfgConfInterval
numResamples <- getConfigItem $ fromLJ cfgResamples
_ <- prolix "analysing with %d resamples\n" numResamples
an@SampleAnalysis{..} <- liftIO $ analyseSample ci times numResamples
let OutlierVariance{..} = anOutlierVar
let wibble = case ovEffect of
Unaffected -> "unaffected" :: String
Slight -> "slightly inflated"
Moderate -> "moderately inflated"
Severe -> "severely inflated"
bs "mean" anMean
summary ","
bs "std dev" anStdDev
summary "\n"
vrb <- getConfigItem $ fromLJ cfgVerbosity
let out = classifyOutliers times
when (vrb == Verbose || (ovEffect > Unaffected && vrb > Quiet)) $ do
noteOutliers out
_ <- note "variance introduced by outliers: %.3f%%\n" (ovFraction * 100)
_ <- note "variance is %s by outliers\n" wibble
return ()
return (times,an,out)
where bs :: String -> Estimate -> Criterion ()
bs d e = do _ <- note "%s: %s, lb %s, ub %s, ci %.3f\n" d
(secs $ estPoint e)
(secs $ estLowerBound e) (secs $ estUpperBound e)
(estConfidenceLevel e)
summary $ printf "%g,%g,%g"
(estPoint e)
(estLowerBound e) (estUpperBound e)
plotAll :: [(String, Sample, SampleAnalysis, Outliers)] -> Criterion ()
plotAll descTimes = do
report (zipWith (\n (d,t,a,o) -> Report n d t a o) [0..] descTimes)
runAndAnalyse :: (String -> Bool)
-> Environment
-> Benchmark
-> Criterion ()
runAndAnalyse p env = plotAll <=< go ""
where go pfx (Benchmark desc b)
| p desc' = do _ <- note "\nbenchmarking %s\n" desc'
summary (show desc' ++ ",")
(x,an,out) <- runAndAnalyseOne env desc' b
return [(desc',x,an,out)]
| otherwise = return []
where desc' = prefix pfx desc
go pfx (BenchGroup desc bs) =
concat `fmap` mapM (go (prefix pfx desc)) bs
prefix "" desc = desc
prefix pfx desc = pfx ++ '/' : desc