DerivKit logo black Workflows#

This page helps you decide which DerivKit tool to use based on your scientific use case. Each section starts from a concrete question and points you to the appropriate workflow and examples.

It is intended as a decision guide, not a full tutorial.

Using DerivKit effectively requires an understanding of the assumptions, strengths, and limitations of the chosen method. Fisher, DALI, Laplace, and sampling-based approaches make different approximations and are suited to different inference regimes. DerivKit does not guarantee the validity of any approximation for a given problem; assessing whether a method is appropriate for a specific model, parameterization, prior choice, and scientific goal requires scientific judgment from the user.

DerivKit does not construct data covariances or define scientific models. Users are expected to provide:

  • their own model mapping parameters to observables

  • their own data covariance (or a function returning one)

DerivKit focuses on derivative evaluation, local likelihood approximations, and fast forecasting utilities built on top of these inputs.

If you are looking for short answers to common questions, you can jump directly to the FAQ / Frequently asked questions section below. Some common mistakes are discussed in the Common mistakes section.

Quick decision guide#

This table provides a fast, high-level guide for choosing a numerical differentiation strategy in DerivKit. It is intended as a quick reference; for detailed workflows and examples, see the sections below.

Situation

Recommended method

Why

Smooth, cheap function

Finite differences

Fast and accurate for smooth functions

Slightly noisy function

Ridders finite differences

Richardson extrapolation improves stability over simple finite differences

Moderate or structured noise

Local polynomial fit

Local regression smooths noise better than finite differences

High noise / messy signal

Adaptive polynomial fit (Chebyshev)

Robust trimming, Chebyshev grid, and fit diagnostics

Expensive function

Adaptive polynomial fit (Chebyshev)

Achieves stable derivatives with fewer function evaluations near x0

Need robustness and diagnostics

Adaptive polynomial fit (Chebyshev)

Provides fit quality metrics, degree adjustment, and suggestions

Unsure / first attempt

Local polynomial fit

Good default when function behavior is not well known

Choose a workflow#

I want Fisher constraints on my parameters#

You have

  • a parameter vector theta0

  • a model mapping parameters to observables

  • a data covariance matrix (or a function returning one)

We compute

  • the Fisher information matrix

  • the Gaussian parameter covariance via inversion

  • approximate posterior samples

  • GetDist-compatible outputs for visualization

Use

  • ForecastKit.fisher

Minimal example

See Fisher matrix and Fisher contours

Notes

  • This assumes the posterior is approximately Gaussian near theta0.

  • If this assumption fails, consider using DALI instead.

  • Parameter-dependent covariances are handled automatically.

I expect non-Gaussian posteriors (banana-shaped, skewed, etc.)#

You have

  • a nonlinear model

  • parameters where Fisher may underestimate uncertainties

  • parameters with physical bounds or informative priors that truncate the Gaussian approximation

  • a data covariance matrix

We compute

  • a DALI expansion up to a chosen order: - order 1: Fisher matrix - order 2: doublet DALI tensors (D1 and D2) - order 3: triplet DALI tensors (T1, T2, and T3)

  • approximate posterior samples

  • GetDist-compatible outputs for visualization

Use

  • ForecastKit.dali(expansion_order=N)

Minimal example

See DALI tensors and DALI contours

Notes

  • Choose the expansion order based on how non-Gaussian you expect the posterior to be.

  • You do not need to manipulate DALI tensors directly.

  • Sampling bounds and informative priors can make posteriors non-Gaussian even when the forward model is close to linear.

  • Fisher/DALI describe the likelihood locally; prior truncation effects are only captured when you sample with explicit priors.

I already have Fisher matrix / DALI tensors. What do I do next?#

You have

  • Fisher F and optionally higher-order tensors D1, D2, …

We compute

  • approximate posterior samples

  • GetDist-compatible outputs for visualization

Use

  • importance sampling (fast)

  • emcee sampling (slower, more robust)

Minimal example

See DALI contours

Notes

  • Importance sampling is extremely fast but may fail for strongly non-Gaussian posteriors.

  • If importance sampling fails, switch to MCMC sampling via emcee.

I want a Gaussian approximation around a MAP#

You have

  • a likelihood or log-posterior

  • a maximum a posteriori (MAP) point

We compute

  • a Laplace (Gaussian) approximation

  • an estimate of the local covariance

Use

  • Laplace approximation utilities

Minimal example

See Laplace approximation and Laplace contours

Notes

  • The Laplace mean is the expansion point (usually the MAP).

  • This is a local approximation and may fail for strongly non-Gaussian posteriors.

I want to include priors#

You have

  • prior information (bounds, Gaussian priors, correlated priors, etc.)

We compute

  • log-prior contributions

  • posterior sampling with explicit priors

Use

  • PriorKit

Minimal example

See the DALI sampling examples with priors in DALI contours

Notes

  • Priors are applied explicitly by design.

  • Sampler bounds truncate the sampled region; informative priors modify the posterior shape.

My model is tabulated or expensive to evaluate#

You have

  • samples of a function on a grid

  • no analytic expression (or an expensive forward model)

We compute

  • numerical derivatives from tabulated data

Use

  • DerivativeKit with x_tab and y_tab inputs

Minimal example

See Tabulated derivatives

Notes

  • This is especially useful when model evaluations are costly.

  • Tabulated models are treated as callables by DerivKit.

I only want numerical derivatives (no forecasting yet)#

You have

  • a function or model

  • a point where derivatives are needed

We compute

  • first and higher-order derivatives

  • gradients, Jacobians, and Hessians

Use

  • DerivativeKit

  • CalculusKit

Minimal example

See Derivatives

Notes

  • Use CalculusKit when you want direct access to gradients/Hessians.

  • Use DerivativeKit for higher-level derivative workflows and diagnostics.

I want Fisher bias / parameter shifts#

You have

  • a mismatched model and data-generating process

  • a Fisher matrix

We compute

  • parameter bias induced by model mismatch

Use

  • Fisher bias utilities in ForecastKit

Minimal example

See Fisher bias

My covariance depends on parameters. What do I do?#

You have

  • a model mapping parameters to observables, and

  • a covariance that depends on parameters, \(C(\theta)\), or

  • noisy inputs and outputs with a block covariance (the X–Y case)

We compute

  • a Gaussian Fisher matrix that includes covariance-derivative terms when \(C(\theta)\) depends on the parameters

  • (optional) X–Y Gaussian Fisher constraints where input uncertainties are propagated into an effective output covariance

Use

Minimal examples

See Gaussian Fisher matrix and X–Y Gaussian Fisher matrix.

Notes

  • ForecastKit.fisher() uses a fixed covariance \(C(\theta_0)\) and computes only the mean-derivative term.

  • ForecastKit.gaussian_fisher() adds the covariance-derivative contribution when a callable covariance \(C(\theta)\) is provided.

  • DALI currently assumes a fixed covariance evaluated at \(\theta_0\); parameter-dependent covariance support for DALI is planned.

I have many parameters and derivative evaluation is expensive#

You have

  • a model with many parameters

  • expensive function evaluations

  • concerns about runtime or scaling

We compute

  • derivatives using parallel execution

  • Jacobians and higher-order tensors efficiently

Use

  • DerivativeKit with n_workers

  • ForecastKit parallel derivative evaluation

Notes

  • DerivKit parallelizes derivative evaluations across parameters and outputs.

  • This is especially useful for large Fisher or DALI expansions.

I want to compare Fisher and DALI forecasts#

You have

  • a Fisher forecast

  • a DALI expansion for the same model

We compute

  • approximate posterior samples from both

  • directly comparable contours and summaries

Use

  • ForecastKit.fisher

  • ForecastKit.dali

  • GetDist-based visualization utilities

Notes

  • This is useful for diagnosing non-Gaussianity.

  • Differences indicate where Fisher assumptions break down.

FAQ / Frequently asked questions#

Why does Fisher underestimate my errors?

Because it assumes the posterior is locally Gaussian. Strong curvature or parameter degeneracies require higher-order (DALI) terms.

Why is the Laplace mean equal to my expansion point?

The Laplace approximation expands around the MAP by construction. It does not estimate a shifted mean.

When should I use DALI instead of Fisher?

When the posterior is visibly non-Gaussian or when Fisher forecasts are known to be biased.

How do I choose the DALI expansion order?

The appropriate expansion order depends on the degree of non-Gaussianity in the posterior. Order 1 corresponds to the Fisher approximation, while higher orders capture increasing levels of skewness and curvature. In practice, comparing results across orders can help diagnose when Fisher assumptions break down.

Why are priors not included automatically?

DerivKit separates likelihood information from prior assumptions by design. This keeps approximations explicit and easier to reason about.

Can I use DerivKit with MCMC samplers?

Yes. DerivKit likelihoods and priors can be used with any sampler that accepts log-posterior functions. We provide examples using emcee and importance sampling, but DerivKit is sampler-agnostic and can be integrated with other sampling frameworks by implementing a thin wrapper around the log-posterior API.

Does DerivKit assume Gaussian likelihoods?

No. Fisher and Laplace methods make local Gaussian approximations, while DALI systematically captures non-Gaussian structure through higher-order terms.

My DALI doublet and triplet contours look identical. Is something wrong?

Usually not. Triplet DALI only modifies the posterior where higher-order corrections are non-negligible at the typical posterior radius. If most of the posterior mass lies close to the expansion point theta0 or if you are using importance sampling, higher-order terms can be strongly suppressed and contours may look identical. Remember that DALI is a local expansion: it captures local skewness and curvature, but cannot reproduce global structure or multiple modes. For triplet DALI to show significant differences from doublet DALI, the posterior must extend far enough from theta0 for cubic terms to become important. We recommend always using emcee sampling for triplet DALI to fully capture its effects. If in doubt, compare results across expansion orders and sampling methods.

Why does DALI behave poorly far from the expansion point?

DALI is based on a Taylor expansion and is only expected to be accurate within a finite neighborhood around theta0. As a result, DALI is not expected to be accurate for values far away from theta0. For this reason, sampling bounds and diagnostic checks are strongly recommended.

Should I always use the highest DALI expansion order available?

No. Higher-order expansions are more expensive and not always informative. If increasing the expansion order does not change posterior summaries, lower orders are usually sufficient and preferable for robustness.

Where do my model and covariance come from?

DerivKit is agnostic to how models and covariances are constructed. Users are expected to supply these based on their scientific application, while DerivKit provides derivative evaluation and inference utilities.

Can I use my own likelihood with DerivKit?

Yes. DerivKit is agnostic to how likelihoods are defined. Users can supply their own likelihood or log-posterior functions, which DerivKit treats as external inputs for derivative evaluation, local approximations, and sampling.

Can I use ``DerivKit`` within an existing inference pipeline?

Yes. DerivKit is designed to integrate with externally defined models, likelihoods, and covariances, and can be used alongside other inference or sampling frameworks.

Are derivatives computed analytically or numerically?

DerivKit computes derivatives numerically using robust finite-difference and polynomial-based methods. Optional automatic differentiation backends may be used for validation, but numerical methods are the default and primary focus.

Where can I find more examples?

See the Examples section of the documentation. Additional extended demos are available at https://github.com/derivkit/derivkit-demos

Who do I contact for support?

Please open an issue on the DerivKit GitHub repository. Go to Contributing for contribution guidelines and support options.

Common mistakes#

Using local approximations for global inference

Fisher, DALI, and Laplace are local approximations around a chosen expansion point. As such, they may not reliably recover global posterior structure, multiple modes, or long nonlocal tails beyond the radius of convergence of the local expansion. If your inference problem is strongly nonlocal, full MCMC or nested sampling is required.

Expanding around a poorly chosen expansion point

All local methods assume that theta0 (or the MAP for Laplace) is close to the region of highest posterior support. Expanding far from the true posterior peak can lead to misleading forecasts or unstable higher-order corrections.

Relying on importance sampling for strongly non-Gaussian posteriors

Importance sampling is fast but fragile. For curved, skewed, or bounded posteriors it can severely underestimate higher-order effects. In these cases, use MCMC sampling (e.g. emcee) instead.

Assuming higher-order expansions always improve results

Increasing the DALI expansion order does not guarantee better accuracy. If higher-order terms are numerically small where the posterior mass lies, results may be unchanged. Unnecessary higher-order terms can also reduce robustness.

Ignoring sampling bounds and priors

Sampling without explicit bounds or priors can lead to unphysical regions where local approximations break down. Always include realistic bounds or priors when sampling Fisher, DALI, or Laplace posteriors.

Expecting priors to be applied automatically

DerivKit treats likelihood information and priors separately by design. Priors must be applied explicitly during sampling; they are not combined with Fisher, DALI, or Laplace objects automatically.

Over-interpreting Laplace approximations

The Laplace approximation provides a Gaussian approximation around the MAP. It does not capture skewness, curvature, or truncation effects and should not be used for strongly non-Gaussian posteriors.

Using very tight data covariances without diagnostics

When the posterior is extremely narrow, higher-order corrections can be numerically suppressed and indistinguishable from Fisher. Always check the typical posterior radius relative to the expansion point.

Assuming numerical derivatives are exact

All derivatives in DerivKit are computed numerically. Poorly scaled parameters, discontinuous models, or insufficient step-size control can degrade derivative accuracy. Diagnostic checks are recommended for sensitive applications.

Assuming automatic differentiation is always correct

Automatic differentiation (e.g. JAX) can silently fail for non-smooth models, conditionals, or interpolations; derivative validation is still recommended, especially for higher-order forecasts.

Confusing likelihood curvature with posterior uncertainty

Fisher, DALI, and Laplace characterize the local curvature of the likelihood. Posterior uncertainties can differ significantly once priors, bounds, or nonlinear transformations are applied. Always interpret results in the context of the full posterior definition.

Using overly informative priors without realizing it

Strong priors can dominate posterior constraints and mask the information content of the likelihood. When comparing Fisher, DALI, or Laplace results, check whether constraints are prior-dominated rather than data-driven.

Interpreting Fisher or DALI contours as exact confidence regions

Contours produced from Fisher or DALI approximations are not guaranteed to have exact frequentist or Bayesian coverage. They should be interpreted as approximate summaries, not as precise confidence regions.

Mixing Bayesian and frequentist interpretations

Fisher matrices originate from frequentist theory, while posterior sampling and priors are Bayesian concepts. Mixing interpretations without care can lead to incorrect conclusions about parameter uncertainties or coverage.

Ignoring parameter reparameterization effects

Local approximations depend on the chosen parameterization. Strong nonlinear transformations can change the apparent degree of non-Gaussianity and affect Fisher or DALI performance. Reparameterization may improve stability and interpretability.