# noqa: D400 D205
"""
Calibration of step sampler
---------------------------
"""
import os
import numpy as np
from ultranest.integrator import ReactiveNestedSampler
def _substitute_log_dir(init_args, nsteps):
"""Append `nsteps` to `log_dir` argument, if set.
Parameters
----------
init_args: dict
arguments passed :py:class:`ReactiveNestedSampler`,
may contain the key `'log_dir'`.
nsteps: int
number of steps
Returns
-------
new_init_args: dict
same as init_args, but if `'log_dir'` was set,
it now has `'-nsteps'+str(nsteps)` appended.
"""
if 'log_dir' in init_args:
args = dict(init_args)
args['log_dir'] = init_args['log_dir'] + '-nsteps%d' % nsteps
return args
return init_args
[docs]
class ReactiveNestedCalibrator():
"""Calibrator for the number of steps in step samplers.
The number of steps in a step sampler needs to be chosen.
A calibration recommended (e.g. https://ui.adsabs.harvard.edu/abs/2019MNRAS.483.2044H)
is to run a sequence of nested sampling runs with increasing number of steps,
and stop when log(Z) converges.
This class automates this. See the :py:meth:`ReactiveNestedCalibrator.run`
for details.
Usage
-----
Usage is designed to be a drop-in replacement for ReactiveNestedSampler.
If your code was::
sampler = ReactiveNestedSampler(my_param_names, my_loglike, my_transform)
sampler.stepsampler = SliceSampler(nsteps=10, generate_direction=region_oriented_direction)
sampler.run(min_num_livepoints=400)
You would change it to::
sampler = ReactiveNestedCalibrator(my_param_names, my_loglike, my_transform)
sampler.stepsampler = SliceSampler(nsteps=10, generate_direction=region_oriented_direction)
sampler.run(min_num_livepoints=400)
The run() command will print the number of slice sampler steps
that appear safe for the inference task.
The initial value for nsteps (e.g. in `SliceSampler(nsteps=...)`)
is overwritten by this class.
"""
def __init__(self,
param_names,
loglike,
transform=None,
**kwargs
):
"""Initialise nested sampler calibrator.
Parameters
-----------
param_names: list of str
Names of the parameters.
Length gives dimensionality of the sampling problem.
loglike: function
log-likelihood function.
transform: function
parameter transform from unit cube to physical parameters.
kwargs: dict
further arguments passed to ReactiveNestedSampler
if `log_dir` is set, then the suffix `-nsteps%d` is added for each
run where %d is replaced with the number of steps (2, 4, 8 etc).
"""
self.init_args = dict(param_names=param_names, loglike=loglike, transform=transform, **kwargs)
self.stepsampler = None
[docs]
def run(self, **kwargs):
"""Run a sequence of ReactiveNestedSampler runs until convergence.
The first run is made with the number of steps set to the number of parameters.
Each subsequent run doubles the number of steps.
Runs are made until convergence is reached.
Then this generator stops yielding results.
Convergence is defined as three consecutive runs which
1) are not ordered in their log(Z) results,
and 2) the consecutive log(Z) error bars must overlap.
Parameters
-----------
**kwargs: dict
All arguments are passed to :py:meth:`ReactiveNestedSampler.run`.
Yields
-------
nsteps: int
number of steps for the current run
result: dict
return value of :py:meth:`ReactiveNestedSampler.run` for the current run
"""
assert self.stepsampler is not None
self.run_args = kwargs
# start with nsteps=d
nsteps = len(self.init_args['param_names'])
self.results = []
self.nsteps = []
self.relsteps = []
while True:
print("running with %d steps ..." % nsteps)
init_args = _substitute_log_dir(self.init_args, nsteps)
sampler = ReactiveNestedSampler(**init_args)
sampler.stepsampler = self.stepsampler.__class__(
nsteps=nsteps, generate_direction=self.stepsampler.generate_direction,
check_nsteps=self.stepsampler.check_nsteps,
adaptive_nsteps=self.stepsampler.adaptive_nsteps,
log=open(init_args['log_dir'] + '/stepsampler.log', 'w') if 'log_dir' in self.init_args else None) # noqa: SIM115
self.sampler = sampler
result = sampler.run(**self.run_args)
print("Z=%(logz).2f +- %(logzerr).2f" % result)
if self.sampler.log_to_disk:
sampler.stepsampler.plot(os.path.join(self.sampler.logs['plots'], 'stepsampler.pdf'))
sampler.stepsampler.plot_jump_diagnostic_histogram(
os.path.join(self.sampler.logs['plots'], 'stepsampler-jumphist.pdf'),
histtype='step', bins='auto')
sampler.stepsampler.print_diagnostic()
if 'jump-distance' in sampler.stepsampler.logstat_labels and 'reference-distance' in sampler.stepsampler.logstat_labels:
i = sampler.stepsampler.logstat_labels.index('jump-distance')
j = sampler.stepsampler.logstat_labels.index('reference-distance')
jump_distances = np.array([entry[i] for entry in sampler.stepsampler.logstat])
reference_distances = np.array([entry[j] for entry in sampler.stepsampler.logstat])
self.relsteps.append(jump_distances / reference_distances)
# TODO: handle population step samplers
self.results.append(result)
self.nsteps.append(nsteps)
yield nsteps, result
if len(self.results) > 2:
last_result = self.results[-2]
last_result2 = self.results[-3]
# check if they agree within the error bars
last_significant = abs(result['logz'] - last_result['logz']) > (result['logzerr']**2 + last_result['logzerr']**2)**0.5
last2_significant = abs(last_result2['logz'] - last_result['logz']) > (last_result2['logzerr']**2 + last_result['logzerr']**2)**0.5
# check if there is order
monotonic_increase = result['logz'] > last_result['logz'] > last_result2['logz']
monotonic_decrease = result['logz'] < last_result['logz'] < last_result2['logz']
if last_significant:
print("not converged: last two Z were significantly different")
elif last2_significant:
print("not yet converged: previous two Z were significantly different")
elif monotonic_increase:
print("not converged: monotonic increase in the last three Z results")
elif monotonic_decrease:
print("not converged: monotonic decrease in the last three Z results")
else:
print("converged! nsteps=%d appears safe" % nsteps)
break
nsteps *= 2
[docs]
def plot(self):
"""Visualise the convergence diagnostics.
Stores into `<log_dir>/plots/` folder:
* stepsampler.pdf: diagnostic of stepsampler, see :py:meth:`StepSampler.plot`
* nsteps-calibration-jumps.pdf: distribution of relative jump distance
* nsteps-calibration.pdf: evolution of ln(Z) with nsteps
"""
self.sampler.stepsampler.plot(os.path.join(self.sampler.logs['plots'], 'stepsampler.pdf'))
# plot U-test convergence run length (at 4 sigma) (or niter) vs nsteps
# plot step > reference fraction vs nsteps
calibration_results = []
import matplotlib.pyplot as plt
plt.figure("jump-distance")
print("jump distance diagnostic:")
for nsteps, relsteps, result in zip(self.nsteps, self.relsteps, self.results):
calibration_results.append([
nsteps, result['logz'], result['logzerr'],
min(result['niter'], result['insertion_order_MWW_test']['independent_iterations']),
result['insertion_order_MWW_test']['converged'] * 1,
np.nanmean(relsteps > 1)])
plt.hist(np.log10(relsteps + 1e-10), histtype='step', bins='auto', label=nsteps)
print(' %-4d: %.2f%% avg:%.2f' % (nsteps, np.nanmean(relsteps > 1) * 100.0, np.exp(np.nanmean(np.log(relsteps)))))
if 'log_dir' in self.init_args:
np.savetxt(
self.init_args['log_dir'] + 'calibration.csv',
calibration_results, delimiter=',', comments='',
header='nsteps,logz,logzerr,maxUrun,Uconverged,stepfrac',
fmt='%d,%.3f,%.3f,%d,%d,%.5f')
plt.xlabel('$log_{10}$(relative step distance)')
plt.ylabel('Frequency')
plt.legend(title='nsteps', loc='best')
if self.sampler.log_to_disk:
plt.savefig(os.path.join(self.sampler.logs['plots'], 'nsteps-calibration-jumps.pdf'), bbox_inches='tight')
plt.close()
plt.figure("logz")
plt.errorbar(
x=self.nsteps,
y=[result['logz'] for result in self.results],
yerr=[result['logzerr'] for result in self.results],
)
plt.title('Step sampler calibration')
plt.xlabel('Number of steps')
plt.ylabel('ln(Z)')
if self.sampler.log_to_disk:
plt.savefig(os.path.join(self.sampler.logs['plots'], 'nsteps-calibration.pdf'), bbox_inches='tight')
plt.close()
self.sampler.logger.debug('Making nsteps calibration plot ... done')