Added multi-thresh simulation to "full" and "short" (currently running).
Added complete "rect-lp" analysis except figure. Added multiple appendix figs. Overhauled normalization options across all condense scripts. Co-authored-by: Copilot <copilot@github.com>
This commit is contained in:
@@ -1,9 +1,9 @@
|
||||
import numpy as np
|
||||
import matplotlib.pyplot as plt
|
||||
from thunderhopper.modeltools import load_data, save_data
|
||||
from thunderhopper.filetools import search_files, crop_paths
|
||||
from thunderhopper.filtertools import find_kern_specs
|
||||
from thunderhopper.model import process_signal
|
||||
from thunderhopper.filters import sosfilter
|
||||
from misc_functions import draw_noise_segment
|
||||
from IPython import embed
|
||||
|
||||
@@ -16,7 +16,7 @@ target_species = [
|
||||
'Gomphocerippus_rufus',
|
||||
'Omocestus_rufipes',
|
||||
'Pseudochorthippus_parallelus',
|
||||
][4]
|
||||
][5]
|
||||
example_file = {
|
||||
'Chorthippus_biguttulus': 'Chorthippus_biguttulus_GBC_94-17s73.1ms-19s977ms',
|
||||
'Chorthippus_mollis': 'Chorthippus_mollis_DJN_41_T28C-46s4.58ms-1m15s697ms',
|
||||
@@ -28,34 +28,26 @@ example_file = {
|
||||
}[target_species]
|
||||
data_paths = search_files(target_species, dir='../data/processed/')
|
||||
noise_path = '../data/processed/white_noise_sd-1.npz'
|
||||
ref_path = '../data/inv/full/ref_measures.npz'
|
||||
thresh_path = '../data/inv/full/thresholds.npz'
|
||||
stages = ['filt', 'env', 'log', 'inv', 'conv', 'feat']
|
||||
pre_stages = stages[:-1]
|
||||
save_path = '../data/inv/full/'
|
||||
|
||||
# ANALYSIS SETTINGS:
|
||||
example_scales = np.array([0.1, 1, 10, 30, 100, 300])
|
||||
scales = np.geomspace(0.01, 10000, 500)
|
||||
scales = np.unique(np.concatenate(([0], scales, example_scales)))
|
||||
thresh_rel = 0.5
|
||||
thresh_rel = np.array([0, 0.5, 1, 1.5, 2, 2.5, 3])
|
||||
|
||||
# SUBSET SETTINGS:
|
||||
kernels = np.array([
|
||||
[1, 0.002],
|
||||
[-1, 0.002],
|
||||
[2, 0.004],
|
||||
[-2, 0.004],
|
||||
[3, 0.032],
|
||||
[-3, 0.032]
|
||||
])
|
||||
kernels = None
|
||||
types = None#np.array([-1])
|
||||
sigmas = None#np.array([0.001, 0.002, 0.004, 0.008, 0.016, 0.032])
|
||||
types = None
|
||||
sigmas = None
|
||||
|
||||
# PREPARATION:
|
||||
pure_noise = np.load(noise_path)['raw']
|
||||
if thresh_rel is not None:
|
||||
# Get threshold values from pure-noise response SD:
|
||||
thresh_abs = np.load(ref_path)['conv'] * thresh_rel
|
||||
thresh_data = dict(np.load(thresh_path))
|
||||
thresh_abs = thresh_rel[:, None] * thresh_data['sds'][None, :]
|
||||
|
||||
# EXECUTION:
|
||||
for data_path, name in zip(data_paths, crop_paths(data_paths)):
|
||||
@@ -66,17 +58,13 @@ for data_path, name in zip(data_paths, crop_paths(data_paths)):
|
||||
data, config = load_data(data_path, files='raw')
|
||||
song, rate = data['raw'], config['rate']
|
||||
|
||||
if thresh_rel is not None:
|
||||
# Set kernel-specific thresholds:
|
||||
config['feat_thresh'] = thresh_abs
|
||||
|
||||
# Reduce to kernel subset:
|
||||
if any(var is not None for var in [kernels, types, sigmas]):
|
||||
kern_inds = find_kern_specs(config['k_specs'], kernels, types, sigmas)
|
||||
config['kernels'] = config['kernels'][:, kern_inds]
|
||||
config['k_specs'] = config['k_specs'][kern_inds, :]
|
||||
config['k_props'] = [config['k_props'][i] for i in kern_inds]
|
||||
config['feat_thresh'] = config['feat_thresh'][kern_inds]
|
||||
thresh_abs = thresh_abs[:, kern_inds]
|
||||
|
||||
# Get song segment to be analyzed:
|
||||
time = np.arange(song.shape[0]) / rate
|
||||
@@ -99,8 +87,8 @@ for data_path, name in zip(data_paths, crop_paths(data_paths)):
|
||||
measure_log=np.zeros(shape_low, dtype=float),
|
||||
measure_inv=np.zeros(shape_low, dtype=float),
|
||||
measure_conv=np.zeros(shape_high, dtype=float),
|
||||
measure_feat=np.zeros(shape_high, dtype=float)
|
||||
)
|
||||
measure_feat=np.zeros(shape_high + (thresh_rel.size,), dtype=float)
|
||||
)
|
||||
if save_detailed:
|
||||
# Prepare optional storage:
|
||||
shape_low = (song.shape[0], example_scales.size)
|
||||
@@ -111,7 +99,7 @@ for data_path, name in zip(data_paths, crop_paths(data_paths)):
|
||||
snip_log=np.zeros(shape_low, dtype=float),
|
||||
snip_inv=np.zeros(shape_low, dtype=float),
|
||||
snip_conv=np.zeros(shape_high, dtype=float),
|
||||
snip_feat=np.zeros(shape_high, dtype=float)
|
||||
snip_feat=np.zeros(shape_high + (thresh_rel.size,), dtype=float)
|
||||
)
|
||||
|
||||
# Execute piecewise:
|
||||
@@ -121,28 +109,40 @@ for data_path, name in zip(data_paths, crop_paths(data_paths)):
|
||||
# Rescale song and add noise:
|
||||
scaled = song * scale + noise
|
||||
|
||||
# Process mixture:
|
||||
signals, rates = process_signal(config, returns=stages,
|
||||
# Process mixture (excluding features):
|
||||
signals, rates = process_signal(config, returns=pre_stages,
|
||||
signal=scaled, rate=rate)
|
||||
# Store results:
|
||||
for stage in stages:
|
||||
# Store non-feature results:
|
||||
for stage in pre_stages:
|
||||
# Log intensity measures:
|
||||
mkey = f'measure_{stage}'
|
||||
if stage == 'feat':
|
||||
measures[mkey][i] = signals[stage][segment, :].mean(axis=0)
|
||||
else:
|
||||
measures[mkey][i] = signals[stage][segment, ...].std(axis=0)
|
||||
measures[f'measure_{stage}'][i] = signals[stage][segment, ...].std(axis=0)
|
||||
|
||||
# Log optional snippet data:
|
||||
if save_detailed and scale in example_scales:
|
||||
scale_ind = np.nonzero(example_scales == scale)[0][0]
|
||||
snippets[f'snip_{stage}'][:, ..., scale_ind] = signals[stage]
|
||||
|
||||
# Execute piecewise again:
|
||||
for j, thresholds in enumerate(thresh_abs):
|
||||
# Finalize processing:
|
||||
feat = sosfilter((signals['conv'] > thresholds).astype(float),
|
||||
rate, config['feat_fcut'], 'lp',
|
||||
padtype='fixed', padlen=config['padlen'])
|
||||
|
||||
# Log intensity measure:
|
||||
measures['measure_feat'][i, :, j] = feat[segment, :].mean(axis=0)
|
||||
|
||||
# Log optional snippet data:
|
||||
if save_detailed and scale in example_scales:
|
||||
snippets['snip_feat'][:, :, scale_ind, j] = feat
|
||||
|
||||
# Save analysis results:
|
||||
if save_path is not None:
|
||||
data = dict(
|
||||
scales=scales,
|
||||
example_scales=example_scales,
|
||||
thresh_rel=thresh_rel,
|
||||
thresh_abs=thresh_abs,
|
||||
)
|
||||
data.update(measures)
|
||||
if save_detailed:
|
||||
|
||||
Reference in New Issue
Block a user