Compare commits
26 Commits
behaviour
...
chirp_body
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
4e7bd40ea4 | ||
|
|
fd2207c8c5 | ||
|
|
ce560bf939 | ||
|
|
ab263d26a2 | ||
|
|
2a32a29d4e | ||
|
|
5763e807d0 | ||
|
|
f36f8606d8 | ||
|
|
fce3503049 | ||
|
|
1064261385 | ||
|
|
fc27fabdb3 | ||
|
|
6193dab97d | ||
|
|
c967a4e5a9 | ||
|
|
87d66dfc2f | ||
|
|
6159121d76 | ||
|
|
a3ddd49040 | ||
|
|
c6facd6f0c | ||
|
|
9bfbe8df5e | ||
|
|
9fcdcfe019 | ||
|
|
d754013aba | ||
|
|
369f26123a | ||
|
|
13b63f4670 | ||
|
|
b755fe8bf3 | ||
|
|
039a027108 | ||
|
|
16e6003958 | ||
|
|
333bb045a6 | ||
|
|
6c2e74a574 |
@@ -18,6 +18,7 @@ from modules.datahandling import (
|
|||||||
purge_duplicates,
|
purge_duplicates,
|
||||||
group_timestamps,
|
group_timestamps,
|
||||||
instantaneous_frequency,
|
instantaneous_frequency,
|
||||||
|
minmaxnorm
|
||||||
)
|
)
|
||||||
|
|
||||||
logger = makeLogger(__name__)
|
logger = makeLogger(__name__)
|
||||||
@@ -26,7 +27,7 @@ ps = PlotStyle()
|
|||||||
|
|
||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
class PlotBuffer:
|
class ChirpPlotBuffer:
|
||||||
|
|
||||||
"""
|
"""
|
||||||
Buffer to save data that is created in the main detection loop
|
Buffer to save data that is created in the main detection loop
|
||||||
@@ -83,6 +84,7 @@ class PlotBuffer:
|
|||||||
q50 + self.search_frequency + self.config.minimal_bandwidth / 2,
|
q50 + self.search_frequency + self.config.minimal_bandwidth / 2,
|
||||||
q50 + self.search_frequency - self.config.minimal_bandwidth / 2,
|
q50 + self.search_frequency - self.config.minimal_bandwidth / 2,
|
||||||
)
|
)
|
||||||
|
print(search_upper, search_lower)
|
||||||
|
|
||||||
# get indices on raw data
|
# get indices on raw data
|
||||||
start_idx = (self.t0 - 5) * self.data.raw_rate
|
start_idx = (self.t0 - 5) * self.data.raw_rate
|
||||||
@@ -94,7 +96,8 @@ class PlotBuffer:
|
|||||||
|
|
||||||
self.time = self.time - self.t0
|
self.time = self.time - self.t0
|
||||||
self.frequency_time = self.frequency_time - self.t0
|
self.frequency_time = self.frequency_time - self.t0
|
||||||
chirps = np.asarray(chirps) - self.t0
|
if len(chirps) > 0:
|
||||||
|
chirps = np.asarray(chirps) - self.t0
|
||||||
self.t0_old = self.t0
|
self.t0_old = self.t0
|
||||||
self.t0 = 0
|
self.t0 = 0
|
||||||
|
|
||||||
@@ -130,7 +133,7 @@ class PlotBuffer:
|
|||||||
data_oi,
|
data_oi,
|
||||||
self.data.raw_rate,
|
self.data.raw_rate,
|
||||||
self.t0 - 5,
|
self.t0 - 5,
|
||||||
[np.max(self.frequency) - 200, np.max(self.frequency) + 200]
|
[np.min(self.frequency) - 100, np.max(self.frequency) + 200]
|
||||||
)
|
)
|
||||||
|
|
||||||
for track_id in self.data.ids:
|
for track_id in self.data.ids:
|
||||||
@@ -145,14 +148,15 @@ class PlotBuffer:
|
|||||||
|
|
||||||
# get tracked frequencies and their times
|
# get tracked frequencies and their times
|
||||||
f = self.data.freq[window_idx]
|
f = self.data.freq[window_idx]
|
||||||
t = self.data.time[
|
# t = self.data.time[
|
||||||
self.data.idx[self.data.ident == self.track_id]]
|
# self.data.idx[self.data.ident == self.track_id]]
|
||||||
tmask = (t >= t0_track) & (t <= (t0_track + dt_track))
|
# tmask = (t >= t0_track) & (t <= (t0_track + dt_track))
|
||||||
|
t = self.data.time[self.data.idx[window_idx]]
|
||||||
if track_id == self.track_id:
|
if track_id == self.track_id:
|
||||||
ax0.plot(t[tmask]-self.t0_old, f, lw=lw,
|
ax0.plot(t-self.t0_old, f, lw=lw,
|
||||||
zorder=10, color=ps.gblue1)
|
zorder=10, color=ps.gblue1)
|
||||||
else:
|
else:
|
||||||
ax0.plot(t[tmask]-self.t0_old, f, lw=lw,
|
ax0.plot(t-self.t0_old, f, lw=lw,
|
||||||
zorder=10, color=ps.gray, alpha=0.5)
|
zorder=10, color=ps.gray, alpha=0.5)
|
||||||
|
|
||||||
ax0.fill_between(
|
ax0.fill_between(
|
||||||
@@ -180,10 +184,11 @@ class PlotBuffer:
|
|||||||
# spec_times[0], spec_times[-1],
|
# spec_times[0], spec_times[-1],
|
||||||
# color=ps.gblue2, lw=2, ls="dashed")
|
# color=ps.gblue2, lw=2, ls="dashed")
|
||||||
|
|
||||||
for chirp in chirps:
|
if len(chirps) > 0:
|
||||||
ax0.scatter(
|
for chirp in chirps:
|
||||||
chirp, np.median(self.frequency) + 150, c=ps.black, marker="v"
|
ax0.scatter(
|
||||||
)
|
chirp, np.median(self.frequency) + 150, c=ps.black, marker="v"
|
||||||
|
)
|
||||||
|
|
||||||
# plot waveform of filtered signal
|
# plot waveform of filtered signal
|
||||||
ax1.plot(self.time, self.baseline * waveform_scaler,
|
ax1.plot(self.time, self.baseline * waveform_scaler,
|
||||||
@@ -318,7 +323,7 @@ def plot_spectrogram(
|
|||||||
aspect="auto",
|
aspect="auto",
|
||||||
origin="lower",
|
origin="lower",
|
||||||
interpolation="gaussian",
|
interpolation="gaussian",
|
||||||
alpha=1,
|
alpha=0.6,
|
||||||
)
|
)
|
||||||
# axis.use_sticky_edges = False
|
# axis.use_sticky_edges = False
|
||||||
return spec_times
|
return spec_times
|
||||||
@@ -431,6 +436,28 @@ def window_median_all_track_ids(
|
|||||||
return frequency_percentiles, track_ids
|
return frequency_percentiles, track_ids
|
||||||
|
|
||||||
|
|
||||||
|
def array_center(array: np.ndarray) -> float:
|
||||||
|
"""
|
||||||
|
Return the center value of an array.
|
||||||
|
If the array length is even, returns
|
||||||
|
the mean of the two center values.
|
||||||
|
|
||||||
|
Parameters
|
||||||
|
----------
|
||||||
|
array : np.ndarray
|
||||||
|
Array to calculate the center from.
|
||||||
|
|
||||||
|
Returns
|
||||||
|
-------
|
||||||
|
float
|
||||||
|
|
||||||
|
"""
|
||||||
|
if len(array) % 2 == 0:
|
||||||
|
return np.mean(array[int(len(array) / 2) - 1:int(len(array) / 2) + 1])
|
||||||
|
else:
|
||||||
|
return array[int(len(array) / 2)]
|
||||||
|
|
||||||
|
|
||||||
def find_searchband(
|
def find_searchband(
|
||||||
current_frequency: np.ndarray,
|
current_frequency: np.ndarray,
|
||||||
percentiles_ids: np.ndarray,
|
percentiles_ids: np.ndarray,
|
||||||
@@ -464,15 +491,17 @@ def find_searchband(
|
|||||||
# frequency window where second filter filters is potentially allowed
|
# frequency window where second filter filters is potentially allowed
|
||||||
# to filter. This is the search window, in which we want to find
|
# to filter. This is the search window, in which we want to find
|
||||||
# a gap in the other fish's EODs.
|
# a gap in the other fish's EODs.
|
||||||
|
current_median = np.median(current_frequency)
|
||||||
search_window = np.arange(
|
search_window = np.arange(
|
||||||
np.median(current_frequency) + config.search_df_lower,
|
current_median + config.search_df_lower,
|
||||||
np.median(current_frequency) + config.search_df_upper,
|
current_median + config.search_df_upper,
|
||||||
config.search_res,
|
config.search_res,
|
||||||
)
|
)
|
||||||
|
|
||||||
# search window in boolean
|
# search window in boolean
|
||||||
search_window_bool = np.ones_like(len(search_window), dtype=bool)
|
bool_lower = np.ones_like(search_window, dtype=bool)
|
||||||
|
bool_upper = np.ones_like(search_window, dtype=bool)
|
||||||
|
search_window_bool = np.ones_like(search_window, dtype=bool)
|
||||||
|
|
||||||
# make seperate arrays from the qartiles
|
# make seperate arrays from the qartiles
|
||||||
q25 = np.asarray([i[0] for i in frequency_percentiles])
|
q25 = np.asarray([i[0] for i in frequency_percentiles])
|
||||||
@@ -480,7 +509,7 @@ def find_searchband(
|
|||||||
|
|
||||||
# get tracks that fall into search window
|
# get tracks that fall into search window
|
||||||
check_track_ids = percentiles_ids[
|
check_track_ids = percentiles_ids[
|
||||||
(q25 > search_window[0]) & (
|
(q25 > current_median) & (
|
||||||
q75 < search_window[-1])
|
q75 < search_window[-1])
|
||||||
]
|
]
|
||||||
|
|
||||||
@@ -492,11 +521,10 @@ def find_searchband(
|
|||||||
q25_temp = q25[percentiles_ids == check_track_id]
|
q25_temp = q25[percentiles_ids == check_track_id]
|
||||||
q75_temp = q75[percentiles_ids == check_track_id]
|
q75_temp = q75[percentiles_ids == check_track_id]
|
||||||
|
|
||||||
print(q25_temp, q75_temp)
|
bool_lower[search_window > q25_temp - config.search_res] = False
|
||||||
|
bool_upper[search_window < q75_temp + config.search_res] = False
|
||||||
search_window_bool[
|
search_window_bool[(bool_lower == False) &
|
||||||
(search_window > q25_temp) & (search_window < q75_temp)
|
(bool_upper == False)] = False
|
||||||
] = False
|
|
||||||
|
|
||||||
# find gaps in search window
|
# find gaps in search window
|
||||||
search_window_indices = np.arange(len(search_window))
|
search_window_indices = np.arange(len(search_window))
|
||||||
@@ -509,6 +537,9 @@ def find_searchband(
|
|||||||
nonzeros = search_window_gaps[np.nonzero(search_window_gaps)[0]]
|
nonzeros = search_window_gaps[np.nonzero(search_window_gaps)[0]]
|
||||||
nonzeros = nonzeros[~np.isnan(nonzeros)]
|
nonzeros = nonzeros[~np.isnan(nonzeros)]
|
||||||
|
|
||||||
|
if len(nonzeros) == 0:
|
||||||
|
return config.default_search_freq
|
||||||
|
|
||||||
# if the first value is -1, the array starst with true, so a gap
|
# if the first value is -1, the array starst with true, so a gap
|
||||||
if nonzeros[0] == -1:
|
if nonzeros[0] == -1:
|
||||||
stops = search_window_indices[search_window_gaps == -1]
|
stops = search_window_indices[search_window_gaps == -1]
|
||||||
@@ -543,16 +574,14 @@ def find_searchband(
|
|||||||
# the center of the search frequency band is then the center of
|
# the center of the search frequency band is then the center of
|
||||||
# the longest gap
|
# the longest gap
|
||||||
|
|
||||||
search_freq = (
|
search_freq = array_center(longest_search_window) - current_median
|
||||||
longest_search_window[-1] - longest_search_window[0]
|
|
||||||
) / 2
|
|
||||||
|
|
||||||
return search_freq
|
return search_freq
|
||||||
|
|
||||||
return config.default_search_freq
|
return config.default_search_freq
|
||||||
|
|
||||||
|
|
||||||
def main(datapath: str, plot: str) -> None:
|
def chirpdetection(datapath: str, plot: str, debug: str = 'false') -> None:
|
||||||
|
|
||||||
assert plot in [
|
assert plot in [
|
||||||
"save",
|
"save",
|
||||||
@@ -560,7 +589,17 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
"false",
|
"false",
|
||||||
], "plot must be 'save', 'show' or 'false'"
|
], "plot must be 'save', 'show' or 'false'"
|
||||||
|
|
||||||
|
assert debug in [
|
||||||
|
"false",
|
||||||
|
"electrode",
|
||||||
|
"fish",
|
||||||
|
], "debug must be 'false', 'electrode' or 'fish'"
|
||||||
|
|
||||||
|
if debug != "false":
|
||||||
|
assert plot == "show", "debug mode only runs when plot is 'show'"
|
||||||
|
|
||||||
# load raw file
|
# load raw file
|
||||||
|
print('datapath', datapath)
|
||||||
data = LoadData(datapath)
|
data = LoadData(datapath)
|
||||||
|
|
||||||
# load config file
|
# load config file
|
||||||
@@ -589,16 +628,16 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
raw_time = np.arange(data.raw.shape[0]) / data.raw_rate
|
raw_time = np.arange(data.raw.shape[0]) / data.raw_rate
|
||||||
|
|
||||||
# good chirp times for data: 2022-06-02-10_00
|
# good chirp times for data: 2022-06-02-10_00
|
||||||
window_start_index = (3 * 60 * 60 + 6 * 60 + 43.5 + 5) * data.raw_rate
|
# window_start_index = (3 * 60 * 60 + 6 * 60 + 43.5 + 5) * data.raw_rate
|
||||||
window_duration_index = 60 * data.raw_rate
|
# window_duration_index = 60 * data.raw_rate
|
||||||
|
|
||||||
# t0 = 0
|
# t0 = 0
|
||||||
# dt = data.raw.shape[0]
|
# dt = data.raw.shape[0]
|
||||||
# window_start_seconds = (23495 + ((28336-23495)/3)) * data.raw_rate
|
# window_start_seconds = (23495 + ((28336-23495)/3)) * data.raw_rate
|
||||||
# window_duration_seconds = (28336 - 23495) * data.raw_rate
|
# window_duration_seconds = (28336 - 23495) * data.raw_rate
|
||||||
|
|
||||||
# window_start_index = 0
|
window_start_index = 0
|
||||||
# window_duration_index = data.raw.shape[0]
|
window_duration_index = data.raw.shape[0]
|
||||||
|
|
||||||
# generate starting points of rolling window
|
# generate starting points of rolling window
|
||||||
window_start_indices = np.arange(
|
window_start_indices = np.arange(
|
||||||
@@ -651,14 +690,14 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
# approximate sampling rate to compute expected durations if there
|
# approximate sampling rate to compute expected durations if there
|
||||||
# is data available for this time window for this fish id
|
# is data available for this time window for this fish id
|
||||||
|
|
||||||
track_samplerate = np.mean(1 / np.diff(data.time))
|
# track_samplerate = np.mean(1 / np.diff(data.time))
|
||||||
expected_duration = (
|
# expected_duration = (
|
||||||
(window_start_seconds + window_duration_seconds)
|
# (window_start_seconds + window_duration_seconds)
|
||||||
- window_start_seconds
|
# - window_start_seconds
|
||||||
) * track_samplerate
|
# ) * track_samplerate
|
||||||
|
|
||||||
# check if tracked data available in this window
|
# check if tracked data available in this window
|
||||||
if len(current_frequencies) < expected_duration / 2:
|
if len(current_frequencies) < 3:
|
||||||
logger.warning(
|
logger.warning(
|
||||||
f"Track {track_id} has no data in window {st}, skipping."
|
f"Track {track_id} has no data in window {st}, skipping."
|
||||||
)
|
)
|
||||||
@@ -750,11 +789,11 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
|
|
||||||
baseline_envelope = -baseline_envelope
|
baseline_envelope = -baseline_envelope
|
||||||
|
|
||||||
baseline_envelope = envelope(
|
# baseline_envelope = envelope(
|
||||||
signal=baseline_envelope,
|
# signal=baseline_envelope,
|
||||||
samplerate=data.raw_rate,
|
# samplerate=data.raw_rate,
|
||||||
cutoff_frequency=config.baseline_envelope_envelope_cutoff,
|
# cutoff_frequency=config.baseline_envelope_envelope_cutoff,
|
||||||
)
|
# )
|
||||||
|
|
||||||
# compute the envelope of the search band. Peaks in the search
|
# compute the envelope of the search band. Peaks in the search
|
||||||
# band envelope correspond to troughs in the baseline envelope
|
# band envelope correspond to troughs in the baseline envelope
|
||||||
@@ -788,25 +827,25 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
# compute the envelope of the signal to remove the oscillations
|
# compute the envelope of the signal to remove the oscillations
|
||||||
# around the peaks
|
# around the peaks
|
||||||
|
|
||||||
baseline_frequency_samplerate = np.mean(
|
# baseline_frequency_samplerate = np.mean(
|
||||||
np.diff(baseline_frequency_time)
|
# np.diff(baseline_frequency_time)
|
||||||
)
|
# )
|
||||||
|
|
||||||
baseline_frequency_filtered = np.abs(
|
baseline_frequency_filtered = np.abs(
|
||||||
baseline_frequency - np.median(baseline_frequency)
|
baseline_frequency - np.median(baseline_frequency)
|
||||||
)
|
)
|
||||||
|
|
||||||
baseline_frequency_filtered = highpass_filter(
|
# baseline_frequency_filtered = highpass_filter(
|
||||||
signal=baseline_frequency_filtered,
|
# signal=baseline_frequency_filtered,
|
||||||
samplerate=baseline_frequency_samplerate,
|
# samplerate=baseline_frequency_samplerate,
|
||||||
cutoff=config.baseline_frequency_highpass_cutoff,
|
# cutoff=config.baseline_frequency_highpass_cutoff,
|
||||||
)
|
# )
|
||||||
|
|
||||||
baseline_frequency_filtered = envelope(
|
# baseline_frequency_filtered = envelope(
|
||||||
signal=-baseline_frequency_filtered,
|
# signal=-baseline_frequency_filtered,
|
||||||
samplerate=baseline_frequency_samplerate,
|
# samplerate=baseline_frequency_samplerate,
|
||||||
cutoff_frequency=config.baseline_frequency_envelope_cutoff,
|
# cutoff_frequency=config.baseline_frequency_envelope_cutoff,
|
||||||
)
|
# )
|
||||||
|
|
||||||
# CUT OFF OVERLAP ---------------------------------------------
|
# CUT OFF OVERLAP ---------------------------------------------
|
||||||
|
|
||||||
@@ -847,25 +886,25 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
# normalize all three feature arrays to the same range to make
|
# normalize all three feature arrays to the same range to make
|
||||||
# peak detection simpler
|
# peak detection simpler
|
||||||
|
|
||||||
baseline_envelope = normalize([baseline_envelope])[0]
|
# baseline_envelope = minmaxnorm([baseline_envelope])[0]
|
||||||
search_envelope = normalize([search_envelope])[0]
|
# search_envelope = minmaxnorm([search_envelope])[0]
|
||||||
baseline_frequency_filtered = normalize(
|
# baseline_frequency_filtered = minmaxnorm(
|
||||||
[baseline_frequency_filtered]
|
# [baseline_frequency_filtered]
|
||||||
)[0]
|
# )[0]
|
||||||
|
|
||||||
# PEAK DETECTION ----------------------------------------------
|
# PEAK DETECTION ----------------------------------------------
|
||||||
|
|
||||||
# detect peaks baseline_enelope
|
# detect peaks baseline_enelope
|
||||||
baseline_peak_indices, _ = find_peaks(
|
baseline_peak_indices, _ = find_peaks(
|
||||||
baseline_envelope, prominence=config.prominence
|
baseline_envelope, prominence=config.baseline_prominence
|
||||||
)
|
)
|
||||||
# detect peaks search_envelope
|
# detect peaks search_envelope
|
||||||
search_peak_indices, _ = find_peaks(
|
search_peak_indices, _ = find_peaks(
|
||||||
search_envelope, prominence=config.prominence
|
search_envelope, prominence=config.search_prominence
|
||||||
)
|
)
|
||||||
# detect peaks inst_freq_filtered
|
# detect peaks inst_freq_filtered
|
||||||
frequency_peak_indices, _ = find_peaks(
|
frequency_peak_indices, _ = find_peaks(
|
||||||
baseline_frequency_filtered, prominence=config.prominence
|
baseline_frequency_filtered, prominence=config.frequency_prominence
|
||||||
)
|
)
|
||||||
|
|
||||||
# DETECT CHIRPS IN SEARCH WINDOW ------------------------------
|
# DETECT CHIRPS IN SEARCH WINDOW ------------------------------
|
||||||
@@ -890,7 +929,7 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
or len(frequency_peak_timestamps) == 0
|
or len(frequency_peak_timestamps) == 0
|
||||||
)
|
)
|
||||||
|
|
||||||
if one_feature_empty:
|
if one_feature_empty and (debug == 'false'):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
# group peak across feature arrays but only if they
|
# group peak across feature arrays but only if they
|
||||||
@@ -911,25 +950,23 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
# check it there are chirps detected after grouping, continue
|
# check it there are chirps detected after grouping, continue
|
||||||
# with the loop if not
|
# with the loop if not
|
||||||
|
|
||||||
if len(singleelectrode_chirps) == 0:
|
if (len(singleelectrode_chirps) == 0) and (debug == 'false'):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
# append chirps from this electrode to the multilectrode list
|
# append chirps from this electrode to the multilectrode list
|
||||||
multielectrode_chirps.append(singleelectrode_chirps)
|
multielectrode_chirps.append(singleelectrode_chirps)
|
||||||
|
|
||||||
# only initialize the plotting buffer if chirps are detected
|
# only initialize the plotting buffer if chirps are detected
|
||||||
chirp_detected = (
|
chirp_detected = (el == (config.number_electrodes - 1)
|
||||||
(el == config.number_electrodes - 1)
|
& (plot in ["show", "save"])
|
||||||
& (len(singleelectrode_chirps) > 0)
|
)
|
||||||
& (plot in ["show", "save"])
|
|
||||||
)
|
|
||||||
|
|
||||||
if chirp_detected:
|
if chirp_detected or (debug != 'elecrode'):
|
||||||
|
|
||||||
logger.debug("Detected chirp, ititialize buffer ...")
|
logger.debug("Detected chirp, ititialize buffer ...")
|
||||||
|
|
||||||
# save data to Buffer
|
# save data to Buffer
|
||||||
buffer = PlotBuffer(
|
buffer = ChirpPlotBuffer(
|
||||||
config=config,
|
config=config,
|
||||||
t0=window_start_seconds,
|
t0=window_start_seconds,
|
||||||
dt=window_duration_seconds,
|
dt=window_duration_seconds,
|
||||||
@@ -954,6 +991,11 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
|
|
||||||
logger.debug("Buffer initialized!")
|
logger.debug("Buffer initialized!")
|
||||||
|
|
||||||
|
if debug == "electrode":
|
||||||
|
logger.info(f'Plotting electrode {el} ...')
|
||||||
|
buffer.plot_buffer(
|
||||||
|
chirps=singleelectrode_chirps, plot=plot)
|
||||||
|
|
||||||
logger.debug(
|
logger.debug(
|
||||||
f"Processed all electrodes for fish {track_id} for this"
|
f"Processed all electrodes for fish {track_id} for this"
|
||||||
"window, sorting chirps ..."
|
"window, sorting chirps ..."
|
||||||
@@ -962,7 +1004,7 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
# check if there are chirps detected in multiple electrodes and
|
# check if there are chirps detected in multiple electrodes and
|
||||||
# continue the loop if not
|
# continue the loop if not
|
||||||
|
|
||||||
if len(multielectrode_chirps) == 0:
|
if (len(multielectrode_chirps) == 0) and (debug == 'false'):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
# validate multielectrode chirps, i.e. check if they are
|
# validate multielectrode chirps, i.e. check if they are
|
||||||
@@ -987,9 +1029,15 @@ def main(datapath: str, plot: str) -> None:
|
|||||||
# if chirps are detected and the plot flag is set, plot the
|
# if chirps are detected and the plot flag is set, plot the
|
||||||
# chirps, otheswise try to delete the buffer if it exists
|
# chirps, otheswise try to delete the buffer if it exists
|
||||||
|
|
||||||
if len(multielectrode_chirps_validated) > 0:
|
if debug == "fish":
|
||||||
|
logger.info(f'Plotting fish {track_id} ...')
|
||||||
|
buffer.plot_buffer(multielectrode_chirps_validated, plot)
|
||||||
|
|
||||||
|
if ((len(multielectrode_chirps_validated) > 0) &
|
||||||
|
(plot in ["show", "save"]) & (debug == 'false')):
|
||||||
try:
|
try:
|
||||||
buffer.plot_buffer(multielectrode_chirps_validated, plot)
|
buffer.plot_buffer(multielectrode_chirps_validated, plot)
|
||||||
|
del buffer
|
||||||
except NameError:
|
except NameError:
|
||||||
pass
|
pass
|
||||||
else:
|
else:
|
||||||
@@ -1049,4 +1097,4 @@ if __name__ == "__main__":
|
|||||||
datapath = "../data/2022-06-02-10_00/"
|
datapath = "../data/2022-06-02-10_00/"
|
||||||
# datapath = "/home/weygoldt/Data/uni/efishdata/2016-colombia/fishgrid/2016-04-09-22_25/"
|
# datapath = "/home/weygoldt/Data/uni/efishdata/2016-colombia/fishgrid/2016-04-09-22_25/"
|
||||||
# datapath = "/home/weygoldt/Data/uni/chirpdetection/GP2023_chirp_detection/data/mount_data/2020-03-13-10_00/"
|
# datapath = "/home/weygoldt/Data/uni/chirpdetection/GP2023_chirp_detection/data/mount_data/2020-03-13-10_00/"
|
||||||
main(datapath, plot="save")
|
chirpdetection(datapath, plot="show", debug="false")
|
||||||
|
|||||||
@@ -1,47 +1,41 @@
|
|||||||
# directory setup
|
# Path setup ------------------------------------------------------------------
|
||||||
dataroot: "../data/"
|
|
||||||
outputdir: "../output/"
|
|
||||||
|
|
||||||
# Duration and overlap of the analysis window in seconds
|
dataroot: "../data/" # path to data
|
||||||
window: 10
|
outputdir: "../output/" # path to save plots to
|
||||||
overlap: 1
|
|
||||||
edge: 0.25
|
|
||||||
|
|
||||||
# Number of electrodes to go over
|
# Rolling window parameters ---------------------------------------------------
|
||||||
number_electrodes: 3
|
|
||||||
minimum_electrodes: 2
|
|
||||||
|
|
||||||
# Search window bandwidth and minimal baseline bandwidth
|
window: 5 # rolling window length in seconds
|
||||||
minimal_bandwidth: 20
|
overlap: 1 # window overlap in seconds
|
||||||
|
edge: 0.25 # window edge cufoffs to mitigate filter edge effects
|
||||||
|
|
||||||
# Instantaneous frequency smoothing usint a gaussian kernel of this width
|
# Electrode iteration parameters ----------------------------------------------
|
||||||
baseline_frequency_smoothing: 5
|
|
||||||
|
|
||||||
# Baseline processing parameters
|
number_electrodes: 2 # number of electrodes to go over
|
||||||
baseline_envelope_cutoff: 25
|
minimum_electrodes: 1 # mimumun number of electrodes a chirp must be on
|
||||||
baseline_envelope_bandpass_lowf: 4
|
|
||||||
baseline_envelope_bandpass_highf: 100
|
|
||||||
baseline_envelope_envelope_cutoff: 4
|
|
||||||
|
|
||||||
# search envelope processing parameters
|
# Feature extraction parameters -----------------------------------------------
|
||||||
search_envelope_cutoff: 5
|
|
||||||
|
|
||||||
# Instantaneous frequency bandpass filter cutoff frequencies
|
search_df_lower: 20 # start searching this far above the baseline
|
||||||
baseline_frequency_highpass_cutoff: 0.000005
|
search_df_upper: 100 # stop searching this far above the baseline
|
||||||
baseline_frequency_envelope_cutoff: 0.000005
|
search_res: 1 # search window resolution
|
||||||
|
default_search_freq: 60 # search here if no need for a search frequency
|
||||||
|
minimal_bandwidth: 10 # minimal bandpass filter width for baseline
|
||||||
|
search_bandwidth: 10 # minimal bandpass filter width for search frequency
|
||||||
|
baseline_frequency_smoothing: 10 # instantaneous frequency smoothing
|
||||||
|
|
||||||
# peak detecion parameters
|
# Feature processing parameters -----------------------------------------------
|
||||||
prominence: 0.005
|
|
||||||
|
|
||||||
# search freq parameter
|
baseline_envelope_cutoff: 25 # envelope estimation cutoff
|
||||||
search_df_lower: 20
|
baseline_envelope_bandpass_lowf: 2 # envelope badpass lower cutoff
|
||||||
search_df_upper: 100
|
baseline_envelope_bandpass_highf: 100 # envelope bandbass higher cutoff
|
||||||
search_res: 1
|
search_envelope_cutoff: 10 # search envelope estimation cufoff
|
||||||
search_bandwidth: 10
|
|
||||||
default_search_freq: 50
|
# Peak detecion parameters ----------------------------------------------------
|
||||||
|
baseline_prominence: 0.00005 # peak prominence threshold for baseline envelope
|
||||||
|
search_prominence: 0.000004 # peak prominence threshold for search envelope
|
||||||
|
frequency_prominence: 2 # peak prominence threshold for baseline freq
|
||||||
|
|
||||||
# Classify events as chirps if they are less than this time apart
|
# Classify events as chirps if they are less than this time apart
|
||||||
chirp_window_threshold: 0.05
|
chirp_window_threshold: 0.02
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
48
code/extract_chirps.py
Normal file
48
code/extract_chirps.py
Normal file
@@ -0,0 +1,48 @@
|
|||||||
|
import os
|
||||||
|
import pandas as pd
|
||||||
|
import numpy as np
|
||||||
|
from chirpdetection import chirpdetection
|
||||||
|
from IPython import embed
|
||||||
|
|
||||||
|
|
||||||
|
def main(datapaths):
|
||||||
|
|
||||||
|
for path in datapaths:
|
||||||
|
chirpdetection(path, plot='show')
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
|
||||||
|
dataroot = '../data/mount_data/'
|
||||||
|
|
||||||
|
datasets = sorted([name for name in os.listdir(dataroot) if os.path.isdir(
|
||||||
|
os.path.join(dataroot, name))])
|
||||||
|
|
||||||
|
valid_datasets = []
|
||||||
|
|
||||||
|
for dataset in datasets:
|
||||||
|
|
||||||
|
path = os.path.join(dataroot, dataset)
|
||||||
|
csv_name = '-'.join(dataset.split('-')[:3]) + '.csv'
|
||||||
|
|
||||||
|
if os.path.exists(os.path.join(path, csv_name)) is False:
|
||||||
|
continue
|
||||||
|
|
||||||
|
if os.path.exists(os.path.join(path, 'ident_v.npy')) is False:
|
||||||
|
continue
|
||||||
|
|
||||||
|
ident = np.load(os.path.join(path, 'ident_v.npy'))
|
||||||
|
number_of_fish = len(np.unique(ident[~np.isnan(ident)]))
|
||||||
|
if number_of_fish != 2:
|
||||||
|
continue
|
||||||
|
|
||||||
|
valid_datasets.append(dataset)
|
||||||
|
|
||||||
|
datapaths = [os.path.join(dataroot, dataset) +
|
||||||
|
'/' for dataset in valid_datasets]
|
||||||
|
|
||||||
|
recs = pd.DataFrame(columns=['recording'], data=valid_datasets)
|
||||||
|
recs.to_csv('../recs.csv', index=False)
|
||||||
|
main(datapaths)
|
||||||
|
|
||||||
|
# window 1524 + 244 in dataset index 4 is nice example
|
||||||
35
code/get_behaviour.py
Normal file
35
code/get_behaviour.py
Normal file
@@ -0,0 +1,35 @@
|
|||||||
|
import os
|
||||||
|
from paramiko import SSHClient
|
||||||
|
from scp import SCPClient
|
||||||
|
from IPython import embed
|
||||||
|
from pandas import read_csv
|
||||||
|
|
||||||
|
ssh = SSHClient()
|
||||||
|
ssh.load_system_host_keys()
|
||||||
|
|
||||||
|
ssh.connect(hostname='kraken',
|
||||||
|
username='efish',
|
||||||
|
password='fwNix4U',
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
# SCPCLient takes a paramiko transport as its only argument
|
||||||
|
scp = SCPClient(ssh.get_transport())
|
||||||
|
|
||||||
|
data = read_csv('../recs.csv')
|
||||||
|
foldernames = data['recording'].values
|
||||||
|
|
||||||
|
directory = f'/Users/acfw/Documents/uni_tuebingen/chirpdetection/GP2023_chirp_detection/data/mount_data/'
|
||||||
|
for foldername in foldernames:
|
||||||
|
|
||||||
|
if not os.path.exists(directory+foldername):
|
||||||
|
os.makedirs(directory+foldername)
|
||||||
|
|
||||||
|
files = [('-').join(foldername.split('-')[:3])+'.csv','chirp_ids.npy', 'chirps.npy', 'fund_v.npy', 'ident_v.npy', 'idx_v.npy', 'times.npy', 'spec.npy', 'LED_on_time.npy', 'sign_v.npy']
|
||||||
|
|
||||||
|
|
||||||
|
for f in files:
|
||||||
|
scp.get(f'/home/efish/behavior/2019_tube_competition/{foldername}/{f}',
|
||||||
|
directory+foldername)
|
||||||
|
|
||||||
|
scp.close()
|
||||||
99
code/modules/behaviour_handling.py
Normal file
99
code/modules/behaviour_handling.py
Normal file
@@ -0,0 +1,99 @@
|
|||||||
|
import numpy as np
|
||||||
|
|
||||||
|
import os
|
||||||
|
|
||||||
|
import numpy as np
|
||||||
|
from IPython import embed
|
||||||
|
|
||||||
|
|
||||||
|
from pandas import read_csv
|
||||||
|
from modules.logger import makeLogger
|
||||||
|
|
||||||
|
|
||||||
|
logger = makeLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
class Behavior:
|
||||||
|
"""Load behavior data from csv file as class attributes
|
||||||
|
Attributes
|
||||||
|
----------
|
||||||
|
behavior: 0: chasing onset, 1: chasing offset, 2: physical contact
|
||||||
|
behavior_type:
|
||||||
|
behavioral_category:
|
||||||
|
comment_start:
|
||||||
|
comment_stop:
|
||||||
|
dataframe: pandas dataframe with all the data
|
||||||
|
duration_s:
|
||||||
|
media_file:
|
||||||
|
observation_date:
|
||||||
|
observation_id:
|
||||||
|
start_s: start time of the event in seconds
|
||||||
|
stop_s: stop time of the event in seconds
|
||||||
|
total_length:
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, folder_path: str) -> None:
|
||||||
|
|
||||||
|
LED_on_time_BORIS = np.load(os.path.join(folder_path, 'LED_on_time.npy'), allow_pickle=True)
|
||||||
|
|
||||||
|
csv_filename = [f for f in os.listdir(folder_path) if f.endswith('.csv')][0]
|
||||||
|
logger.info(f'CSV file: {csv_filename}')
|
||||||
|
self.dataframe = read_csv(os.path.join(folder_path, csv_filename))
|
||||||
|
|
||||||
|
self.chirps = np.load(os.path.join(folder_path, 'chirps.npy'), allow_pickle=True)
|
||||||
|
self.chirps_ids = np.load(os.path.join(folder_path, 'chirp_ids.npy'), allow_pickle=True)
|
||||||
|
|
||||||
|
self.ident = np.load(os.path.join(folder_path, 'ident_v.npy'), allow_pickle=True)
|
||||||
|
self.idx = np.load(os.path.join(folder_path, 'idx_v.npy'), allow_pickle=True)
|
||||||
|
self.freq = np.load(os.path.join(folder_path, 'fund_v.npy'), allow_pickle=True)
|
||||||
|
self.time = np.load(os.path.join(folder_path, "times.npy"), allow_pickle=True)
|
||||||
|
self.spec = np.load(os.path.join(folder_path, "spec.npy"), allow_pickle=True)
|
||||||
|
|
||||||
|
for k, key in enumerate(self.dataframe.keys()):
|
||||||
|
key = key.lower()
|
||||||
|
if ' ' in key:
|
||||||
|
key = key.replace(' ', '_')
|
||||||
|
if '(' in key:
|
||||||
|
key = key.replace('(', '')
|
||||||
|
key = key.replace(')', '')
|
||||||
|
setattr(self, key, np.array(self.dataframe[self.dataframe.keys()[k]]))
|
||||||
|
|
||||||
|
last_LED_t_BORIS = LED_on_time_BORIS[-1]
|
||||||
|
real_time_range = self.time[-1] - self.time[0]
|
||||||
|
factor = 1.034141
|
||||||
|
shift = last_LED_t_BORIS - real_time_range * factor
|
||||||
|
self.start_s = (self.start_s - shift) / factor
|
||||||
|
self.stop_s = (self.stop_s - shift) / factor
|
||||||
|
|
||||||
|
|
||||||
|
def correct_chasing_events(
|
||||||
|
category: np.ndarray,
|
||||||
|
timestamps: np.ndarray
|
||||||
|
) -> tuple[np.ndarray, np.ndarray]:
|
||||||
|
|
||||||
|
onset_ids = np.arange(
|
||||||
|
len(category))[category == 0]
|
||||||
|
offset_ids = np.arange(
|
||||||
|
len(category))[category == 1]
|
||||||
|
|
||||||
|
woring_bh = np.arange(len(category))[category!=2][:-1][np.diff(category[category!=2])==0]
|
||||||
|
if onset_ids[0] > offset_ids[0]:
|
||||||
|
offset_ids = np.delete(offset_ids, 0)
|
||||||
|
help_index = offset_ids[0]
|
||||||
|
woring_bh = np.append(woring_bh, help_index)
|
||||||
|
|
||||||
|
category = np.delete(category, woring_bh)
|
||||||
|
timestamps = np.delete(timestamps, woring_bh)
|
||||||
|
|
||||||
|
# Check whether on- or offset is longer and calculate length difference
|
||||||
|
if len(onset_ids) > len(offset_ids):
|
||||||
|
len_diff = len(onset_ids) - len(offset_ids)
|
||||||
|
logger.info(f'Onsets are greater than offsets by {len_diff}')
|
||||||
|
elif len(onset_ids) < len(offset_ids):
|
||||||
|
len_diff = len(offset_ids) - len(onset_ids)
|
||||||
|
logger.info(f'Offsets are greater than onsets by {len_diff}')
|
||||||
|
elif len(onset_ids) == len(offset_ids):
|
||||||
|
logger.info('Chasing events are equal')
|
||||||
|
|
||||||
|
|
||||||
|
return category, timestamps
|
||||||
@@ -1,9 +1,10 @@
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
from typing import List, Any
|
from typing import List, Any
|
||||||
from scipy.ndimage import gaussian_filter1d
|
from scipy.ndimage import gaussian_filter1d
|
||||||
|
from scipy.stats import gamma, norm
|
||||||
|
|
||||||
|
|
||||||
def norm(data):
|
def minmaxnorm(data):
|
||||||
"""
|
"""
|
||||||
Normalize data to [0, 1]
|
Normalize data to [0, 1]
|
||||||
|
|
||||||
@@ -18,7 +19,7 @@ def norm(data):
|
|||||||
Normalized data.
|
Normalized data.
|
||||||
|
|
||||||
"""
|
"""
|
||||||
return (2*((data - np.min(data)) / (np.max(data) - np.min(data)))) - 1
|
return (data - np.min(data)) / (np.max(data) - np.min(data))
|
||||||
|
|
||||||
|
|
||||||
def instantaneous_frequency(
|
def instantaneous_frequency(
|
||||||
@@ -167,6 +168,9 @@ def group_timestamps(
|
|||||||
]
|
]
|
||||||
timestamps.sort()
|
timestamps.sort()
|
||||||
|
|
||||||
|
if len(timestamps) == 0:
|
||||||
|
return []
|
||||||
|
|
||||||
groups = []
|
groups = []
|
||||||
current_group = [timestamps[0]]
|
current_group = [timestamps[0]]
|
||||||
|
|
||||||
@@ -209,6 +213,117 @@ def flatten(list: List[List[Any]]) -> List:
|
|||||||
return [item for sublist in list for item in sublist]
|
return [item for sublist in list for item in sublist]
|
||||||
|
|
||||||
|
|
||||||
|
def causal_kde1d(spikes, time, width, shape=2):
|
||||||
|
"""
|
||||||
|
causalkde computes a kernel density estimate using a causal kernel (i.e. exponential or gamma distribution).
|
||||||
|
A shape of 1 turns the gamma distribution into an exponential.
|
||||||
|
|
||||||
|
Parameters
|
||||||
|
----------
|
||||||
|
spikes : array-like
|
||||||
|
spike times
|
||||||
|
time : array-like
|
||||||
|
sampling time
|
||||||
|
width : float
|
||||||
|
kernel width
|
||||||
|
shape : int, optional
|
||||||
|
shape of gamma distribution, by default 1
|
||||||
|
|
||||||
|
Returns
|
||||||
|
-------
|
||||||
|
rate : array-like
|
||||||
|
instantaneous firing rate
|
||||||
|
"""
|
||||||
|
|
||||||
|
# compute dt
|
||||||
|
dt = time[1] - time[0]
|
||||||
|
|
||||||
|
# time on which to compute kernel:
|
||||||
|
tmax = 10 * width
|
||||||
|
|
||||||
|
# kernel not wider than time
|
||||||
|
if 2 * tmax > time[-1] - time[0]:
|
||||||
|
tmax = 0.5 * (time[-1] - time[0])
|
||||||
|
|
||||||
|
# kernel time
|
||||||
|
ktime = np.arange(-tmax, tmax, dt)
|
||||||
|
|
||||||
|
# gamma kernel centered in ktime:
|
||||||
|
kernel = gamma.pdf(
|
||||||
|
x=ktime,
|
||||||
|
a=shape,
|
||||||
|
loc=0,
|
||||||
|
scale=width,
|
||||||
|
)
|
||||||
|
|
||||||
|
# indices of spikes in time array:
|
||||||
|
indices = np.asarray((spikes - time[0]) / dt, dtype=int)
|
||||||
|
|
||||||
|
# binary spike train:
|
||||||
|
brate = np.zeros(len(time))
|
||||||
|
brate[indices[(indices >= 0) & (indices < len(time))]] = 1.0
|
||||||
|
|
||||||
|
# convolution with kernel:
|
||||||
|
rate = np.convolve(brate, kernel, mode="same")
|
||||||
|
|
||||||
|
return rate
|
||||||
|
|
||||||
|
|
||||||
|
def acausal_kde1d(spikes, time, width):
|
||||||
|
"""
|
||||||
|
causalkde computes a kernel density estimate using a causal kernel (i.e. exponential or gamma distribution).
|
||||||
|
A shape of 1 turns the gamma distribution into an exponential.
|
||||||
|
|
||||||
|
Parameters
|
||||||
|
----------
|
||||||
|
spikes : array-like
|
||||||
|
spike times
|
||||||
|
time : array-like
|
||||||
|
sampling time
|
||||||
|
width : float
|
||||||
|
kernel width
|
||||||
|
shape : int, optional
|
||||||
|
shape of gamma distribution, by default 1
|
||||||
|
|
||||||
|
Returns
|
||||||
|
-------
|
||||||
|
rate : array-like
|
||||||
|
instantaneous firing rate
|
||||||
|
"""
|
||||||
|
|
||||||
|
# compute dt
|
||||||
|
dt = time[1] - time[0]
|
||||||
|
|
||||||
|
# time on which to compute kernel:
|
||||||
|
tmax = 10 * width
|
||||||
|
|
||||||
|
# kernel not wider than time
|
||||||
|
if 2 * tmax > time[-1] - time[0]:
|
||||||
|
tmax = 0.5 * (time[-1] - time[0])
|
||||||
|
|
||||||
|
# kernel time
|
||||||
|
ktime = np.arange(-tmax, tmax, dt)
|
||||||
|
|
||||||
|
# gamma kernel centered in ktime:
|
||||||
|
kernel = norm.pdf(
|
||||||
|
x=ktime,
|
||||||
|
loc=0,
|
||||||
|
scale=width,
|
||||||
|
)
|
||||||
|
|
||||||
|
# indices of spikes in time array:
|
||||||
|
indices = np.asarray((spikes - time[0]) / dt, dtype=int)
|
||||||
|
|
||||||
|
# binary spike train:
|
||||||
|
brate = np.zeros(len(time))
|
||||||
|
brate[indices[(indices >= 0) & (indices < len(time))]] = 1.0
|
||||||
|
|
||||||
|
# convolution with kernel:
|
||||||
|
rate = np.convolve(brate, kernel, mode="same")
|
||||||
|
|
||||||
|
return rate
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
|
||||||
timestamps = [
|
timestamps = [
|
||||||
|
|||||||
87
code/plot_chirp_bodylegth.py
Normal file
87
code/plot_chirp_bodylegth.py
Normal file
@@ -0,0 +1,87 @@
|
|||||||
|
import numpy as np
|
||||||
|
|
||||||
|
import os
|
||||||
|
|
||||||
|
import numpy as np
|
||||||
|
import matplotlib.pyplot as plt
|
||||||
|
from thunderfish.powerspectrum import decibel
|
||||||
|
|
||||||
|
from IPython import embed
|
||||||
|
from pandas import read_csv
|
||||||
|
from modules.logger import makeLogger
|
||||||
|
from modules.plotstyle import PlotStyle
|
||||||
|
from modules.behaviour_handling import Behavior, correct_chasing_events
|
||||||
|
|
||||||
|
ps = PlotStyle()
|
||||||
|
|
||||||
|
logger = makeLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def main(datapath: str):
|
||||||
|
|
||||||
|
foldernames = [datapath + x + '/' for x in os.listdir(datapath) if os.path.isdir(datapath+x)]
|
||||||
|
path_to_csv = ('/').join(foldernames[0].split('/')[:-2]) + '/order_meta.csv'
|
||||||
|
meta_id = read_csv(path_to_csv)
|
||||||
|
meta_id['recording'] = meta_id['recording'].str[1:-1]
|
||||||
|
|
||||||
|
chirps_winner = []
|
||||||
|
chirps_loser = []
|
||||||
|
|
||||||
|
for foldername in foldernames:
|
||||||
|
# behabvior is pandas dataframe with all the data
|
||||||
|
if foldername == '../data/mount_data/2020-05-12-10_00/':
|
||||||
|
continue
|
||||||
|
bh = Behavior(foldername)
|
||||||
|
# chirps are not sorted in time (presumably due to prior groupings)
|
||||||
|
# get and sort chirps and corresponding fish_ids of the chirps
|
||||||
|
category = bh.behavior
|
||||||
|
timestamps = bh.start_s
|
||||||
|
# Correct for doubles in chasing on- and offsets to get the right on-/offset pairs
|
||||||
|
# Get rid of tracking faults (two onsets or two offsets after another)
|
||||||
|
category, timestamps = correct_chasing_events(category, timestamps)
|
||||||
|
|
||||||
|
folder_name = foldername.split('/')[-2]
|
||||||
|
winner_row = meta_id[meta_id['recording'] == folder_name]
|
||||||
|
winner = winner_row['winner'].values[0].astype(int)
|
||||||
|
winner_fish1 = winner_row['fish1'].values[0].astype(int)
|
||||||
|
winner_fish2 = winner_row['fish2'].values[0].astype(int)
|
||||||
|
if winner == winner_fish1:
|
||||||
|
winner_fish_id = winner_row['rec_id1'].values[0]
|
||||||
|
loser_fish_id = winner_row['rec_id2'].values[0]
|
||||||
|
elif winner == winner_fish2:
|
||||||
|
winner_fish_id = winner_row['rec_id2'].values[0]
|
||||||
|
loser_fish_id = winner_row['rec_id1'].values[0]
|
||||||
|
else:
|
||||||
|
continue
|
||||||
|
|
||||||
|
print(foldername)
|
||||||
|
all_fish_ids = np.unique(bh.chirps_ids)
|
||||||
|
chirp_loser = len(bh.chirps[bh.chirps_ids == loser_fish_id])
|
||||||
|
chirp_winner = len(bh.chirps[bh.chirps_ids == winner_fish_id])
|
||||||
|
chirps_winner.append(chirp_winner)
|
||||||
|
chirps_loser.append(chirp_loser)
|
||||||
|
|
||||||
|
|
||||||
|
fish1_id = all_fish_ids[0]
|
||||||
|
fish2_id = all_fish_ids[1]
|
||||||
|
print(winner_fish_id)
|
||||||
|
print(all_fish_ids)
|
||||||
|
|
||||||
|
|
||||||
|
fig, ax = plt.subplots()
|
||||||
|
ax.boxplot([chirps_winner, chirps_loser], showfliers=False)
|
||||||
|
ax.scatter(np.ones(len(chirps_winner)), chirps_winner, color='r')
|
||||||
|
ax.scatter(np.ones(len(chirps_loser))*2, chirps_loser, color='r')
|
||||||
|
ax.set_xticklabels(['winner', 'loser'])
|
||||||
|
for w, l in zip(chirps_winner, chirps_loser):
|
||||||
|
ax.plot([1,2], [w,l], color='r', alpha=0.5, linewidth=0.5)
|
||||||
|
|
||||||
|
ax.set_ylabel('Chirpscounts [n]')
|
||||||
|
plt.show()
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
|
||||||
|
# Path to the data
|
||||||
|
datapath = '../data/mount_data/'
|
||||||
|
|
||||||
|
main(datapath)
|
||||||
111
code/plot_event_timeline.py
Normal file
111
code/plot_event_timeline.py
Normal file
@@ -0,0 +1,111 @@
|
|||||||
|
import numpy as np
|
||||||
|
|
||||||
|
import os
|
||||||
|
|
||||||
|
import numpy as np
|
||||||
|
import matplotlib.pyplot as plt
|
||||||
|
from thunderfish.powerspectrum import decibel
|
||||||
|
|
||||||
|
from IPython import embed
|
||||||
|
from pandas import read_csv
|
||||||
|
from modules.logger import makeLogger
|
||||||
|
from modules.plotstyle import PlotStyle
|
||||||
|
from modules.behaviour_handling import Behavior, correct_chasing_events
|
||||||
|
|
||||||
|
ps = PlotStyle()
|
||||||
|
|
||||||
|
logger = makeLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def main(datapath: str):
|
||||||
|
|
||||||
|
foldernames = [datapath + x + '/' for x in os.listdir(datapath) if os.path.isdir(datapath+x)]
|
||||||
|
for foldername in foldernames:
|
||||||
|
if foldername == '../data/mount_data/2020-05-12-10_00/':
|
||||||
|
continue
|
||||||
|
# behabvior is pandas dataframe with all the data
|
||||||
|
bh = Behavior(foldername)
|
||||||
|
|
||||||
|
category = bh.behavior
|
||||||
|
timestamps = bh.start_s
|
||||||
|
# Correct for doubles in chasing on- and offsets to get the right on-/offset pairs
|
||||||
|
# Get rid of tracking faults (two onsets or two offsets after another)
|
||||||
|
category, timestamps = correct_chasing_events(category, timestamps)
|
||||||
|
|
||||||
|
# split categories
|
||||||
|
chasing_onset = (timestamps[category == 0]/ 60) /60
|
||||||
|
chasing_offset = (timestamps[category == 1]/ 60) /60
|
||||||
|
physical_contact = (timestamps[category == 2] / 60) /60
|
||||||
|
|
||||||
|
all_fish_ids = np.unique(bh.chirps_ids)
|
||||||
|
fish1_id = all_fish_ids[0]
|
||||||
|
fish2_id = all_fish_ids[1]
|
||||||
|
# Associate chirps to inidividual fish
|
||||||
|
fish1 = (bh.chirps[bh.chirps_ids == fish1_id] / 60) /60
|
||||||
|
fish2 = (bh.chirps[bh.chirps_ids == fish2_id] / 60) /60
|
||||||
|
fish1_color = ps.red
|
||||||
|
fish2_color = ps.orange
|
||||||
|
|
||||||
|
fig, ax = plt.subplots(4, 1, figsize=(10, 5), height_ratios=[0.5, 0.5, 0.5, 6], sharex=True)
|
||||||
|
# marker size
|
||||||
|
s = 200
|
||||||
|
ax[0].scatter(physical_contact, np.ones(len(physical_contact)), color='firebrick', marker='|', s=s)
|
||||||
|
ax[1].scatter(chasing_onset, np.ones(len(chasing_onset)), color='green', marker='|', s=s )
|
||||||
|
ax[2].scatter(fish1, np.ones(len(fish1))-0.25, color=fish1_color, marker='|', s=s)
|
||||||
|
ax[2].scatter(fish2, np.zeros(len(fish2))+0.25, color=fish2_color, marker='|', s=s)
|
||||||
|
|
||||||
|
|
||||||
|
freq_temp = bh.freq[bh.ident==fish1_id]
|
||||||
|
time_temp = bh.time[bh.idx[bh.ident==fish1_id]]
|
||||||
|
ax[3].plot((time_temp/ 60) /60, freq_temp, color=fish1_color)
|
||||||
|
|
||||||
|
freq_temp = bh.freq[bh.ident==fish2_id]
|
||||||
|
time_temp = bh.time[bh.idx[bh.ident==fish2_id]]
|
||||||
|
ax[3].plot((time_temp/ 60) /60, freq_temp, color=fish2_color)
|
||||||
|
|
||||||
|
#ax[3].imshow(decibel(bh.spec), extent=[bh.time[0]/60/60, bh.time[-1]/60/60, 0, 2000], aspect='auto', origin='lower')
|
||||||
|
|
||||||
|
# Hide grid lines
|
||||||
|
ax[0].grid(False)
|
||||||
|
ax[0].set_frame_on(False)
|
||||||
|
ax[0].set_xticks([])
|
||||||
|
ax[0].set_yticks([])
|
||||||
|
ps.hide_ax(ax[0])
|
||||||
|
|
||||||
|
|
||||||
|
ax[1].grid(False)
|
||||||
|
ax[1].set_frame_on(False)
|
||||||
|
ax[1].set_xticks([])
|
||||||
|
ax[1].set_yticks([])
|
||||||
|
ps.hide_ax(ax[1])
|
||||||
|
|
||||||
|
ax[2].grid(False)
|
||||||
|
ax[2].set_frame_on(False)
|
||||||
|
ax[2].set_yticks([])
|
||||||
|
ax[2].set_xticks([])
|
||||||
|
ps.hide_ax(ax[2])
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
ax[3].axvspan(0, 3, 0, 5, facecolor='grey', alpha=0.5)
|
||||||
|
ax[3].set_xticks(np.arange(0, 6.1, 0.5))
|
||||||
|
|
||||||
|
labelpad = 40
|
||||||
|
ax[0].set_ylabel('Physical contact', rotation=0, labelpad=labelpad)
|
||||||
|
ax[1].set_ylabel('Chasing events', rotation=0, labelpad=labelpad)
|
||||||
|
ax[2].set_ylabel('Chirps', rotation=0, labelpad=labelpad)
|
||||||
|
ax[3].set_ylabel('EODf')
|
||||||
|
|
||||||
|
ax[3].set_xlabel('Time [h]')
|
||||||
|
ax[0].set_title(foldername.split('/')[-2])
|
||||||
|
|
||||||
|
plt.show()
|
||||||
|
embed()
|
||||||
|
|
||||||
|
# plot chirps
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
# Path to the data
|
||||||
|
datapath = '../data/mount_data/'
|
||||||
|
main(datapath)
|
||||||
29
recs.csv
Normal file
29
recs.csv
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
recording
|
||||||
|
2020-03-13-10_00
|
||||||
|
2020-03-16-10_00
|
||||||
|
2020-03-19-10_00
|
||||||
|
2020-03-20-10_00
|
||||||
|
2020-03-23-09_58
|
||||||
|
2020-03-24-10_00
|
||||||
|
2020-03-25-10_00
|
||||||
|
2020-03-31-09_59
|
||||||
|
2020-05-11-10_00
|
||||||
|
2020-05-12-10_00
|
||||||
|
2020-05-13-10_00
|
||||||
|
2020-05-14-10_00
|
||||||
|
2020-05-15-10_00
|
||||||
|
2020-05-18-10_00
|
||||||
|
2020-05-19-10_00
|
||||||
|
2020-05-21-10_00
|
||||||
|
2020-05-25-10_00
|
||||||
|
2020-05-27-10_00
|
||||||
|
2020-05-28-10_00
|
||||||
|
2020-05-29-10_00
|
||||||
|
2020-06-02-10_00
|
||||||
|
2020-06-03-10_10
|
||||||
|
2020-06-04-10_00
|
||||||
|
2020-06-05-10_00
|
||||||
|
2020-06-08-10_00
|
||||||
|
2020-06-09-10_00
|
||||||
|
2020-06-10-10_00
|
||||||
|
2020-06-11-10_00
|
||||||
|
Reference in New Issue
Block a user