83 lines
3.6 KiB
Python
83 lines
3.6 KiB
Python
import os #compability with windows
|
|
from IPython import embed
|
|
import numpy as np
|
|
|
|
def parse_dataset(dataset_name):
|
|
assert(os.path.exists(dataset_name)) #see if data exists
|
|
f = open(dataset_name, 'r') #open data we gave in
|
|
lines = f.readlines() #read data
|
|
f.close() #?
|
|
|
|
# metadata lists for every loop
|
|
eodfs = []
|
|
deltafs = []
|
|
stimulusfs = []
|
|
|
|
# data itself
|
|
times = []
|
|
frequencies = []
|
|
amplitudes = []
|
|
|
|
# temporary lists with data we put in the lists above
|
|
time = []
|
|
ampl = []
|
|
freq = []
|
|
|
|
for i in range(len(lines)):
|
|
l = lines[i].strip() #all lines of textdata, exclude all empty lines (empty () default for spacebar)
|
|
if "#" in l and "EODf" in l: #if line starts with # EODf:
|
|
eodfs.append(float(l.split(':')[-1].strip()[:-2])) #append: line splitted by ':' the 2nd part ([-1],
|
|
if "#" in l and "Delta f" in l: #which got striped so we sure there is no space at the end,
|
|
deltafs.append(float(l.split(':')[-1].strip()[:-2])) #from that all expect the last two signs (Hz unit)
|
|
if "#" in l and "StimulusFrequency" in l: #this for different metadata in different lists
|
|
stimulusfs.append(float(l.split(':')[-1].strip()[:-2]))
|
|
|
|
if '#Key' in l:
|
|
#print('KEY')
|
|
if len(time) != 0: #therefore empty in the first round
|
|
times.append(time) #2nd loop means time != 0, so we put the times/amplitudes/frequencies to
|
|
amplitudes.append(ampl) #the data of the first loop
|
|
frequencies.append(freq)
|
|
|
|
time = [] #temporary lists to overwrite the lists with the same name we made before
|
|
ampl = [] #so they are empty again
|
|
freq = []
|
|
print(len(times))
|
|
|
|
if len(l) > 0 and l[0] is not '#': #line not empty and doesnt start with #
|
|
temporary = list(map(float, l.split())) #temporary list where we got 3 index splitted by spacebar, map to find them
|
|
time.append(temporary[0]) #temporary lists with the data at that place, respectively
|
|
freq.append(temporary[1])
|
|
ampl.append(temporary[2])
|
|
|
|
times.append(time) #append data from one list to another
|
|
amplitudes.append(ampl) #these append the data from the first loop to the final lists, because we overwrite them (?)
|
|
frequencies.append(freq)
|
|
|
|
return times, frequencies, amplitudes, eodfs, deltafs, stimulusfs #output of the function
|
|
|
|
|
|
|
|
def noise_reduce(dataset_name, n):
|
|
assert (os.path.exists(dataset_name)) # see if data exists
|
|
f = open(dataset_name, 'r') # open data we gave in
|
|
lines = f.readlines() # read data
|
|
f.close()
|
|
#len of frequencies is 10 time shorter than before, so worked?
|
|
#put in frequencies instead of dataset?
|
|
#2nd loop cut frequencies by this function?
|
|
cutf = []
|
|
frequencies = []
|
|
for i in range(len(lines)):
|
|
l = lines[i].strip()
|
|
|
|
if len(l) > 0 and l[0] is not '#':
|
|
temporary = list(map(float, l.split()))
|
|
frequencies.append(temporary[1])
|
|
|
|
for k in np.arange(0, len(frequencies), n): # sollte nach k+n weitergehen?
|
|
f = frequencies[k:k+n]
|
|
mean = np.mean(f)
|
|
cutf.append(mean)
|
|
|
|
return cutf |