Updated projects

This commit is contained in:
2015-11-02 09:36:45 +01:00
parent 0bdc7eab98
commit ce2345acdb
10 changed files with 115 additions and 88 deletions

View File

@@ -6,10 +6,10 @@
\pagestyle{headandfoot}
\runningheadrule
\firstpageheadrule
\firstpageheader{Scientific Computing}{Project Assignment}{11/05/2014
-- 11/06/2014}
\firstpageheader{Scientific Computing}{Project Assignment}{11/02/2014
-- 11/05/2014}
%\runningheader{Homework 01}{Page \thepage\ of \numpages}{23. October 2014}
\firstpagefooter{}{}{{\bf Supervisor:} Fabian Sinz}
\firstpagefooter{}{}{{\bf Supervisor:} Jan Benda}
\runningfooter{}{}{}
\pointsinmargin
\bracketedpoints
@@ -33,7 +33,7 @@
\begin{questions}
\question A subject was presented two possible objects for a very
brief time ($50$ms). The task of the subject was to report which of
brief time ($50$\,ms). The task of the subject was to report which of
the two objects was shown. In {\tt decisions.mat} you find an array
that stores which object was presented in each trial and which
object was reported by the subject.
@@ -50,6 +50,10 @@
information $$I[x:y] = \sum_{x\in\{1,2\}}\sum_{y\in\{1,2\}} P(x,y)
\log_2\frac{P(x,y)}{P(x)P(y)}$$ that the answers provide about the
actually presented object.
The mutual information is a measure from information theory that is
used in neuroscience to quantify, for example, how much information
a spike train carries about a sensory stimulus.
\part What is the maximally achievable mutual information (try to
find out by generating your own dataset which naturally should
yield maximal information)?