diff --git a/demos/3_behavioural/demo_QlearningAsymmetric.m b/demos/3_behavioural/demo_QlearningAsymmetric.m index ea923cc6..b6c37378 100644 --- a/demos/3_behavioural/demo_QlearningAsymmetric.m +++ b/demos/3_behavioural/demo_QlearningAsymmetric.m @@ -117,7 +117,7 @@ theta = [Inf; 0]; % for each trial, recompute the state evolution for t = 1 : n_t - PE(t) = sum (f_QlearningAsym (Qvalues(:,t), theta, u(:,t+1), struct) - Qvalues(:,t)); + posterior.PE(t) = sum (f_QlearningAsym (Qvalues(:,t), theta, u(:,t+1), struct) - Qvalues(:,t)); end end