You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Are you missing a factor 1/(T_trial*n_trial) or in other words 1/t_total, where t_total is the total time since last weight update? (see eqn (7) in supplement)?
The other thing I don't understand when basing my knowledge on the Nature Comms paper and supplementary is the additional decay of e_ij through efiltered here:
... why the (1-alpha) rather than 1? (probably makes no difference though)
More generally, it would probably be safer to just do the full average over the previous learning period and keep constant during each period as suggested originally in the supplementary note 2
I have the same questions for the ALIF eprop implementation regarding regularisation and additional filtering of e_ij here:
500.0); // Firing rate averaging time constant [ms]
is much too short as it might resolve the presumably strong differences in the different phases of the experiment. It should probably be a scale that encompasses at least one, if not several trials (of some 2000 to 3000 ms each) - so probably more like >> 10000?
I was unable to find information about initial weight matrices. You seem to use G(0, weight0/sqrt(n)) for the input-RSNN connections, G(0,weight0/sqrt(2n_RSNN)) and G(0,weight0/sqrt(2n_RSNN)) ... where are those from (or guessed?)
I think the maxdelay here should be 1500 according to the paper?
I think in the paper there is an initial 50ms settling time before the first cue comes in, while I think you start on timestep 0 with the first cue according to this:
How confident are you that the Adam optimizer works exactly as it should. I was struggling a little to decipher & while it now seems structurally understandable to me I don't know enough about how it's supposed to work including "standard parameters". If that wasn't working just right, it could (to state the obvious) have fatal consequences.
The text was updated successfully, but these errors were encountered:
In this recording of delta g updates:
genn_examples/eprop/common/eprop_models.h
Line 143 in 6b24c2a
Are you missing a factor 1/(T_trial*n_trial) or in other words 1/t_total, where t_total is the total time since last weight update? (see eqn (7) in supplement)?
The other thing I don't understand when basing my knowledge on the Nature Comms paper and supplementary is the additional decay of e_ij through
efiltered
here:genn_examples/eprop/common/eprop_models.h
Line 142 in 6b24c2a
Equation (23) in the main paper seems to suggest the only decays happens through
ZFilter
I am not sure I understand your moving average of the firing rate here:
genn_examples/eprop/common/eprop_models.h
Line 149 in 6b24c2a
... why the (1-alpha) rather than 1? (probably makes no difference though)
More generally, it would probably be safer to just do the full average over the previous learning period and keep constant during each period as suggested originally in the supplementary note 2
I have the same questions for the ALIF eprop implementation regarding regularisation and additional filtering of e_ij here:
genn_examples/eprop/common/eprop_models.h
Lines 201 to 203 in 6b24c2a
If you do a moving average of spike rate for the regulariser, I wonder whether 500 ms here:
genn_examples/eprop/evidence_accumulation/model.cc
Line 122 in 6b24c2a
is much too short as it might resolve the presumably strong differences in the different phases of the experiment. It should probably be a scale that encompasses at least one, if not several trials (of some 2000 to 3000 ms each) - so probably more like >> 10000?
I was unable to find information about initial weight matrices. You seem to use G(0, weight0/sqrt(n)) for the input-RSNN connections, G(0,weight0/sqrt(2n_RSNN)) and G(0,weight0/sqrt(2n_RSNN)) ... where are those from (or guessed?)
I think the maxdelay here should be 1500 according to the paper?
genn_examples/eprop/evidence_accumulation/parameters.h
Line 24 in 6b24c2a
I think in the paper there is an initial 50ms settling time before the first cue comes in, while I think you start on timestep 0 with the first cue according to this:
genn_examples/eprop/evidence_accumulation/simulator.cc
Lines 80 to 84 in 6b24c2a
(not sure whether it matters)
How confident are you that the Adam optimizer works exactly as it should. I was struggling a little to decipher & while it now seems structurally understandable to me I don't know enough about how it's supposed to work including "standard parameters". If that wasn't working just right, it could (to state the obvious) have fatal consequences.
The text was updated successfully, but these errors were encountered: