Dear NEST users,
I am trying to debug my network and I found that my spike_recorder with *ascii* option isn't working as expected. I am simulating a network for a few milliseconds (say 1000. ms) and then do some structure changes and then re-run for another phase of a few ms (say 4000. ms) In the compact version, I am trying to see how this sample code would work. import nest nest.ResetKernel() nest.SetKernelStatus({ "overwrite_files": True, "data_path": './Debug-Log', "data_prefix": 'DEBUG-SPIKE-REC', }) noise_device = nest.Create("poisson_generator",params={'start': 0.,'rate' : 100.}) sd_params = { "record_to": 'ascii', "label": "Input_spike_recorder" } Input_SD = nest.Create("spike_recorder",1, params=sd_params) nest.Connect(noise_device,Input_SD) nest.Simulate(100.) # do something here nest.Simulate(100.)
*Issue:* I only get the data recorded from 100th millisecond but for memory, I get all the data from 0-200 ms. This behavior is consistent for both overwrite_files as True and False. In the previous version (v2.20) I also used the same approach and had no issues. Is there a newer way to do such a simulation paradigm? I didn't find anything on this on the comparison page of NESTv2.x and NESTv3.x
Am I missing something? It's really important for my work, I hope I could find a solution.