text
stringlengths 0
1.96k
|
---|
"It is probable that revolutionary computational systems can be created in this way with only moderate expenditure of resources and effort"" Of course whole fields are working on this problem." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"Hardly what I'd call moderate effort" "['con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"Claims of efficiency of more brain-like approaches compared to AI are disingenuous" "['con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"A major draw-back of spiking models is that they are much more costly than ANNs, because of the small time-steps required." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"Its more a series of statements than a cleverly woven argument" "['con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"But the individual statements are sometimes seductive" "['non', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro']" "paper quality"
|
"When it hears an incoming pattern of spikes that matches a pattern it knows, it responds with a spike of its own." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"Repeat this process recursively tens to trillions of times, and suddenly you have a brain controlling a body in the world or doing something else equally clever." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"Our challenge is to understand how this occurs." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"We require a new class of theories that dispose of the simplistic stimulus-driven encode/ transmit/decode doctrine. """ "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"I feel this statement: ""Our challenge is to understand how this occurs." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"We require a new class of theories that dispose of the simplistic stimulus-driven encode/ transmit/decode doctrine. """ "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"There was an absence of nuance" "['con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"While it covers important ground , I think the arguments need more refinement and focus before they can inspire productive discussion" "['non', 'pro', 'pro', 'pro', 'pro', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"The authors consider how biologically motivated synaptic eligibility traces can be used for backpropagation-like learning, in particular by approximating local gradient computations in recurrent neural networks." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"This sheds new light on how artificial network algorithms might be implementable by the brain" "['pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro']" "paper quality"
|
"Space is of course limited, but the mathematics presented seem to pass all sanity checks and gives sufficiently rigor to the authors' approach" "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro']" "paper quality"
|
"It would have been nice to present a figure showing how e-prop yields eligibility traces resembling STDP, as this is one of the key connections of this work to biology" "['con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"Gives important new results about how eligibility traces can be used to approximate gradients when adequately combined with a learning signal" "['pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro']" "paper quality"
|
"While eligibility traces have received some attention in neuroscience their relevance to learning has not been thoroughly explored, so this paper makes a welcome contribution that fits well within the workshop goals" "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro']" "paper quality"
|
"One part that would have been nice to clarify is the relative role of random feedback vs eligibility traces in successful network performance" "['con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"The authors state three high-level improvements they want to make to CNN-based models of neural systems: 1 & 2) Capturing computational mechanisms and extracting conceptual insights." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"Operationally, I'm not quite sure how these are different, so, to me this goal is roughly ""be explainable"", and progress towards it could be measured e.g. in MDLs." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"3) Suggest testable hypotheses." "['non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"The technical aspects of the paper seem correct , though I have some higher-level conceptual concerns" "['pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"Is the attribution analysis stable for different stimulus frequencies?" "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"If not, is it really an explanation of the OSR?" "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"2) I agree with a concern raised by reviewer 3: It's difficult to see a 1-layer network as a ""mechanistic explanation"" of a 3-layer network" "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"Explanations are mostly complete , though some details are missing" "['pro', 'pro', 'pro', 'pro', 'non', 'non', 'con', 'con', 'con', 'con']" "paper quality"
|
"e.g. what was the nonlinearity used in the model CNN" "['non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"Also, do the CNN layers correspond to cell populations , and if so, why is it reasonable to collapse the time dimension after the first layer" "['non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'non', 'non', 'non', 'non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"I believe this paper is addressing questions that many of the workshop attendees will find interesting" "['non', 'non', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro']" "paper quality"
|
"The question of how networks maintain memory over long timescales is a longstanding and important one, and to my knowledge this question hasn't been thoroughly explored in spiking, trained recurrent neural networks (RNN)." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"The importance is tempered by the findings only covering what is to be expected, and not pushing beyond this or describing a path to push beyond this" "['con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"The work would benefit from more detailed discussion of the training algorithm that provides some indication that the results aren't unduly sensitive to these details" "['con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"The statistical tools are fairly well described and appear to be well-suited for illustrating the phenomena of interest" "['pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro']" "paper quality"
|
"I feel that more tools should have been used to further support or push the results" "['non', 'non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"For instance, while the heatmaps in Figure 3 provide visual evidence for their claims (except see my comments below), the work could have benefitted from a quantification of this evidence" "['non', 'non', 'non', 'non', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"For instance, the claim of ""stronger cue-specific differences across the cue stimulus window"" between fast and slow intrinsic timescale neurons in the RNN model isn't clearly supported by the heatmap in Figure 3 -- the cue-specific differences for the short instrinsic timescale group to me appears to be at least as great as that of the long intrinsic timescale group within the cue stimulus window" "['non', 'non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"It seems that one of the main points of the work is that ""longer intrinsic timescales correspond to more stable coding"", but I didn't find that this point was made very convincingly" "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"The work would have benefited from a discussion of the implications of longer intrinsic timescale neurons retaining task-relevant information for longer -- in particular, this finding feels a bit ""trivial"" without the case being made for why this should push understanding in the field" "['con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'non', 'non', 'non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"I think the interesting part may be in quantifying just how much of a difference there is between short and long timescale neurons -- for instance, does task-relevant information in both neuron groups fall off in a way that can be well predicted by their intrinsic time constants" "['non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'non', 'non', 'non', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"How does this relate to their synaptic time constants" "['con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"Does limiting the synaptic time constants limit the intrinsic time constants, and if so by how much" "['con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con']" "paper quality"
|
"The same type of comments apply to the second part of the results, which demonstrates that a task that doesn't require working memory results in neurons with shorter intrinsic timescales compared to the working memory task." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"The authors use an artificial network model to shed light on the biological mechanisms enabling and shaping working memory in the brain." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"The paper in the process reveals some (expected) results about how spiking RNNs behave on a working memory task" "['pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro']" "paper quality"
|
"The proof-of-concept work (among others) that this can be done with spiking RNN may inspire more work in this area" "['pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro', 'pro']" "paper quality"
|
"The work is a basic proof-of-concept of results that may not do much to advance understanding since they are what one would expect to see (i.e. the antithesis of their thesis seems very unlikely)." "['con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'non', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'con', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
"Looking into the nuances of the explored phenomena may provide new information for the field." "['non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non', 'non']" "paper quality"
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.