Benjamin Scellier
Independent
Core Metrics
Total Citations
1,642
H-Index
9
Publications
28
i10-Index
9
2-Year Citedness
2.9
avg citations per work
Ability Dimensions
1,642 citations, h=9
2yr mean: 2.9
28 papers (3.1/year)
59 cites/paper
2 unique research topics
5 topic areas
Top 3 papers: 90% of citations
* Percentile scores are calculated relative to all scholars in the computational neuroscience dataset. Tags are assigned based on dimension combinations. Hover over the radar chart for details.
Scholar Profile Analysis
Benjamin Scellier is a emerging scholar in computational neuroscience, currently affiliated with Independent Research.
Over a 9-year academic career, published 28 papers (averaging 3.1 per year), with 1,642 citations.
Outstanding early career performance: first 5 years account for 96.9% of total citations (1,591), showing strong early impact.
Primary research areas include Statistics, Statistics, Statistics.
Key Findings
Signature Work
"A deep learning framework for neuroscience" is the most influential work, with 1,018 citations, published in 2019.
Early Breakthrough
Over half of citations come from the first 5 years, possibly from pioneering work or classic textbooks.
Early Career Analysis (First 5 Years)
Career Start
2016 - 2020
Early Citations
1,591
Early Works
16
Early Impact %
96.9%
Top Early Career Paper
A deep learning framework for neuroscience
Publication Timeline
Research Topics
Top Publications
A deep learning framework for neuroscience
1,018
Citations
Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation
414
Citations
Training End-to-End Analog Neural Networks with Equilibrium Propagation
38
Citations
Towards a Biologically Plausible Backprop
31
Citations
Generalization of Equilibrium Propagation to Vector Field Dynamics
19
Citations
Feedforward Initialization for Fast Inference of Deep Generative Networks is biologically plausible
18
Citations
Equilibrium Propagation with Continual Weight Updates
17
Citations
Equilibrium Propagation: Bridging the Gap Between Energy-Based Models and Backpropagation
15
Citations
Scaling Equilibrium Propagation to Deep ConvNets by Drastically Reducing Its Gradient Estimator Bias
10
Citations
Updates of Equilibrium Prop Match Gradients of Backprop Through Time in\n an RNN with Static Input
8
Citations