Yigit Demirag

I am a PhD candidate at the Institute of Neuroinformatics of ETH Zürich. I am interested in gradient-based learning on neural network accelerators. My PhD advisor is Giacomo Indiveri.

My PhD is focused on algorithm-hardware codesign for efficient training on neural network accelerators built with in-memory computing technologies. I had pleasure to work with Melika Payvand, Emre Neftci, Charlotte Frenkel, Angeliki Pantazi (IBM Research Zürich) and Elisa Vianello (CEA-LETI) on several projects. I designed the course material and teached Spiking Neural Network on Neuromorphic Processors for undergraduate students at ETH Zürich.

As of July 2023, I am a Student Researcher at Google. My previous internships include bio-inspired quantized training of neural networks with Blake Richards at MILA in 2023, modeling memory technologies at Advanced Technology Development Group of Samsung Electronics in 2018, working on backprop implementation on crossbar arrays at EPFL in 2017 and SIMD optimization of monte carlo simulations at CERN in 2015.

Email  /  CV  /  GitHub  /  Google Scholar  /  Twitter

profile photo

Publications

project image

Mosaic: in-memory computing and routing for small-world spike-based neuromorphic systems


Thomas Dalgaty*, Filippo Moro*, Yigit Demirag*, Alessio De Pra, Giacomo Indiveri, Elisa Vianello, Melika Payvand
Nature Communications, 2024
paper / code

project image

Network of biologically plasuible neuron models can solve motor tasks through heterogeneity


Yigit Demirag, Giacomo Indiveri
Computational and Systems Neuroscience (COSYNE), 2024

project image

DenRAM: Neuromorphic Dendritic Architecture with RRAM for Efficient Temporal Processing with Delays


Simone D'Agostino*, Filippo Moro*, Tristan Torchet*, Yigit Demirag, Laurent Grenouillet, Niccolo Castellani, Giacomo Indiveri, Elisa Vianello, Melika Payvand
Nature Communications, 2024
paper / code

project image

Overcoming phase-change material non-idealities by meta-learning for adaptation on the edge


Yigit Demirag, Regina Dittmann, Giacomo Indiveri, Emre Neftci
NeuMatDeCas (best run up oral contribution), 2023
code

project image

Reconfigurable halide perovskite nanocrystal memristors for neuromorphic computing


Rohit Abraham John*, Yigit Demirag*, Yevhen Shynkarenko, Yuliia Berezovska, Natacha Ohannessian, Melika Payvand, Peng Zeng, Maryna I Bodnarchuk, Frank Krumeich, Gökhan Kara, Ivan Shorubalko, Manu V Nair, Graham A Cooke, Thomas Lippert, Giacomo Indiveri, Maksym V Kovalenko
Nature Communications, 2022
paper

project image

Biologically-inspired training of spiking recurrent neural networks with neuromorphic hardware


Thomas Bohnstingl, Anja Surina, Maxime Fabre, Yigit Demirag, Charlotte Frenkel, Melika Payvand, Giacomo Indiveri, Angeliki Pantazi
AICAS, 2022
paper / code

project image

PCM-Trace: Scalable synaptic eligibility traces with resistivity drift of phase-change materials


Yigit Demirag, Filippo Moro, Thomas Dalgaty, Gabriele Navarro, Charlotte Frenkel, Giacomo Indiveri, Elisa Vianello, and Melika Payvand
ISCAS, 2021
paper

project image

Online training of spiking recurrent neural networks with phase-change memory synapses


Yigit Demirag, Charlotte Frenkel, Melika Payvand, Giacomo Indiveri
arXiv, 2021
paper / code

project image

Analog weight updates with compliance current modulation of binary ReRAMs for on-chip learning


Melika Payvand, Yigit Demirag, Thomas Dalgaty, Elisa Vianello, Giacomo Indiveri
ISCAS, 2020
paper

project image

Modeling electrical resistance drift with ultrafast saturation of OTS selectors


Yigit Demirag, Ekmel Ozbay, Yusuf Leblebici
arXiv, 2018
paper

project image

Multiphysics modeling of GST-based synaptic devices for brain inspired computing


Yigit Demirag
Bilkent University, 2018
paper




Open Source

project image

Gradients without Backpropagation

JAX implementation of the Gradients without Backpropagation paper, which gets rid of backpropagation of errors and estimate unbiased gradient of loss function during single inference pass via perturbed Jacobian-vector product.

code

project image

e-prop local learning rule for training recurrent networks

JAX implementation of e-prop algorithm. It is written to be simple, clean and fast. I replicated the pattern generation task as described in the paper.

code

project image

Dynap-SE1 Neuromorphic Processor Simulator

A minimal and interpretable Brian2 based DYNAP neuromorphic processor simulator for educational purposes.

code

project image

Spiking Neural Network Simulator for TPU

A simple Colab notebook for the training of recurrent spiking neural networks on TPU or multi-GPU settings. On a single A100 GPU, it can train a small size network of 512 LIF neurons via SHD dataset under <50 sec.

code

project image

Physical model of HfO2/TiOx Bilayer ReRAM memory cell

Compact model simulation of ReRAM cell. The model simulates IV characteristics of the cell, incorporating gradual SET and abrupt RESET processes. It considers filament geometry as a combination of plug and disc, and employs the Schottky barrier to represent the metal/oxide interface’s high work function, which depends on the oxide’s defect concentration.

code





2023 🪴