Skip to content

Commit b076726

Browse files
authored
Merge pull request #284 from neural-loop/student-talk
Student talk
2 parents 0a3d309 + e501dfc commit b076726

File tree

7 files changed

+50
-2
lines changed

7 files changed

+50
-2
lines changed

.github/workflows/main.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@ name: Deploy Hugo site to Pages
22

33
on:
44
push:
5-
branches: ["main", "cache-images"]
5+
branches: ["main", "student-talk"]
66

77
permissions:
88
contents: read
@@ -63,7 +63,7 @@ jobs:
6363
6464
- name: Generate OG Images
6565
run: npm run og-images
66-
66+
6767
- name: Fetch GitHub Stars
6868
env:
6969
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}

.gitignore

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -26,3 +26,4 @@ yarn.lock
2626
content/**/*-og-*.jpg
2727
static/images/og-image-*.jpg
2828
tmp/og-cache-manifest.json
29+
/data/github_stars.json
Lines changed: 16 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,16 @@
1+
---
2+
title: "Matei Stan"
3+
description: "Matei Stan is a PhD student at the University of Manchester, focusing on State Space Models (SSMs) and their application in Spiking Neural Networks for long-range sequential tasks."
4+
image: "matei-stan.jpg"
5+
social:
6+
- icon: "fa-brands fa-linkedin"
7+
link: "https://www.linkedin.com/in/matei-stan"
8+
title: "linkedin"
9+
draft: false
10+
---
11+
12+
Matei Stan is a third-year PhD student in the Department of Computer Science at the University of
13+
Manchester, UK. He is supervised by Dr Oliver Rhodes in the Advanced Processor Technologies
14+
(APT) group. In his PhD work, Matei has primarily focused on the applications of deep State Space
15+
Models (SSMs), such as S4, in neuromorphic computing, and their potential in scaling
16+
energy-efficient algorithms for long-range sequential tasks.
539 KB
Loading
Loading
Lines changed: 31 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,31 @@
1+
---
2+
title: "Student Talk: Learning Long Sequences in Spiking Neural Networks with Matei Stan"
3+
author:
4+
- "Matei Stan"
5+
date: 2025-07-27
6+
start_time: "08:30"
7+
end_time: "09:45"
8+
time_zone: "EST"
9+
description: "Explore how State Space Models (SSMs) combined with Spiking Neural Networks (SNNs) can outperform Transformers on long-sequence tasks, and learn about a novel feature mixing layer that challenges assumptions about binary activations."
10+
upcoming: true
11+
upcoming_url: "https://teams.microsoft.com/l/meetup-join/19%3Ameeting_OTBkNTY5MjgtMjE3Ni00OTFmLWEwNzktN2QwZTU1NWIxNDc2%40thread.v2/0?context=%7B%22Tid%22%3A%22c152cb07-614e-4abb-818a-f035cfa91a77%22%2C%22Oid%22%3A%223f444780-d657-4917-993e-0f42adeff90e%22%7D"
12+
video: ""
13+
image: "banner.png"
14+
speaker_photo: "matei-stan.jpg"
15+
type: "student-talks"
16+
speaker_bio: "Matei Stan is a third-year PhD student in the Department of Computer Science at the University of Manchester, UK. He is supervised by Dr Oliver Rhodes in the Advanced Processor Technologies (APT) group. In his PhD work, Matei has primarily focused on the applications of deep State Space Models (SSMs), such as S4, in neuromorphic computing, and their potential in scaling energy-efficient algorithms for long-range sequential tasks."
17+
---
18+
19+
Matei’s published work, “Learning Long Sequences in Spiking Neural Networks” [1], systematically investigates, for the first time, the intersection of the State‑of‑The‑Art State Space Models (SSMs) with Spiking Neural Networks (SNNs) for long‑range sequence modelling. Results suggest that SSM‑based SNNs can outperform the Transformer on all tasks of a well‑established long‑range sequence modelling benchmark - the “Long-Range Arena” [2]. It is also shown that the SSM‑based SNNs can outperform current State‑of‑The‑Art SNNs with fewer parameters on sequential image classification. Finally, a novel feature mixing layer is introduced, improving SNN accuracy while challenging assumptions about the role of binary activations in SNNs. This work paves the way for deploying powerful SSM-based architectures, such as Large Language Models, on neuromorphic hardware for energy-efficient long-range sequence modelling.
20+
21+
This talk will highlight, at a high level, the similarities in computational primitives between SSMs and the existing neuromorphic standards such as Leaky Integrate-and-Fire (LIF) neurons. It will also focus on the specific drawbacks brought about by the introduction of binary activations in SSMs, as well as the extent to which these can be mitigated by the development of more accurate surrogate gradient methods that account for non-differentiability. Finally, arguments will be presented in favour of separating biological plausibility from energy efficiency in attempting to create scalable neuromorphic solutions.
22+
23+
**Contents of the talk:**
24+
- Introduction to long-range sequence modelling
25+
- Link between SSMs and SNNs
26+
- Binary S4D and the Gated Spiking Unit
27+
- Q&A
28+
29+
**References**
30+
[1]: Stan, M.I. and Rhodes, O., 2024. Learning long sequences in spiking neural networks. Scientific Reports, 14(1), p.21957.
31+
[2]: Tay, Y., Dehghani, M., Abnar, S., Shen, Y., Bahri, D., Pham, P., Rao, J., Yang, L., Ruder, S. and Metzler, D., 2020. Long range arena: A benchmark for efficient transformers. arXiv preprint arXiv:2011.04006.
Loading

0 commit comments

Comments
 (0)