Abstract
Artificial intelligence (AI) researchers currently believe that the main approach to building more general model problems is the big AI model, where existing neural networks are becoming deeper, larger and wider. We term this the big model with external complexity approach. In this work we argue that there is another approach called small model with internal complexity, which can be used to find a suitable path of incorporating rich properties into neurons to construct larger and more efficient AI models. We uncover that one has to increase the scale of the network externally to stimulate the same dynamical properties. To illustrate this, we build a Hodgkin–Huxley (HH) network with rich internal complexity, where each neuron is an HH model, and prove that the dynamical properties and performance of the HH network can be equivalent to a bigger leaky integrate-and-fire (LIF) network, where each neuron is a LIF neuron with simple internal complexity.
| Original language | English |
|---|---|
| Pages (from-to) | 584-599 |
| Number of pages | 16 |
| Journal | Nature Computational Science |
| Volume | 4 |
| Issue number | 8 |
| DOIs | |
| Publication status | Published - Aug 2024 |
| Externally published | Yes |
ASJC Scopus subject areas
- Computer Science (miscellaneous)
- Computer Science Applications
- Computer Networks and Communications
Fingerprint
Dive into the research topics of 'Network model with internal complexity bridges artificial intelligence and neuroscience'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver