Skip to content

weird idea for more general purpose model of computation and how to train it

Notifications You must be signed in to change notification settings

shariq/local_loopy_networks

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

58 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Locally Learned Loopy Networks

goal: beat some significant state of the art metric in 4 months using a local "loopy" network

Overview

modern recurrent models tend to have a single "loop" of learning/recursion hypothesis: this is not "good" for representing certain kinds of functions e.g, long/medium term memory and empirically not "good" at learning certain kinds of functions (but we don't have a good theory of learnability)

instead if we have a model with many "loops", which all touch each other... so kind of like a big randomly connected directed graph, with a lot of cycles let's call this a loopy network then maybe this is "good" at representing long/medium term memory and the crazy idea: maybe there exists a local learning rule to learn these kinds of functions the rule being local is biologically and physically inspired

with a computational prior this would be formalized approximately as a random LSTM is further away from random code than a random loopy network

with a physical prior this would be formalized approximately as a random LSTM is further away from random physical function than loopy network

anyways this formalization has a lot of holes, so let's ignore it for now; more importantly is to get this working, then go back and try to find out why

Author

Shariq Hashme

About

weird idea for more general purpose model of computation and how to train it

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages