Submitted by ch1214ch t3_121ozof in askscience
adventuringraw t1_jdnl4sd wrote
Hm... There's not a detailed answer here yet, so I guess I'll jump in.
First thing to realize: in reach retina, there's over 100 million photoreceptors, but only 1 million axons in the optic nerve. How's that work? The answer, is that it's completely accurate to say that the earliest part of visual processing is in the eye. There's 5 layers to the retina. The million ganglion cells sending the optic nerve to central command, a middle 'processing' layer with three main types of interneurons, and the actual photoreceptor cell layer (the other two layers are 'in between' layers made up of the cabling connecting these three cell body layers).
There's actually about 20 different kinds of ganglion cells, so you can look at it as there being 20 different image filters sent in to central. Some with blue/yellow contrast, others for pure luminance contrast, etc. Some of these 'views' have very tiny 'receptive fields' (the part of the retina their signal contains information for). So-called 'P' ganglions for example can have one single cone that causes them to depolarize and fire, though they all have surrounding parts/colors that inhibits firing. Each ganglion cell actually has a fairly complex 'optimal' pattern that makes them fire... Usually either 'light in the center, dark on the edges', the reverse of that, or a color version (blue light but no yellow light, for example). So even by the time the signal's leaving the eye, you're already getting various size views containing different kinds of contrast information. It should be said too, the closer to the fovea (center of the view) the smaller the receptive field gets, so the more detail you can perceive... So the 'resolution' of your view is actually not consistent even just across the eye.
Anyway. So: the optic nerve. These million axons aren't sending pixel information, they're sending 20 pictures, each with larger or smaller receptive fields depending on type and distance from the fovea, and different activation patterns they're 'looking for' to fire. This tract splits off on the way to the central switchboard (the LGN in the thalamus). For the left hemisphere of the LGN (say), you get the right eye's ear half of the view, and the left eye's nose half of the view. Opposite that for the right hemisphere.
So, the left hemisphere switchboard gets input from the right half of the visual field. But these two views don't perfectly line up... Maybe 80% of that view does, but the most peripheral part of the view only gets a signal from the outside edge, not the nose-side edge, so your peripheral vision doesn't have a binocular signal to put together in the first place.
Input from different eyes is totally separate still in this central switchboard. Each hemisphere's LGN has 6 main layers, 3 from each eye, with thin middle layers in between carrying the comparatively small amount of color information (koniocellular layers), also still separated by eye.
Among other places (pupil and eye muscle autonomic circuits in particular) this tract then gets sent to the primary visual cortex in V1, at the very back of your head. Here, it's still separated by eye. You've got these stripes running from the back to the front, with alternating left eye, right eye input. Perpendicular to that, you've got simple edge detectors going through different orientations, and mixed into all that, you've got these barrel shaped blobs of cells that respond to color information. You can see a picture of what I mean here. One 'cycle' of left eye/right eye, and 180 degrees of orientation preference for edges marks out a roughly 1mm x 1mm 'hypercolumn', that takes in input from a chunk of the retina. These hypercolumn are the basic processing unit here in V1, and they tile over the visual field. The receptive field here is larger than a single cone or rod certainly, but it's still fairly small. You can see that the same parts of the visual field are at least nearby now though, that's what those stripes are... Lined up regions from the same part of the visual field from different eyes.
Once it gets to V2, the next layer of processing, this is where you start to have 'binocular integration', neurons that selectively fire based on visual input from both eyes.
As you climb up in levels, you'll see larger and larger receptive fields. By the time you get deep into the ventral visual stream (very loosely speaking, ventral stream is identifying things you're seeing, dorsal stream is for helping to guide hands and stuff for grabbing things and so on) you're seeing cells that selectively fire given complex input from anywhere in the visual field. A 'Jennifer Aniston' neuron for example that might fire anytime you're seeing her face... Or her name written, or a drawing of her, or so on, from anywhere in the visual field.
But anyway. You get the gist. The full complex view of even the early visual system is hilariously intricate, and there's no really simple description that captures all the detail. But maybe a partway answer that's close enough... Peripheral vision has no binocular component, since only one eye captures that outside edge. For the bulk of your visual field though, yes... Things get tied together eventually, but it's not until V2... After many layers of processing in the retina, LGN, and V1. By then, you're talking about receptive fields with hundreds of input photoreceptors, and you're already talking about integrating signals for fairly high level information... Direction of movement, edge orientations, color information and so on, all in separate parallel feeds, to be integrated into even more high level, abstract tracts of information with even larger receptive fields as you continue climbing up towards the dorsal (hand eye coordination) and ventral (object recognition) tracts.
Note too: every step here is more tightly interconnected than I'm describing. In particular, there's 10x connections coming BACKWARDS from downstream than there are coming in from upstream towards the retina, so you probably do have some binocular signal affecting neuron firing even before you get to the binocular integration part in V2. Those incredibly numerous backpropagating connections aren't well understood, but it does definitely complicate the question of where you can start pointing to neurons influenced by input from both eyes in the same part of the visual field.
So anyway... There you go, haha. This is largely cobbled together from Kandel's 'principles of Neuroscience, 6th edition', there's a half dozen chapters in the low 20's going through how the brain processes visual information in pretty heavy detail.
_AlreadyTaken_ t1_jdnuw2l wrote
I've heard the retina described as an extension of the brain.
It is amazing to think the foundations of those processing pathways can develop from genes.
Blakut t1_jdo7dkp wrote
afaik the first convolutional neural networks in AI were modelled to mimic the retina (cows in particular? idk)
Fenrisvitnir t1_jdpqicw wrote
No. Convolutional networks are simply fully connected all-combinations of every pixel in the image (under a sliding window, usually). They are not modeled after any brain, they are modeled after signal processing convolution filters (pre-neural network) for 2D signals. The learning epochs of the convolution network teach the network which pixels to pay attention to at the meta level (features), and the further levels combine those features.
Blakut t1_jdpy8xc wrote
> Convolutional networks are simply fully connected a
uhm no.
https://en.wikipedia.org/wiki/Convolutional_neural_network
Convolutional networks were inspired by biological processes[10][11][12][13] in that the connectivity pattern between neurons resembles the organization of the animal visual cortex. Individual cortical neurons respond to stimuli only in a restricted region of the visual field known as the receptive field. The receptive fields of different neurons partially overlap such that they cover the entire visual field.
Fenrisvitnir t1_jdq1qvn wrote
Um no. References 10-13 don't establish the fact, if you look at them. Convolution kernels long predate their use in neural networks as a convolution layer.
The sliding NxM convolution window is the "receptive field" but it isn't analogous to the field in the eye. The kernel matrix existed long before it was used in NNs, and is the mapping mechanism to the fully connected convolution input layer.
https://en.wikipedia.org/wiki/Kernel_(image_processing)
Thanks for being interested, but there is a lot of fluffery in ML discussions. The neurons of a NN are not remotely the same as biological neurons - the only thing they share in common is the activation function, and even then they are only symbolically similar.
adventuringraw t1_jdrnbmp wrote
Um no (we have to keep the comment chain going).
You're actually being overly dismissive of what they're saying I think. The key word they used was 'inspired'. I tried to dig up the origin of convolutional image kernels, and while I couldn't find much in five minutes of digging, I'm sure you're right, that they predate deep learning certainly, and possibly even digital computing entirely given that their historical origin was probably in signal processing.
Their comment though wasn't whether or not CNNs directly imitate biology, or that the way they did it was entirely novel... They were just pointing out that biology was an inspiration for trying it this way, and that part's unambiguously true. To my knowledge, the first paper introducing the phrase 'convolutional neural network' was from Yann LeCun. This one I believe, from 1989. If you look at the references, you'll note Hubel and Wiesel's 1962 paper introducing a crude model of biological vision processing is in the references. More importantly, Fukushima, 1980 is referenced (and mentioned in the text as a direct inspiration). This 'Neocognitron' is generally accepted to be the first proto-CNN. The architecture is a bit different than we're used to, but it's where things started... And as the author puts it in the abstract:
> A neural network model for a mechanism of visual pattern recognition is proposed in this paper. The network is self-organized by "learning without a teacher", and acquires an ability to recognize stimulus patterns based on the geometrical similarity (Gestalt) of their shapes without affected by their positions. This network is given a nickname "neocognitron". After completion of self-organization, the network has a structure similar to the hierarchy model of the visual nervous system proposed by Hubel and Wiesel.
So... Yes. CNNs weren't inspired by cow vision or something... Hubel and Wiesel's most famous work involved experiments on kittens. but CNN origins are unambiguously tied into Hubel and Wiesel's work in biological visual processing, so the person you're responding to is actually the one that was right. I just noticed even, some of the papers referenced from Wikipedia that you said didn't show biological inspiration are the same ones I mentioned even, so they were the correct papers to cite.
If I may be a bit rude for my own Sunday morning amusement: 'Thanks for being interested, but there is a lot of fluffery in ML discussions.'
Seriously though, it's an interesting topic for sure, and historical image processing techniques are certainly equally important to the history of CNNs... They were the tool reached for given the biological inspiration, so in all seriousness you're not entirely wrong from another perspective, even if you're not justified in shooting down a biological inspiration.
[deleted] t1_jdtdnbo wrote
[removed]
Coomb t1_jdrsw29 wrote
Did anyone say that the matrix operation of convolution, or the idea of smearing it across an image, was invented via inspiration from experimental explication of image processing as performed by animals? I don't think they did, and I would be surprised if that were true. But those references do show that the "neocognitron" was explicitly inspired by actual physical neural networks used by animals for image processing, because among other things they include the original neocognitron paper, which is very clear about its inspiration. This is relevant because review papers of convolutional neural networks like this one from University College London almost universally identify the neocognitron as the direct precursor to modern convolutional neural networks.
Fenrisvitnir t1_jdsp8nn wrote
https://glassboxmedicine.com/2019/04/13/a-short-history-of-convolutional-neural-networks/
"The popular press often talks about how neural network models are “directly inspired by the human brain.” In some sense, this is true, as both CNNs and the human visual system follow a “simple-to-complex” hierarchical structure. However, the actual implementation is totally different; brains are built using cells, and neural networks are built using mathematical operations."
Coomb t1_jdsre81 wrote
Is this supposed to be responsive to my point?
[deleted] t1_jdqwcy1 wrote
[removed]
[deleted] t1_jdnxe4x wrote
[removed]
affordable_firepower t1_jdqdzoi wrote
Thank you for this explanation.
It's blown my mind a bit - I have a servered optic nerve on my right side and it's amazing to think that my brain processes the left and right side of what I see with my remaining eye and then stitches the images together seamlessly.
Obviously I have no binocular vision which causes issues with close up depth perception.
adventuringraw t1_jdxn4xh wrote
That reminds me of one of the diagrams from the book I got this from, I took a screenshot and posted it if you're curious.
And yeah, if the severed nerve is between your right eye and the optic chiasm, then it seems that is what happens. Half your left eye's view gets sent to one hemisphere, the other gets sent to the other, and then they get stitched together again upstream. Though I suppose if it happened when you were young enough, it could be a fair bit different... injuries when you're still in the 'critical period' can rewire in really unusual ways.
That diagram shows what's lost from severed optic tract at different points through the pipeline, thought you might think it was interesting. For every one of those, there's probably a bunch of people living that life. Sounds like you're number '1'. I think I'd actually prefer that to '2' or '3'.
Anyway, cheers... glad I could share something you found interesting. I've got ADHD, so I've got my own version of neuroscience adding to my understanding of myself, haha.
affordable_firepower t1_jdxodk0 wrote
Oh wow. Thanks for that.
The cut is definitely before the optic Chiasm. In fact, it's not far behind the eyeball so yeah, I'm a No.1
The accident happened when I was around six months old, so definitely still in the developmental stage. Now I'm wondering how my optics are wired 🤣
adventuringraw t1_jdxpfd3 wrote
Interesting, yeah. I bet that'd be an interesting thing to find out about even, maybe eventually brain scan technology will be cheap and powerful enough that you could look into it for a lark:).
That book mentioned that approximately 50% of the brain (or 50% of the cortex at least?) Is dedicated to vision, and there's evidence I guess for tissue that'd normally take on one function to end up doing something else if the normal input feed's down for some reason. With only half the visual input coming in when you were that young, seems like that's a lot of computational hardware that's freed up for something else. Maybe you've got some only vaguely noticed superpower you'd be surprised other people don't have, who knows?
Edit: one last thing you might find interesting. Elsewhere in this thread actually, there was a discussion about biological inspiration behind convolutional neural networks from the field of machine learning and artificial intelligence. The inspiration was from Hubel and Wiesel, two really foundational Neuroscience researchers in the late 1950's and 1960's. They won the Nobel prize for their work, one critical experiment of which involved keeping one eye of a kitten closed and seeing how it changed their development. I don't know the details of their findings, but given the historical significance of that research, I bet your case actually has a lot of understanding behind it. Just wondering out loud more than sharing anything specific, but interesting that Hubel and Wiesel more or less came up in two comment threads here.
[deleted] t1_jdobjcv wrote
[removed]
awhildsketchappeared t1_jdnt3rv wrote
Thank you for a much better answer than mine!!
[deleted] t1_jdo7b2x wrote
[removed]
afcagroo t1_jdteoe6 wrote
Wow! Thank you for that. I had no idea about almost any of this.
Viewing a single comment thread. View all comments