Connected to dev server

Vision

Article curated by

Vision is arguably a human's primary sense - we use it to spot danger and navigate our environment. Despite this, there is still plenty we don't know about how this sense works - both in us and in other animals. In the future, this understanding will be vital to create robots that can see and analyse visual information as well as we can.

Mysteries remain about the role of the eye and the brain in vision Image credit: Public domain

There is a huge amount we don't know about the human sense of vision. As we seek to develop better and better computational vision we become more aware of some of these unknowns. For example, the ability to recognise a blurry face despite the loss of detailed information suggests we must commonly rely on low resolution data, using our memories to fill in the blanks. We also don't fully understand the way in which our brain 'cancels out' the effects of blinking, eye movements and the fact that only a tiny region of the retina can see in high resolution to provide a seemingly static and clear image of the world.
Learn more about Vision.

 Lightbulb Institutes Delving Deeper
List of Institutes Researching Vision
Things We Don’t Know (1 researchers)
2

We know that cells in the eye of many animals, from mice to humans, can detect motion - this is essential to avoid predators or track prey. But even this seemingly simple ability isn't fully understood. The retina of the eye is packed with photoreceptor cells - the rods and cones that allow us to detect light. Behind these, there are layers of cells that convert this information into a form that the brain can understand - these include amacrine cells and ganglion cells. Using mice, researchers from Washington University discovered that a subtype of amacrine cells excite the ganglion cells when motion is detected. This information can then be relayed to the brain.

Despite this step forward, Daniel Kerschensteiner - leader of the research team - highlighted just how little we still know about the retina; There are many elements in the retinal circuitry that we haven’t figured out yet. We know the signals from the rods and cones are transmitted to the retina — where the amacrine and ganglion cells are located — and that’s really where the ‘magic’ happens that allows us to see what we see. Unfortunately, we still have a very limited understanding of what most of the cells in the inner retina actually do.

 Lightbulb
Public Domain via jonbonsilver (Pixabay)
How our retina detects movement isn't fully understood Image credit: Public domain

Once the signals from the eye begin their journey to the brain, things become even more complicated. The classical view says there are two streams of visual processing known as the dorsal and ventral or ‘what & where’ streams. The ventral or ‘what’ stream provides us with conscious recognition of the object and its details, but is slow and prone to illusions. The dorsal ‘where’ stream, meanwhile, drives our actions towards an object, often quickly and subconsciously. This idea is supported by studies on patients with damage to areas involved in one stream but not the other, who may, for example, be able to see an object but not use it to guide motion. Some studies have also shown that while we fall for illusions such as the Titchener illusion (see below) when asked about them consciously, we use the actual (rather than illusory) size to drive movements.

However, more recent studies have raised questions about this distinction - showing that if visual tasks and those requiring actions are appropriately matched, there is no different in the size of the illusion seen. The dissociations seen in patients also seem to be less clear cut than originally thought. Other research has supported the idea that these processing pathways are less distinct than first thought, working together as well as separately. Studies in primates suggest there might even be three processing systems - one for shape, one for colour and one for movement, location, and spatial organisation.

 Lightbulb Institutes Delving Deeper
List of Institutes Researching dorsal & ventral stream in vision
Things We Don’t Know (1 researchers)

New research has also indicated that visual processing might be intimately linked with audio processing. Observations of changes in ear canal pressure, which scientists think is the middle-ear muscles moving the eardrum, have been matched up with eye movements: when our eyes move to an object, so do our ears. Scientists have suggested that both motions are brain-directed, but it’s also possible that the eyes send a signal to the brain that then sends a message to the ears to follow, the opposite, or both in a big feedback loop. It’s thought that syncing the ears and eyes facilitates visual processing, because it helps us identify what is making what sound. However, the discovery is still new and there remains lots to find out about it.
Learn more about Synchronised ears and eyes.

 Lightbulb Institutes Delving Deeper
List of Institutes Researching Synchronised ears and eyes
Things We Don’t Know (1 researchers)
2
"BodyParts3D, © The Database Center for Life Science CC-BY-SA-2.1-jp via Wikimedia Commons
Visual areas of the brain Image credit: "BodyParts3D, © The Database Center for Life Science CC-BY-SA-2.1-jp via Wikimedia Commons 

Whatever route it takes, information from the eyes is processed in the visual cortex, located in the occipital lobe at the back of the brain. The primary visual area (V1) is the first part of the cortex to receive information from the eye, which travels via the lateral geniculate nucleus. From here, the areas involved become more complicated and less clear cut. It used to be thought that the information then travelled through various other areas, which processed increasingly more complicated elements of the scene - starting with edges and shapes, then moving on to colour, movement, and areas that respond to attention. However, we now know that rather than a simple hierarchy, these areas connect in complex patterns - feeding back to lower areas as well as forward to higher ones. V1, for example, can be modulated by attention, something previously thought to only be taken into account by V4. There are also disagreements about the exact location and extent of each area, which is bringing into debate whether it is useful to divide the visual cortex up or whether a network approach would be more useful in helping us understand how this complex system works.

 Lightbulb Institutes Delving Deeper
List of Institutes Researching Visual hierarchy in the brain
Things We Don’t Know (1 researchers)

Vision Problems

Hans (CC0) via Pixabay
Is too much reading responsible for the increase in short-sightedness? Image credit: Public domain
Over the last 60 years or so, the prevalence of short-sightedness, or myopia, have increased dramatically. In China, for example, rates have gone from 10-20% of the population to around 90% in just a couple of generations. So what is going on? This change is far to fast to be the result of evolution, so there must be an environmental reason.

It has long been thought that there is a link between reading or studying and vision problems, but recent research suggests the important factor is not how long children spend staring at something close, but how long they spend outside. The mechanism behind this is as yet unknown, but one idea is that plenty of bright, natural light is protective for the developing eye. Others argue that it is the viewing of landscapes and objects a long way away that has the effect. The mechanism behind either of these theories is yet to be confirmed.

Research is ongoing in the area, to determine whether outdoor time really can prevent myopia developing. Scientists are also looking into alternatives, such as bright, daylight-mimicing lamps, for children in places where natural light is limited or where outdoor play is impossible.
Learn more about Myopia Boom.

 Lightbulb Institutes Delving Deeper
List of Institutes Researching Myopia Boom
Things We Don’t Know (1 researchers)

Visual Illusions

In this example of the Titchener/Ebbinghaus illusion, the orange circles are the same size but appear to be different. Image credit: Public domain
Illusions can tell us a lot about how the brain processes visual information, but in many cases we don't know exactly why we fall for them. The Titchener or Ebbinghaus illusion, for example, is widely known but poorly understood. In this illusion, two circles of equal size appear to differ greatly when one is surrounded by smaller circles, and the other by larger ones. It is commonly explained as being due to a simple comparison - you process the size of the middle circle in relation to those around it. But the distance of the surrounding circles from the middle circle also has an effect - the closer they are, the more different the middle ones seem. Changing the colour contrast of the circles can also change the size of the illusion, and neither of these findings can be explained by the simple comparison theory.

Instead, some people think it is a version of a perspective illusion - we think the small circles are far away, so the one in the middle must be huge, as it looks so big from such a distance. But more research is needed to figure out exactly what is going on in our brains when we see this illusion.

 Lightbulb

Public Domain via Dunc (Pixabay)
The moon often looks larger when it is on the horizon- but this is an illusion Image credit: Public domain
A real-life mystery that may be linked to the Titchener illusion is the moon illusion. You may have noticed that the moon appears larger when it is near the horizon than when it is directly overhead, but no-one knows why this is. As far back as Aristotle, people beleived the earth's atmosphere magnified the moon when it was on the horizon, but we now know that isn't the case. In fact, the size of the moon on a viewer's retina doesn't change as its position in the sky changes, so the illusion must occur in our brain.

One theory is that when we view objects like birds or clouds in the sky, those on the horizon are further away than those above us, so we make the same assumption about the moon. Our brain then 'corrects' for the assumed distance of the moon on the horizon, making it seem bigger. This is an example of size constancy. However when asked, few people consciously perceive the horizon moon as being further away, casting doubt on this interpretation.

Another idea is that it is similar to the Titchener illusion - when on the horizon, the moon is surrounded by small objects like trees and houses, making it seem bigger. The large expanses of sky around the moon when it is overhead, however, have the opposite effect. Interestingly, looking at the moon with your head upside-down (e.g. by bending over and looking at it through your legs) reduces the size of the illusion, though no-one knows why!

 Lightbulb

© Cmglee (CC BY-SA 3.0) via wikimedia commons
Motion is seen in this static image Image credit: © Cmglee (CC BY-SA 3.0) via wikimedia commons 
Another illusion that scientists cant explain is the 'Rotating Snakes' illusion. In this image, a number of striped spirals (or ‘snakes’) appear to move as the viewer’s eyes move over them, despite there being no actual motion. The illusion is fairly robust, working on the majority of people, and has even been shown to work on fish and (anecdotally) on some cats. But why this particular pattern of colours and shapes should induce illusory motion isn’t understood. Interestingly, while it is normally shown in colour (blue, black, yellow and white stripes) greyscale versions still invoke the illusion, and studies have found no difference in strength. Motion is almost universally perceived in the direction of Black→Dark Grey→White→Light Grey ( or Black→Blue→White→Yellow for the colour version), suggesting the brightness of the stripes is important. However there are some suggestions that an isolated block or even a line of repeated blocks do not evoke a strong motion illusion - the concentric circles are vital for the illusion to work. The illusion also doesn’t work on everyone - some people don’t perceive motion at all, or perceive it in the opposite direction to the majority.

One suggestion for the mechanism is that areas with a greater difference in luminance (black & yellow or blue & white) are processed more quickly by the brain than those with more similar luminance (white & yellow or black & blue). This difference in processing speed could create the illusion of motion. However this mechanism can’t explain other illusions that also produce illusory motion, so for now their mechanism remains a mystery.

 Lightbulb Institutes Delving Deeper
List of Institutes Researching rotating snakes motion illusion
Things We Don’t Know (1 researchers)

In February 2015, the internet went crazy over a picture that wasn’t of a cat or Kim Kardashian’s bottom. The offending picture was of a bodycon dress with lace trim. But the question that was dividing opinion wasn’t whether it was flattering or suitable for a wedding but what colour it was. Some people looked at the image and saw a blue dress with black detail while others swore blind it was white with gold trim. So what was going on? How can an innocent picture cause such a divide? Scientists were puzzled, and began to investigate.

It was soon confirmed that the dress itself was blue & black, but in the image, the pixels were actually light blue and brown. A study of 1,400 people discovered that 57% saw it as blue and black, 30% saw it as white and gold, 11% saw it as light blue and brown. However Buzzfeed’s poll found that 67% of respondents see it as white & gold, although this difference could be down to that fact that only 2 options were given in the buzzfeed poll. While the exact proportions are unclear, what we do know is that people see things very differently when looking at the image.

Scientists think that the difference is down to what assumption your brain makes about lighting. The picture itself doesn’t make it clear what kind of lighting the dress is under, so you effectively have to guess. If you assume you are viewing the dress in cool daylight, it is likely to look white/gold as you ‘factor out’ the blue tinge as down to lighting; If, however, you assume a warm incandescent light, you will see blue/black. These guesses aren’t conscious, and most people can’t will themselves to see the opposite colour-way, but scientists discovered that they could change what colour people saw by giving them cues as to the kind of lighting involved.

Which of these camps you fall into could be down to a number of things. One suggestion is that ‘Larks’ who wake early and spend more waking hours in daylight, are more likely to assume that is the kind of lighting they are seeing the dress under. This fits with the study’s finding that women & older people (who are more likely to be larks) saw the dress as white & gold more often. However others think it might be down to the yellowing of the corneas that happens with age, and some experts argue that neither of these suggestions fits, and there must be some other, as yet unknown factor involved.

So what made this particular image of the dress so confusing for our poor brains? It seems it was photographed bathed in both natural light from a nearby window and artificial light in the shop. The best explanation is that this means your brain received ambiguous cues, so had to rely on past experience. Which experience your brain picks affects which colour combination you see. However this is far from being a definitive explanation - more research is needed to discover if this is actually what is going on.

 Lightbulb Institutes Delving Deeper
List of Institutes Researching The dress
Things We Don’t Know (1 researchers)
© xkcd.com (CC BY-NC 2.5)
If you assume it is in cool, blueish light, the dress will appear white & gold. An assumption of warm, yellowish light leads to a blue & black appearance. Image credit: © xkcd.com (CC BY-NC 2.5) 

Animal Vision

There are many aspects of animal vision we don't understand, and scientists hope that improving our knowledge might be useful in creating artificial vision. One mystery is the reason some birds, like chickens, bob their heads as they walk. To find out more, researchers put pigeons on a treadmill, and found that the bobbing stops if their surroundings are stable. This suggests the birds are using the head movements to keep the appearance of a still world as they move - something we do via tiny, unconscious eye movements. The bobbing means that the bird's head only moves in short bursts (which are easy for the brain to cancel out), and is then still as the body moves to catch up. This is similar to the technique of 'spotting' used by dancers - the head is left behind as they start a pirouette, and then snapped round at the last minute, so it is still for as long as possible.

This much is well supported by evidence, but the question remains - why do birds use this method of keeping the whole head still, when we, and other animals, instead cancel out movement with just our eyes?
Learn more about Bird Head Bobbing.

 Lightbulb Institutes Delving Deeper
List of Institutes Researching Bird Head Bobbing
Things We Don’t Know (1 researchers)
2

Do sharks see in colour? We just don't know. Image credit: Public domain
There are also mysteries about animal vision in the sea. Shark eyes are highly developed and in many ways quite similar to human eyes. Their brains also allocate a huge amount of space to vision. Yet some researchers believe that sharks are colourblind. This hypothesis follows years of speculation and observation of behaviour, and recent research in the Royal Society Biology Letters.

Sequencing the photoreceptor proteins revealed wobbegongs (a group of shark species) only have one kind of colour detecting cone cell. However, other researchers have observed sharks being attracted or repulsed by certain colours. For example, Gruber trained a lemon shark to select coloured targets from a mixture of ones of the same brightness and has identified the presence of colour-sensitive cones in the eyes of several shark species. The evidence is conflicting, and fundamentally, nobody knows whether or not sharks can see in colour.
Learn more about Whether sharks see colour.

 Lightbulb Institutes Delving Deeper
List of Institutes Researching Whether sharks see colour
Things We Don’t Know (1 researchers)
2

Computer Vision

imgbucket
For robots to interact naturally with humans, they need to be able to see like us. Image credit: Public domain
For robots to be adaptable to human environments and become part of our daily lives, they need to be able to successfully perceive the world around them. For this, they don't just need to be able to 'see' like humans, their eyes need to move like human eyes too.

Researchers have attempted to replicate the muscle motion of the human eye using piezoelectric ceramics which change in size slightly when electricity is passed through them. This allows them to more closely imitate the movements of the human eye which will be useful for research studies on human eye movement as well as for making video feeds from robots more intuitive.

 Lightbulb Institutes Delving Deeper
List of Institutes Researching Could robots soon have 'human-like' vision?
Things We Don’t Know (1 researchers)

This article was written by the Things We Don’t Know editorial team, with contributions from Ginny Smith, Johanna Blee, Rowena Fletcher-Wood, Joshua Fleming, and Holly Godwin.

This article was first published on and was last updated on 2018-01-27.

References
why don’t all references have links?

Tahnbee, K., Florentina, S., Kerschensteiner, D., (2015) An excitatory amacrine cell detects object motion and provides feature-selective input to ganglion cells in the mouse retina eLife 4 DOI: 10.7554/eLife.08025
Lafer-Sousa, R., Hermann, K,L., Conway, B,R., (2015) Striking individual differences in color perception uncovered by ‘the dress’ photograph Current Biology 25(13):R545-R546 DOI: 10.1016/j.cub.2015.04.053
Schultz, J. and Ueda, J., (2013) Nested Piezoelectric Cellular Actuators for a Biologically Inspired Camera Positioning Mechanism IEEE 29.5:1125 - 1138 DOI: 10.1109/TRO.2013.2264863

Recent vision News

Get customised news updates on your homepage by subscribing to articles

atom icon
Neural scene representation and rendering
18th June, 2018
Google's DeepMind algorithm can now teach itself to see, using two networks that play a collaborative game with each other. In the game, one describes an image as concisely as possible and the other replicates it from a new viewpoint. After comparison, the networks adjust their descriptions and visualisations and move on to the next description. This takes a step away from traditional computer vision algorithms, which focus on specific tasks, and towards a more flexible, organic visual system.
atom icon
'Blind spots' in peripheral vision vary from person to person
11th April, 2017
Areas of strength & weakness in peripheral vision vary from person to person, small study finds. A busy visual field also makes it harder to spot items in the periphery, which may have implications when driving in a crowded city.


Follow TWDK

Mailing list

Sign up for our Newsletter
constant contact safe subscribe logo

Fundraising

Easyfundraising banner

Creative Commons License
Except where otherwise noted, content on this site by Things We Don’t Know C.I.C. is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. | Privacy & Cookies
Things We Don’t Know C.I.C. is registered in England and Wales. Company Number 8109669.
Registered address at 34B York Way, London, N1 9AB.