Optimize Data Visualizations with the Right “Visual Features”
As a follow-up to a prior blog post on the Top 7 Data Visualization Best Practices, I wanted to go into a bit more detail about “visual features”. In the prior blog, we discussed “intuitive” data visualizations and how that stems from your brain pre-attentively processing visuals. While it’s not technically accurate, many people think of pre-attentive processing as subconsciously understanding a visual. However you describe it, it is fast! On the order of 40-50 milliseconds for the average person. To put that in perspective, a single flap of a hummingbird’s wings is about 20 milliseconds. Alternatively, the human brain perceives anything around 30 milliseconds as “continuous”, which is how “motion pictures” work. So 40 milliseconds would be around the speed of a single frame in a movie.
Unfortunately, not all things can be processed pre-attentively. Text and numbers cannot be processed pre-attentively, nor can 3D images. Of course, this begs the question, what *CAN* be processed pre-attentively?
Categories of Visual Features
Visual features that can be processed pre-attentively fall into four basic categories: form, colour, position, and motion.
- Form: How shapes appear in terms of shape, size, orientation, etc.
- Colour: How shapes appear to eyes in terms of combinations of red, blue, and green.
- Position: Where shapes appear in relation to each other.
- Motion: The direction in which shapes move compared to each other.
Naturally these four categories of visual features help us to classify pre-attentive visual features, but they don’t really help us understand exactly what each one is. Luckily we have a list of the various visual features that belong to each type. Everything in this list can be processed very quickly by your end users.
For all of the following visual features, the average person can spot the “difference” in 40-50 milliseconds. How fast can you see the "odd one out?"
For some of these visual features the term used may not be self-explanatory, so a brief description is provided next to the term. You'll have to forgive the odd spacing of the following images, but LinkedIn's post editing abilities leaves a LOT to be desired and provides very little ability to properly lay these images out.
Form
The majority of pre-attentive visual features fall into this category. This category includes the following:
Orientation
Length / Width
Size
Enclosure
Density
Curvature
Shape
Convex / Concave
Of this list, concave and convex are not normally used due to their three dimensional appearance. So while concavity/convexity can be processed pre-attentively, it is not generally best practice to include this when visualizing data.
Colour
There are only two visual features related to colour that can be processed pre-attentively, and they are intensity and hue.
Intensity (lighter variation of a colour already present)
Hue (a colour different from other colours already present)
Position
In positional visual features, there are only two that can be processed pre-attentively, and they are two-dimensional (2D) position and stereoscopic depth.
2D Position
Sterioscopic Depth (appearance of one object being “deeper” or “further away” than another)
Motion
While it is possible to pre-attentively process motion, the use of motion can also be very distracting when visualizing data. The visual features in this type include flicker, direction of motion, and velocity. These three are presented for completeness of theory, but use of motion should be kept to a minimum, if used at all. This is mainly due to the fact that motion when combined with other pre-attentive visual features, normally results in post-attentive processing. While some visual features can be combined without difficulty, motion cannot generally combine with other visual features and remain pre-attentive.
* Note the following images should be animated, but it appears that LinkedIn doesn't handle animated gifs very well. Feel free to pop over to my blog on Unilytics' website to see the images in action. You'll find it here.
Velocity
Direction of Motion
Flicker
Caveat
Be careful when using pre-attentive visual features, as not all go well together, and use of too many pre-attentive visual features actually forces the brain to process the visual post-attentively, which is considerably slower. For example, look at the following images:
Image A
Image B
In image A, is the boundary between the red area and blue area horizontal or vertical? Pretty easy to see that it is horizontal, correct? So even though shape and hue, two visual features, are being used in the display, it is still obvious where the boundary is. Contrast this with Image B. Image B also uses shape and colour, but can you tell where the boundary is and whether it is horizontal or vertical? It’s vertical, but it takes some focus to detect it.
As another example, in the following image, find the two red squares.
Image C
These last two questions take considerably longer than 40-50 milliseconds, so they would be considered post-attentive, meaning they take focused attention to view, perceive, and comprehend.
Conclusion
There you have it… the official list of pre-attentive visual features. I’m not normally one to memorize things, but when it comes to data visualization, I recommend committing the list of pre-attentive features to memory. In addition, memorizing the list of visual feature effectiveness by data type from my previous blog is also a huge help when designing data visualizations.