Next Article in Journal
Branding Practices on Four Dairies in Kantale, Sri Lanka
Next Article in Special Issue
Thermography as a Non-Invasive Measure of Stress and Fear of Humans in Sheep
Previous Article in Journal
Spontaneous Learning of Visual Structures in Domestic Chicks
Previous Article in Special Issue
The Use of Qualitative Behaviour Assessment for the On-Farm Welfare Assessment of Dairy Goats
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Affect-Driven Attention Biases as Animal Welfare Indicators: Review and Methods

1
Institute for Global Food Security, School of Biological Sciences, Queen’s University Belfast, Medical Biology Centre, 97 Lisburn Road, Belfast BT9 7BL, UK
2
Research Centre in Brain and Behaviour, School of Natural Sciences and Psychology, Liverpool John Moores University, James Parsons Building, Byrom Street, Liverpool L3 3AF, UK
*
Author to whom correspondence should be addressed.
Animals 2018, 8(8), 136; https://doi.org/10.3390/ani8080136
Submission received: 1 June 2018 / Revised: 2 August 2018 / Accepted: 4 August 2018 / Published: 7 August 2018
(This article belongs to the Special Issue Animal Emotion)

Abstract

:

Simple Summary

Good animal welfare requires minimizing suffering and promoting positive experiences. To achieve this, we need reliable indicators of animals’ psychological states. In humans, different moods and emotions (“affects”) are associated with changes in visual attention (“attention bias”). We review studies investigating whether attention biases are also indicators of affect in animals. Although research is limited, evidence for affect-driven attention biases has been found in several species, especially primates and livestock. These studies are discussed in relation to tasks developed for measuring attention in humans. We identify additional findings from human psychology that might be applied to animals, particularly species not studied before, and conclude that affect-driven attention bias is a promising welfare indicator. However, it may be more useful for studying responses to specific stimuli, rather than general wellbeing. With further study, we hope these findings contribute to fulfilling society’s ethical obligations towards animals.

Abstract

Attention bias describes the differential allocation of attention towards one stimulus compared to others. In humans, this bias can be mediated by the observer’s affective state and is implicated in the onset and maintenance of affective disorders such as anxiety. Affect-driven attention biases (ADABs) have also been identified in a few other species. Here, we review the literature on ADABs in animals and discuss their utility as welfare indicators. Despite a limited research effort, several studies have found that negative affective states modulate attention to negative (i.e., threatening) cues. ADABs influenced by positive-valence states have also been documented in animals. We discuss methods for measuring ADAB and conclude that looking time, dot-probe, and emotional spatial cueing paradigms are particularly promising. Research is needed to test them with a wider range of species, investigate attentional scope as an indicator of affect, and explore the possible causative role of attention biases in determining animal wellbeing. Finally, we argue that ADABs might not be best-utilized as indicators of general valence, but instead to reveal specific emotions, motivations, aversions, and preferences. Paying attention to the human literature could facilitate these advances.

1. Introduction

Animal welfare science aims to reduce suffering and improve wellbeing [1,2]. To achieve this, we need objective measures and standardised tests that quantify the psychological component of affect. However, animals cannot communicate verbally and we cannot measure their subjective experience directly [3]. Welfare scientists instead rely on indirect indicators, such as behaviour and physiology. This can be problematic, though, because behaviour is often species-specific, difficult to interpret, and varies between individuals (personality). It may only highlight extremes of welfare and can become dissociated from affective state, as in stereotypies [4]. Physiology, meanwhile, fluctuates with activity level and circadian rhythms, often signalling affective intensity (arousal) rather than valence (see Reference [5]). Moreover, behavioural and physiological welfare indicators have traditionally focused on negative affect, but good welfare also requires recognizing and promoting positive states [2]. Additional measures are therefore needed.
A promising avenue of research for measuring affective states in animals comes from cognitive psychology (see Reference [6]). In humans, theory and methods to investigate the relationship between affect, cognition, and subjectively experienced feelings are well established. For example, people in negative-valence states recall more negative memories, interpret ambiguous information more pessimistically, and allocate more attention to potential threats than happier people [7]. In animal welfare science, affect-modulated cognition is termed cognitive bias.
Most cognitive bias research on animals has focused on judgement biases (reviewed by [8,9,10,11]). These are usually measured by training animals to respond differently to two unidimensional stimuli, which yield relatively positive- and negative-valence outcomes. For experimental trials, ambiguous intermediate “probe” stimuli are introduced. Responding to probes as though to the positive stimulus is interpreted as an optimistic judgement bias characteristic of positive-valence states; responding as though to the negative stimulus is interpreted as a pessimistic judgement bias indicative of negative-valence states. For instance, Harding et al. (2004) [12] trained rats (Rattus norvegicus) to press a lever after one tone for food but not to press it after a different tone to avoid a burst of white noise. When they played an intermediate tone, rats housed in stressful conditions pressed the lever in significantly fewer trials than non-stressed individuals, suggesting a stronger inclination to expect the unpleasant outcome. Variants of the judgement bias task (JBT) have since been developed for other mammals (e.g., rhesus macaques [13], Macaca mulatta, sheep [14], Ovis aries, dogs [15], Canis lupus familiaris, and bottlenose dolphins [16], Tursiops truncatus), birds (e.g., chickens [17], Gallus gallus domesticus, and starlings [18], Sturnus vulgaris), and insects (e.g., fruit flies [19], Drosophila melanogaster, and honeybees [20], Apis mellifera).
However, JBTs have limitations. Lengthy training periods are time-consuming for researchers, impractical in applied settings, and lead to attrition of subjects. The effect of stress on learning [21,22,23] may also cause a selection bias, with animals in negative-valence states less likely to meet the inclusion criteria [10]. Furthermore, subjects tested repeatedly can learn the probes are unreinforced, making them less likely to respond [24]. Finally, JBTs require well-designed controls for non-valence confounds, such as arousal, motivation, distraction, and general activity (see References [9,10] for further discussion). Cognitive bias paradigms that require less training and fewer controls would help to resolve these issues.
Another class of cognitive bias, attention bias, describes the differential allocation of attentional resources towards one stimulus compared to others. Unlike JBTs, measuring attention biases often requires little or no training and does not depend on interpreting optimistic and pessimistic responses. More fundamentally, external stimuli determine internal valence via attentional processes. In humans, attention biases prioritizing negative information are implicated in the onset and persistence of affective disorders, such as anxiety and social phobia [25]. The stimuli animals attend to also underpin their experience of the world and, ultimately, their wellbeing. Therefore, attention biases warrant investigation [9,26,27].
Attention researchers distinguish between different aspects of attention: initial engagement (attentional capture or orienting, which is enhanced for threat-relevant stimuli [28]), maintenance of attention towards a stimulus (permitting detailed processing), and disengagement (which facilitates shifting to other stimuli) [29]. Human clinical studies have found that affect impacts attention at each stage, from faster engagement to enhanced maintenance and facilitated [30,31] or impaired [32] disengagement. Overall, the evidence shows that clinically anxious people look towards threatening information faster and for longer than non-anxious people [33].
In animal welfare science, attention modulated by the observer’s affective state is typically referred to as “attention bias”. However, the human literature also recognizes attention biases unrelated to affect. For example, people locate inverted letters amongst upright letters faster than upright letters amongst inverted letters [34]. This attention bias towards novelty is not induced by either stimulus valence or the observer’s affective state. To maintain clarity, then, we suggest welfare scientists adopt the term affect-driven attention biases (ADAB). This is measured using attention bias tasks (ABTs; reviewed by [35]). Developed for human research and adapted for animals, ABTs quantify attention allocation to experimental presentations of stimuli. Gaze might be tracked directly or reaction-times recorded to specific cues.
In this review, we evaluate ADAB as a welfare indicator and summarize studies conducted with animals to date. We first outline a nomenclature based on common terminology in human psychology, which will facilitate integration between the literatures (see Table 1). Next, we review studies exploring ADABs in a welfare assessment context and introduce the most widely used ABTs, focusing on their potential as welfare indicators. Finally, we suggest future directions for ADAB research. In this paper, “attention” refers to overt visual attention, typically based on eye gaze. We aim to comprehensively review studies using overt visual attention as the response variable in ADAB tasks, while acknowledging that attentional resources may be attributed to information acquired through other sensory modalities.

2. Literature Review: Methods

We aimed to exhaustively review the ADAB literature to date. In March 2018, the Web of Science database was searched with the term “attention bias animal welfare” (26 results). This was updated at the end of July 2018. References to “attention” or “attention bias” were also identified in previous reviews of cognitive bias [7,8,9,10,11] and ABTs for animals [36,37]. In addition, the references in papers identified through these methods and the papers citing them were systematically searched, as well as papers citing the reviews. Titles and abstracts were read to ascertain the relevance of search results. Selection criteria were that subjects must have been tested in different valence conditions, and their attention measured towards an emotional stimulus or stimuli. Animal welfare researchwhere the authors described their findings as an attention bias was also included. Table 2 summarises studies that met these criteria.

3. Literature Review: Results and Discussion

We identified 12 ADAB studies, which investigated eight species and used five ABT methodologies. These involved negative-/positive-valence treatments (five studies), negative/control treatments (five studies), inferences of valence from behaviour (one study), and testing animals with different affect-related personality profiles (one study). Of the 10 studies that manipulated affective state, two compared long-term differences in housing, three used pharmacological treatments, and five implemented non-pharmacological manipulations in the week before testing. Stimulus presentations were either negative-valence (one study), negative/neutral (three studies), negative/positive (five studies), negative/neutral/positive (one study), positive (one study) or positive/neutral (one study). All but one study identified significant treatment differences potentially attributable to ADABs. We now discuss this body of research in the context of ABTs from cognitive psychology and other attention bias studies on animals. In particular, we focus on state ADABs, rather than trait affect, and experiments where attention biases were not confounded with judgement biases, which have been reviewed elsewhere [7,8,9,10,11].

3.1. Looking Time Tasks

The simplest ABTs are looking time tasks (reviewed by [37]). Originally developed for preverbal human infants [48], they directly measure subjects’ gaze patterns towards visual stimuli, which can be presented either singly or simultaneously. Single-presentation tasks compare looking times between successive trials and reveal which aspects of a stimulus are attended or avoided in the absence of distractions. By contrast, dual-presentation methodologies (also known as the preferential looking paradigm or visual paired comparison) introduce competition between stimuli for processing [49].
Although the preferential looking paradigm had previously been used to address questions related to social attention (e.g., [50]), the first study measuring ADAB was conducted by Bethell et al. (2012) [26] with rhesus macaques. Monkeys were shown two images of conspecifics simultaneously (one with an aggressive expression, the other with a neutral expression) and attention bias was quantified as more time spent looking at one image than the other. Monkeys were tested after a negative-valence manipulation (veterinary health-check) and during a positive-valence manipulation (enhanced enrichment). The macaques showed an attention bias towards the aggressive expression: they looked towards the aggressive face faster than the neutral face. However, maintenance of attention towards the aggressive face was mediated by the affect manipulation. Monkeys continued to look at the aggressive expression during enrichment but were faster to look away following the vet check (and continued to avoid the face for the rest of the trial).
Another ADAB looking time paradigm has recently been validated for sheep [41,42] and cattle [43] (Bos taurus; see also Reference [51]). Subjects were moved to a test arena with food available and a hatch was opened for 10s to reveal a dog (a threatening predator stimulus). The response variables were looking time towards the dog and towards the closed hatch after the dog’s removal, as well as latency to feed. In both sheep and cattle, looking time towards the hatch increased with the administration of anxiogenic drugs and decreased with anxiolytics.
Two further studies have investigated ADAB in sheep (see also Reference [52]). Verbeek et al. (2014) [39] demonstrated that food-deprived sheep spent more time interacting with an empty bucket associated with food, which the authors interpreted as enhanced attention. In a longer-term study, Vögeli et al. (2015) [40] kept flocks in two housing conditions, either enriched and predictable to cause positive-valence mood states or unenriched and unpredictable to induce negative-valence mood states. Subjects were then shown videos of other sheep engaged in aggressive, affiliative, and non-social behaviours. Both treatments spent the most time oriented towards the aggressive behaviours and the least towards the affiliative behaviours. Negative-valence sheep spent longer oriented towards the stimuli overall, however, which may have been an ADAB to social information. A video-based preferential looking paradigm was also recently developed for sheep [53], although this study did not include an affect manipulation. Unlike in Vögeli et al., no significant differences were found between the positive (conspecifics) and negative stimuli (dogs).
An additional looking time experiment explored ADAB in starlings. Brilot et al. (2009) [38] switched off the cage lights, added food, and exposed the birds to either a negative-valence treatment (alarm calls, predator calls, and white noise) or a control treatment (conspecific calls). When the lights came back on, stimuli mimicking predator eyespots also appeared and competed with the food for attention. However, there were no treatment differences in time spent orienting towards the stimuli. Brilot et al. attributed this to eyespots not being inherently aversive.
As measures of ADAB, looking time tasks have many benefits. Quantifying eye gaze directly avoids potentially confounding proxies and allows different aspects of attention to be distinguished. Moreover, gaze patterns across stimuli can be observed. For instance, Somppi et al. (2016) [54] found that dogs fixate on certain facial features, such as the eyes, and this effect was influenced by the valence of the facial expression. Measuring untrained looking behaviour is also useful when conditioning would be impractical or impossible, such as research with wild animals and marine species. By suspending objects from a ship, for example, Siniscalchi et al. (2012) [55] measured the looking time of striped dolphins (Stenella coeruleoalba) in the Mediterranean Sea. Other paradigms have been successful with free-ranging macaques [56,57,58,59,60]. Future research might also investigate neglected taxa, like recent studies measuring looking behaviour towards conspecific [61] and predator [62] stimuli in lizards.
However, as an indicator of valence, looking time can be difficult to interpret. In Bethell et al.’s rhesus macaque study [26], the authors predicted that stressed monkeys would allocate more attention towards the threatening faces, but looking times were actually shorter in this treatment. Human studies have also associated negative-valence states with both attention to threat (e.g., [34]) and threat-avoidance (for a review, see Reference [63]). While this directionality issue is problematic for attempts to measure valence, it may be overcome by varying stimulus intensity. Human research has identified avoidance of low-level threat and attention to high-level threat in nonclinical populations [64,65]. Demonstrating this effect in macaques may require more objective methods for classifying stimuli. For example, subjective experimenter ratings could be replaced by the Macaque Facial Action Coding System (MaqFACS [66]).
Furthermore, some ADAB studies have used crude measures of looking time, such as how often and for how long the subject’s head is up (e.g., [38]). More precise methods involve manually coding eye gaze from video footage (e.g., [26,67,68]) and automated eye-tracking (see References [37,69,70]; human studies reviewed by References [71,72,73,74]). The latter is fast, objective, and accurate, but also expensive and needs modifying for new species (e.g., marmosets [75], Callithrix jacchus, peafowl [76], Pavo cristatus, and archerfish [77], Toxotes chatareus). Although impractical outside controlled conditions, eye-trackers have been successfully mounted to freely-moving ring-tailed lemurs [78,79] (Lemur catta). While existing tasks typically measure attention to static images, responses to photographs are often quantitatively weaker or qualitatively different than responses to moving images or the objects themselves [80,81]. Researchers could, therefore, experiment with videos [82,83,84,85,86], computer animations [87,88], models [89], and real animals or objects [41,42,43,54].
Collectively, these studies demonstrate the potential of looking time tasks to investigate ADAB, with research on macaques, sheep, and cattle suggesting it is affect-modulated. Similar paradigms would be suitable for any animal with measurable eye gaze, including birds and reptiles, and the simplest methods do not require training. Looking time tasks could, therefore, be adapted to a range of species and situations.

3.2. Emotional Stroop Tasks

The emotional Stroop task [90] (reviewed by Reference [91]) measures how much an individual is distracted by emotional information while they perform an otherwise neutral cognitive task. Typically, participants are instructed to name, as quickly and accurately as possible, the colour in which words appear on a screen. Anxious populations are slower to name the colour of negative-valence words (e.g., pain) than neutral words (e.g., gain; see Reference [92] for a review), an effect not seen in non-anxious people. Variants of the task using facial expressions (neutral, aggressive, and happy) instead of words produce similar results [93]. The emotional Stroop effect is interpreted as negative affect enhancing attentional capture by negative distractor content [94,95,96], although the task does not distinguish capture, maintenance or disengagement of attention. It also does not rule out alternative explanations, such as freeze response [97].
An emotional Stroop task was recently developed for chimpanzees [44] (Pan troglodytes), which were trained to press a blue-framed square on a touchscreen, but not a yellow-framed square. The chimpanzees’ response-times were then recorded to blue-framed squares containing images of either the veterinarian or non-threatening humans. Subjects were slower to touch the blue frame when it contained a picture of the vet, a slowing effect that was stronger when they had recently undergone a veterinary procedure. The authors attributed this result to ADABs; specifically, attentional capture by stimuli associated with negative-valence states (see also Reference [98]).
However, this paradigm required extensive training. Of 16 chimpanzees conditioned on the blue-frame/yellow-frame discrimination task, only seven reached the study’s inclusion criteria. Even those needed 900–6700 trials [98]. This extended training period and attrition of subjects suggests the paradigm may be impractical for welfare assessment. More fundamentally, reaction-times in the emotional Stroop paradigm can reflect motor action biases rather than attention biases, so changes in response are difficult to interpret in terms of attention. This is less of an issue in other ABTs, which measure reaction-times to valence-neutral targets after the emotional stimuli have disappeared.
Other studies have quantified emotional Stroop effects by observing how emotional stimuli distract from an attention-demanding task. Performance decrements were observed in amazon parrots (Amazona amazonica) exposed to humans during a foraging task, and this effect correlated with subjective ratings of the birds’ neuroticism [45]. Landman et al. (2014) [99] found that threatening facial expressions distracted macaques from a visual task, whereas Bellegarde et al. (2017) [100] reported that sheep learnt a discrimination task faster when it involved negative-valence facial stimuli than valence-neutral images.
Stroop effects have also been studied in the context of learned helplessness. A classic model of depression, learned helplessness describes the unresponsiveness of animals that cannot escape repeated uncontrollable stressors, typically electric shocks (for a review, see Reference [101]). One consequence is an attention bias towards goal-irrelevant external stimuli [102]. In studies on rats, for example, the performance of subjects in learned helplessness treatments was equivalent to controls on an automated cognitive task, but they were slower and made more errors when the experimenter was present as a distraction [103,104]. Eliminating the trained task completely, Rodd et al. (1997) [105] investigated whether learned helplessness-induced attentional shifts influenced an innate behaviour in chickens. Duration of tonic immobility, a catatonic response to restraint, increases with perceived threat. In the presence of eyespots, chickens in a learned helplessness condition were motionless for longer than control birds, but they recovered faster in the presence of unaffected conspecifics. Despite indicating a putatively negative-valence state, we do not consider these examples of ADAB because they are attention biases towards external stimuli rather than emotional stimuli per se [102]. Learned helplessness research nonetheless demonstrates that attentional shifts are associated with depression-like states as well as anxiety, and similar experiments could investigate ADAB.
The conventional emotional Stroop task has only been tested with chimpanzees, although emotional Stroop effects are observed in various species. Given the time requirement of training and issues in interpretation, the human paradigm is unlikely to translate to the applied setting. However, neutral task performance or behavioural shifts in the presence of a threatening stimulus are a simple, adaptable approach to measuring ADAB in animals.

3.3. Dot-Probe Tasks

The dot-probe paradigm [106] (reviewed by Reference [36]; metanalyses by References [34,107]) presents participants with two stimuli on a screen. Stimuli may be words (e.g., threatening/neutral pairs [106,108]) or images (e.g., different facial expressions [109,110]). After a fixed duration, both stimuli disappear, and one is replaced by the “dot-probe”—a neutral target subjects must respond to. Faster reaction-times to the target indicate the participant’s attention was already fixed on that location, whereas slower reaction-times suggest their attention shifted from the other location. In humans, dot-probe studies pairing negative and neutral stimuli have found attention to threat is stronger in anxious [34,111] and depressed [112] people, and during high-stress situations [113,114].
To our knowledge, no published studies have tested for ADAB in animals using the dot-probe paradigm. However, Kret et al. (2016) [115] identified attention biases to positive-valence social cues in bonobos (Pan paniscus) with a dot-probe task. When presented with image pairs of conspecifics performing emotion-regulatory and neutral behaviours, subjects responded faster if targets replaced the emotional stimuli, but only for certain behaviour classes. Effects were significant for grooming and sex, but not play or distress. Kret et al. explained the attention biases towards affiliative social stimuli as an adaptation to facilitate bonobos’ characteristic conflict-resolution and emotion-regulation strategies [116]. While effect sizes may be smaller [117], ADAB studies could further explore responses to positive stimuli, such as play faces [118,119], gentle handlers [51], and food [39].
Several authors have also trained macaques on dot-probe tasks (e.g., [120,121]). King et al. (2012) [122] observed a baseline attention bias towards threatening stimuli (open-mouth conspecifics), which was unaffected by testosterone administration. In a study with Japanese macaques (Macaca fuscata), Koda et al. (2013) [123] used stimulus pairs of new-borns and adults, but did not find a significant difference in reaction-times.
The dot-probe task provides a snapshot of attention allocation, and questions about orienting and maintenance of attention can be answered by adjusting the length of stimulus presentation. For example, work with humans has shown attention bias for threatening stimuli presented below the temporal threshold for awareness. Testing with multiple durations offers a better understanding of the aspects of attention involved, with shorter durations measuring engagement (e.g., [123]) and longer durations measuring disengagement (e.g., [120]).
Future dot-probe studies could investigate whether affect manipulations influence reaction-times and explore the aspects of attention underpinning ADAB. While this task requires relatively little training, touchscreens are best suited to laboratory settings and dexterous subjects, especially primates.

3.4. Emotional Spatial Cueing Tasks

Spatial cueing tasks [124] also quantify attention biases through reaction-times to a valence-neutral target. Subjects first fixate on a point in the centre of a screen. Their objective is to respond as quickly as possible to the target, which can appear on either side. Before the target’s arrival, a cue signals its position. This is usually located where the target will be, but is on the other side of the screen in a minority of trials. Reaction-times when the cue correctly predicts the target’s location indicate attentional engagement towards the cue, whereas reaction-times when the target and cue appear in different locations indicate disengagement from the cue [125,126]. Hence, unlike the dot-probe task, spatial cueing automatically distinguishes between different aspects of attention.
The emotional spatial cueing paradigm involves manipulating the cue’s affective content. For example, Fox et al. (2001) [30] tested anxious and non-anxious people on a task with threatening words and faces as cues. There were no treatment differences in engagement to the cues, but a significant difference in disengagement. Anxious individuals were slower to reallocate their attention from the threatening cue to the valence-neutral target.
No studies, to our knowledge, have adapted the emotional spatial cueing task for animals. However, non-valenced predictive cue paradigms have been successful with macaques [127], rats [128], chickens [129], honeybees [130], and archerfish [131]. In the latter study, a touchscreen was suspended above the tank and a symbolic cue predicted the location of a food-delivering target. Fish were trained to hit the targets using mouth-propelled water jets, a natural behaviour for this species. Their reaction-times were faster when the target appeared in the same location as the cue.
To create an emotional spatial cueing task, it would be a small step to introduce affective stimuli into the predictive cue paradigm and test subjects in different valence states. This could be effective with diverse taxa, albeit under controlled conditions.

3.5. Visual Search Tasks

In the visual search task, participants are instructed to locate a target stimulus in an array of distractor stimuli (e.g., [28,132]). Faster target detection reveals attention bias for the stimulus, whereas slower detection suggests either the target does not capture attention or does so less than other images in the array.
Marzouki et al. (2014) [46] used abstract shapes with conditioned valence in a visual search task for Guinea baboons (Papio papio). Subjects were trained to locate a T-shaped target among seven L-shape distractors. To investigate the influence of affective state, the authors only analysed trials preceded by ostensibly valence-informative behaviours. Reaction-times in trials following negative-valence behaviours were slower than those following valence-neutral or positive-valence behaviours.
However, observational studies inferring affective state from behaviour can be open to interpretation. For example, while Marzouki et al. categorized self-grooming as positive-valence, this displacement activity has been widely linked to stress [133,134]. Resting, analysed as negative-valence, is a biological necessity. It is also a low-arousal activity, suggesting the observed effect could be attributed to affective intensity rather than valence. Well-designed affect-induction experiments can avoid such confounds and demonstrate causality.
Primate response-times have also been recorded towards non-symbolic, emotional images, albeit in studies that did not manipulate subjects’ affect. Chimpanzees located conspecific faces faster than neutral objects [135], and human faces faster when their gaze was directly forward rather than averted [136]. In Japanese macaques, median reaction-times were faster towards target aggressive faces amongst neutral face distractors than vice versa [137] (see also Reference [138]). Macaques also detected snakes quicker than fear-irrelevant stimuli in arrays of neutral distractors [139,140]. Instead of using reaction-times as the response variable, Simpson et al. (2017) [141] measured infant macaques’ gaze across visual search arrays with an eye-tracker. This eliminated training and potential motor response biases.
Other taxa have also been trained on visual search tasks, including recent work with barn owls [142,143] (Tyto alba), zebrafish [144] (Danio rerio), and bumblebees [145] (Bombus terrestris). The latter study conditioned bees to associate specific colours with rewards, which they could detect in arrays of distractor colours. As far as we are aware, no non-primate studies have investigated whether affective state influences performance on visual search tasks. However, research on judgement bias has validated affect manipulations for bees [20,146].
Visual search tasks have recently been criticised in cognitive psychology, because searching is heavily influenced by the intensity of emotional stimuli, as well as their valence [147,148,149]. In a systematic reanalysis of human studies, Lundqvist et al. (2013) [150] concluded that happy facial expressions captured attention faster than angry expressions when they were rated higher on arousal indices, but this effect was reversed when the angry expressions were rated as higher arousal. Another study exposed subjects to either an arousing negative-valence sound, an arousing positive-valence sound or a neutral sound before they completed a visual search task [151]. Both high-arousal sounds had similar effects on performance, regardless of valence. To avoid these issues, visual search experiments should control for the effects of arousal (see Reference [152]).
The complexity of visual search tasks, coupled with the issue of arousal, suggests they are not an immediate priority for ADAB research. However, primate studies could investigate the influence of different affect manipulations on search behaviour.

3.6. Related Paradigms

In addition to standard ABTs, other tests and metrics might also measure ADAB. Based on human paradigms [153], the five-choice serial reaction time task [154] (5-CSRTT; reviewed by Reference [155]) presents subjects with five apertures and requires them to approach the one just illuminated. It is framed in terms of measuring attention. In a study on pain in rats, Boyette-Davis et al. (2008) [156] tested subjects injected with formalin on the 5-CSRTT. Formalin-treated rats made fewer approaches, which the authors interpreted as a failure to attend to the task when in pain (see also Reference [157,158]; pain-induced cognitive impairment reviewed by Reference [159]). Behavioural data confirmed that subjects that did not receive morphine were in the most pain, and they showed the highest rates of locomotion in open field tests, suggesting their failure to respond was not due to reduced activity. Additionally, data from trials when they did approach indicated responses were no slower in the pain group.
JBTs may also measure ADABs. Like Harding et al. (2004) [12], Parker et al. (2014) [47] tested rats on tones intermediate between two reference tones associated with differently valenced outcomes, but they also recorded responses to dual presentations of both trained tones. The authors suggested the latter approach was a novel ABT. In addition, results from JBTs using biologically relevant stimuli might be attributed to ADABs. For example, chicks moved slower down an alley containing predator silhouettes and conspecific/predator morphs following social isolation [160], an effect reversed by anxiolytic treatment [161]. This may indicate an ADAB away from threatening stimuli. Other studies have tested ambiguous eyespots on starlings [38] (no effect) and morphed facial expressions on goats [162] (Capra aegagrus hircus; no affect manipulation).
However, there is insufficient evidence to infer that attention underpins “judgement” biases. In the dual-presentation task, for instance, responses relied on the rats hearing one or both tones (attention), categorizing the presentation as either positive or negative (judgement), remembering the lever associated with each outcome (memory), and choosing which to press (decision-making). More research is needed to identify which cognitive faculties contribute to observed biases, bearing in mind they may act synergistically [163,164]. Parker et al. did not specify why their dual-presentation JBT should measure attention biases any more than the ambiguous-probe paradigm, but presumably reasoned that valence-conditioned stimuli would influence attention more than unfamiliar ambiguous stimuli.
Paul et al. (2005) [7] suggested some personality tests might also quantify ADAB. For example, novel object and human reactivity tests [165] often record looking time, latency to first touch, and subsequent interactions, which could be measures of attention and potentially affect-modulated. Startle tests [166,167], which expose subjects to sudden acoustic or visual stimuli, also measure looking behaviour towards the source. Stress potentiates the startle reflex in humans [168], and it is associated with negative-valence conditions such as clinical anxiety [169] and chronic pain [170]. Destrez et al. (2013) [171] explored the potential of these paradigms as indicators of ADAB by subjecting both chronically stressed and control lambs to a battery of personality tests. The negative-valence lambs made fewer contacts with a human entering their pen and looked towards a novel object for shorter bouts, although total looking time did not differ and there were no treatment differences in a startle test. Negative-valence states increase the startle response in rodents and macaques, however [172,173].
Finally, vigilance, or scanning the environment for threats, is influenced by fear and anxiety [174]. For example, Brilot et al. (2012) [27] discovered that, following an alarm call, starlings with access to a water bath spent less time vigilant and more time with their heads down compared to those without. The authors attributed this to the importance of bathing in maintaining feather condition, which improved the birds’ flying ability and lowered their vulnerability to predators. In livestock, anxiolytic treatment reduces vigilant behaviour during isolation tests [175] and after exposure to a potential predator [41,42,43]. Vigilance in cows is also lower in the presence of gentle, compared to aversive, stock people [51]. Furthermore, wild animals display heightened vigilance under dangerous or stressful conditions (reviewed by [176]). In playback experiments, coots (Fulica atra) spent longer scanning their surroundings after dog barks than control sounds [177]. Observational studies have also associated vigilance with predation risk in African ungulates [178], human disturbance in Japanese cranes [179] (Grus japonensis), and proximity to neighbouring territories in spider monkeys [180] (Ateles geoffroyi). Although these findings do not meet our stimulus-directed definition of ADAB, they demonstrate how attention biases might be measured in the field and provide an insight into their adaptive function.

4. Outstanding Questions and Future Directions

4.1. Other Sensory Modalities

Although we have focused on visual attention, other sensory modalities warrant investigation (see Reference [181]). This is illustrated by judgement bias studies, which have used auditory [12], olfactory [20], gustatory [182], and haptic [183] cues. Previous ABTs have also experimented with acoustic stimuli, such as starling alarm calls [38] and threatening dog barks [184], but most measured looking time towards the source. For some species, however, other response variables might be more conspicuous and accurate. Ruminants have a wide field of vision, for instance, rendering head orientation a potentially unreliable proxy for eye gaze. Instead, ear posture and movement signals affective state in several livestock species (e.g., cattle [185,186], sheep [53,187,188,189], goats [190], and pigs [191], Sus scrofa domesticus), as well as dogs [192] and mice [193] (Mus spicilegus). A preferential hearing paradigm might replace the competing images with a positive-valence conspecific vocalization and a negative-valence predator vocalization. Ear position would indicate ADAB. Like eye-trackers, an automated ear tracking system has even been developed for sheep [187]. However, ear postures can be purely communicative and indicate arousal as well as valence [185].
Non-visual ABTs would also be more biologically relevant for animals where vision is not the primary sense. Olfaction in particular is often overlooked by animal welfare scientists [194], despite being a dominant sensory modality for species widely used by humans, such as dogs [195], rats [196], and chickens [197]. Moreover, it is integral for communication and information-gathering in arthropods [198]. Cárdenas et al. (2012) [199] presented predatory spiders (Zodarion rubidum) with a control chamber and an experimental chamber, which airborne odours from different prey species were channelled into. Approaches into the experimental chamber therefore indicated attractive kairomones. While olfactory stimuli are inherently difficult to work with [194], ADAB researchers might use this method to investigate whether affect influences responses to food, conspecific, and predator odours. Such non-visual ABTs may facilitate research on commercially important and poorly studied taxa.

4.2. Effect Specificity

In humans, attention biases are often stimulus- or motivation-specific [117]. Combat veterans with post-traumatic stress disorder exhibit facilitated engagement and impaired disengagement when viewing war-related stimuli, but not disgust, neutral or positive-valence stimuli [200]. In emotional Stroop tasks, anxiety sufferers concerned about physical threats are slower to name the colour of words like “attack” and “illness”, whereas those with social worries are distracted by “incompetent” and “stupid” [64,65]. Insomniacs struggle to disengage from sleep-related stimuli [201]; addicts are biased towards opiates [202], cigarettes [203], and alcohol [204]; and, in healthy populations, food cues are more salient to hungry people [205,206,207]. These findings suggest blanket valence-based interpretations of attention biases may be inappropriate in animal welfare science, as well. Whereas JBT results can be understood as a correlation between optimism and valence, ABTs defy simple, overarching explanations.
Nonetheless, the studies reviewed herein demonstrate that ADABs can reveal specific emotions, motivations, aversions, and preferences. In the aforementioned sheep and cattle ABT, for example, anxiogenic and anxiolytic drugs modulated attention allocation towards a dog [41,42,43]. This threat-based task did not measure negative valence generally, but fear and anxiety specifically. Verbeek et al. (2014) [39] found that food-motivated sheep attended to a food-delivering bucket, while the emotional Stroop task for chimpanzees [44] used a contextual aversive stimulus (images of the veterinarian) that potentiated ADABs when subjects had recently encountered the vet.
Dawkins (2003) [208] defined animal welfare without recourse to affect, arguing it could be distilled into two questions: Is the animal physically healthy and does the animal have what it wants? Although other variables can increase attention towards a stimulus [41], ABTs might be a “quick-and-dirty” method to identify promising avenues for labour-intensive preference, motivation or aversion tests (reviewed by [209]). Hence, ADABs might help to answer Dawkins’ second question: Do animals have what they want?

4.3. Attentional Scope

Whereas ADABs towards specific stimuli might not indicate general valence, attentional scope may do so. The broaden-and-build theory of positive emotions [210,211] (reviewed by Reference [212]) proposes that, since positive-valence states often reflect overall wellbeing, contented individuals can devote resources to exploration, learning, and building up resources. Positive affects are, therefore, associated with a broad attentional scope and more attention allocated towards the periphery of the visual field (i.e., seeing the forest rather than the trees). On the other hand, negative-valence states are typically directed towards specific threats, so they narrow attentional scope and maximize attention towards the centre of the visual field [213] (i.e., seeing the trees rather than the forest). This has been demonstrated using various attentional tasks. In the Kimchi test [214], for instance, a target shape is presented, followed by two comparison shapes. Subjects are instructed to select the comparison most like the target. To test broaden-and-build theory, Gasper and Clore (2002) [215] showed participants a larger shape consisting of different smaller shapes as the target. This was either a triangle consisting of three squares or a square consisting of four triangles. After a positive-valence affect manipulation, subjects chose the larger shape as a closer match than the component shape more than subjects in a negative-valence condition, suggesting a broader attentional scope. Similar experiments might also indicate positive emotions and general wellbeing in animals (although, see Reference [216]).

4.4. Attention Bias Modification

Animal welfare scientists study attention biases as a symptom of negative-valence states, but some cognitive models of affective disorders identify them as a proximate cause (e.g., References [217,218,219]; reviewed by [220]). For instance, exaggerated attention to threat in anxious populations generates a positive feedback loop that reinforces existing biases [221]. A controversial new treatment aims to disrupt this harmful bidirectional relationship: attention bias modification (reviews and metaanalyses by References [25,222,223,224,225,226,227]). Using simple computer-based tasks, attention bias modification repeatedly presents subjects with valence/neutral stimulus pairs, such as angry and neutral facial expressions. For this example, correct responses would always require focusing on the neutral stimulus, training participants to divert attention away from threats through operant conditioning. In a modified dot-probe task, for instance, probes only ever appear behind neutral stimuli (e.g., Reference [228]). Although several recent studies have reported null effects (e.g., References [229,230,231,232]), the more promising tasks could be adapted for animals (e.g., the positive search paradigm [233,234]). For anxious individuals or situations where chronic stress is unavoidable, such as some primate laboratories, attention bias modification might be a cost-effective way to enhance mood states. However, this is not a substitute for good housing and husbandry.

5. Conclusions

We reviewed the literature on affect-driven attention bias as a welfare indicator and identified 12 studies to date. The initial results are promising. In chimpanzees, macaques, sheep, and cattle, affect manipulations have modulated attention towards or away from emotional stimuli, as well as the speed and duration of fixation. Both positive- and negative-valence states have been studied, although most animal research has concentrated on fear, anxiety, and threat biases. While welfare scientists were quick to recognise the potential of judgement biases, however, research on affect-driven attention biases remains comparatively limited. Nonetheless, methods might be developed for a range of taxa, including birds, reptiles, fish, and insects, and tested in both captive and free-range settings. Different attentional tasks measure different aspects of attention, but the looking time, dot-probe, and spatial cueing paradigms are especially promising. Future studies could use them to distinguish engagement and disengagement of attention, investigate effect specificity, and explore attentional scope as a welfare indicator. Attention bias modification might also ameliorate negative-valence moods in chronically stressed animals. By describing potential methodologies and evaluating the existing literature, we hope this review stimulates such research in diverse settings and species.

Author Contributions

Conceptualization: A.C. and E.J.B.; Methodology: A.C.; Writing-Original Draft Preparation: A.C. and E.J.B.; Writing-Review & Editing: G.A.; Supervision: G.A. and E.J.B.

Funding

This work was supported by a Postgraduate Studentship from the Department for the Economy (DfE), Northern Ireland.

Acknowledgments

We thank two anonymous reviewers for their thoughtful and constructive feedback on this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Duncan, I.J. A Concept of Welfare Based on Feelings. In The Well-Being of Farm Animals: Challenges and Solutions, 1st ed.; Benson, G.J., Rollin, B.E., Eds.; Blackwell: Oxford, UK, 2004; pp. 85–101. ISBN 9780470791752. [Google Scholar]
  2. Boissy, A.; Manteuffel, G.; Jensen, M.B.; Moe, R.O.; Spruijt, B.; Keeling, L.J.; Winckler, C.; Forkman, B.; Dimitrov, I.; Langbein, J.; et al. Assessment of positive emotions in animals to improve their welfare. Physiol. Behav. 2007, 92, 375–397. [Google Scholar] [CrossRef] [PubMed]
  3. Paul, E.S.; Mendl, M.T. Animal emotion: Descriptive and prescriptive definitions and their implications for a comparative perspective. Appl. Anim. Behav. Sci. 2018. [Google Scholar] [CrossRef]
  4. Higham, J.P.; MacLarnon, A.M.; Heistermann, M.; Ross, C.; Semple, S. Rates of self-directed behaviour and faecal glucocorticoid levels are not correlated in female wild olive baboons (Papio hamadryas anubis). Stress 2009, 12, 526–532. [Google Scholar] [CrossRef] [PubMed]
  5. Mendl, M.; Burman, O.H.; Paul, E.S. An integrative and functional framework for the study of animal emotion and mood. Proc. R Soc. Lond. Biol. 2010, 277, 2895–2904. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Shettleworth, S.J. Cognition, Evolution, and Behaviour, 1st ed.; Oxford University Press: New York, NY, USA, 1998; p. 0195319842. [Google Scholar]
  7. Paul, E.S.; Harding, E.J.; Mendl, M. Measuring emotional processes in animals: The utility of a cognitive approach. Neurosci. Biobehav. Rev. 2005, 29, 469–491. [Google Scholar] [CrossRef] [PubMed]
  8. Baciadonna, L.; McElligott, A.G. The use of judgement bias to assess welfare in farm livestock. Anim. Welf. 2015, 24, 81–91. [Google Scholar] [CrossRef]
  9. Bethell, E.J. A “how-to” guide for designing judgment bias studies to assess captive animal welfare. J. Appl. Anim. Welf. Sci. 2015, 18 (Suppl. 1), S18–S42. [Google Scholar] [CrossRef]
  10. Mendl, M.; Burman, O.H.; Parker, R.M.; Paul, E.S. Cognitive bias as an indicator of animal emotion and welfare: Emerging evidence and underlying mechanisms. Appl. Anim. Behav. Sci. 2009, 118, 161–181. [Google Scholar] [CrossRef]
  11. Roelofs, S.; Boleij, H.; Nordquist, R.E.; van der Staay, F.J. Making decisions under ambiguity: Judgment bias tasks for assessing emotional state in animals. Front Behav. Neurosci. 2016, 10, 119. [Google Scholar] [CrossRef] [PubMed]
  12. Harding, E.J.; Paul, E.S.; Mendl, M. Animal behaviour: Cognitive bias and affective state. Nature 2004, 427, 312. [Google Scholar] [CrossRef] [PubMed]
  13. Bethell, E.J.; Holmes, A.; MacLarnon, A.; Semple, S. Cognitive bias in a non-human primate: Husbandry procedures influence cognitive indicators of psychological well-being in captive rhesus macaques. Anim. Welf. 2012, 21, 185. [Google Scholar] [CrossRef]
  14. Doyle, R.E.; Fisher, A.D.; Hinch, G.N.; Boissy, A.; Lee, C. Release from restraint generates a positive judgement bias in sheep. Appl. Anim. Behav. Sci. 2010, 122, 28–34. [Google Scholar] [CrossRef]
  15. Burman, O.; McGowan, R.; Mendl, M.; Norling, Y.; Paul, E.; Rehn, T.; Keeling, L. Using judgement bias to measure positive affective state in dogs. Appl. Anim. Behav. Sci. 2011, 132, 160–168. [Google Scholar] [CrossRef]
  16. Clegg, I.L.; Delfour, F. Cognitive judgement bias is associated with frequency of anticipatory behavior in bottlenose dolphins. Zoo Biol. 2018, 37, 1–7. [Google Scholar] [CrossRef] [PubMed]
  17. Hernandez, C.E.; Hinch, G.; Lea, J.; Ferguson, D.; Lee, C. Acute stress enhances sensitivity to a highly attractive food reward without affecting judgement bias in laying hens. Appl. Anim. Behav. Sci. 2015, 163, 135–143. [Google Scholar] [CrossRef]
  18. Matheson, S.M.; Asher, L.; Bateson, M. Larger, enriched cages are associated with ‘optimistic’ response biases in captive European starlings (Sturnus vulgaris). Appl. Anim. Behav. Sci. 2008, 109, 374–383. [Google Scholar] [CrossRef]
  19. Deakin, A.; Mendl, M.; Browne, W.J.; Paul, E.S.; Hodge, J.J.L. State-dependent judgement bias in Drosophila: Evidence for evolutionarily primitive affective processes. Biol. Lett. 2018, 14, 20170779. [Google Scholar] [CrossRef] [PubMed]
  20. Bateson, M.; Desire, S.; Gartside, S.E.; Wright, G.A. Agitated honeybees exhibit pessimistic cognitive biases. Curr. Biol. 2011, 21, 1070–1073. [Google Scholar] [CrossRef] [PubMed]
  21. Vogel, S.; Schwabe, L. Learning and memory under stress: Implications for the classroom. Sci. Learn. 2016, 1, 1–10. [Google Scholar] [CrossRef]
  22. Sandi, C. Stress and cognition. Wiley Interdiscip. Rev. Cogn. Sci. 2014, 4, 245–261. [Google Scholar] [CrossRef] [PubMed]
  23. Conrad, C.D. A critical review of chronic stress effects on spatial learning and memory. Prog. Neuropsychopharmacol. Biol. Psychiatry 2010, 34, 742–755. [Google Scholar] [CrossRef] [PubMed]
  24. Doyle, R.E.; Vidal, S.; Hinch, G.N.; Fisher, A.D.; Boissy, A.; Lee, C. The effect of repeated testing on judgement biases in sheep. Behav. Proc. 2010, 83, 349–352. [Google Scholar] [CrossRef] [PubMed]
  25. McNally, R.J. Attentional bias for threat: Crisis or opportunity? Clin. Psychol. Rev. 2018. [Google Scholar] [CrossRef] [PubMed]
  26. Bethell, E.J.; Holmes, A.; MacLarnon, A.; Semple, S. Evidence that emotion mediates social attention in rhesus macaques. PLoS ONE 2012, 7, e44387. [Google Scholar] [CrossRef] [PubMed]
  27. Brilot, B.O.; Bateson, M. Water bathing alters threat perception in starlings. Biol. Lett. 2012, 8, 379–381. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Öhman, A.; Flykt, A.; Esteves, F. Emotion drives attention: Detecting the snake in the grass. J. Exp. Psychol. Gen. 2001, 130, 466–478. [Google Scholar] [CrossRef] [PubMed]
  29. Posner, M.I.; Petersen, S.E. The attention system of the human brain. Annu. Rev. Neurosci. 1990, 13, 25–42. [Google Scholar] [CrossRef] [PubMed]
  30. Fox, E.; Russo, R.; Bowles, R.; Dutton, K. Do threatening stimuli draw or hold visual attention in subclinical anxiety? J. Exp. Psychol. Gen. 2001, 130, 681–700. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Amir, N.; Elias, J.; Klumpp, H.; Przeworski, A. Attentional bias to threat in social phobia: Facilitated processing of threat or difficulty disengaging attention from threat? Behav. Res. Ther. 2003, 41, 1325–1335. [Google Scholar] [CrossRef]
  32. Rudaizky, D.; Basanovic, J.; MacLeod, C. Biased attentional engagement with, and disengagement from, negative information: Independent cognitive pathways to anxiety vulnerability? Cogn. Emot. 2014, 28, 245–259. [Google Scholar] [CrossRef] [PubMed]
  33. Bar-Haim, Y.; Lamy, D.; Pergamin, L.; Bakermans-Kranenburg, M.J.; Van Ijzendoorn, M.H. Threat-related attentional bias in anxious and nonanxious individuals: A meta-analytic study. Psychol. Bull. 2007, 133, 1. [Google Scholar] [CrossRef] [PubMed]
  34. Reicher, G.M.; Snyder, C.R.R.; Richards, I.T. Familiarity of background characters in visual scanning. J. Exp. Psychol. 1976, 2, 522–530. [Google Scholar] [CrossRef]
  35. Yiend, J. The effects of emotion on attention: A review of attentional processing of emotional information. Cogn. Emot. 2010, 24, 3–47. [Google Scholar] [CrossRef]
  36. Van Rooijen, R.; Ploeger, A.; Kret, M.E. The dot-probe task to measure emotional attention: A suitable measure in comparative studies? Psychon. Bull. Rev. 2017, 24, 1686–1717. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Winters, S.; Dubuc, C.; Higham, J.P. Perspectives: The looking time experimental paradigm in studies of animal visual perception and cognition. Ethology 2015, 121, 625–640. [Google Scholar] [CrossRef]
  38. Brilot, B.O.; Normandale, C.L.; Parkin, A.; Bateson, M. Can we use starlings’ aversion to eyespots as the basis for a novel ‘cognitive bias’ task? Appl. Anim. Behav. Sci. 2009, 118, 182–190. [Google Scholar] [CrossRef] [Green Version]
  39. Verbeek, E.; Ferguson, D.; Lee, C. Are hungry sheep more pessimistic? The effects of food restriction on cognitive bias and the involvement of ghrelin in its regulation. Physiol. Behav. 2014, 123, 67–75. [Google Scholar] [CrossRef] [PubMed]
  40. Vögeli, S.; Lutz, J.; Wolf, M.; Wechsler, B.; Gygax, L. Valence of physical stimuli, not housing conditions, affects behaviour and frontal cortical brain activity in sheep. Behav. Brain Res. 2015, 267, 144–155. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Lee, C.; Verbeek, E.; Doyle, R.; Bateson, M. Attention bias to threat indicates anxiety differences in sheep. Biol. Lett. 2016, 12, 20150977. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Monk, J.E.; Doyle, R.E.; Colditz, I.G.; Belson, S.; Cronin, G.M.; Lee, C. Towards a more practical attention bias test to assess affective state in sheep. PLoS ONE 2018, 13, e0190404. [Google Scholar] [CrossRef] [PubMed]
  43. Lee, C.; Cafe, L.M.; Robinson, S.; Doyle, R.E.; Lea, J.; Small, A.H.; Colditz, I.G. Anxiety influences attention bias but not flight speed and crush score in beef cattle. Appl. Anim. Behav. Sci. 2018, 205, 210–215. [Google Scholar] [CrossRef]
  44. Allritz, M.; Call, J.; Borkenau, P. How chimpanzees (Pan troglodytes) perform in a modified emotional Stroop task. Anim. Cogn. 2016, 19, 435–449. [Google Scholar] [CrossRef] [PubMed]
  45. Cussen, V.A.; Mench, J.A. Personality predicts cognitive bias in captive psittacines, Amazona Amazonica. Anim. Behav. 2014, 89, 123–130. [Google Scholar] [CrossRef]
  46. Marzouki, Y.; Gullstrand, J.; Goujon, A.; Fagot, J. Baboons’ response speed is biased by their moods. PLoS ONE 2014, 9, e102562. [Google Scholar] [CrossRef] [PubMed]
  47. Parker, R.M.; Paul, E.S.; Burman, O.H.; Browne, W.J.; Mendl, M. Housing conditions affect rat responses to two types of ambiguity in a reward–Reward discrimination cognitive bias task. Behav. Brain Res. 2014, 274, 73–83. [Google Scholar] [CrossRef] [PubMed]
  48. Fantz, R.L. Pattern vision in young infants. Psychol. Rec. 1958, 8, 43–47. [Google Scholar] [CrossRef]
  49. Desimone, R.; Duncan, J. Neural mechanisms of selective visual-attention. Annu. Rev. Neurosci. 1995, 18, 193–222. [Google Scholar] [CrossRef] [PubMed]
  50. Waitt, C.; Gerald, M.S.; Little, A.C.; Kraiselburd, E. Selective attention toward female secondary sexual color in male rhesus macaques. Am. J. Primatol. 2006, 68, 738–744. [Google Scholar] [CrossRef] [PubMed]
  51. Welp, T.; Rushen, J.; Kramer, D.L.; Festa-Bianchet, M.; De Passille, A.M.B. Vigilance as a measure of fear in dairy cattle. Appl. Anim. Behav. Sci. 2004, 87, 1–13. [Google Scholar] [CrossRef]
  52. McBride, S.D.; Morton, A.J. Visual attention and cognitive performance in sheep. Appl. Anim. Behav. Sci. 2018, 206, 52–58. [Google Scholar] [CrossRef]
  53. Raoult, C.M.; Gygax, L. Valence and intensity of video stimuli of dogs and conspecifics in sheep: Approach-avoidance, operant response, and attention. Animals 2018, 8, 121. [Google Scholar] [CrossRef] [PubMed]
  54. Somppi, S.; Törnqvist, H.; Kujala, M.V.; Hänninen, L.; Krause, C.M.; Vainio, O. Dogs evaluate threatening facial expressions by their biological validity-evidence from gazing patterns. PLoS ONE 2016, 11, e0143047. [Google Scholar] [CrossRef] [PubMed]
  55. Siniscalchi, M.; Dimatteo, S.; Pepe, A.M.; Sasso, R.; Quaranta, A. Visual lateralization in wild striped dolphins (Stenella coeruleoalba) in response to stimuli with different degrees of familiarity. PLoS ONE 2012, 7, e30001. [Google Scholar] [CrossRef] [PubMed]
  56. Dubuc, C.; Allen, W.L.; Cascio, J.; Lee, D.S.; Maestripieri, D.; Petersdorf, M.; Winters, S.; Higham, J.P. Who cares? Experimental attention biases provide new insights into a mammalian sexual signal. Behav. Ecol. 2015, 27, 68–74. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Mandalaywala, T.M.; Parker, K.J.; Maestripieri, D. Early experience affects the strength of vigilance for threat in rhesus monkey infants. Psychol. Sci. 2014, 25, 1893–1902. [Google Scholar] [CrossRef] [PubMed]
  58. Mandalaywala, T.M.; Petrullo, L.A.; Parker, K.J.; Maestripieri, D.; Higham, J.P. Vigilance for threat accounts for inter-individual variation in physiological responses to adversity in rhesus macaques: A cognition×environment approach. Dev. Psychol. 2017, 59, 1031–1038. [Google Scholar] [CrossRef] [PubMed]
  59. Pfefferle, D.; Kazem, A.J.; Brockhausen, R.R.; Ruiz-Lambides, A.V.; Widdig, A. Monkeys spontaneously discriminate their unfamiliar paternal kin under natural conditions using facial cues. Curr. Biol. 2014, 24, 1806–1810. [Google Scholar] [CrossRef] [PubMed]
  60. Schell, A.; Rieck, K.; Schell, K.; Hammerschmidt, K.; Fischer, J. Adult but not juvenile Barbary macaques spontaneously recognize group members from pictures. Anim. Cogn. 2011, 14, 503–509. [Google Scholar] [CrossRef] [PubMed]
  61. Frohnwieser, A.; Pike, T.W.; Murray, J.C.; Wilkinson, A. Lateralized eye use towards video stimuli in bearded dragons. Anim. Behav. Cogn. 2017, 4, 340–348. [Google Scholar] [CrossRef]
  62. Bonati, B.; Csermely, D.; Sovrano, V.A. Looking at a predator with the left or right eye: Asymmetry of response in lizards. Laterality 2013, 18, 329–339. [Google Scholar] [CrossRef] [PubMed]
  63. Cisler, J.M.; Koster, E.H. Mechanisms of attentional biases towards threat in anxiety disorders: An integrative review. Clin. Psychol. Rev. 2010, 30, 203–216. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  64. Mogg, K.; McNamara, J.; Powys, M.; Rawlinson, H.; Seiffer, A.; Bradley, B.P. Selective attention to threat: A test of two cognitive models of anxiety. Cogn. Emot. 2000, 14, 375–399. [Google Scholar] [CrossRef]
  65. Wilson, E.; MacLeod, C. Contrasting two accounts of anxiety-linked attentional bias: Selective attention to varying levels of stimulus threat intensity. J. Abnorm. Psychol. 2003, 112, 212–218. [Google Scholar] [CrossRef] [PubMed]
  66. Parr, L.A.; Waller, B.M.; Burrows, A.M.; Gothard, K.M.; Vick, S.J. Brief communication: MaqFACS: A muscle-based facial movement coding system for the rhesus macaque. Am. J. Phys. Anthropol. 2010, 143, 625–630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  67. Bogartz, R.S.; Shinskey, J.L.; Speaker, C.J. Interpreting infant looking: The event set × event set design. Dev. Psychol. 1997, 33, 408. [Google Scholar] [CrossRef] [PubMed]
  68. Hughes, K.D.; Higham, J.P.; Allen, W.L.; Elliot, A.J.; Hayden, B.Y. Extraneous color affects female macaques’ gaze preference for photographs of male conspecifics. Evol. Hum. Behav. 2015, 36, 25–31. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  69. Machado, C.J.; Nelson, E.E. Eye-tracking with nonhuman primates is now more accessible than ever before. Am. J. Primatol. 2011, 73, 562–569. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  70. Kimmel, D.L.; Mammo, D.; Newsome, W.T. Tracking the eye non-invasively: Simultaneous comparison of the scleral search coil and optical tracking techniques in the macaque monkey. Front. Behav. Neurosci. 2012, 6, 49. [Google Scholar] [CrossRef] [PubMed]
  71. Armstrong, T.; Olatunji, B.O. Eye tracking of attention in the affective disorders: A meta-analytic review and synthesis. Clin. Psychol. Rev. 2012, 32, 704–723. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  72. Hansen, D.W.; Ji, Q. In the eye of the beholder: A survey of models for eyes and gaze. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 478–500. [Google Scholar] [CrossRef] [PubMed]
  73. Gredeback, G.; Johnson, S.; von Hofsten, C. Eye-tracking in infancy research. Dev. Neuropsychol. 2009, 35, 1–19. [Google Scholar] [CrossRef] [PubMed]
  74. Mele, M.L.; Federici, S. Gaze and eye-tracking solutions for psychological research. Cogn. Process. 2012, 13, 261–265. [Google Scholar] [CrossRef] [PubMed]
  75. Kotani, M.; Shimono, K.; Yoneyama, T.; Nakako, T.; Matsumoto, K.; Ogi, Y.; Konoike, N.; Nakamura, K.; Ikeda, K. An eye tracking system for monitoring face scanning patterns reveals the enhancing effect of oxytocin on eye contact in common marmosets. Psychoneuroendocrinology 2017, 83, 42–48. [Google Scholar] [CrossRef] [PubMed]
  76. Yorzinski, J.L.; Patricelli, G.L.; Babcock, J.S.; Pearson, J.M.; Platt, M.L. Through their eyes: Selective attention in peahens during courtship. J. Exp. Biol. 2013, 216, 3035–3046. [Google Scholar] [CrossRef] [PubMed]
  77. Ben-Simon, A.; Ben-Shahar, O.; Segev, R. Measuring and tracking eye movements of a behaving archer fish by real-time stereo vision. J. Neurosci. Methods 2009, 184, 235–243. [Google Scholar] [CrossRef] [PubMed]
  78. Shepherd, S.V.; Platt, M.L. Noninvasive telemetric gaze tracking in freely moving socially housed prosimian primates. Methods 2006, 38, 185–194. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  79. Shepherd, S.V.; Platt, M.L. Spontaneous social orienting and gaze following in ringtailed lemurs (Lemur catta). Anim. Cogn. 2008, 11, 13–20. [Google Scholar] [CrossRef] [PubMed]
  80. Bovet, D.; Vauclair, J. Picture recognition in animals and humans. Behav. Brain Res. 2000, 109, 143–165. [Google Scholar] [CrossRef]
  81. Mustafar, F.; De Luna, P.; Rainer, G. Enhanced visual exploration for real objects compared to pictures during free viewing in the macaque monkey. Behav. Proc. 2015, 118, 8–20. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  82. Waitt, C.; Buchanan-Smith, H.M. Perceptual considerations in the use of colored photographic and video stimuli to study nonhuman primate behavior. Am. J. Primatol. 2006, 68, 1054–1067. [Google Scholar] [CrossRef] [PubMed]
  83. D’Eath, R.B. Can video images imitate real stimuli in animal behaviour experiments? Biol. Rev. 1998, 73, 267–292. [Google Scholar] [CrossRef]
  84. Oliveira, R.F.; Rosenthal, G.G.; Schlupp, I.; McGregor, K.; Cuthill, I.C.; Endler, J.A.; Fleishman, L.J.; Zeil, J.; Barata, E.; Burford, F.; et al. Considerations on the use of video playbacks as visual stimuli: The Lisbon workshop consensus. Acta Ethol. 2000, 3, 61–65. [Google Scholar] [CrossRef]
  85. Nelson, X.J.; Fijn, N. The use of visual media as a tool for investigating animal behaviour. Anim. Behav. 2013, 85, 525–536. [Google Scholar] [CrossRef]
  86. Fleishman, L.J.; McClintock, W.J.; D’Eath, R.B.; Brainards, D.H.; Endler, J.A. Colour perception and the use of video playback experiments in animal behaviour. Anim. Behav. 1998, 56, 1035–1040. [Google Scholar] [CrossRef] [PubMed]
  87. Woo, K.L.; Rieucau, G. From dummies to animations: A review of computer-animated stimuli used in animal behavior studies. Behav. Ecol. Sociobiol. 2011, 65, 1671–1685. [Google Scholar] [CrossRef]
  88. Chouinard-Thuly, L.; Gierszewski, S.; Rosenthal, G.G.; Reader, S.M.; Rieucau, G.; Woo, K.L.; Gerlai, R.; Tedore, C.; Ingley, S.J.; Stowers, J.R.; et al. Technical and conceptual considerations for using animated stimuli in studies of animal behavior. Curr. Zool. 2017, 63, 5–19. [Google Scholar] [CrossRef] [PubMed]
  89. Tinbergen, N.; Perdeck, A.C. On the stimulus situation releasing the begging response in the newly hatched herring gull chick (Larus argentatus argentatus Pont.). Behaviour 1951, 3, 1–39. [Google Scholar] [CrossRef]
  90. Stroop, J.R. Studies of interference in serial verbal reactions. J. Exp. Psychol. 1935, 18, 643–662. [Google Scholar] [CrossRef]
  91. MacLeod, C.M. Half a century of research on the stroop effect: An integrative review. Psychol. Bull. 1991, 109, 163–203. [Google Scholar] [CrossRef] [PubMed]
  92. Williams, J.M.G.; Mathews, A.; MacLeod, C. The emotional stroop task and psychopathology. Psychol. Bull. 1996, 120, 3–24. [Google Scholar] [CrossRef] [PubMed]
  93. Mauer, N.; Borkenau, P. Temperament and early information processing: Temperament-related attentional bias in emotional stroop tasks. Personal. Individ. Differ. 2007, 43, 1063–1073. [Google Scholar] [CrossRef]
  94. Mathews, A.; MacLeod, C. Selective processing of threat cues in anxiety states. Behav. Res. Ther. 1985, 23, 563–569. [Google Scholar] [CrossRef]
  95. Mogg, K.; Mathews, A.; Weinman, J. Selective processing of threat cues in anxiety states: A replication. Behav. Res. Ther. 1989, 27, 317–323. [Google Scholar] [CrossRef]
  96. Reynolds, M.G.; Langerak, R.M. Emotional stroop dilution: The boundary conditions of attentional capture by threat words. Acta. Psychol. 2015, 159, 108–115. [Google Scholar] [CrossRef] [PubMed]
  97. Algom, D.; Chajut, E.; Lev, S. A rational look at the emotional stroop phenomenon: A generic slowdown, not a stroop effect. J. Exp. Psychol. Gen. 2004, 133, 323–338. [Google Scholar] [CrossRef] [PubMed]
  98. Bethell, E.J.; Holmes, A.; MacLarnon, A.; Semple, S. Emotion evaluation and response slowing in a non-human primate: New directions for cognitive bias measures of animal emotion? Behav. Sci. 2016, 6, 2. [Google Scholar] [CrossRef] [PubMed]
  99. Landman, R.; Sharma, J.; Sur, M.; Desimone, R. Effect of distracting faces on visual selective attention in the monkey. Proc. Natl. Acad. Sci. USA 2014, 111, 18037–18042. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  100. Bellegarde, L.G.; Erhard, H.W.; Weiss, A.; Boissy, A.; Haskell, M.J. Valence of Facial cues influences sheep learning in a Visual Discrimination Task. Front. Vet. Sci. 2017, 4, 188. [Google Scholar] [CrossRef] [PubMed]
  101. Maier, S.F.; Seligman, M.E. Learned helplessness at fifty: Insights from neuroscience. Psychol. Rev. 2016, 123, 349–367. [Google Scholar] [CrossRef] [PubMed]
  102. Lee, R.K.; Maier, S.F. Inescapable shock and attention to internal versus external cues in a water discrimination escape task. J. Exp. Psychol. Anim. Behav. Proc. 1988, 14, 302–310. [Google Scholar] [CrossRef]
  103. Minor, T.R.; Jackson, R.L.; Maier, S.F. Effects of task-irrelevant cues and reinforcement delay on choice-escape learning following inescapable shock: Evidence for a deficit in selective attention. J. Exp. Psychol. Anim. Behav. Proc. 1984, 10, 543–556. [Google Scholar] [CrossRef]
  104. Jackson, R.L.; Alexander, J.H.; Maier, S.F. Learned helplessness, inactivity, and associative deficits: Effects of inescapable shock on response choice escape learning. J. Exp. Psychol. Anim. Behav. Proc. 1980, 6, 1–20. [Google Scholar] [CrossRef]
  105. Rodd, Z.A.; Rosellini, R.A.; Stock, H.S.; Gallup, G.G., Jr. Learned helplessness in chickens (Gallus gallus): Evidence for attentional bias. Learn Motiv. 1997, 28, 43–55. [Google Scholar] [CrossRef]
  106. MacLeod, C.; Mathews, A.; Tata, P. Attentional bias in emotional disorders. J. Abnorm. Psychol. 1986, 95, 15–20. [Google Scholar] [CrossRef] [PubMed]
  107. Winer, E.S.; Salem, T. Reward devaluation: Dot-probe meta-analytic evidence of avoidance of positive information in depressed persons. Psychol. Bull. 2016, 142, 18–78. [Google Scholar] [CrossRef] [PubMed]
  108. Mogg, K.; Mathews, A.; Eysenck, M. Attentional bias to threat in clinical anxiety states. Cogn. Emot. 1992, 6, 149–159. [Google Scholar] [CrossRef]
  109. Bradley, B.P.; Mogg, K.; Millar, N.H. Covert and overt orienting of attention to emotional faces in anxiety. Cogn. Emot. 2000, 14, 789–808. [Google Scholar] [CrossRef]
  110. Mathews, A.; Fox, E.; Yiend, J.; Calder, A. The face of fear: Effects of eye gaze and emotion on visual attention. Vis Cogn 2003, 10, 823–835. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  111. Bradley, B.P.; Mogg, K.; Falla, S.J.; Hamilton, L.R. Attentional bias for threatening facial expressions in anxiety: Manipulation of stimulus duration. Cogn. Emot. 1998, 12, 737–753. [Google Scholar] [CrossRef]
  112. Peckham, A.D.; McHugh, R.K.; Otto, M.W. A meta-analysis of the magnitude of biased attention in depression. Depress. Anxiety 2010, 27, 1135–1142. [Google Scholar] [CrossRef] [PubMed]
  113. Bar-Haim, Y.; Holoshitz, Y.; Eldar, S.; Frenkel, T.I.; Muller, D.; Charney, D.S.; Pine, D.S.; Fox, N.A.; Wald, I. Life-threatening danger and suppression of attention bias to threat. Am. J. Psychiatry 2010, 167, 694–698. [Google Scholar] [CrossRef] [PubMed]
  114. Sipos, M.L.; Bar-Haim, Y.; Abend, R.; Adler, A.B.; Bliese, D. Postdeployment threat-related attention bias interacts with combat exposure to account for PTSD and anxiety symptoms in soldiers. Depress. Anxiety 2014, 31, 124–129. [Google Scholar] [CrossRef] [PubMed]
  115. Kret, M.E.; Jaasma, L.; Bionda, T.; Wijnen, J.G. Bonobos (Pan paniscus) show an attentional bias toward conspecifics’ emotions. Proc. Natl. Acad. Sci. USA 2016, 113, 3761–3766. [Google Scholar] [CrossRef] [PubMed]
  116. Clay, Z.; de Waal, F.B. Bonobos respond to distress in others: Consolation across the age spectrum. PLoS ONE 2013, 8, e55206. [Google Scholar] [CrossRef] [PubMed]
  117. Pool, E.; Brosch, T.; Delplanque, S.; Sander, D. Attentional bias for positive emotional stimuli: A meta-analytic investigation. Psychol. Bull. 2016, 142, 79–106. [Google Scholar] [CrossRef] [PubMed]
  118. Palagi, E.; Mancini, G. Playing with the face: Playful facial “chattering” and signal modulation in a monkey species (Theropithecus gelada). J. Comp. Psychol. 2011, 125, 11–21. [Google Scholar] [CrossRef] [PubMed]
  119. Palagi, E. Sharing the motivation to play: The use of signals in adult bonobos. Anim. Behav. 2008, 75, 887–896. [Google Scholar] [CrossRef]
  120. Lacreuse, A.; Schatz, K.; Strazzullo, S.; King, H.M.; Ready, R. Attentional biases and memory for emotional stimuli in men and male rhesus monkeys. Anim. Cogn. 2013, 16, 861–871. [Google Scholar] [CrossRef] [PubMed]
  121. Parr, L.A.; Modi, M.; Siebert, E.; Young, L.J. Intranasal oxytocin selectively attenuates rhesus monkeys’ attention to negative facial expressions. Psychoneuroendocrinology 2013, 38, 1748–1756. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  122. King, H.M.; Kurdziel, L.B.; Meyer, J.S.; Lacreuse, A. Effects of testosterone on attention and memory for emotional stimuli in male rhesus monkeys. Psychoneuroendocrinology 2012, 37, 396–409. [Google Scholar] [CrossRef] [PubMed]
  123. Koda, H.; Sato, A.; Kato, A. Is attentional prioritisation of infant faces unique in humans? Comparative demonstrations by modified dot-probe task in monkeys. Behav. Proc. 2013, 98, 31–36. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  124. Posner, M.I. Orienting of attention. Q. J. Exp. Psychol. 1980, 32, 3–25. [Google Scholar] [CrossRef] [PubMed]
  125. Yiend, J.; Mathews, A. Anxiety and attention to threatening pictures. Q. J. Exp. Psychol. (Hove) 2001, 54, 665–681. [Google Scholar] [CrossRef] [PubMed]
  126. Stormark, K.M.; Nordby, H.; Hugdahl, K. Attentional shifts to emotionally charged cues: Behavioural and ERP data. Cogn. Emot. 1995, 9, 507–523. [Google Scholar] [CrossRef]
  127. Cook, E.P.; Maunsell, J.H. Dynamics of neuronal responses in macaque MT and VIP during motion detection. Nat. Neurosci. 2002, 5, 985–994. [Google Scholar] [CrossRef] [PubMed]
  128. Marote, C.F.O.; Xavier, G.F. Endogenous-like orienting of visual attention in rats. Anim. Cogn. 2011, 14, 535–544. [Google Scholar] [CrossRef] [PubMed]
  129. Sridharan, D.; Ramamurthy, D.L.; Schwarz, J.S.; Knudsen, E.I. Visuospatial selective attention in chickens. Proc. Natl. Acad. Sci. USA 2014, 111, e2056–e2065. [Google Scholar] [CrossRef] [PubMed]
  130. Eckstein, M.P.; Mack, S.C.; Liston, D.B.; Bogush, L.; Menzel, R.; Krauzlis, R.J. Rethinking human visual attention: Spatial cueing effects and optimality of decisions by honeybees, monkeys and humans. Vision Res. 2013, 85, 5–19. [Google Scholar] [CrossRef] [PubMed]
  131. Saban, W.; Sekely, L.; Klein, R.M.; Gabay, S. Endogenous orienting in the archer fish. Proc. Natl. Acad. Sci. USA 2017, 114, 7577–7581. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  132. Wieser, M.J.; Hambach, A.; Weymar, M. Neurophysiological correlates of attentional bias for emotional faces in socially anxious individuals-evidence from a visual search task and N2pc. Biol. Psychol. 2018, 132, 192–201. [Google Scholar] [CrossRef] [PubMed]
  133. Troisi, A. Displacement activities as a behavioral measure of stress in nonhuman primates and human subjects. Stress 2002, 5, 47–54. [Google Scholar] [CrossRef] [PubMed]
  134. Castles, D.L.; Whiten, A. Post-conflict behaviour of wild olive baboons. II. Stress and self-directed behaviour. Ethology 1998, 104, 148–160. [Google Scholar] [CrossRef]
  135. Tomonaga, M.; Imura, T. Efficient search for a face by chimpanzees (Pan troglodytes). Sci. Rep. 2015, 5, 11437. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  136. Tomonaga, M.; Imura, T. Visual search for human gaze direction by a chimpanzee (Pan Troglodytes). PLoS ONE 2010, 5, e9131. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  137. Kawai, N.; Kubo, K.; Masataka, N.; Hayakawa, S. Conserved evolutionary history for quick detection of threatening faces. Anim. Cogn. 2016, 19, 655–660. [Google Scholar] [CrossRef] [PubMed]
  138. Nakata, R.; Eifuku, S.; Tamura, R. Crucial information for efficient face searching by humans and Japanese macaques. Anim. Cogn. 2018, 21, 155–164. [Google Scholar] [CrossRef] [PubMed]
  139. Shibasaki, M.; Kawai, N. Rapid detection of snakes by Japanese monkeys (Macaca fuscata): An evolutionarily predisposed visual system. J. Comp. Psychol. 2009, 123, 131–135. [Google Scholar] [CrossRef] [PubMed]
  140. Kawai, N.; Koda, H. Japanese monkeys (Macaca fuscata) quickly detect snakes but not spiders: Evolutionary origins of fear-relevant animals. J. Comp. Psychol. 2016, 130, 299–303. [Google Scholar] [CrossRef] [PubMed]
  141. Simpson, E.A.; Jakobsen, K.V.; Damon, F.; Suomi, S.J.; Ferrari, F.; Paukner, A. Face Detection and the Development of Own-Species Bias in Infant Macaques. Child Dev. 2017, 88, 103–113. [Google Scholar] [CrossRef] [PubMed]
  142. Lev-Ari, T.; Gutfreund, Y. Interactions between top-down and bottom-up attention in barn owls (Tyto alba). Anim. Cogn. 2018, 21, 197–205. [Google Scholar] [CrossRef] [PubMed]
  143. Orlowski, J.; Ben-Shahar, O.; Wagner, H. Visual search in barn owls: Task difficulty and saccadic behavior. J. Vis. 2018, 18, 4. [Google Scholar] [CrossRef] [PubMed]
  144. Proulx, M.J.; Parker, M.O.; Tahir, Y.; Brennan, C.H. Parallel mechanisms for visual search in zebrafish. PLoS ONE 2014, 9, e111540. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  145. Nityananda, V.; Pattrick, J.G. Bumblebee visual search for multiple learned target types. J. Exp. Biol. 2013, 216, 4154–4160. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  146. Perry, C.J.; Baciadonna, L.; Chittka, L. Unexpected rewards induce dopamine-dependent positive emotion–like state changes in bumblebees. Science 2016, 353, 1529–1531. [Google Scholar] [CrossRef] [PubMed]
  147. Lundqvist, D.; Bruce, N.; Öhman, A. Finding an emotional face in a crowd: Emotional and perceptual stimulus factors influence visual search efficiency. Cogn. Emot. 2015, 29, 621–633. [Google Scholar] [CrossRef] [PubMed]
  148. Lee, T.H.; Sakaki, M.; Cheng, R.; Velasco, R.; Mather, M. Emotional arousal amplifies the effects of biased competition in the brain. Soc. Cogn. Affect Neurosci. 2014, 9, 2067–2077. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  149. Mather, M.; Sutherland, M.R. Arousal-biased competition in perception and memory. Perspect. Psychol. Sci. 2011, 6, 114–133. [Google Scholar] [CrossRef] [PubMed]
  150. Lundqvist, D.; Juth, P.; Öhman, A. Using facial emotional stimuli in visual search experiments: The arousal factor explains contradictory results. Cogn. Emot. 2013, 12, 1–13. [Google Scholar] [CrossRef] [PubMed]
  151. Sutherland, M.R.; Mather, M. Arousal (but not valence) amplifies the impact of salience. Cogn. Emot. 2017, 32, 616–622. [Google Scholar] [CrossRef] [PubMed]
  152. Zsido, A.N.; Bernath, L.; Labadi, B.; Deak, A. Count on arousal: Introducing a new method for investigating the effects of emotional valence and arousal on visual search performance. Psychol. Res. 2018, 1–14. [Google Scholar] [CrossRef] [PubMed]
  153. Wilkinson, R.T. Interaction of noise with knowledge of results and sleep deprivation. J. Exp. Psychol. 1963, 66, 135–153. [Google Scholar] [CrossRef]
  154. Carli, M.; Robbins, T.W.; Evenden, J.L.; Everitt, B.J. Effects of lesions to ascending noradrenergic neurones on performance of a 5-choice serial reaction task in rats; implications for theories of dorsal noradrenergic bundle function based on selective attention and arousal. Behav. Brain Res. 1983, 9, 361–380. [Google Scholar] [CrossRef]
  155. Fizet, J.; Cassel, J.C.; Kelche, C.; Meunier, H. A review of the 5-Choice Serial Reaction Time (5-CSRT) task in different vertebrate models. Neurosci. Biobehav. Rev. 2016, 71, 135–153. [Google Scholar] [CrossRef] [PubMed]
  156. Boyette-Davis, J.A.; Thompson, C.D.; Fuchs, N. Alterations in attentional mechanisms in response to acute inflammatory pain and morphine administration. Neuroscience 2008, 151, 558–563. [Google Scholar] [CrossRef] [PubMed]
  157. Pais-Vieira, M.; Lima, D.; Galhardo, V. Sustained attention deficits in rats with chronic inflammatory pain. Neurosci. Lett. 2009, 463, 98–102. [Google Scholar] [CrossRef] [PubMed]
  158. Freitas, K.C.; Hillhouse, T.M.; Leitl, M.D.; Negus, S.S. Effects of Acute and Sustained Pain Manipulations on Performance in a Visual-Signal Detection Task of Attention in Rats. Drug. Dev. Res. 2015, 76, 194–203. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  159. Moriarty, O.; McGuire, B.E.; Finn, D.P. The effect of pain on cognitive function: A review of clinical and preclinical research. Prog. Neurobiol. 2011, 93, 385–404. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  160. Salmeto, A.L.; Hymel, K.A.; Carpenter, E.C.; Brilot, B.O.; Bateson, M.; Sufka, K.J. Cognitive bias in the chick anxiety–depression model. Brain Res. 2011, 1373, 124–130. [Google Scholar] [CrossRef] [PubMed]
  161. Hymel, K.A.; Sufka, K.J. Pharmacological reversal of cognitive bias in the chick anxiety-depression model. Neuropharmacology 2012, 62, 161–166. [Google Scholar] [CrossRef] [PubMed]
  162. Bellegarde, L.G.; Haskell, M.J.; Duvaux-Ponter, C.; Weiss, A.; Boissy, A.; Erhard, H.W. Face-based perception of emotions in dairy goats. Appl. Anim. Behav. Sci. 2017, 193, 51–59. [Google Scholar] [CrossRef]
  163. Kress, L.; Bristle, M.; Aue, T. Seeing through rose-colored glasses: How optimistic expectancies guide visual attention. PLoS ONE 2018, 13, e0193311. [Google Scholar] [CrossRef] [PubMed]
  164. Everaert, J.; Tierens, M.; Uzieblo, K.; Koster, E.H. The indirect effect of attention bias on memory via interpretation bias: Evidence for the combined cognitive bias hypothesis in subclinical depression. Cogn. Emot. 2013, 27, 1450–1459. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  165. Forkman, B.; Boissy, A.; Meunier-Salaün, M.C.; Canali, E.; Jones, R.B. A critical review of fear tests used on cattle, pigs, sheep, poultry and horses. Physiol. Behav. 2007, 92, 340–374. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  166. Grillon, C.; Baas, J. A review of the modulation of the startle reflex by affective states and its application in psychiatry. Clin. Neurophysiol. 2003, 114, 1557–1579. [Google Scholar] [CrossRef]
  167. Lanier, J.L.; Grandin, T.; Green, R.D.; Avery, D.; McGee, K. The relationship between reaction to sudden, intermittent movements and sounds and temperament. J. Anim. Sci. 2000, 78, 1467–1474. [Google Scholar] [CrossRef] [PubMed]
  168. Schmitz, A.; Merikangas, K.; Swendsen, H.; Cui, L.; Heaton, L.; Grillon, C. Measuring anxious responses to predictable and unpredictable threat in children and adolescents. J. Exp. Child. Psychol. 2011, 110, 159–170. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  169. Bakker, M.J.; Tijssen, M.A.; van der Meer, J.N.; Koelman, J.H.; Boer, F. Increased whole-body auditory startle reflex and autonomic reactivity in children with anxiety disorders. J. Psychiatry Neurosci. 2009, 34, 314–322. [Google Scholar] [PubMed]
  170. Alfvén, G.; Grillner, S.; Andersson, E. Children with chronic stress-induced recurrent muscle pain have enhanced startle reaction. Eur. J. Pain 2017, 21, 1561–1570. [Google Scholar] [CrossRef] [PubMed]
  171. Destrez, A.; Deiss, V.; Leterrier, C.; Boivin, X.; Boissy, A. Long-term exposure to unpredictable and uncontrollable aversive events alters fearfulness in sheep. Animal 2013, 7, 476–484. [Google Scholar] [CrossRef] [PubMed]
  172. Brown, J.S.; Kalish, H.I.; Farber, I.E. Conditioned fear as revealed by magnitude of startle response to an auditory stimulus. J. Exp. Psychol. 1951, 41, 317–328. [Google Scholar] [CrossRef] [PubMed]
  173. Davis, M.; Antoniadis, E.A.; Amaral, D.G.; Winslow, J.T. Acoustic startle reflex in rhesus monkeys: A review. Rev. Neurosci. 2008, 19, 171–186. [Google Scholar] [CrossRef] [PubMed]
  174. Beauchamp, G. What can vigilance tell us about fear? Anim. Sent. 2017, 2, 1–31. [Google Scholar]
  175. Destrez, A.; Deiss, V.; Belzung, C.; Lee, C.; Boissy, A. Does reduction of fearfulness tend to reduce pessimistic-like judgment in lambs? Appl. Anim. Behav. Sci. 2012, 139, 233–241. [Google Scholar] [CrossRef]
  176. Elgar, M.A. Predator vigilance and group size in mammals and birds: A critical review of the empirical evidence. Biol. Rev. Camb. Philos. Soc. 1989, 64, 13–33. [Google Scholar] [CrossRef] [PubMed]
  177. Randler, C. Disturbances by dog barking increase vigilance in coots Fulica atra. Eur. J. Wildlife Res. 2006, 52, 265–270. [Google Scholar] [CrossRef]
  178. Creel, S.; Schuette, P.; Christianson, D. Effects of predation risk on group size, vigilance, and foraging behavior in an African ungulate community. Behav. Ecol. 2014, 25, 773–784. [Google Scholar] [CrossRef] [Green Version]
  179. Wang, Z.; Li, Z.; Beauchamp, G.; Jiang, Z. Flock size and human disturbance affect vigilance of endangered red-crowned cranes (Grus japonensis). Biol. Cons. 2011, 144, 101–105. [Google Scholar] [CrossRef]
  180. Busia, L.; Schaffner, C.M.; Aureli, F. Watch out or relax: Conspecifics affect vigilance in wild spider monkeys (Ateles geoffroyi). Behaviour 2016, 153, 107–124. [Google Scholar] [CrossRef]
  181. Nielsen, B.L. Making sense of it all: The importance of taking into account the sensory abilities of animals in their housing and management. Appl. Anim. Behav. Sci. 2018, 205, 175–180. [Google Scholar] [CrossRef]
  182. d’Ettorre, P.; Carere, C.; Demora, L.; Le Quinquis, P.; Signorotti, L.; Bovet, D. Individual differences in exploratory activity relate to cognitive judgement bias in carpenter ants. Behav. Proc. 2017, 134, 63–69. [Google Scholar] [CrossRef] [PubMed]
  183. Brydges, N.M.; Hall, L. A shortened protocol for assessing cognitive bias in rats. J. Neurosci. Methods 2017, 286, 1–5. [Google Scholar] [CrossRef] [PubMed]
  184. Albuquerque, N.; Guo, K.; Wilkinson, A.; Savalli, C.; Otta, E.; Mills, D. Dogs recognize dog and human emotions. Biol. Lett. 2016, 12, 20150883. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  185. Proctor, H.S.; Carder, G. Can ear postures reliably measure the positive emotional state of cows? Appl. Anim. Behav. Sci. 2014, 161, 20–27. [Google Scholar] [CrossRef]
  186. De Oliveira, D.; Keeling, L.J. Routine activities and emotion in the life of dairy cows: Integrating body language into an affective state framework. PLoS ONE 2018, 13, e0195674. [Google Scholar] [CrossRef] [PubMed]
  187. Vögeli, S.; Wechsler, B.; Gygax, L. Welfare by the ear: Comparing relative durations and frequencies of ear postures by using an automated tracking system in sheep. Anim. Welf. 2014, 23, 267–274. [Google Scholar] [CrossRef]
  188. Reefmann, N.; Kaszàs, F.B.; Wechsler, B.; Gygax, L. Ear and tail postures as indicators of emotional valence in sheep. Appl. Anim. Behav. Sci. 2009, 118, 199–207. [Google Scholar] [CrossRef]
  189. Boissy, A.; Aubert, A.; Désiré, L.; Greiveldinger, L.; Delval, E.; Veissier, I. Cognitive sciences to relate ear postures to emotions in sheep. Anim. Welf. 2011, 20, 47–56. [Google Scholar]
  190. Briefer, E.F.; Tettamanti, F.; McElligott, A.G. Emotions in goats: Mapping physiological, behavioural and vocal profiles. Anim. Behav. 2015, 99, 131–143. [Google Scholar] [CrossRef]
  191. Reimert, I.; Bolhuis, J.E.; Kemp, B.; Rodenburg, T.B. Indicators of positive and negative emotions and emotional contagion in pigs. Physiol. Behav. 2013, 109, 42–50. [Google Scholar] [CrossRef] [PubMed]
  192. Racca, A.; Guo, K.; Meints, K.; Mills, D.S. Reading faces: Differential lateral gaze bias in processing canine and human facial expressions in dogs and 4-year-old children. PLoS ONE 2012, 7, e36076. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  193. Lecorps, B.; Féron, C. Correlates between ear postures and emotional reactivity in a wild type mouse species. Behav. Proc. 2015, 120, 25–29. [Google Scholar] [CrossRef] [PubMed]
  194. Nielsen, B.L.; Jezierski, T.; Bolhuis, J.E.; Amo, L.; Rosell, F.; Oostindjer, M.; Christensen, J.W.; McKeegan, D.; Wells, D.L.; Hepper, P. Olfaction: An overlooked sensory modality in applied ethology and animal welfare. Front. Vet. Sci. 2015, 2, 69. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  195. Gazit, I.; Terkel, J. Domination of olfaction over vision in explosives detection by dogs. Appl. Anim. Behav. Sci. 2003, 82, 65–73. [Google Scholar] [CrossRef] [Green Version]
  196. Kroon, J.A.; Carobrez, A.P. Olfactory fear conditioning paradigm in rats: Effects of midazolam, propranolol or scopolamine. Neurobiol. Learn Mem. 2009, 91, 32–40. [Google Scholar] [CrossRef] [PubMed]
  197. Jones, R.B.; Roper, T.J. Olfaction in the domestic fowl: A critical review. Physiol. Behav. 1997, 62, 1009–1018. [Google Scholar] [CrossRef]
  198. Hansson, B.S. (Ed.) Insect Olfaction; Springer Science & Business Media: Berlin, Germany, 1999. [Google Scholar]
  199. Cárdenas, M.; Jiroš, P.; Pekár, S. Selective olfactory attention of a specialised predator to intraspecific chemical signals of its prey. Naturwissenschaften 2012, 99, 597–605. [Google Scholar] [CrossRef] [PubMed]
  200. Olatunji, B.O.; Armstrong, T.; McHugo, M.; Zald, D.H. Heightened attentional capture by threat in veterans with PTSD. J. Abnorm. Psychol. 2013, 122, 397–405. [Google Scholar] [CrossRef] [PubMed]
  201. Akram, U.; Beattie, L.; Ypsilanti, A.; Reidy, J.; Robson, A.; Chapman, A.J.; Barclay, N.L. Sleep-related attentional bias for tired faces in insomnia: Evidence from a dot-probe paradigm. Behav. Res. Ther. 2018, 103, 18–23. [Google Scholar] [CrossRef] [PubMed]
  202. Lubman, D.I.; Peters, L.A.; Mogg, K.; Bradley, B.P.; Deakin, J.F.W. Attentional bias for drug cues in opiate dependence. Psychol. Med. 2000, 30, 169–175. [Google Scholar] [CrossRef] [PubMed]
  203. Ehrman, R.N.; Robbins, S.J.; Bromwell, M.A.; Lankford, M.E.; Monterosso, J.R.; O’Brien, C.P. Comparing attentional bias to smoking cues in current smokers, former smokers, and non-smokers using a dot-probe task. Drug Alcohol Depend. 2002, 67, 185–191. [Google Scholar] [CrossRef]
  204. Townshend, J.; Duka, T. Attentional bias associated with alcohol cues: Differences between heavy and occasional social drinkers. Psychopharmacology 2001, 157, 67–74. [Google Scholar] [CrossRef] [PubMed]
  205. Castellanos, E.H.; Charboneau, E.; Dietrich, M.S.; Park, S.; Bradley, B.P.; Mogg, K.; Cowan, R.L. Obese adults have visual attention bias for food cue images: Evidence for altered reward system function. Int. J. Obes. 2009, 33, 1063. [Google Scholar] [CrossRef] [PubMed]
  206. Tapper, K.; Pothos, E.M.; Lawrence, A.D. Feast your eyes: Hunger and trait reward drive predict attentional bias for food cues. Emotion 2010, 10, 949. [Google Scholar] [CrossRef] [PubMed]
  207. Davidson, G.R.; Giesbrecht, T.; Thomas, A.M.; Kirkham, T.C. Pre-and postprandial variation in implicit attention to food images reflects appetite and sensory-specific satiety. Appetite 2018, 125, 24–31. [Google Scholar] [CrossRef] [PubMed]
  208. Dawkins, M.S. Behaviour as a tool in the assessment of animal welfare. Zoology 2003, 106, 383–387. [Google Scholar] [CrossRef] [PubMed]
  209. Kirkden, R.D.; Pajor, E.A. Using preference, motivation and aversion tests to ask scientific questions about animals’ feelings. Appl. Anim. Behav. Sci. 2006, 100, 29–47. [Google Scholar] [CrossRef]
  210. Fridrickson, B. The role of positive emotion in positive psychology: The broaden-and-build theory of positive emotion. Am. Psychol. 2001, 56, 218–226. [Google Scholar] [CrossRef]
  211. Fredrickson, B.L. The value of positive emotions. Am. Sci. 2003, 91, 330–335. [Google Scholar] [CrossRef]
  212. Vanlessen, N.; De Raedt, R.; Koster, E.H.; Pourtois, G. Happy heart, smiling eyes: A systematic review of positive mood effects on broadening of visuospatial attention. Neurosci. Biobehav. Rev. 2016, 68, 816–837. [Google Scholar] [CrossRef] [PubMed]
  213. Easterbrook, J.A. The effect of emotion on cue utilization and the organization of behavior. Psychol. Rev. 1959, 66, 183–201. [Google Scholar] [CrossRef] [PubMed]
  214. Kimchi, R.; Palmer, S.E. Form and texture in hierarchically constructed patterns. J. Exp. Psychol. Hum. Percept Perform. 1982, 8, 521–535. [Google Scholar] [CrossRef] [PubMed]
  215. Gasper, K.; Clore, G.L. Attending to the big picture: Mood and global versus local processing of visual information. Psychol. Sci. 2002, 13, 34–40. [Google Scholar] [CrossRef] [PubMed]
  216. Harmon-Jones, E.; Gable, A.; Price, T.F. Does negative affect always narrow and positive affect always broaden the mind? Considering the influence of motivational intensity on cognitive scope. Curr. Dir. Psychol. Sci. 2013, 22, 301–307. [Google Scholar] [CrossRef]
  217. Beck, A.T.; Clark, D.A. An information processing model of anxiety: Automatic and strategic processes. Behav. Res. Ther. 1997, 35, 49–58. [Google Scholar] [CrossRef]
  218. Eysenck, M.W.; Derakshan, N.; Santos, R.; Calvo, M.G. Anxiety and cognitive performance: Attentional control theory. Emotion 2007, 7, 336–353. [Google Scholar] [CrossRef] [PubMed]
  219. Mogg, K.; Bradley, B.P. A cognitive-motivational analysis of anxiety. Behav. Res. Ther. 1998, 36, 809–848. [Google Scholar] [CrossRef]
  220. Van Bockstaele, B.; Verschuere, B.; Tibboel, H.; De Houwer, J.; Crombez, G.; Koster, E.H. A review of current evidence for the causal impact of attentional bias on fear and anxiety. Psychol. Bull. 2014, 140, 682–721. [Google Scholar] [CrossRef] [PubMed]
  221. Mathews, A. Why worry? The cognitive function of anxiety. Behav. Res. Ther. 1990, 28, 455–468. [Google Scholar] [CrossRef]
  222. Beard, C.; Sawyer, A.T.; Hofmann, S.G. Efficacy of attention bias modification using threat and appetitive stimuli: A meta-analytic review. Behav. Ther. 2012, 43, 724–740. [Google Scholar] [CrossRef] [PubMed]
  223. Grafton, B.; MacLeod, C.; Rudaizky, D.; Holmes, E.A.; Salemink, E.; Fox, E.; Notebaert, L. Confusing procedures with process when appraising the impact of cognitive bias modification on emotional vulnerability. Br. J. Psychiatry 2017, 211, 266–271. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  224. Jones, E.B.; Sharpe, L. Cognitive bias modification: A review of meta-analyses. J. Affect. Disord. 2017, 223, 175–183. [Google Scholar] [CrossRef] [PubMed]
  225. Mogg, K.; Bradley, B.P. Anxiety and Threat-Related Attention: Cognitive-Motivational Framework and Treatment. Trends Cogn. Sci. 2018, 22, 225–240. [Google Scholar] [CrossRef] [PubMed]
  226. Price, R.B.; Wallace, M.; Kuckertz, J.M.; Amir, N.; Graur, S.; Cummings, L.; Popa, P.; Carlbring, P.; Bar-Haim, Y. Pooled patient-level meta-analysis of children and adults completing a computer-based anxiety intervention targeting attentional bias. Clin. Psychol. Rev. 2016, 50, 37–49. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  227. Krebs, G.; Pile, V.; Grant, S.; Degli Esposti, M.; Montgomery, P.; Lau, J.Y. Research Review: Cognitive bias modification of interpretations in youth and its effect on anxiety: A meta-analysis. J. Child Psychol. Psychiatry 2018, 59, 831–844. [Google Scholar] [CrossRef] [PubMed]
  228. Amir, N.; Beard, C.; Burns, M.; Bomyea, J. Attention modification program in individuals with generalized anxiety disorder. J. Abnorm. Psychol. 2009, 118, 28–33. [Google Scholar] [CrossRef] [PubMed]
  229. Enock, M.; Hofmann, S.G.; McNally, R.J. Attention bias modification training via smartphone to reduce social anxiety: A randomized, controlled multi-session experiment. Cognit. Ther. Res. 2014, 38, 200–216. [Google Scholar] [CrossRef]
  230. Carlbring, P.; Apelstrand, M.; Sehlin, H.; Amir, N.; Rousseau, A.; Hofmann, S.G.; Andersson, G. Internet-delivered attention bias modification training in individuals with social anxiety disorder-a double blind randomized controlled trial. BMC Psychiatry 2012, 12, 66. [Google Scholar] [CrossRef] [PubMed]
  231. Julian, K.; Beard, C.; Schmidt, N.B.; Powers, M.B.; Smits, J.A. Attention training to reduce attention bias and social stressor reactivity: An attempt to replicate and extend previous findings. Behav. Res. Ther. 2012, 50, 350–358. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  232. Schoorl, M.; Putman, P.; Van Der Does, W. Attentional bias modification in posttraumatic stress disorder: A randomized controlled trial. Psychother. Psychosom. 2013, 82, 99–105. [Google Scholar] [CrossRef] [PubMed]
  233. De Voogd, E.L.; Wiers, R.W.; Prins, J.M.; Salemink, E. Visual search attentional bias modification reduced social phobia in adolescents. J. Behav. Ther. Exp. Psychiatry 2014, 45, 252–259. [Google Scholar] [CrossRef] [PubMed]
  234. Waters, A.; Zimmer-Gembeck, M.; Craske, M.; Pine, D.; Bradley, B.P.; Mogg, K. A preliminary evaluation of a home-based, computer-delivered attention training treatment for anxious children living in regional communities. J. Exp. Psychopathol. 2016, 7, 511–527. [Google Scholar] [CrossRef]
Table 1. Glossary.
Table 1. Glossary.
TermDefinition
AffectUmbrella term for all valenced experiences, including moods, emotions, and feelings.
Affect-driven attention bias (ADAB)An attention bias towards or away from emotional information that is influenced by the observer’s affective state. Often labelled “attention bias” in the animal welfare literature.
Affective stateA temporary affect, e.g., emotions or moods.
AttentionThe selective allocation of cognitive resources to particular information.
Attention biasThe preferential allocation of attentional resources towards one form of information over another.
Attention bias task (ABT)An experimental paradigm that presents subjects with stimuli and records how their attention is allocated. Examples covered here include looking time, emotional Stroop, dot-probe, emotional spatial cueing, and visual search tasks.
Attention to emotionAttention allocated towards emotional stimuli.
Attention to threatAttention allocated towards threatening stimuli.
Avoidance of threatAttention allocated away from threatening stimuli.
CognitionThe mechanisms, such as attention and judgement, that animals use to gather, process, and store information.
Cognitive biasIn the animal welfare literature, an umbrella term for cognitive processes influenced by affect, e.g., attention and judgement biases.
Disengagement (of attention)The allocation of attention away from a stimulus previously attended to.
EmotionStimulus-directed affective state. Consists of behavioural, physiological, and cognitive components, and may occur outside awareness (cf. “Feeling”).
Engagement (of attention)The initial allocation of attention towards a stimulus. Limited attentional resources mean engagement to one stimulus may draw resources away from other tasks.
FeelingSubjective, experiential element of affect. Because animals’ feelings cannot be reported directly, we rely on indirect indicators that can be objectively measured, e.g., behaviour, physiology, and cognitive biases.
Judgement BiasA cognitive bias where affect influences judgements about the affective value of ambiguous stimuli. Positive affect is associated with optimistic judgements; negative affect is associated with pessimistic judgements.
Judgement bias task (JBT)A task that uses judgements of ambiguous stimuli as an indicator of affect. Typically, subjects are trained to react differently to two stimuli to achieve relatively positive- and negative-valence outcomes. Responses to subsequent presentations of intermediate “probe” stimuli indicate whether subjects judge them more positively (optimistic responses) or negatively (pessimistic responses).
MoodA long-lasting affective state that reflects the cumulative impact of emotion over preceding days, weeks or months.
MotivationDrives arising from internal signals that compel behaviour to meet basic biological needs, e.g., hunger and thirst.
Overt attentionA measurable proxy for attention, such as movements of the eye with respect to stimuli.
PersonalityBehavioural and psychological traits with inter-individual variation but intra-individual consistency across time and contexts.
Trait affectAffect stable within individuals over time. A personality trait that does not encompass transient emotions or moods.
VigilanceScanning the environment for potential threats (may occur in the absence of threatening stimuli).
Table 2. Affect-driven attention bias studies on animals.
Table 2. Affect-driven attention bias studies on animals.
SpeciesRef.NStimuliMeasure/Manipulation of AffectMeasure of AttentionFindings
Starling[38]32Eyespots, ambiguous eyespots, ctrlsNV: predator call, alarm call, white noiseOrienting towards stimuliNo effect
Sheep[39]41Empty food bucketNV: food-deprivationDetection/approach latency, object interactionNo effect for detection/approach latency; NV sheep interacted longer
[40]29Aggressive, affiliative, and non-social behaviours (video)NV: unpredictable, unenriched housing; PV: predictable, enriched housingOrienting towards stimuliTime oriented towards stimuli (all subjects): aggressive > neutral > affiliative. NV: oriented towards stimuli longer overall
Dual-Presentation Looking Time Task
Rhesus macaque[26]7Aggressive/neutral facesNV: post-vet health-check; PV: enrichmentEye gazeAll monkeys initially oriented faster towards aggressive faces; NV monkeys disengaged faster from aggressive faces
Starling[27]14Alarm call/foodNV: no water bathHead up/down durationNV birds longer head-up bout and shorter head-down bout duration
Sheep[41]60Dog/foodNV: anxiogenic; PV: anxiolyticLooking time, head up duration, latency to eatLooking duration/head up/latency to eat: NV > ctrl > PV
[42]60Dog/foodNV: anxiogenic; PV: anxiolyticLooking time, head up duration, latency to eatLooking duration/head up/latency to eat: NV > ctrl > PV
Cattle[43]36Dog/foodNV: anxiogenic; PV: anxiolyticLooking time, head up duration, latency to eatNV looking duration/head up/latency to feed > ctrl; no effect for PV
Emotional Stroop Task
Chimpanzee[44]7Vet (negative-valence) and other humansNV: post-vet health-checkColour discrimination task RTsAll subjects: RTs slower to touch correct colour when it contained image of the vet than non-threatening humans. NV subjects: slower than ctrls to touch the correct colour when it contained image of the vet
Orange-winged amazon[45]20HumanSubjective personality assessmentSpatial memory task RTsNegative correlation between neuroticisim ratings and task performance (suggests greater distraction from human present)
Visual Search Task
Guinea baboon[46]6T-/L-shapes (conditioned valence)NV and PV behaviours (observational)RT to the targetRT: NV > ctrl > PV
Dual-Presentation Judgement Bias Task
Brown rat[47]16Tones (conditioned valence)NV: unpredictable housingLever pressed (binary) and RT to lever pressAmbiguous-probe and dual-presentation JBT congruent: NV rats pressed positive lever (optimistic responses) more than ctrls. Suggests AB towards negative-valence stimulus
Abbreviations: Negative valence (NV), positive valence (PV), reaction-time (RT), attention bias (AB), control (ctrl).

Share and Cite

MDPI and ACS Style

Crump, A.; Arnott, G.; Bethell, E.J. Affect-Driven Attention Biases as Animal Welfare Indicators: Review and Methods. Animals 2018, 8, 136. https://doi.org/10.3390/ani8080136

AMA Style

Crump A, Arnott G, Bethell EJ. Affect-Driven Attention Biases as Animal Welfare Indicators: Review and Methods. Animals. 2018; 8(8):136. https://doi.org/10.3390/ani8080136

Chicago/Turabian Style

Crump, Andrew, Gareth Arnott, and Emily J. Bethell. 2018. "Affect-Driven Attention Biases as Animal Welfare Indicators: Review and Methods" Animals 8, no. 8: 136. https://doi.org/10.3390/ani8080136

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop