Professional Documents
Culture Documents
B9 Modes
B9 Modes
When we say that different modes have different affordances and constraints, we
mean both that they have different sets of ‘meaning potential’ and that they
allow us to take different kinds of actions. For example, in written text and
spoken language we must present information in a sequential way governed by
the logic of time. Thus, an author or speaker can manipulate the order and speed
at which information is given out, perhaps withholding certain facts until later in
the text or conversation for strategic purposes. Images, on the other hand, are
governed by the logic of space. The producer of the image presents all of the
elements in the image all at once and has limited control over the order in which
viewers look at those elements. Similarly, images allow for the communication of
very fine gradations of meaning when it comes to things like shape and color –
the exact shade of pink in someone’s cheeks, for example – whereas language
forces us to represent things in terms of types – the word ‘pink’ for example,
cannot represent an exact color, but only a range of colors within a particular
class.
The fact that different modes make different kinds of meanings more possible
and others less possible is one of the reasons why people strategically mix
different modes when they are communicating, so that the constraints of one
mode are balanced out by the affordances of others. While there are some things
that ‘just cannot be expressed in words,’ it might be possible to express them
with a carefully timed facial expression or a carefully placed image.
89
fulfill these three functions, but do so in a rather different way than language
(see Section D9).
Ideational function
In narrative images, action processes are usually represented by what Kress and
van Leeuwen call vectors, compositional elements that indicate the directionality
of an action. In figure B9.1, for example, the arm of the boxer on the left
extending rightward towards the head of the other boxer portrays the process of
‘hitting’. There are also other processes portrayed. For example, the upward
gazes of the figures in the background create vectors connecting the spectators
with the fighters.
TM
90
Like this image, many images actually represent multiple processes
simultaneously. Figure B9.2, for example, also involves action processes, the face
on the left (representing a library user) joined to the different kinds of resources
he or she ‘consumes, uses, evaluates, creates, combines, and shares’. This image,
however, is more abstract, and so the vectors are represented as labeled arrows
rather than visual representations of these actions. At the same time, the image
also contains classificatory relationships – the objects portrayed under the
headings ‘Information Literacy’, ‘Media Literacy’ and ‘Digital Literacy’
representing distinct classes of things ‐‐ and analytical relationships ‐‐ the
smaller faces in the lower right corner, for example, portrayed as parts of a
larger social network.
Figure B9.2 Using information, media and digital literacy (credit Karin Dalziel)
Interpersonal function
91
In images, viewers are placed into relationships with the figures in the image,
and, by extension, the producers of the image, through devices like perspective
and gaze. The image of the child in figure B9.3 illustrates both of these devices.
The camera angle positions the viewer above the child rather than on the same
level, creating the perspective of an adult, and the child’s direct gaze into the
camera creates a sense of intimacy with the viewer, though the expression on the
child’s face does denote some degree of uncertainty. Another important device
for expressing the relationship between the viewer and the figures in an image is
how close or far away they appear. Long shots tend to create a more impersonal
relationship, whereas close‐ups tend to create a feeling of psychological
closeness along with physical closeness.
‘Modality’ in images is partially realized by how ‘realistic’ the image seems to the
viewer. Photographs, for example, generally attest more strongly to the ‘truth’ of
a representation than drawings or paintings. However, this is not always the
case. Scientific diagrams and sketches, for example, are often regarded as having
even more ‘authority’ than photographs, and black and white images like those
often found in newspapers are often regarded as more ‘realistic’ than highly
saturated color images in magazine advertisements.
TM
Textual function
92
producers of images have much less control than producers of written texts over
how viewers ‘read’ the image, they can create pathways for the viewer’s gaze by,
for example, placing different figures in different places within the frame and
making some more prominent and others less prominent.
The distinction between the upper part of an image and the lower part is related
to the strong metaphorical connotations of ‘up’ and ‘down’ in many cultures
(Lakoff and Johnson, 1980). According to Kress and van Leeuwen, the top part of
the image is often used for more ‘ideal’, generalized or abstract information, and
the bottom for ‘real’, specific and concrete information. They give as an example
advertisements in which the upper section usually shows ‘the “promise of the
product”, the status of glamour it can bestow on its users’ and the lower section
tends to provide factual information such as where the product can be obtained
(2006: 186).
Both of these principles can be seen in figure B9.4. In order to make sense of the
‘narrative’ of HIV transmission that the text tells, one must begin at the far left of
the advertisement and move to the far right. The figures of the man and the
woman on the left of the image constitute ‘given’ information, while the virus on
the right of the image constitutes the ‘new’ information. There is also a clear
demarcation between the upper half of the text and the lower half. While the
upper half does not portray the positive ‘promise’ of a particular product aTs M
many advertisements do, it does represent a kind of idealized hypothetical
situation which the viewer is invited to imagine. Rather than a ‘promise’,
however, it is something more akin to a ‘threat’. And the lower half of the image,
rather than giving information about where the product portrayed in the upper
93
half can be obtained, it gives information on how this hypothetical situation can
94
be avoided, along with specific information about such things as the name of the
organization that produced the ad and the condom company that sponsored it.
This text also illustrates how images and words often work together. The words
‘Joy Stick’, ‘Play Station’ and ‘Game Over’ tell the viewer how the images are to be
interpreted, and the slogan in the lower half of the text (‘You only have one life:
use a condom’), explains the image of the condom above it. Finally, this text
shows how multimodality can be effective in getting viewers to make
connections between different ‘Discourses’. While the images belong to the
‘Discourse of biomedicine’, the words invite the viewer to interpret these images
within the framework of the ‘Discourse of video games’.
Multimodality in Interaction
Modes in face‐to‐face interaction such as gaze and gesture also fulfill these three
functions. The mode of gaze, for example, has an obvious interpersonal function,
creating a relationship between the gazer and whomever or whatever is the
object of the gaze. It also carries ideational meaning, conveying that the gazer is
looking at, watching or paying attention to something. Finally, gaze is oftenTaMn
important textual resource, helping people to manage things like turn‐taking in
conversations.
While the ‘inter‐modal’ relationships (the ways multiple modes work together)
in static texts like the advertisement analyzed above can be complicated, they
can be even more complicated in dynamic interactions. One of the problems
with
95
analyzing real time, face‐to‐face interactions is that participants have so
many modes available to them to make meaning. There are what Norris
(2005) calls ‘embodied’ modes such as gaze, gesture, posture, head
movement, proxemics (the distance one maintains from his or her
interlocutor), spoken language, and prosody (features of stress and
intonation in a person’s voice). And there are also ‘disembodied’ modes
like written texts, images, signs, clothing, the layout of furniture and the
architectural arrangement of rooms and other spaces in which the
interaction takes place. All of these different modes organize meaning
differently. Some, like spoken language and gaze tend to operate
sequentially, while others like gesture and prosody tend to operate
globally, often helping to create the context in which other modes like
spoken language are to be interpreted (see Section B6). Not all of these
modes are of equal importance to participants at any given moment in
the interaction. In fact, different modes are likely to take on different
degrees of importance at different times. How then is the analyst to
determine which modes to focus on in a multimodal analysis?