Multi Modal Conversations 09-11-2010
Multi Modal Conversations 09-11-2010
conversations
Abstract
Online conversations are now ubiquitous, and language learning researchers are one community
with a particular need for a better understanding of data collected from such exchanges. This paper
addresses the lack of formalised methodology for analysing research data created in
technologically-mediated conversations. First we show the importance of conversations in language
learning and establish that appropriate methodologies for investigating them have as yet emerged
neither from the field of computer-assisted language learning nor from conversation analysis (CA).
We analyse three brief text and voice-based conversations among language learners engaged in
online tutorials via a multimodal platform. From this we conclude that the analysis and
interpretation of such exchanges can be improved by a crossdisciplinary approach which consists of
augmenting constructs drawn from CA with selected constructs from social semiotics.
Keywords
1
Introduction
by learners seeking to enhance their oral and written fluency in their second or foreign language
(L2). The aim of the study is to establish methodological principles to address a gap in the analysis
of online multimodal data. The paper, which focuses on such phenomena as turn-taking and face-
saving in an online course, takes Conversational Analysis (CA) as its theoretical starting point but
adopts a multimodal perspective that appears to shed greater light on the meaning-making processes
To this aim, we will start by establishing the need for an understanding of what learners are
doing when they converse via multimodal electronic systems. We will then discuss the applicability
multimodal social semiotics, we will identify three frameworks which may provide guiding
principles for a methodology with the potential to answer questions raised by our data.
Conversations are seen as beneficial to L2 learners for three reasons. Firstly they have been part of
the communicative model of language teaching for half a century, as ‘conversation classes’ in
which learners, in the safety of the classroom, are invited to experience the ‘pressure of
conversation’ (Cook, 1996 [1991]: 61) that it is assumed they will face when called upon to talk to
native speakers in the target country. Secondly, work carried out since the mid-80s within Second
Language Acquisition theory has established that, providing they are structured so as to require
2
andVaronis 1995; Long 1998; Gass, Mackeyand Pica 1998; Chapelle 2004). Thirdly, for
researchers such as Belz andThorne (2005) and O’Dowd (2006), who have investigated
intercultural aspects of telecollaborative transnational online projects, conversations are the means
whereby learners from different cultures construct their knowledge of the conversational strategies
To attain these learning outcomes, research has shown that conversations have to satisfy
certain criteria. They need to be part of a constraining task (for example a group of learners might
be asked to address a problem then reach a negotiated consensus). However, to help learners cope
with the spontaneity of speech, conversations also have to offer them sufficient time and freedom to
develop new topics or to attend to unexpected conversational moves made by their discourse
partners. Tutorial software can offer such opportunities: it may allow learners to negotiate in groups
online (through the use of grouping tools and ‘breakaway rooms’), to respond fast (by textchat or
audio) or more reflectively (with written documents co-created during the interaction). Such
software may also allow the sharing of visual objects or sound files, that act as stimuli for talk.
Multimodal voice-enabled platforms are thus of great interest to language learning researchers.
Some preliminary thinking about how to approach the study of conversations in these
environments was nourished by reflections on our earlier piece of work (Author, 2006: 260),
focused on a course involving both written text (asynchronous and synchronous) and recorded
speech (asynchronous only). We had observed that prescribing the use of specific communication
channels for particular discursive goals was ineffective, because users re-appropriated the
environment in their own way, successfully developing conversations in phases of the educational
task, and in ‘spaces’ of the environment, that were not originally dedicated to this purpose. For
example, although we had structured the asynchronous forum into two separate ‘threads’, one for
task-completion and one for social interaction, users exchanged messages about both topics
3
everywhere in the environment, making no attempt to select the prescribed thread. Building on this
experience, we based the design of the course from which the data below is extracted on the
environment’s modalities would provide us with opportunities to understand how they achieved
their discourse aims. We also expected that learners would reach some of these aims by creative re-
Central to our project was a focus on multimodality, defined for the purpose of this study as
the co-availability of several modalities, this term itself being understood as the material through
which the conditions for interaction are created when humans use an online platform. For example
any technological feature seen in its relationship with a semiotic resource is a modality, so that a
semiotic resource may be associated with a single modality (as when spoken language is associated
solely with the output of an audio channel), or with several modalities (as when written language is
associated both with user-generated messages (real-time and asynchronous) as well as with system-
generated ones.
Introduced by Herring in 2004, it has been influential within the computer-assisted language
learning research community but its application is limited to text-based exchanges. To account for
interactions, Sacks, Schegloff and Jefferson (1974) used CA to analyse everyday conversations in a
technologically-mediated setting, the telephone. Hutchby (2001) further extended the field of
4
involving surgical teams using both mother tongue and L2 (Mondada, 2003). There has been some
(1996) querying it and Seedhouse (1998, 2004) convincingly countering his views. Garcia and
Jacobs (1999) demonstrated CA’s appropriateness for analysing online L2 conversations in a paper
on turn-taking in written chat, as long as the research instruments included videos of the screen of
each learner – which today we would call ‘dynamic screen capture’ – in order that hesitations and
self-corrections could be evidenced in users’ messages prior to their being sent. Also focused on
textchat, Payne and Ross (2005) used elements of CA to argue that there is a relationship between
written chat proficiency and oral proficiency in L2. But to date there has been no use of CA for
investigations that integrate the multimodality of online settings and human interactivity in a L2
learning project, with the exception of our own work Author (2004) and Author and Another
Among the tenets of CA, two are pertinent to the data to be introduced shortly. Like all CA
principles, they can account for conversational material both when participants abide by them and
equally when they breach them. In a simplified form they can be summed up as follows: (a) turns of
speech alternate and interlink, since the basic principle of ‘conditional relevance’ asserts that given
a question, you may expect an answer; given an apology, expect an acknowledgement; given a
topic, expect that it will be pursued; (b) face-saving is a major preoccupation in any conversation,
because conversants are in a relationship of perpetually converging and diverging interests with
their fellow-conversants - you may try to save your own face or to protect others’, and it is also
5
In the computer-assisted multimodal communication context under study, discourse partners
use many semiotic systems, including linguistic (written and spoken), iconic and symbolic systems.
They may use these systems in rapid succession (type then draw), quasi-simultaneously (speak
while hitting the ‘send’ button to post a written message) or they may choose among systems to
make meaning in particular ways. For example, a user may close a conversation by typing ‘Bye for
now’, by clicking a specific button, or by announcing their withdrawal orally. Different semiotic
systems are also involved in responding to such a move. For example, if users of a synchronous
environment signal their impending departure orally, their discourse partners will receive aural
input, made up of linguistic and non-linguistic information, e.g. a phone heard ringing in the
departing person’s home will suggest reasons for their terminating the conversation. Or a user may
type a valedictory message into a box and send it. In this case, their message will be displayed on
the shared screens and will remain a part of the ongoing conversation even after they have
disappeared from the environment. Or again users might avail themselves of the system’s
telepresence indicators, e.g. icons of different colours signalling that users are ‘offline’, ‘online’,
‘on standby’ etc, or display symbolic changes to the objects on the screen, such as the greying-out
of a name, or fading-out of a photo. Choosing to leave the conversation by activating such a tool is
a strategy likely to influence the direction of the conversation in a way that is distinct from the
effect produced by disconnecting altogether (in which case the name or image would vanish), or
from the effect of leaving all connections ostensibly active but walking away from the computer for
interlinking of semiotic systems and effects can become extremely complex, and there is scope for
observing conditional relevance and face-saving at work, through studying the users’ conventional
as well as catachrestic use of the tools. In the rest of this paper we seek theoretical support when
attempting to grasp this complex reality, by studying three brief conversations carried out through
6
the Lyceum environment as part of an English-for-Special-Purposes project. The next section briefly
describes the system and the pedagogical setting, and then introduces the three data sets.
Lyceum is a synchronous audiographic groupware system used at the University of XXXX between
1995 and 2009, designed to facilitate social learning in distance tutorials. Figure 1 shows the
characteristic tools and spaces of this tutorial platform: a facility for creating ‘rooms’ for small-
group work, a space for textual and graphic displays (whiteboards or other shared documents), a
space representing the participants’ conversational status (speaking, listening quietly or seeking
permission to take a turn), a space for text-chatting and icons for indicating agreement or
time, the logistics of turn-taking on this platform is such that the optimal number of participants is
10 to 15 per tutorial room. However the subdividing of plenary groups into separate sub-groups of 3
to 5 students, and back again to plenary spaces, is easy and almost instantaneous. Within each room,
different tools and shared spaces can be displayed and collaboratively used. A turn-taking
management system based on clickable icons is another feature of such platforms that has relevance
for our study. As shown in Figure 1, a raised hand icon symbolises an intention to take a turn, while
obstacle to ignoring the turn-taking tool altogether (although in practice no more than three people
7
The extracts discussed below originate from a project concerned with teacher-learner
partnership between the University of XXXX and the University of YYYY. Sixteen French-
speaking students studying for a Professional Masters in Open and Distance Teaching (ODT)
worked in two groups of eight, connecting from their homes in various parts of France. Each group
had an English native speaker tutor connecting from the United Kingdom. The groups met during
10 sessions of over an hour each. The course had a dual objective, linguistic and professional,
which was the development of competences in ODT through spoken and written English. The less
proficient group (false beginners with wide variations in knowledge) provided the extracts
discussed below.
The Lyceum environment and the Copéas project are presented in detail in Vetter and
Chanier (2006). The environment supports semiotic resources for constructing discourse in
interaction, such as natural language in its written and spoken forms, as well as visual resources
The three extracts occur within the last seven minute of the last session in the course. Both Tables 1
and 2 give simplified versions of a tabular transcription of the screen videos. Reading the top row of
Table 1 from left to right, the transcripts show: participants’ initials (User), the speech turns (T,
characterising all inputs, rather than simply those in ‘speech’), chronological sequences (Time) and
data collected from the four modalities (in the four rightmost columns).
Table 1
8
[INSERT TABLE 1 EXTRACT 1]
When the first extract begins, the learners are engaged in an evaluation-type conversation.
Three students, H, C and A are communicating orally negotiating an agreed heading, summing up
what they think they mostly learned during the course. The heading then needs to be typed on the
shared document screen by one of the group. In Extract 1 (Table 1), H twice tries to pronounce the
phrase technical vocabulary (at T1 and T7) but stutters and hesitates each time. As C attempts to
settle the answer by typing vocabulary learning in the shared document (T3) and eliciting H’s
approval of this formulation (T5), H holds to his original wording by writing it in the chat window
(T6). This results in C modifying (T8) what he had typed on the shared document, which H finally
approves (T9) by clicking the Tick icon. Arrows show topic-maintenance by H across modalities.
strategy that takes advantage of the environment’s multimodality. His strategy cannot be made
visible through conversation analysis conducted in its classical form since CA relies on the
sequentiality principle, and the conversation, if read vertically down the second column of Table 1,
has no sequencing that could be sensibly interpreted. Sequentiality is in fact present, but can only be
Table 2
Just before the start of Extract 2, the tutor had launched the conversation by asking the
members of the group (A, P, G and C) to introduce themselves orally. The extract in Table 2 shows
the conversation proceeding orally, with some input typed into the text chat window (highlighted in
grey-blue), while two of the learners (A and P) carry out a completely different conversation in text
9
chat (highlighted in green), concerned with an auditory difficulty. In the recording which was used
for the transcription, A speaks quite audibly but, possibly due to her home setup or for server-
related reasons, P does not hear her. So A and P run a conversation in parallel with the main
conversation initiated by the tutor. They use chat to construct a dialogue on a different theme,
without apparently disturbing the ‘main’ conversation which is proceeding in the audio modality.
Thus, their attitude maybe seen as non-transgressive in terms of saving the tutor’s face. In a non-
virtual classroom, such a move would probably have been seen as face-threatening for the tutor,
who may have taken steps to stop it, had it persisted for more than a few seconds. The face-saving
principle helps explain A and P’s conversational moves. However, the example shows how this
In the last example, Figure 2, we see how learner C uses the four modalities at his disposal:
[INSERT FIGURE 2]
The chart in Figure 2 highlights discrepancies in the ways each participant (H, C and A) uses
these modalities: the horizontal axis shows the modalities available while the vertical axis gives the
number of speech turns per individual. Learner C’s inputs are mainly via the audio channel (45
audio speech turns, 4 text-chat turns). Of the group, C has the highest level of proficiency, as
evidenced by his score in the pre-experiment test and by his ease with target structures such as
‘would you repeat?’, ‘we must choose’, ‘we can’t answer’, ‘do you want to add something?’, ‘don’t
you think so?’ ‘are you OK with what I write?’ etc. Content analysis of his input shows that in over
half his spoken turns (28 out of 44) he is asking for others’ opinions. Yet all his text-chat inputs are
language accuracy checks. Additionally, C makes much more use of the shared document than do
his two colleagues. In data collected from the parts of the session that precede and follow Extract 3,
10
90% of C’s turns in the shared document are directly preceded by a turn elsewhere (mostly in the
audio channel) seeking confirmation of the group’s approval of what he has written.
We hypothesize that C uses the text-chat modality when he is checking the accuracy of the
lexicogrammatical forms he is using rather than moving through the conversational agenda of the
class. He prefers to progress through the class agenda via the shared document modality, but only
after obtaining consensus from his peers orally. This pattern of use could relate to two different yet
converging factors: C may be specialising particular discourse aims to particular screen elements,
and his representation of himself as a communicator may also play a role in which of these he uses.
For example, in relation to the elements in the screen layout, analysts of his conversational
strategies would need to assess the degree of salience of the text-chat window tucked away
unobtrusively (and indicated in the bottom right-hand corner of Figure 1 by an arrow pointing to
‘chat’) in contrast to the intrusiveness of the sound of his voice coming through each group
member’s headset. As for self-representation, we may posit that face-saving issues are involved: for
example C might be prepared to give an image of himself as a confident language speaker in the
audio channel while specialising the chat window for more risk-taking face-threatening activities,
such as asking for help with English forms. A similar explanation might be offered for his use of the
shared document: the visual organisation of the screen when that document is uploaded to it (see
Figure 3 further down), the central disposition of that document and its status as the ‘official’ record
of the group’s collaboratively negotiated view may explain both his self-appointed guardianship
Discussion
In this section and on the basis of the findings from the extracts presented earlier we suggest that
multimodal settings, it needs to be rearticulated. Our suggestion is that such reframing can be done
11
in the light of three theoretical frameworks drawing from affordance theory as developed by Gibson
(1979), multimodal social semiotics and geosemiotics respectively. Our reading of the data showed
that two of the principles of traditional CA, sequentiality (Extract 1) and face management (Extract
conversations, albeit redefined, because when more than one modality was available, both
sequentiality and face management operated across modalities. Hutchby’s (2001) work is, we
Hutchby (2001) reminds us through the example of the early history of telephone communication
that technology is in a reciprocal relationship with its users. The telephone was originally sold as an
instrument for transacting business or for exchanging practical information. Early marketing
stressed this functional use to the extent of advising subscribers to wait until late at night if they
really needed to use the machine for the purpose of personal chat. However users were not
persuaded and appropriated the telephone as a medium for family or other intimate conversations,
[...] while designers may be said to have some control over the features they design into an
artefact, and while they may have some idea about the range of uses to which the artefact
should be put, they have little control over the artefact’s communicative affordances – over
the range of things it turns out to enable people to do (Hutchby, 2001: 123).
We witnessed this mechanism at work in the project extracts, where learners appropriated
modalities in diverse ways. For example the text-chat was re-appropriated in the contexts of
different conversations in order to support control over content (Extract 1), to provide technical
assistance (Extract 2) and to enact face management (Extract 3). We also observed that individual
12
needs. For example legitimation of self as a turn-taker (and therefore control over the conversation)
was achieved via moves from audio to text-chat then to the Tick icon by H (in Extract 1), but for C
(in Extract 3) by toggling between the audio channel and the shared document. The construct of
communicative affordance – as described by Hutchby in the last line of the quotation above - also
helps to understand the different ways in which human actors in different technological settings
solve a single problem such as (to stay with the example of the telephone) answering a landline call,
a mobile phone or a computer bleep from an Internet-telephony system. The same conversational
problem is involved in each case (how to respond to a conversational invitation from a remote
caller), but different discursive solutions are appropriate. For instance, a person called on a landline
without user ID will in most ordinary circumstances pick up the handset and initiate their side of the
conversation by uttering a conventional query (e.g. hello, allô, pronto etc.) in a rising tone followed
by a pause, in the expectation that the caller will identify themselves. With user ID, or in Internet
telephony where the caller’s name and sometimes photograph appears on screen, the person called
is likely to respond instead with a greeting usually followed by the caller’s name. A pause, if
provided by the person called, will not be heard by the caller as an invitation to identify him/herself;
instead it will require interpretation and negotiation so that the conversation may progress.
Although a close textual analysis of our learners’ input is not the focus of the present
The detailed ways in which people have taken up the affordance for verbal intimacy across
distance and shaped it for communicative ends can only be revealed once we move beyond
mediated multimodal settings. However only by paying attention to communicative affordances will
we be able to do the necessary work of re-interpreting its principles in terms of their manifestations
13
2. The influence of design on discourse in multimodal environments
The materiality of the environment impacts on the dynamics of conversations. In Extract 2, two
conversations were progressing in parallel. It is not accidental that the tutor-initiated one was
carried out orally while the students’ exchange about sound levels was conducted in text-chat: all
we need to do to become persuaded of this is to mentally invert the modal choices and imagine that
the tutor led his tutorial via postings in the text-chat while students talked about other topics in the
audio channel. It is unlikely that the group would accept such a position for the tutor, and we draw
Kress and van Leeuwen’s (2001) work on the semiotics of multimodal pages in newspapers
and books identifies the four dimensions of structuring for such artefacts: discourse, production,
dissemination and design. While we believe that all four are relevant to understanding meaning-
making in electronic learning environments (see for example a discussion of the impact of
dissemination on learner progress in Author, 2004: 523-524), for the purpose of the current
discussion we concentrate on one of them, the dimension of design. The designers of the software
that our learners used can appositely be compared with the architect in Kress and van Leeuwen’s
An architect, for instance, designs (but does not build) a house or a block of apartments.
The discourse provides a certain view of how houses are lived in the way they do, and
arguments which critique or defend this way of life. The design of the house then
conceptualises how to give shape to this discourse in the form of a house, or a type of
14
The relevance of this architectural example will become clear as we look at two very
different designs of environments that have been used for language learning, Lyceum and Traveler,
through comments made by learners. Lyceum (Figure 3), as we have seen, is an academic tool
designed to look like a university campus building, which is a traditional design choice for
electronic learning environments as pointed out by Bayne (2008). Traveler (Figure 4) is an avatar-
based system with a ‘fantasy’ feel inherited from the world of games.
[INSERT FIGURE 3]
Fig.3: The design of a Lyceum screen, with whiteboard displaying learners’ work
[INSERT FIGURE 4]
Just as the architect has provided a shape for the cultural discourses of human habitats, so
the designers of electronic environments and virtual worlds conceptualise the form that the
environments take on the screen into interpretable signs. For example here are responses from two
of our users (for Lyceum) and two of Örnberg Berglund’s (2005) users (for Traveler), upon being
Lyceum user 1: ‘Quand le prof rentre dans la salle, cela ne dérange pas. Je sais pas comment
l’expliquer’. (When the teacher enters the room, it’s not intrusive. I don’t know how to
explain it).
15
Lyceum user 2: ‘Le style du prof joue, mais le fait qu’il est invisible, il ne peut pas s’imposer
de la même façon qu’en présentiel’. (The teacher’s style is a factor but the fact that he’s
Traveler user 1: ‘It took me to another world and was a real adrenaline buzz. It was on my
screen and I was conscious of it always, but I was definitely virtually gone from my usual
habitat’.
Traveler user 2: ‘I’m always immersed.[…] It doesn’t matter that the environment is
Whereas these two Traveler users produce a discourse of emotions and escapism, the
classes. We make two comments here. Firstly, although these differences in perception may not be
surprising given the strongly contrasted visual identities of Figures 3 and 4, the question is whether
two groups using these environments for achieving the same language-learning objectives might
have very different types of conversations in each environment. The second observation relates to
designers and the uses made of their designs. Lyceum’s design was underpinned by a democratic
and participative pedagogical posture: ‘We have imposed minimal technical constraints on ‘floor
control’: anyone can speak anytime’ (Buckingham-Shum, Marshall, Brier and Evans, 2001:4). Yet
the users’ comments show a preoccupation with teacher control. It is likely that this is part of their
pre-existing non-virtual educational culture, in which case the question can be asked: to what extent
and in what ways can the design features of interactive learning environments transform the users’
representations of self? The answer to this question is another determiner of sense-making in these
environments. Finally, regarding the device which introduced this section, i.e. the proposal that the
Copéas tutor could conduct core tutorial business in the chat box while the students conversed
orally, evidence from Lyceum users’ perceptions supports the view that the system’s design
16
provides a shape for the cultural discourses of traditional teacher-centered classrooms. But based on
multimodal social semiotics’ understanding of design, there is no in-principle reason why other
types of design could not work to support other cultural discourses, producing distinct types of
conversations.
While Scollon and Scollon (2003) acknowledge the importance of Kress and van Leeuwen’s design
dimension, they re-interpret and re-inforce it in order to account for interaction. We have found
their insights useful for understanding the way that our users perform conversations by choosing
among the different spaces offered to them within the interface. Scollon et al stress the spatial
dimension of social semiotic resources, presenting a framework which they call ‘geosemiotics’:
Geosemiotics is the study of meaning systems by which language is located in the material
world. This includes not just the location of words on the page you are reading now but also
the location of the book in your hands and your location as you stand or sit reading this
The authors structure geosemiotics into three sub-sets: the interaction order, visual semiotics
(on which we will not elaborate here, as this concept comes close to Kress and van Leeuwen’s
notion of design mentioned earlier) and space semiotics. The understanding of space that is of
The interaction order provides a construct for understanding how individuals perceive the
interactional value of the space they choose to use. In their description of the interaction order, the
authors include perceptual spaces and interpersonal distances. Dominant perceptual spaces are
17
visual and auditory (‘less noticed’ ones are olfactory, thermal and tactile). The addition of the
proxemics – allows geosemiotics to ask questions about the relationship between space, sound,
embodiment and socialisation. For example, the auditory space which I perceive and my perceived
intimacy or distance with the individual vocalising the sound that I am hearing, together create the
semiotic resource through which I embody meanings. Applying this framework to Copéas
participants, in particular to the parallel conversation mechanism in Extract 2 and to the multimodal
preferences of the learner in Extract 3, the question becomes: how do they co-construct
interpersonal values (intimate, personal, social, public) into conversations which proceed
simultaneously through visual spaces of varying salience and through an auditory space defined by
Space semiotics, in Scollon et al’s words, is the most fundamental part of geosemiotics,
because it asks ‘Where in the world is the sign or image located?’ and because it aims to account for
‘any aspect of the meaning that is predicated on the placement of the sign in the material world’
(2003: 146, our italics). The distinction between well-rehearsed debates within visual semiotics on
the subversive placement of images for artistic purposes (e.g. Warhol soup cans), and the focus of
space semiotics, is that the latter is looking at the material world as a whole, and not simply at
materials used for the bearing of signs, such as paper, canvas or brick. In terms of multimodal
electronic environments, space semiotics provides the basis for asking questions such as: how do
users decode and encode meanings in a material situation involving their computer and its various
peripherals (keyboard, mouse or keypad, webcam) as well as other stimuli around them (possibly
another computer, a video screen, a person physically present, who is talking, writing or drawing,
etc.)?
Conclusion
18
Two considerations should be offered in conclusion, one reflective and the other more concerned
with application. Firstly, due to the complexity of the phenomena under study and the
pluridisciplinary nature of the analysis, we have only touched on principles for a future
methodology. A clear direction for further work, then, is to complement and operationalise these
principles. Secondly, we are aware of some use of social semiotics for researching some online
environments, such as the semiotics of multiplayer immersion games (Lemke, 2006), and of social
tagging (Huang and Chuang, 2009), but not yet for investigating the use of the multimodal
communication tools at the core of this study. Similarly, in the educational field, although Jewitt,
and Kress (2003) have used social semiotics in their work as part of a research area sometimes
called ‘multimedia semiotics’ to lay the foundation for an understanding of learning via multimodal
texts and via educational software (Jewitt 2004), neither on-screen interaction nor issues related to
As a response to both the reflective and the application dimensions just mentioned, we see a
useful research agenda emerging: to test out, with a large volume of learner interaction data
collected from interactive multimodal technologies (different platforms, social networks, mobile
devices) in settings both educational and non-educational, the methodological claims made in this
paper according to which such data can be better understood through the synergistic use of
Acknowledgement
19
References
Author, (2006) ‘Interactive Task Design: Metachat and the Whole Learner’, in M. del Pilar García
Matters, 242-264.
Author and Another, (forthcoming) ‘The multimodal mediation of semiotic resources in an online
Belz, J. and Thorne S.(eds.) (2005). Internet-Mediated Intercultural Foreign Language Education.
Bayne, S. (2008) ‘Higher Education as a Visual Practice: Seeing Through the Virtual Learning
Buckingham-Shum, S., Marshall, S. and Evans, T. (2001) ‘Lyceum: Internet Voice Groupware for
Chapelle, C. (2004). ‘Technology and Second Language Learning: Expanding Methods and
Cook, V. (1996 [1991]). Second Language Teaching and Language Learning. London: Arnold.
Garcia, A. and Jacobs, J. (1999). ‘The Eyes of the Beholder: Understanding the Turn-taking System
Gass, S., Mackey, A. and Pica, T.(1998). ‘The Role of Input and Interaction in Second Language
Gass, S. and Varonis, E. (1994). ‘Input, Interaction, and Second Language Production’, Studies in
20
Gibson, J. J. (1979). The Ecological Approach to Visual Perceptions. Boston: Haughton Mifflin.
House.
Behavior’, in S. Barab, R. Kling and J. Gray. (eds), Designing for Virtual Communities in the
Huang A.W. and Chuang, T.R. (2009) ‘Social Tagging, Online Communication, and Peircean
Hutchby, I. (2001). Conversation and Technology: From the Telephone to the Internet. Cambridge:
Polity.
Jewitt, C. and Kress, G. (2003). Multimodal Literacy. Bern: Peter Lang Publishing.
Kress, G. and Van Leeuwen, T. (2001). Multimodal Discourse: the Modes and Media of
and D. Reinking. (eds). International Handbook of Literacy and Technology Volume III.
Mondada, L. (2003). ‘Describing Surgical Gestures: the View from Researcher’s and Surgeons’
Video Recordings’, in Proceedings of the First International Gesture Conference, Austin, July
2002.
21
O'Dowd, R. (2006). Telecollaboration and the Development of Intercultural Communicative
Göteborg University.
Payne, J. S. and Ross, B. M. (2005) ‘Synchronous CMC, Working memory, and L2 Oral
Rabardel, P. (1995) Les Hommes et les Technologies, Approche Cognitive des Instruments
Sacks, H., Schegloff, E. and Jefferson, G. (1974). ‘A Simplex Systematics for the Organisation of
Scollon, R. and Scollon, S. W. (2003). Discourses in Place: Language in the Material World.
Seedhouse, P. (1998) ‘CA and the Analysis of Foreign language Interaction: a Reply to Wagner’,
van Leeuwen, T. and Jewitt, C. (2001) A Handbook of Visual Analysis. London: Sage.
Vetter, A. and Chanier, T. (2006). ‘Supporting Oral Production for Professional Purposes, in
Wagner, J. (1996) ‘Language Acquisition through Foreign Language Interaction: A Critical Review
Online reference
22