Types of Depiction in ASL Paul Dudis 1. Introduction The body and the surrounding space are often brought into ASL discourse to represent entities of the scene being depicted. Close examination of depiction using cognitive linguistic frameworks suggests that additional components figure in these iconic representations, namely the subject (or the self), vantage point, and temporal progression. Identification of these components contributes towards greater precision with which to describe various types of depiction observed to occur in ASL discourse. This in turn lends to a greater appreciation of issues involving depiction. One issue of particular interest is the relationship between the depiction of an event involving a subject and signs produced during the depiction. The following analysis discusses one pattern in which, when an action of a subject is being depicted, any sign or gesture produced is necessarily associated with the representation of temporal progression through which the depicted subject exists. 2. Depicting vs. non-depicting signs Many words in signed languages exhibit iconic mappings, but only one set of these words have the ability to visually represent semantic components. Following Liddell (2003), I use the term depiction to describe this ability. The distinction between signs that depict and those that do not are discussed below, followed by a description of an informal test with which to determine the status of an ASL sign as an event-depicting verb. The ASL noun BIRD (Figure 1) exemplifies an iconic but non-depicting sign. The iconic correspondences it exhibits are likely to be discerned without great difficulty by anyone who knows what birds are and what the sign means. The manual articulator corresponds to the beak, its location to the location on the bird’s head, and so forth. However, the sign does not function to describe what a bird looks like, nor does it function to describe the actions of a bird. To use a somewhat popular term, it is “frozen.” Moreover, the iconicity of many signs like BIRD belies the schematicity of the concept that the sign symbolizes. As noted by Taub (2001), such signs designate a general category of diverse but related things—any bird with beak morphology different from the prototype can still be called a BIRD. Figure 1 1 Many verbs in ASL are also iconic but non-depicting. Two such verbs are the indicating verbs GIVEy and EXPLAINy, partly illustrated in Figures 2 and 3. When we consider the handshape of GIVEy and what the sign symbolizes, it becomes clear that there is a “hand-forhand” iconicity. In contrast, it is not readily apparent what the articulators of EXPLAINy are iconic for. Directionality is iconic in both verbs. Taub (2001) describes how conceptual paths— the action-chain path, the literal path or the metaphorical path—motivate directionality in these verbs. Yet, given these and other iconic elements that can be discerned, indicating verbs do not depict events. GIVEy encodes the transfer of any of a large range of objects, e.g. a paper document or a kitchen appliance, e.g. a kitchen blender. These objects are held differently, some requiring the use of two hands, and many of these objects cannot be held with the hand configuration similar to that in GIVEy. Moreover, the wrist-only movement in GIVEy typically produced during informal signing does not resemble any of the usual ways of physically handing an object to someone. Figure 2 Figure 3 Depiction is a type of iconicity different from that exhibited by nouns and indicating verbs. As described in Liddell (2003), some verbs have, in addition to their usual function as verbs, the 2 ability to depict the event they encode. HAND-TO is an example of an event-depicting verb. It is not a derivative of GIVEy but a verb in its own right. (The two verbs are also contrasted in Padden 1986 and P. Wilcox 1998.) HAND-TO can be used to describe only the transfer of objects that can be held between the thumb and the four fingers of the handshape—a paper document or credit card, but certainly not a kitchen blender. This is one way that the verb’s iconicity constrains its usage. Additionally, the palm’s continuously upward orientation and the path of the hand created via the elbow emulate the physical motion of the transfer event. However, the high degree to which the verb’s form maps onto the scene it encodes does not necessarily make the sign a depicting verb. Rather, it is the verb’s ability to portray a dynamic and visual representation of a transfer, which is a demonstration rather than “plain” description. One way the verb can be used is akin to a re-enactment by an actor, but with just the signer’s upper body used to create the only visible part of the depiction, the “giver.” Instead of a visible object, the transferred entity is represented by a small portion of space occupied in part by the signer’s hand, and instead of another actor, the recipient is represented by a larger portion of space in front of the signer. The ability of a verb to depict an event can be determined by an informal test involving a depiction of an event sequence performed by the signer who is representing an animate event participant throughout the sequence. If the verb in question can be signed while the event participant is continually being represented, then it is likely a depicting verb. For example, TAPSHOULDER depicts a prototypical way of getting someone’s attention within the U.S. Deaf community. Since the giver typically has the recipient’s attention prior to the act of transfer, it is an ideal verb to be used in the first part of the event sequence depiction. The scene being depicted then, has someone getting another individual’s attention and giving that person an object. Partly illustrated in Figure 4a is the depiction of the attention-getting portion of the scene. Here the signer is conceptualized to be the attention-getter and the location towards which the hand is directing is imagined to be someone’s shoulder. GIVEy would not be able immediately follow this if the signer intends to depict a transfer. Since HAND-TO can (Figure 4b), its depicting verb status is confirmed. This test’s utility is evident in that it also confirms the non-depicting status of EXPLAINy as well as other indicating verbs such as SHOWy and TELLy. Figure 4a Figure 4b 3 With GIVEy and HAND-TO, we see that a certain scene may be encoded into either a non-depicting or depicting verb. Other scenes are similarly encoded into these two types of verbs, including one in which an object is shown to someone. SHOWy (Figure 5) is a verb of showing in ASL. Figure 5 From the semantics of the verb and related cultural knowledge, its handshapes can be seen to be iconic for a flat object and a finger pointing to something on that object, e.g. information on a document. To determine whether this verb can depict an act of showing, a event sequence depiction test similar to the one above can be used. The first part of the test depicts an individual being approached by someone, e.g. a police officer (Figure 6a). SHOWy cannot follow this depiction if the signer wishes to continue to represent the individual. The extended eye contact that would be made between the participants in the event sequence cannot be depicted during the production of SHOWy in this case, demonstrating that the signer is not representing any event participant. Should the signer wish to represent this person while depicting the showing of the object, a different verb would be used. If the object were a document, the verb’s handshape would be similar to that of HAND-TO (Figure 6b). Figure 6a Figure 6b 4 If the document is understood to be already in hand, this verb can immediately follow the verb that depicts the approach of an individual and be directed towards the imagined presence of that individual (otherwise a depiction of how the document came to be held would likely to intervene between the two verbs). Throughout this depiction the signer’s eye gaze is continually directed towards the individual, even during the production of the signs. Moreover, after being directed towards the individual, the hand can remain in place, with eye gaze remaining fixed towards this individual, depicting the length of the presentation. This is typically not possible with SHOWy. Also, the indicating verb is directed with the fingertips of the non-dominant hand (the “document”) more or less pointing towards the person being shown the object. A closely related verb has the palm facing this person, and can remain in place with eye gaze towards the same location. This is a different verb—one that passes the event sequence depiction test—in which the hands do not just represent but depict a document and a finger pointing to something on it. There are some scenes that are encoded into indicating verbs that are absent from depicting verbs. Some examples are scenes that are encoded in TELLy and EXPLAINy. Both verbs fail exhibit a hand-for-hand iconicity in the way that other indicating verbs do. TELLy has an index finger moving from the chin towards the person getting the information. This finger does not seem to represent any finger that is part of the event. EXPLAINy has F-handshapes moving bidirectionally towards an event participant, and these handshapes are also not iconic for any component within the encoded event. While both TELLy and EXPLAINy appear to lack depicting verb counterparts, there does seem to be a way to depict an event in which an individual is explaining something to someone that does not make use of constructed dialogue. In the event sequence test, EXPLAINy cannot follow TAP-SHOULDER or another sign depicting getting someone’s attention by a quick waving of the hand (Figure 7a). However, what appears to be an aspectual form of the sign can. In Figure 7b, the signer, representing the explainer, demonstrates how eye contact is made, which facial expression is used, and so on. During this demonstration, the signer also produces a sign related to EXPLAINy. This sign can be produced in a span of time similar to the time it takes to produce the citation form of the indicating verb, but it can be produced for a longer time span. This suggests that the sign is an aspectual form of the indicating verb. Figure 7a Figure 7b 5 The ability of the aspectual form of EXPLAINy to be produced as part of a event sequence depiction in which the signer represents an event participant would seem to invalidate the test as a way to identify depicting verbs. As it turns out, it rather demonstrates the greater applicability of the test. The test can be used to determine not only whether a given verb may be produced within a depiction of an event, but also whether other linguistic units or conventions may be produced within event depictions. In the rest of this paper I attempt to explicate circumstances that allow for different types of depictions to be performed in ASL. Section 3 reviews the blending model (Fauconnier & Turner 1996, 2002) that is used here to analyze the conceptual process underlying depiction. Section 4 describes the variety of depictions observed to occur in ASL discourse. Section 5 examines the restrictions on the use of linguistic items when an animate event participant is being represented by the signer, a product of conceptual integration that is labeled here as |subject|. 3. The depiction of dialogue in ASL The theory of conceptual blending (Fauconnier & Turner 1996, 2002) provides an elegant way to describe depictions in ASL. In this section, the basic blending analysis is applied to depiction of dialogue in ASL (see Liddell and Metzger 1998 for a comparable blending analysis); other types of depictions are analyzed in the next section. The depiction of dialogue is popularly known as a type of “roleshifting” and is also known as “constructed dialogue” (see Tannen 1989 for discussion of constructed dialogue in spoken languages; see Roy 1989, Winston 1991, and Metzger 1995 for discussion of constructed dialogue in ASL). Constructed dialogue in ASL is often more than just the representation of an utterance; as can be seen in exemplars of constructed dialogue in ASL, the interlocutor producing the utterance is also being represented. In Figure 8, the signer is producing a second-person singular pronoun and the non-manual yes-no question signal. The signer is also attending to an area above the location towards which the pronoun is being directed. The signer’s apparent change of attention away from the addressee while continuing to sign is one of the cues signaling that constructed dialogue is in effect. Throughout the constructed dialogue the signer is understood to represent an interlocutor and his signing represents what was said in the dialogue being depicted. A portion of the space to the signer’s right is in this depiction understood to represent the second interlocutor. Figure 8 6 Constructed dialogue is not simply a one-to-one representation. It is a creative act in which both the signer and addressee imagine the interlocutors of the dialogue being depicted as present. That this imagination is involved is clear when one sees that where the addressee would identify the “signer-as-interlocutor,” individuals without access to the discourse would identify the signer as just the signer. One way to illustrate how there can be two different conceptualizations associated with the signer is by using the conceptual blending model, which I now describe. The concepts that are blended during constructed dialogue are counterparts of two distinct mental spaces. One mental space is built up through discourse, where linguistic items introduce elements and establish relations between them. As the signer is describing a conversation between two individuals, this mental space has two elements, interlocutor 1 and interlocutor 2. If these interlocutors were introduced by name, then these elements can be labeled in the model accordingly, e.g. Tracy and Dana. This mental space is labeled Event Space in the diagrams to follow. The other mental space is different in that it is not created via linguistic means but is a conceptualization of the surrounding environment. This is Real Space (Liddell 1995). As a quick demonstration of the nature of some of the elements within this mental space, suppose someone shows you a writing instrument. You would have a Real-Space element that is a conceptualization of the instrument, but it is not the actual instrument itself. You are aware of its presence in front of you because your perceptual system takes in the external environment and creates a cognitive representation of it. Since the signer’s and the addressee’s respective Real Spaces of course differ, they have different Real-Space elements. The former has a conceptualization of the addressee, and the latter a conceptualization of the signer. Another Real-Space element is the empty space surrounding the two. In recent investigations of depiction in ASL (e.g. Liddell 2003), the Real-Space signer and surrounding space were virtually the only elements described as contributing to the depiction of dialogue (or other actions depicted to occur on a human scale). Liddell (2003:151) does not seem to make a distinction between the signer and the body, but as we will see in the next section there is good reason to do so. Moreover, other Real-Space elements described below also come into play in the depiction of dialogue and other events, and in the next section it will be clear how identification of these other elements have descriptive benefits. Part of the process in depicting the dialogue involves the creation of counterpart connections between elements within Event Space on one hand and Real Space on the other. In order to facilitate these connections, a schematization process is needed as well as a Generic Space that contains what relevant elements the inputs have in common. The Real-Space signer and the Event-Space interlocutor 1 are different concepts. The differences between the two could be that of age, gender, race, physical characteristics, and so on. Additionally, only the Real-Space signer is visible and has specific dimensions. It is possible to abstract these differences away so that the two elements are seen to be counterparts of one another. What the signer and interlocutor 1 have in common is that they are conceptualizers who are capable of experiencing thoughts, physical sensations, and so forth. There are several candidate terms with which to describe both counterparts, including “self” (see Cutrer 1994 for discussion of strong and weak versions of V-POINT). In describing the relationship between a speech act participant and the meaning of expressions produced during discourse, Langacker (2000) labels the former as subject of conception. I will use subject here to describe the Real-Space signer and potential counterparts in the other input space (and this is not to be confused with the grammatical relation with the same name). In his discussion of the viewing effects exhibited in various expressions, 7 Langacker observes how notions associated with perception have analogues to those of conception, and uses “viewer” to describe both the individual who perceives entities and the individual who apprehends the meaning of expressions. “Viewer” is as good a candidate term as “subject” is, but the latter appears to require less disambiguation. Interlocutor 2 is also a subject, but since the Real-Space subject is already mapped with interlocutor 1, there is no Real-Space subject available to integrate with it. Instead, what is recruited is a portion of empty space near the signer, labeled for convenience “spatial portion 2.” The physical space surrounding the signer is a Real-Space element that is partitionable into many different portions. The counterpart mapping between interlocutor 2 and “spatial portion 2” is in part motivated by how both can be seen to be occupy an area of space within a larger space. The particular Real-Space location that is selected is often motivated by the knowledge about the locative relationships not only between the interlocutors but within the more general setting from the perspective of interlocutor 1. Here we can see that another counterpart mapping takes place, which involves the setting elements within the input spaces. The Event-Space setting and the Real-Space setting, while distinct, are obviously ideal counterparts. Sometimes the specifics of the setting are not relevant in the depiction of dialogue, but as any dialogue, and in fact all events, must take place within a setting, the blend network underlying constructed dialogue will invariably have setting counterpart mappings. Moreover, the interaction between interlocutors is both spatial and energetic. Thus, while the signer can elect to not include details about circumstances of the event being depicted, such as the distance between the interlocutors, the signer will still need to depict a spatial relationship between the interlocutors, no matter how schematic its construal is. The side towards which the canonical “body shift” moves then can be seen here to be less motivated but not completely arbitrary. In order for the Real-Space subject to take part in depiction, a temporary (but of course not complete) disengagement with the addressee is necessary. Equally necessary is establishing a spatial relationship distinct from that between the Real-Space interlocutors, or at the very least, the relationship of the individual to a setting. For example, the signer can report to a co-host of a summer party at a home about how one of the guests is reacting to the strength of the new air conditioner: FUNNY DANA [COLD COLD] “Funny -- Dana is going, ‘It’s cold, it’s cold!’” Here at the very least, the co-host must be aware of certain circumstances existing prior to the depicted utterance, e.g. the air conditioner is currently on at full blast, but they do not need to include the specific location in which the utterance was made--Dana could be anywhere in the house (but not the yard) where the cold air is known to likely exist. Depiction does not arise simply through counterpart connections between inputs. In neither input space is there anything being represented by the signer. It is in a fourth mental space, the blend, where the dialogue and the interlocutors are represented. The counterpart elements are integrated into the blend, creating new elements. Since the signer is visible, this property is inherited by |interlocutor 1|, the only visible element in the blend. For as long as this blend is activated and is functioning to depict dialogue, the actions performed by the signer, i.e. his signing, is understood to be enacted by |interlocutor 1|. |Interlocutor 2| is the result of integrating interlocutor 2 and “spatial portion 2”, so this element is not visible. However, it does have a conceptual presence within the blend. Other than the attention given by the |interlocutor 1| to |interlocutor 2|, evidence for its presence is in the ability for the signer to direct signs towards the location where the |interlocutor 2| is conceptualized to be, whether during constructed dialogue or not. The blend also has an element, the |setting|, existing via integration of the setting elements in Event Space and Real Space, both serving as inputs to the blend. 8 Figure 9 illustrates the four-space network model of constructed dialogue in ASL. (Lines connecting counterpart elements not included for diagrammatic clarity.) Figure 9 As described above, there is more to the ASL depiction of face-to-face dialogue than just the representation of a set of utterances. At the minimum, the signer will direct his face and eye gaze to a selected portion of physical space to depict the attention given by one interlocutor to another. An example of a minimal depiction of dialogue occurs as part of a larger chunk of discourse. Here the only depiction is of someone asking “why,” followed by continued discourse which lacks any depiction: PRO-1 [WHY], PRO DON’T-KNOW “I asked ‘Why?’, and she said she didn’t know.” This is contrasted with other instances of constructed dialogue, typically with longer duration, where the signer also simultaneously depicts facial expression, body posture, and even mannerisms. These, along with the depiction of attention, complement the depiction of the dialogue itself and are actions rather than dialogue. The production of signs is also an action, but we understand the event as a whole to be one of communicating something to someone else through language. A depiction of action other than dialogue using facial expression, body posture, mannerisms, and nonlinguistic manual action might be identified as constructed action (Winston 1991, Metzger 1995). Anticipating the discussion of multiple visible blended elements in the next section, it is worth noting that the same type of blend network underlies both minimal 9 and elaborate depictions. The depictions of dialogue discussed here have only one visible element in the blend, the |interlocutor|. Any manual actions, facial expressions, and body posture that are intended as visual demonstrations are understood to be actions made by the |interlocutor|. 4. Selective Projection of Real-Space Components In the last section, constructed dialogue was described as one type of depiction in ASL, and the process of creating the depiction was explicated using the conceptual blending model. Two Real-Space elements were also introduced: the subject and the setting. What follows are descriptions of other types of depiction. As will be demonstrated, the existence of different types of depiction is a consequence of the availability of other Real-Space elements taking part in depiction as well as the selective projection (Fauconnier & Turner 1998) of these elements into the blend. It is worth detailing further what is meant by “Real-Space element.” Relevant here is the ground, which is used “to indicate the speech event, its participants, and its setting” (Langacker 1987:126). Liddell (1995:22) describes Real Space as a “grounded mental space.” In this way, Real Space is contrasted with the other input space which, in Liddell’s terms, is a non-grounded mental space containing elements of its own. As mentioned above, one useful distinction between the two input spaces is whether or not the mental space is established and structured via linguistic means. Real Space is not, as it is an individual’s mental space that is continually active within or without discourse. This mental space arises from continuously taking in of external information through the visual and the auditory systems (Langacker 1987:112). One’s conceptualization of any object perceived to be present is a Real-Space element, and this includes other individuals with whom one is interacting face-to-face. The stream of linguistic signals is also a Real-Space entity, but the resulting meaning constructions the signals prompt require an array of mental spaces containing elements that are not (at least not strictly) RealSpace elements. Other components of the ground are also Real-Space components. The signer’s selfawareness is certainly part of the ground. The signer isn’t always fully aware of the self, as is often the case during episodes of daydreaming, but when returning to the present circumstances, the signer becomes more aware of the self. This difference in awareness is a difference in how the subject of conception is construed. At one extreme the signer is fully aware of the self—and thus construed as fully objective—and at the other extreme the signer is least aware of the self— and thus construed as fully subjective (Langacker 2000). In any case, the signer always has some awareness of the self that exists as part of the ground, being the “locus of experience” (Lakoff 1996:93). This subject/self is part of one’s Real Space, unique to that individual. Also unique to the conceptualizer is the vantage point, “the position from which a scene is viewed” (Langacker 1987:123). An object within a room may be a Real-Space element to individuals, but from unique vantage points. The notion of subject necessarily subsumes a unique vantage point, as one always finds oneself at a particular location within a larger setting. We now turn to the descriptive advantage of distinguishing between the subject and vantage point where depiction is concerned. 10 Figure 10 First, it will be convenient to use simple illustrations like Figure 10 to represent blended components existing within different types of depiction. Figure 10 represents the blended components in just about any constructed dialogue involving two interlocutors. Representing the |setting| is a box-like container, and within it are two figures. One of them is the |subject|, and its visibility is indicated by the figure with a shaded fill. The second figure is the other |interlocutor|, and its lack of a fill indicates the non-visibility of the blended element. (The arrow below the diagram will be explained later below.) It should be recognized that this diagram and others below do not do much more than indicate the blended components and their schematic relation to one another, e.g. the |subject| exists within a |setting| with an |interlocutor| in front. The figure representing the |subject| is a stick figure that is standing, but signers can of course produce utterances sitting down or lying down, even during depiction. They typically also do not move away from their position when signing, even when a path movement of the |subject| is being described. In fact, someone running can be depicted by a signer who is comfortably seated. This means that often only a portion of the signer’s body takes part in the mapping creating the |subject|. Pattern completion, another cognitive process involved in blending (Fauconnier & Turner 1998), supplies whatever is necessary to successfully interpret the blend despite minimal visual information. Evidence that the Real-Space subject and vantage point are distinct components comes from a type of depiction that makes use of a setting blend. Such blends are created when signers wish to talk about objects within an environment other than their current one. One would be created when, for example, the signer is talking with an addressee about a new light fixture in a mutual friend’s kitchen. In this situation, a mental space would have been previously established, containing certain elements associated with the kitchen being described. Elements in this Kitchen Space would include the kitchen and the light fixture. Since the kitchen is an element in this mental space, signers also have access to general knowledge, or frame, pertaining to kitchens, e.g. it is a type of room with the usual walls, ceiling, floor, entrances, and so forth. Should the signer wish to describe the location of this light fixture, the ASL grammatical convention is to not produce a periphrastic construction akin to the English “on the ceiling.” Rather, the convention is to direct a single depicting verb (often with a glance) towards the imagined ceiling above the signer, as illustrated in Figure 11. 11 Figure 11 The verb in Figure 11 is used here to simultaneously depict several features of the light fixture. The handshape depicts its general bowl-like shape. The orientation of the sign depicts the direction towards which certain sides of the light fixture are facing. In this case, since the palm of the hand is facing upward, the light fixture is depicted as being convex-shaped. If it is possible for the downward-facing palm to be used in a similar depiction, we would understand the light fixture as being concave-shaped. The location of the light fixture is depicted via the directionality of the verb: the verb is directed upwards and slightly away from the signer. To be able to direct this sign, typically a Real-Space blend needs to have been previously created (see Liddell 2003:154 for a similar observation). Once the signer imagines the presence of a |kitchen ceiling|, the verb can then be directed towards the blended element. Otherwise it is not possible to use the verb (except in cases where the signer is talking about an actual ceiling directly accessible to the interlocutors). The use of this type of depicting verb is often an indicator of a Real-Space blend, and in this case it is clear that the signer has created a Kitchen blend. The Kitchen blend is an instance of a setting blend. It is created by integrating the kitchen from the Kitchen Space input with a portion of the Real-Space setting into a blend, creating a |kitchen|. The |ceiling| in this blend is the result of integrating the ceiling (accessible via the kitchen frame) in the Kitchen Space input with a portion of space above the signer. The set-up of the |kitchen| and |ceiling| is not possible without the signer’s selection of a location within the kitchen from which to locate the light fixture. There are many possible locations within the kitchen that can be selected for this purpose. Once a location is selected, it integrates not with the Real-Space subject, but with the Real-Space vantage point alone. The signer does happen to be standing in what is understood to be the |kitchen| and is also glancing at the |kitchen ceiling| while directing a verb towards it. However, this does not mean that the Kitchen blend contains a |subject|. There is no animate being in the Kitchen Space input that is a candidate counterpart to the Real-Space subject. Moreover, the blend is created to depict a setting rather than an event, and as discussed below, setting blends appear to preclude the existence of a |subject|. Since the Real-Space vantage point is a type of location, virtually any location within the kitchen is a suitable counterpart. The vantage point is not merely a location on the floor, but a threedimensional concept. To quickly illustrate, consider the differences in the articulation of a verb produced by a very young child and a tall adult to depict the location of a painting on the kitchen wall. The child would produce the sign above eye level referencing the actual painting’s location in space whereas the adult would produce the sign at eye level. This is obviously a consequence of the different, three-dimensional vantage points that the two have, which is integrated with a 12 location in the kitchen to create the respective |vantage points|. Should the child or adult depict the other’s experience of the kitchen, their articulations would change accordingly—the adult would raise both arms to depict the location of the |painting|, but the child wouldn’t need to. Figure 12 Figure 12 represents setting blends that have life-sized dimensions. As with the constructed dialogue representation in Figure 10, the |setting| is represented by a box. Instead of the figure with a shaded fill used to represent a |subject|, a dotted figure icon is used here to represent the |vantage point|. These distinct representations should not be taken to mean that the |subject| lacks a |vantage point|, but as mentioned above, the former actually subsumes the latter. It is not possible for a subject to lack a vantage point because the conceptualization of one’s location within a larger setting is an essential component of self-awareness. However, as we have seen, it is possible to project the Real-Space vantage point independently of the Real-Space subject. This is an example of selective projection: in certain types of depiction, only a number of Real-Space components are selected to integrate with its counterparts. Constructed dialogue has a |subject| with a concomitant |vantage point|, but life-sized setting blends only have the latter. Since both types of depiction have |settings| with life-sized dimensions, they are examples of what Emmorey and Falgier (1999) call viewer space. Surrogate space is the term used by Liddell (1995), although in my assessment, it has been used to describe only similarly-scaled blends that depict dialogue or action. Setting blends can also be created with a smaller portion of space in front of the signer. In this blend, the dimensions of the |setting| and entities within it are scaled down. This is the result of the cognitive process of compression (Fauconnier and Turner 2002). The depicted setting is compressed with the smaller portion of space into the blend. Because this physical space does not include the area where the signer is located, the Real-Space vantage point is not available to take part in the mappings that create this blend, so the blend lacks a |vantage point|. The signer’s Real-Space vantage point remains in effect. In contrast to viewer space, this is a diagrammatic space (Emmorey and Falgier ibid.), represented by the small box in Figure 13. The benefit (and necessity) of compression can be seen in the depiction of the spatial relations of planets or galaxies. Components of the entity being depicted are compressed with its smaller Real-Space counterparts into the blend, allowing for a human-scale conceptualization that would otherwise not be possible. 13 Figure 13 It appears that the blended element |setting| is a constant among the various types of depiction explicated in this paper. Since dialogue necessarily takes place within a setting, the depiction of dialogue requires a |setting| element. This is also true not just for constructed action but also for any event depiction. In addition to |setting|, what is also constant among the various types of event blends to be described shortly is a |temporal progression| element (and this also appears to be what basically distinguishes setting and event blends). As described in Dudis (2004b), this element is the result of integrating temporal progression elements from two inputs. The temporal element in Real Space is the currently experienced progression of time. Similar to the varying degrees of awareness we have of our selves in any given moment, we are also not continuously aware of the progression of time. There are certain experiences where it does come to the forefront, as when one wishes to adjust the blinking rate of the cursor on a computer monitor. The counterpart to this Real-Space temporal component is Event Time, i.e. the temporal progression associated with the event being described. |Temporal progression| is represented in the diagrams in this section by an arrow below the box, as in Figure 10 above. Events are often depicted from the viewpoint of a |subject agent|, the result of integrating the Real-Space subject with the participant with the status of agent in the Event-Space input. Since other animate event participants are also subjects, they are potential counterparts to the Real-Space subject. In fact, we do see depictions from the viewpoint of, say, the |patient|. What is interesting in these non-|agentive| viewpoint depictions is that there is typically another visible blended component (where in, say, constructed dialogue there is only one visible blended element). For example, in a depiction of a punch, the signer could direct a fist towards his chin. Considered in isolation, there are two ways to interpret this depiction. One is that the signer is depicting someone punching himself. In this case, there is only one visible blended element, the |self-puncher|. Another interpretation is that the signer is understood to be the |victim|, and that the |fist| is thrown by an |attacker|. This interpretation requires partitioning (Dudis 2004a) of the signer’s fist from the rest of the signer’s body. The fist then becomes available to integrate with the attacker’s fist, while the Real-Space subject is integrated with the victim. One of the |victim’s| hands is not visible, but through pattern completion, it remains conceptually present in the blend. With partitioning and pattern completion, the signer has the ability to create distinct visible components in event blends while allowing a single viewpoint to be maintained throughout the depiction. It is also possible to partition the portion of the face that is used to create facial expressions as well as the lower region of the face that includes the mouth and cheeks. The whole face (excepting the eyes) can take part in the creation of a visible |facial expression| that is distinct from the |subject’s| (which would not be visible, at least not fully), as 14 can be seen in the depiction of someone noticing that a glare is being directed towards him. The partitioned-off mouth can be used to produce what appear to be onomatopoeic units depicting aural/vibratory-related phenomena within events, e.g. one produced when two objects collide. Other types of event depiction are possible through the selection of fewer Real-Space components. Events that do not have an animate participant, e.g. a lightning striking a tree in a woods during a storm, can be depicted with the Real-Space setting, vantage point, and temporal progression. It is not necessary for the signer to witness the event firsthand to depict it; a reconstruction of the event in the aftermath of the storm or even imagination alone is sufficient. The Event-Space tree and woods would be integrated with its spatial counterparts in Real Space to create |tree| and |woods|. A location v (this is just a label for a selected location within the setting) would be integrated with the Real-Space vantage point, resulting in a |vantage point| from which the lightning strike is described. The lightning bolt’s action is depicted by a verb in which at the |lightning bolt’s| path is visible. This blend parallels the viewer kitchen blend in the following ways: both have a |setting| and a |vantage point|, and both depict a relationship between a |figure| and a |ground| (|light fixture & |ceiling|; |lightning bolt| & |tree|). A key difference between the event blend and the setting blend is the existence of a blended temporal component in the former. The extent to which the lightning strike blend, represented in Figure 14, and the kitchen blend parallel can be seen in the comparison of the diagrams. Both Figure 14 and Figure 11 are basically identical save for the arrow in the former, which represents an integrated temporal component. Figure 14 There is an association between the Real-Space temporal progression and subject comparable to that between the Real-Space setting and vantage point. The Real-Space vantage point is a particular location within a larger Real-Space setting, and the former clearly cannot exist outside of the latter. The adoption of an imagined vantage point is not possible without an imagined setting. Thus, if the signer creates a viewer space blend, we know that the signer has integrated not only setting components but also the Real-Space vantage point with a location within the depicted setting. Such attendant integrations do not occur in diagrammatic setting blends (Figure 13), in which a setting is depicted without integration of the signer’s vantage point into the blend. The Real-Space subject is characterized above as having self-awareness, and being self-aware is an activity that takes place over time. More generally, a span of time is needed to entertain any concept. Langacker 1987 distinguishes this processing time from conceived time. For example, one can visualize leaves on a tree changing from green to red and orange. The rate of change can be adjusted, resulting in visualizing colors as bursting out all at 15 once or gradually emanating leaf by leaf. These different rates of change involve different conceived times, extending through the conceptualizer’s processing time. Processing time is required for the signer to imagine setting blends, but since no event occurs within these blends, conceived time would not have a role within them. A conceived time, |temporal progression|, is necessarily a component in event blends containing a |subject|, who is aware of the self and other entities. This awareness or any activity cannot be depicted independently of a |temporal progression|. Thus, when it is clear that a blend has a |subject|, we know that there is attendant integration of temporal progression components. |Temporal progression|, as illustrated with the depiction of the lightning strike above, can be created independently of a |subject|. Three other types of event blends observed in ASL involve compression. Figure 15 is an example of a diagrammatic event blend used to depict a car passing another car. Figure 15 Here the two manual articulators are linguistic units and in this instance are used to depict the action of the cars. While there are |drivers| in this blend, conceptualized to be inside the |cars|, there is no |subject|. In fact, it does not appear possible for any diagrammatic event blend to have a |subject|. The only Real-Space elements that take part in this depiction are the depicting units (the manual articulators), the setting, and temporal progression. Because the depiction makes use of only a portion of the space in front of the signer, there is no evidence suggesting that the signer’s vantage point is integrated with a particular location in the setting of the event. As seen in Figure 15, the signer’s eye gaze is on the interaction of the two cars, making it clear that the signer is not representing the driver of either car. Had the eye gaze instead been consistently directed straight ahead instead, this would likely to be interpreted as the actions of a driver and thus constitute evidence suggesting the existence of a |subject|. The manual articulators do not have the capacity for self-awareness, so they are not regarded as subjects and their integration with the cars in the Event-Space input result in blended elements that are not subjects. This is true even if the manual articulator that is conventionally used to depict an upright human (the index finger) integrates with an Event-Space subject. Only when the Real-Space subject integrates with an Event-Space counterpart will there be a |subject|, and this is possible only within a viewer space. Apparently, that only one |subject| can exist within any instance of depiction is also true even when a viewer event blend and a diagrammatic event blend exist concurrently. The present analysis suggests that this involves the partitioning off of the manual articulators to create visible 16 blended elements existing only within the diagrammatic space. To illustrate, the signer in Figure 16a is depicting someone riding uphill on a motorcycle. Figure 16a Figure 16b Only one blend is used to depict this scene, and it contains a |motorcyclist subject| who is understood to be steering a non-visible |motorcycle|. Then to add further details to the ongoing depiction, the signer creates a diagrammatic blend. The diagrammatic blend is visually manifested when the signer partitions off the manual articulators to depict a motorcycle moving up on a hill. The 3-handshape in Figure 16b is similar to those in Figure 15, but is here used to represent the motorcycle. The flat-B handshape is used to represent a portion of the hill. Based in part on the placement of the 3-handshape, which is close and perpendicular to the signer’s chest, the |motorcycle| is understood to be the visual blended counterpart to the non-visible |motorcycle| in the viewer blend. This relationship would not be possible had the |motorcycle| been placed even a few inches further away from the signer or to either side of the signer. The visible |motorcycle| in the diagrammatic blend would then be another motorcycle different from the one that the |motorcyclist subject| is on. This blend serves as a global perspective counterpart to the participant perspective within the viewer blend. The actions of the visible |motorcycle| in this blend provides direct information whereas related information in the viewer blend can be obtained only indirectly. Figure 17 diagrams the co-existence of the viewer and diagrammatic spaces and the blended components within them. Since both spaces are event blends, an arrow representing |temporal progression| is included in both blend representations. Figure 17 17 The other two event blends are viewer and diagrammatic blends in which the Event Time is compressed with a shorter span of the Real-Space temporal progression. It is probably the case that many event blends exhibit temporal compression, but determining whether there is temporal compression is not as easy as it is with setting compression. However, the use of temporal aspectual constructions or certain head movements accompanied by facial non-manual signals (e.g. quick nodding of the head with eyes squinting and mouth slightly open) are good candidates for indicators of temporal compression as described here. Dudis (2004b) describes how some conventional head movements in ASL accompany gradual change-of-state depictions. Soon after depicting the causing event, the signer makes a slow, backward head movement, and then produces a quick return of the head to at or near its neutral position while simultaneously producing a sign describing the change. In the ASL expression more or less equivalent to the English expression “I practiced until I became good at it” (Figure 18), an aspectual form of PRACTICE is produced with non-manual signals suggesting the existence of a |subject|. During the production of the aspectual form, the head moves slowly back, a gesture suggesting the recognition that a change is underway. In the last part of the expression, the production of the sign SKILL and the head’s return to neutral position or near it occur simultaneously. An actual building of a skill requires more than a few seconds, but this span of time is exactly how long it takes to produce the expression. This is clear evidence of a compression of temporal progression. As this construction can be used to depict other changes-of-state, we can take the head non-manual signal just described as an indicator of this type of compression. Figure 18 The last Real-Space element to be described in this paper is the body itself. Liddell 2003:141-142 describes how a signer makes use of a Real-Space blend, replicated in Figure 19, to describe the preparation of a fish before cooking it. The signer places the tip of a B-hand on the top of her chest near her throat and moves it down to her waist. Since the signer is specifically talking about slicing the fish, we understand that the B-hand is used to depict something related to the slicing, either a part of the knife and its slicing motion or the path of the slice and perhaps its depth. We also interpret that the signer’s chest is used to depict the underside of the fish. The underbelly of the fish in the Fish Space is mapped onto the Real-Space signer’s chest, and their integration results in the second visible blended element, |fish underbelly|. What about the signer’s looking downwards? Aside from those in cartoons, fish can’t watch themselves being sliced, so it is safe to say that there is no |subject| in this blend. So 18 here the signer’s head is just the signer’s head. Because fish take up space and events take place in settings, a |setting| is posited, albeit an abstract one (it seems highly doubtful that signers would imagine a surface of, say, a table on one of the signer’s sides). Since the blend created is not a diagrammatic space, but appears to be a larger-than-life-scale viewer space, it is clear that there exists a |vantage point| in this blend. Similar examples of viewer space blends containing blended |bodies| but not |subjects| are often found in descriptions of how people have been injured, received scars, and so forth. In this section, I have demonstrated that the various types of depiction observed to occur in ASL discourse are the result of the selective projection of Real-Space elements, namely the setting, vantage point, temporal progression, the subject, and the body, along with cognitive abilities including the ability to partition the body into several meaningful zones, to compress the setting and time of the scene being depicted, and to create simultaneous blends. The different types of depiction are listed in Table 1 below. Since any event blend can have a compressed |temporal progression|, no separate item indicating temporal compression is necessary. Also, the |subject| is not listed with diagrammatic blends because this blended element can only be part of a viewer event blend. Types of Real-Space blend used in depiction Viewer event blend with |subject| Viewer event blend without |subject| Viewer setting blend Diagrammatic setting blend Diagrammatic event blend Table 1 5. Restrictions on the use of linguistic items when a |subject| is active Conceptual blending analyses of ASL expressions also have shed further light on the relationship of ASL linguistic units to event blends. Of interest here is the set of restrictions that is typically in effect when a |subject| is active. As described below, these restrictions are related to the activation of a |temporal progression| and the type of event blend created. These demonstrate why the informal event sequence depiction test described in Section 2 is useful in identifying whether a linguistic unit is one that actually depicts an event: if the sign is associated with |temporal progression|, then the sign further contributes to the overall depiction of the event. Fauconnier (1997) describes discourse as comprising several mental spaces within a single lattice. The discourse participants are described as navigators through this lattice, the producer of the discourse building spaces and guiding addressees through the lattice via a variety of cues, both linguistic and non-linguistic. Discourse begins with a base space (Fauconnier ibid.), a “privileged set of mental-space structures which [the speaker] understands as corresponding to her actual experience and/or to situations which she believes actually prevail or took place in the past” (Dancygier and Sweetser 2005:31). (Real Space could be regarded as part of the conceptualizer’s base space.) Distinct mental spaces are created relative to not just other mental spaces but to this base space as well. Where constructed dialogue is concerned, the Event Space 19 that consists of elements related to the dialogue is established relative to the base space (the term “Event Space” used in this paper is used as a label for input spaces that are parents of event blends and is not the “discourse primitive” discussed in Cutrer 1994 and Fauconnier 1997). Thus (at least) three mental spaces have roles in the depiction of dialogue. Among the different roles mental spaces have in discourse is one in which a mental space is being structured. This is the focus space, “the current, most active space; the space which an utterance is ‘about’” (Cutrer 1994:71). An expression in which dialogue is depicted appears have both the Event Space and the event blend as active mental spaces. I will leave in-depth examination of focus spaces in dialogue involving depiction to future investigation. It is sufficient to note the following: when an event blend is created, it becomes part of the lattice of mental spaces created within a discourse; having been created, it remains accessible throughout the discourse in question; and the moment the signer ceases depicting an event, the event blend is deactivated. This event blend can be reactivated (or perhaps becomes a focus space once again), allowing the signer to continue the discourse with further depiction. It is also possible to have two different event blends in one stretch of discourse (Liddell and Metzger 1998; Liddell 2003). Figure 19 is a diagram of a discourse that I analyze as having two event blends created to depict a dialogue between Dana and Tracy. Figure 19 20 There is an Event Space set up relative to a Base Space. Because the interlocutors are known to the signer, they are elements of the Base Space. The “same” interlocutors are being talked about in a context different from the Base Space, one that is contained within the past Event Space. Once one of these interlocutors are identified, perhaps via a noun phrase, the signer can change posture from one facing the addressee to the posture in the video still near the blend labeled Event Blend 1 (EB1). For as long as the |subject interlocutor| is understood to be active and producing dialogue, EB1 (or perhaps more precisely, the |temporal progression| element) remains active. A variety of cues signal the deactivation of the |temporal progression| element. The eyegaze with the addressee is re-established and, simultaneously, the signer’s posture is no longer aligned with the |subject interlocutor|. If the posture now assumed is one previously held by the signer prior to the creation of EB1, then it is clear that depiction has ceased. A new posture, as seen in the video still near Event Blend 2 (EB2), would signal the creation of a new blend, and this blend may be preceded by linguistic units identifying interlocutor 2 as the dialogue participant being depicted. These three postures and associated cues assist the signer in moving between mental spaces. Often the signer alternates between the postures associated with the event blends without returning to the neutral signer posture, a deactivation-reactivation process between EB1 and EB2. When an event blend created to depict dialogue is active, virtually any portion of the expression the signer makes is understood to be the dialogue of the |interlocutor|. As discussed previously, since a portion of the Real-Space body, especially from the head to the lower torso, is integrated with the Event-Space interlocutor, any action of the Real-Space signer is understood to be the action of the |interlocutor|. Whether the signs being produced are depicting is irrelevant in this case. What is being depicted is the act of dialogue making, which includes both depicting and non-depicting signs. This is one way that signs are associated with |temporal progression|. This does not require more than one event blend, and so compared to the other types of event blends discussed below, the depiction of dialogue is a relatively straightforward process. Also exhibiting a straightforward blending process is the type of event blend in which a manual action is performed by a |subject|. The manual action depicted may or may not be considered to be an instance of a depicting verb. Figure 20 illustrates a verb that depicts the action of opening a capped bottle. Figure 20 21 The manual articulators appear to be the |hand holding opener| and the |hand holding (neck of) bottle|, respectively. The |bottle| and |opener| are blended elements that are not visible but are nevertheless conceptually present. Note that the two |hands| contact each other. This contact doesn’t typically happen in actual opening of bottles. This does not make the verb a nondepicting verb, but it can be taken to indicate the verb’s unit status rather than simply a nonlinguistic gesture. The opening-bottlecap depicting verb just described is an example of what is known as a handling classifier construction. “Handling” is used to label this type of verb because these verbs have a visible |hand| that is understood to be holding or handling |something|. The latter blended element is non-visible, but the configuration of the |hand| indicates part of the |held object’s| shape. Figure 21 is another verb of this type, one that depicts a gun being held. A different depicting verb is illustrated in Figure 22. Figure 21 Figure 22 The verb in Figure 22 is similar in many ways to its counterpart except that it depicts the object held rather than the hand that holds it: here we have two visible elements, the |subject shooter| and the |gun|. This type of verb is known as an instrument classifier construction, and using it within an event blend with a |subject| requires the partitioning of the signer’s hand. The blend described in Section 4 in which a |subject victim| receives a |punch| also involves the partitioning of the hand, but the partitioning is only required if the punch is made by someone other than the |subject|; this is also true for “hand-depicting” verbs. “Instrument-depicting” verbs necessitates partitioning when produced with a |subject| active, regardless of whether the |subject| is understood to be holding the object or not. The differences between these two types of depicting verbs notwithstanding, only a single event blend is necessary for either of these two verbs to be used. In the case of “hand-depicting” verbs, the |hand| is part of the |body| that the |subject| has, and any |held object| is not visible but is present as an element within the blend. In the case of “instrument-depicting” verbs, the |instrument| is visible, but the |hand| holding it is not. We are already beginning to see the differences between expressions depicting dialogue and expressions depicting manual or instrument actions. While it is not obvious how different the event blends within these expressions are from each other, it is clear that they differ in the type of linguistic unit used. Again, any sign produced within a blend created to depict dialogue will be understood to be part of a |subject’s| dialogue. This includes nouns like GUN (which is a fingerspelled loan sign) and relations like SILVER (as it is not absolutely clear that this sign is 22 an adjective, I instead use “relation,” a Cognitive Grammar term that describes the class of verbs, adjectives, adverbs, and prepositions). Since both GUN and SILVER do not depict anything, they would not be able to be produced in a blend created to depict events other than dialogue. The two signs can be produced in the event sequence test only as constructed dialogue. The verbs in Figures 21 and 22 can be used in event blends, not just because they are depicting verbs. The verb illustrated in Figure 11 (repeated below as Figure 23) is produced to depict the shape and location of the light fixture, but not any action. Hand- and instrument-depicting verbs are event-depicting verbs. Since they depict events, they have a |temporal progression| element. In the same way that the verb in Figure 23 requires a setting blend to be active for it to be produced, these verbs require an event blend to be active for them to be produced. The event blend need not be made manifest to the addressee prior to the use of event-depicting verbs. The appearance of such verbs, with concomitant non-linguistic cues, is sufficient for the addressee to create or activate an event blend within the mental space lattice. Figure 23 It is possible to produce the non-depicting sign SILVER with the hand representing a gun or a hand holding a gun remains in place. Whether an event blend is still active depends on what the signer is doing with SILVER. If the signer is depicting how the person with the gun is describing the gun, e.g. “Here’s the gun, it’s silver all over,” then SILVER is produced as part of constructed dialogue, and this requires activation of an event blend. Only one event blend is needed to depict both the dialogue and the gun in hand. If the signer and not the |subject| is describing the gun, then |temporal progression| has been deactivated. Deactivation of |temporal progression| does not necessarily deactivate other elements in the blend. While the |subject| would be deactivated, the visible |hand| or the visible |gun| can remain active because they are the result of integrating an Event-Space component with the Real-Space body, which is conceptually independent from temporal progression. The signer can then talk about the gun using signs like SILVER and directing signs towards the |gun|. This is an example of a discourse strategy in which |temporal progression| is deactivated while keeping not the event-depicting verbs active, but the visible blended elements associated with them. The relationship between the event-depicting verb that signers know and the actual use of the verb within an event blend can be described in Cognitive Grammar terms (Langacker 1987, 1991) as a schema-instance relation. In the view of usage-based models of grammar, of which Cognitive Grammar is one, linguistic units are obtained in part through recurrence of actual 23 usages of these units. Non-recurrent features are abstracted away during the process in which expressions gain unit status. In the manual- and instrument-depicting verbs described above, it appears that the unit(s) which they instantiate do not specify where the hands should be directed. Rather, directionality is schematic, even though it is an essential feature of the event-depicting verb schema. The actual use of these verbs are instances of the verb schema, and their directionality is for the most part motivated by the signer’s knowledge of the event being depicted (e.g. whether the gun is aimed to a side, towards the side, etc.). It appears that the schematic features of these verbs are what Liddell (2003) is referring to in his discussion of gradience in depicting verbs. Since instances of manual- and instrument-depicting verbs consistently activate |temporal progression|, this is evidence that when it comes to these verb types, this activation is not optional. This means that |temporal progression| is a component in the event-depicting verb: to use an event-depicting verb is to create or maintain an event blend with at least one visible blended element. In the case of manual-depicting verbs, at the very least the |agent’s hand| is visible. The |subject agent| can be visible, but through partitioning the signer’s Real-Space hand can be the only visible part of the |agent| acting on the |subject patient|, as in the depiction of someone receiving a punch (Figure 24). In the case of instrument-depicting verbs, the |instrument| is the only visible element, but a |subject| can be, and often is, visible, for a total of two visible blended elements in the actual depiction. Figure 24 Figure 25 24 These two types of event-depicting verbs are also known as classifier constructions, and for those classifier constructions that are used to depict events, a similar analysis would apply. The expression partly illustrated in Figure 15 (repeated here as Figure 25) has been labeled as “whole-entity” classifier constructions. Since their use involves the activation of a blend with a |temporal progression| element, the schematic unit that sanctions the use of these constructions has this element as a component. As these constructions activate a diagrammatic blend, they do not have a |subject| component. These verbs could be regarded as “previously partitioned.” If an animate participant is also part of the scene being depicted, this allows the creation of a blend with a |subject|, resulting in multiple visible blended elements. Figure 26 ASL also appears to have event-depicting verbs that would not traditionally be regarded as classifier constructions. BE-TAKEN-ABACK and DO-THINGS (Figure 26) are two such verbs. When either sign is produced, the various non-manual cues (eye-gaze shift, posture shift, change in facial expression, etc.) are exhibited, signaling the existence of a |subject|. BE-TAKENABACK seems to be iconic for an individual experiencing something so intense that she throws up her arms in order to regain her balance. In contrast, it is not immediately obvious what the manual articulators of DO-THINGS are iconic for, much less what they depict. It is not sufficient to just consider these articulators as partitioned off from the |subject|. This would not reconcile the manual sign with the |subject| in accordance to the claim put forth in this section. My preliminary analysis of DO-THINGS is that it is a verb that depicts someone doing a number of things (some if not all involving the use of the hands). There are some features of the manual sign that, upon closer examination, do appear to exhibit iconic mappings. The palms of the manual articulators are facing away from the signer, suggesting contact or interaction with some entity. The sign’s movement to one side and then towards the other suggests interaction with more than one entity. The wiggling of the fingers in this sign is also found in a sign that has the meaning “some time later.” The length of the sign is relatively protracted compared to nondepicting signs, and is itself one indicator of |temporal progression|. If this analysis were borne out, this sign would then be evidence that—in addition to body partitioning and compression of setting—temporal progression plays a role in the creation of signs. Interestingly, DO-THINGS would also be a candidate sign that exhibits the compression of setting into a viewer blend: the schematic activity being depicted is not necessarily limited to the space in front of the |subject|, but may be understood to encompass a whole room or a variety of settings. 25 Aside from those expressions that create diagrammatic blends with distinct, visible elements (e.g. Figure 25), most of the signs described thus far are likely to be viewed as monomorphemic if one follows Liddell’s 2003 approach to the analysis of depicting verbs. Again, such linguistic units have schematic components that prompt the mapping of semantic components onto Real-Space elements. Given the Cognitive Grammar view, we can then say that depiction has a role in the lexicon of ASL, where concepts such as |setting| and |subject| are essential components of certain units. Symbolic units are viewed as comprising a network in which units at the lowest level have the greatest semantic and phonological specifications and those at the higher levels are successively more schematic with respect to semantic and phonological information. Does the role of depiction extend into the grammar of ASL? Cognitive Grammar and constructional approaches to grammar (e.g. Goldberg 1995) do not assume a strict division between lexicon and grammar, maintaining that phrasal and clausal units are to be treated as form-meaning pairings as well. This means that if we find depiction to have a role in multi-morphemic or multi-word expressions, this would suggest that depiction has a role in grammar as well. Figure 27 Figure 7b, repeated here as Figure 26, partly illustrates the aspectual form of EXPLAINy. As discussed in Section 2, it is not readily apparent as to what the manual articulators of EXPLAINy are iconic for, and this is still true with this form. However, there is a visible |subject explainer|, the result of integrating the Real-Space signer and the explainer, one of the event participants encoded in the verb EXPLAINy. This |subject| and other features, including the protracted length, suggest that the aspectual form is comparable to DO-THINGS, except that the aspectual form is analyzable as being multi-morphemic. As mentioned earlier, it is not possible to produce EXPLAINy in an event blend aside from instances of constructed dialogue, so this sign would not be analyzed as having a |subject| or a |temporal progression| component. The aspectual form can, and the obvious source is the aspectual constructional schema that the form instantiates. This constructional schema is a productive one, as there are several instances of this schema, e.g. the durational forms of WAIT and USE-KEYBOARD, all exhibiting cues associated with either the |subject| or |temporal progression|. (Klima and Bellugi 1979 and Liddell 2003 use “durational aspect” to label one type of aspectual construction, which seems appropriate for the aspectual form being described here.) Other aspectual constructional schemas, including the unrealized inceptive constructional schema (Liddell 1984), also are analyzed as having at the very least a |temporal progression| component. As this is where 26 morphological processes come in, this suggests that depicting components are found not only in the lexicon but grammar as well. Returning to the claim being developed in this section, the non-iconicity of the manual articulators of EXPLAINy or other component sign involved in the creation of aspectual forms does not prevent an event blend from being active. This is because the articulators are now part of a depicting expression produced via integration of a verb with the aspectual constructional schema. Thus this is a case of the schema-instance relation which satisfies the provision that linguistic units or gestures produced with a |subject| active be associated with |temporal progression|. The non-depicting nature of the articulators does not mean that they by themselves contribute nothing to the blend. Rather, they can be seen as temporal markers whose extended presence maps iconically onto the duration of the event being depicted. Further evidence that depiction has a role in the grammar of ASL is found in ASL changeof-state constructions (Dudis 2004b). Figure 18, repeated here as Figure 28, partly illustrates a depiction of an event in which an individual practices some skill, e.g. shooting baskets or fingerspelling, until a higher level of competency is achieved. I analyze this expression as being an instance of a gradual change-of-state construction. Figure 28 The first part of the construction has an aspectual form, here a durational form of PRACTICE. While continuing to produce the aspectual form, the signer slowly moves his head backward, which depicts not the actual moving of a head, but a recognition that a change is beginning to take place. Near the end of the expression the signer produces BE-SKILLED with the head returning to neutral position. Since neither PRACTICE nor BE-SKILLED are depicting units, the depiction in Figure 27 is attributed to the complex constructional schema. The association the articulators of PRACTICE have to |temporal progression| is similar to the one the articulators of EXPLAINy have in the aspectual form. This suggests that an aspectual construction is part of this change-of-state construction. BE-SKILLED describes the resulting change and comprises the final part of the manual portion of the construction. Its production can be seen to mark the point in the |temporal progression| when the anticipated change is complete. The set of non-manual signals is a component of the change-of-state constructional schema. It appears that the phonological information for the non-manual signal during the initial phase is schematic, at least where the head movement is concerned, but is specified for the following phases (the head moves back slowly, etc.). Again, the head movement is associated with the 27 appearance of the anticipated change. In all, there is ample evidence that this constructional schema is a depicting unit in the grammar of ASL. That this construction depicts a schematic change-of-state explains why its instances can be produced with a |subject| active. Thus far we have seen that non-depicting items can be produced when a |subject| is active as long as they are associated with |temporal progression|, either as part of constructed dialogue or as an instance of a linguistic unit that has a |temporal progression| component. A third possibility has already been hinted in the description of the non-manual component of the change-of-state construction, which is associated with the recognition of the appearance of a change and the tracking of its progress. In addition to depicting dialogue and external events, it is also possible to depict psychological events. Figure 29 partly illustrates an expression depicting someone noticing, upon arriving at a shopping plaza, that two stores are unexpectedly closed. Figure 29 Here a |subject| clearly is present. The sign BE-CLOSED is directed first towards the signer’s left and then towards his right, both accompanied by eye gaze towards the same respective direction. It is also clear that the signs are directed where the two |stores| are conceptualized to be. Following Liddell (2003:179), the BE-CLOSED sign is analyzed here as not an instance of indicating verb or a depicting verb, but a plain verb that the signer simply points towards a blended element. Here the |subject| is not signing to anyone in particular. If this is instead a depiction of inner speech, intuitively it is a sparse depiction compared to other instances of inner speech depictions. There is an alternative analysis. BE-CLOSED is produced to represent the |subject’s| perception of the scene. Since the sign BE-CLOSED is not part of constructed dialogue (or an event-depicting verb), the manual articulators are understood to be partitioned off from the |subject|. The mental space it is part of is separate from the viewer space in a way similar to the diagrammatic blended space’s independence from the viewer space in the depiction of a motorcyclist riding uphill. The difference is that BE-CLOSED is not a diagrammatic space, but a perceptual space associated with the |subject|. BE-CLOSED is then associated with |temporal progression|, in spite of its being a non-depicting unit. A perceptual space relative to an event blend is also created to depict the |subject’s| focusing on an |object| within a |setting|. Figure 30 partly illustrates an expression depicting someone seeing newly-installed light fixtures on a ceiling. Here it is clearer that dialogue is not being depicted. A manual articulator is partitioned off from the |subject| to produce an instance of the same verb described in Section 4, one that does not depict an event, but the shape and 28 location of an object, in this case a light fixture. As this verb does not depict an event, it lacks a |temporal progression| component. However, as it is understood to depict what the |subject| is seeing, it then has an association with |temporal progression|. Figure 30 The event blend here has elements similar to those found the setting blend created to depict the location of the light fixture (described in Section 4). Without access to previous information, it would not be possible to determine whether an event or a setting alone is being depicted. A related ambiguity is whether the viewer blend contains a |subject| or simply a |vantage point|. This demonstrates not only how context is an essential part of the expression, but also the utility of identifying the distinct conceptual components involved in the depiction of setting and events. I hope to have demonstrated in the above that when a |subject| is active, specific restrictions on the use of signs are in effect. Table 2 lists these restrictions. Signs produced when a |subject| is active Sign is part of dialogue (or gesture) being depicted Sign depicts psychological event experienced by |subject| Sign is an instance of a schematic unit containing a |temporal progression| component Table 2 6.0 Conclusion The depiction of settings, objects, and events explicated in this paper are clearly staples of ASL discourse. When ASL verbs and constructions are shown to have components that depict semantic features, depiction becomes a focus of grammatical analysis from the cognitive linguistic perspective discussed here. The specific approach being established here, begun by other cognitive linguists investigating signed languages, demonstrates the significant potential through further analysis for elucidating of the role depiction has in ASL grammar. This approach lends support to the view that while there are some facets of ASL grammar that “submerge” iconicity (Klima and Bellugi 1979), other facets exist where iconicity “emerges” (S. Wilcox 2004). 29 Notes This research was supported in part by the National Science Foundation under grant number SBE-0541953. Any opinions, findings, and conclusions or recommendations expressed are those of the author and do not necessarily reflect the views of the National Science Foundation. Paul G. Dudis Department of Linguistics Gallaudet University 800 Florida Avenue NE Washington, DC 20002 paul.dudis@gallaudet.edu 30 References Cutrer, Michele. 1994. Time and Tense in Narratives and Everyday Language. Doctoral dissertation, University of California at San Diego. Dancygier, Barbara and Eve Sweetser. 2005. Mental Spaces in Grammar: Conditional Constructions. Cambridge: Cambridge University Press. Dudis, Paul G. 2004a. Body partitioning and real-space blends. Cognitive Linguistics 15:2, 223238. Dudis, Paul G. 2004b. Depiction of Events in ASL: Conceptual Integration of Temporal Components. Doctoral dissertation, University of California at Berkeley. Emmorey, Karen and Falgier, Brenda. 1999. Talking about Space with Space: Describing Environments in ASL. In E.A. Winston (ed.), Story Telling and Conversations: Discourse in Deaf Communities. Washington, D.C.: Gallaudet University Press, 3-26. Fauconnier, Gilles. 1997. Mappings in Thought and Language. Cambridge: Cambridge University Press. Fauconnier, Gilles and Mark Turner. 1996. Blending as a central process of grammar. In Adele Goldberg (ed.), Conceptual Structure, Discourse and Language. Stanford, CA: CSLI Publications, 113-130. Fauconnier, Gilles and Mark Turner. 1998. Conceptual integration networks. Cognitive Science 22:2, 133-188. Fauconnier, Gilles and Mark Turner. 2002. The Way We Think: Conceptual Blending and The Mind’s Hidden Complexities. New York: Basic Books. Goldberg, Adele. 1995. Constructions. Chicago: The University of Chicago Press. Klima, Edward and Ursula Bellugi. 1979. The signs of language. Cambridge: Harvard University Press. Lakoff, George. 1996. Sorry, I’m not myself today. In Gilles Fauconnier and Eve Sweetser (eds.), Spaces, Worlds, and Grammar. Chicago: Chicago University Press, 91-123. Langacker, Ronald W. 1987. Foundations of Cognitive Grammar. Vol. I, Theoretical Prerequisites. Stanford, CA: Stanford University Press. Langacker, Ronald W. 1991. Foundations of Cognitive Grammar. Vol. II, Descriptive Application. Stanford, CA: Stanford University Press. Langacker, Ronald W. 2000. Grammar and Conceptualization. New York: Mouton de Gruyter. Liddell, Scott K. 1984. Unrealized-inceptive aspect in American Sign Language: Feature insertion in syllabic frames. In J. Drogo, V. Mishra, and D. Teston (eds.), Papers from the 31 20th Regional Meeting of the Chicago Linguistic Society. Chicago: University of Chicago Press, 257-270. Liddell, Scott K. 1995. Real, surrogate, and token space: Grammatical consequences in ASL. In Karen Emmorey and Judy Reilly (eds.), Language, Gesture, and Space. Hillsdale, NJ: Lawrence Erlbaum Associates, 19-41. Liddell, Scott K. 2003. Grammar, gesture and meaning in American Sign Language. Cambridge: Cambridge University Press. Liddell, Scott K. and Melanie Metzger. 1998. Gesture in sign language discourse. Journal of Pragmatics 30, 657-697. Metzger, Melanie. 1995. Constructed Dialogue and Constructed Action in American Sign Language. In Ceil Lucas (ed.), Sociolinguistics in Deaf Communities. Washington, DC: Gallaudet University Press, 255-271. Padden, Carol. 1986. Verbs and role-shifting in ASL. in C. Padden (ed.), Proceedings of the Fourth National Symposium on Sign Language Research and Teaching. National Association of the Deaf: Silver Spring, MD, 44-56. Roy, Cynthia. 1989. Features of Discourse in an American Sign Language Lecture. In Ceil Lucas (ed.), The Sociolinguistics of the Deaf Community. San Diego: Academic Press. 231-251. Tannen, Deborah. 1989. Talking voices: repetition, dialogue, and imagery in conversational discourse. Cambridge: Cambridge University Press. Taub, Sarah F. 2001. Language from the body: iconicity and metaphor in American Sign Language. Cambridge: Cambridge University Press. Wilcox, Phyllis P. 1998. GIVE: Acts of giving in American Sign Language. In John Newman (ed.), The Linguistics of Giving. Philadelphia: J. Benjamins, 175-207. Wilcox, Sherman. 2004. Cognitive iconicity: Conceptual spaces, meaning, and gesture in signed languages. Cognitive Linguistics 15:2, 119-147. Winston, Elizabeth. 1991. Spatial referencing and cohesion in an American Sign Language text. Sign Language Studies 73:397-410. 32