key: cord-230430-38fkbjq0 authors: Williams, Tom; Johnson, Torin; Culpepper, Will; Larson, Kellyn title: Toward Forgetting-Sensitive Referring Expression Generationfor Integrated Robot Architectures date: 2020-07-16 journal: nan DOI: nan sha: doc_id: 230430 cord_uid: 38fkbjq0 To engage in human-like dialogue, robots require the ability to describe the objects, locations, and people in their environment, a capability known as"Referring Expression Generation."As speakers repeatedly refer to similar objects, they tend to re-use properties from previous descriptions, in part to help the listener, and in part due to cognitive availability of those properties in working memory (WM). Because different theories of working memory"forgetting"necessarily lead to differences in cognitive availability, we hypothesize that they will similarly result in generation of different referring expressions. To design effective intelligent agents, it is thus necessary to determine how different models of forgetting may be differentially effective at producing natural human-like referring expressions. In this work, we computationalize two candidate models of working memory forgetting within a robot cognitive architecture, and demonstrate how they lead to cognitive availability-based differences in generated referring expressions. Effective human-robot interaction requires human-like natural language and dialogue capabilities that are sensitive to robots' embodied nature and inherently situated context (Mavridis, 2015; Tellex et al., 2020) . In this paper we explore the role that models of Working Memory can play in enabling such capabilities in integrated robot architectures. While Working Memory has long been understood to be a core feature of human cognition, and thus a central component of cognitive architectures, recent evidence from psychology suggests a conception of working memory that is subtly different from what is implemented in most cognitive architectures. Specifically, while most models of working memory in computational cognitive architectures maintain a single central working memory store, Converging evidence from different communities suggests that humans have different resource limitations for different types of information. Moreover, recent psychological evidence suggests that Working Memory may be a limited resource pool, with resources consumed on the basis of the number and type of features retained. This suggests that forgetting in Working should be modeled in cognitive architectures as a matter of systematic removal (on the basis of decay or interference) of entity features with sensitivity to the resource limitations imposed for the specific type of information represented by that feature. Of course robot cognition need not directly mirror human cognition, and indeed robots have both unique knowledge representational needs and increased flexibility in how resource limitations are implemented. In this work, we present a robot architecture in which (1) independent resource pools are maintained for different robot-oriented types of information; (2) WM resources are maintained at the feature level rather than entity level; and (3) both interferenceand decay-based forgetting procedures may be used. This architecture is flexibly configurable both in terms of what type of forgetting procedure is used, and how that model is parameterized. For robot designers, this choice of parameterization may be made in part on the basis of facilitation of interaction. In this paper we specifically consider how the use of different models of forgetting within this architecture lead to different information being retained in working memory, which in turn leads to different referring expressions being generated by the robot, which in turn can produce interactive alignment effects purely through Working Memory dynamics. While in future work it will be important to identify exactly which parameterizations lead to selection of referring expressions that are optimal for effective human-robot interaction and teaming, in this work we take the critical first step of demonstrating, as a proof-of-concept, that decay-and interference-based forgetting mechanisms can be flexibly used within this architecture, and that those policies do indeed produce different natural language generation behavior. "Referring" has been referred to as the "fruit fly" of language due to the amount of research it has attracted (Van Deemter, 2016; Gundel & Abbott, 2019) . In this work, we focus specifically on Referring Expression Generation (REG) (Reiter & Dale, 1997) in which a speaker must choose words or phrases that will allow the listener to uniquely identify the speaker's intended referent. REG includes two constituent sub-problems (Gatt & Krahmer, 2018) : referring form selection and referential content determination. While referring form selection (in which the speaker decides whether to use a definite, indefinite, or pronominal form (Poesio et al., 2004; McCoy & Strube, 1999) (see also Pal et al., 2020) ) has attracted relatively little attention, referential content determination is one of the most well-explored sub-problems within Natural Language Generation, in part due to the logical nature of the problem that enables it to be studied in isolation, to the point where "REG" is typically used to refer to the referential content determination phase alone. In this section we will briefly define and describe the general strategies that have been taken in the computational modelling of referential content determination; for a more complete account we recommend the recent book by Van Deemter (2016), which provides a comprehensive account of work on this problem. Referential content determination, typically employed when generating definite descriptions, is the process by which a speaker seeks to determine a set of constraints on known objects that if communicated will distinguish the target referent from other candidate referents in the speaker and listener's shared environment. These constraints most often include attributes of the target referent, but can also include relationships that hold between the target and other salient entities that can serve as "anchors", as well as attributes of those anchors themselves (Dale & Haddock, 1991) . Three referential content determination models have been particularly influential: the Full Brevity Algorithm (Dale, 1989) , in which the speaker selects the description of minimum length, in order to straightforwardly satisfy Grice's Maxim of Quantity (Grice, 1975) ; the Greedy Algorithm, in which the speaker incrementally adds to their description whatever property rules out the largest number of distractors (Dale, 1992) ; and the Incremental Algorithm (IA), in which the speaker incrementally adds properties to their description in order of preference so long as they help to rule out distractors (Dale & Reiter, 1995) . A key aspect of the IA is its ability to overspecify through its inclusion of properties that are not strictly needed from a logical perspective to single out the target referent, but are nevertheless included due to being highly preferred; a behavior also observed in human speakers (Engelhardt et al., 2006) . Because the IA's behavior is highly sensitive to preference ordering (Gatt et al., 2007) , there has been substantial research seeking to determine what properties are in general psycholinguistically preferred over others (Belke & Meyer, 2002) , or to automatically learn optimal preference orderings (Koolen et al., 2012) . As highlighted by Goudbeek & Krahmer (2012) , however, this focus on a uniform concept of "preference" obscures a much more complex story that relates to fundamental debates over the extent to which speakers leverage listener knowledge during sentence production. A notion of "preference" as encoded in the IA could be egocentrically grounded, with speakers "prefer" concepts that are easy for themselves to assess or cognitively available to themselves (Keysar et al., 1998) ; it could be allocentrically grounded, with speakers intentionally seeking to facilitate the listener's ease of processing (Janarthanam & Lemon, 2009) ; or a hybrid model could be used, in which egocentric and allocentric processes compete (Bard et al., 2004) , with egocentrism vs. allocentrism "winning out" on the basis of factors such as cognitive load (Fukumura & van Gompel, 2012) . These approaches, which require accounting for listener knowledge to be slow and intentional, stand in contrast to memory-oriented account of referential content determination in which such accounting can naturally occur as a result of priming. Pickering & Garrod (2004)'s Interactive Alignment model of dialogue suggests that dialogue is a highly negotiated process (see also Clark & Wilkes-Gibbs (1986) ) in which priming mechanisms lead interlocutors to influence each others' linguistic choices at the phonetic, lexical, syntactic and semantic levels, through mutual activation of phonetic, lexical, syntactic, and semantic structures and mental representations, as in the case of lexical entrainment (Brennan & Clark, 1996) . While there has been extensive evidence for lexical and syntactic priming, research on semantic or conceptual priming in dialogue has only relatively recently become a target of substantial investigation (Gatt et al., 2011) . A theory of dialogue including semantic or conceptual priming would suggest that the properties or attributes that speakers choose to highlight in their referring expressions (e.g., when a speaker chooses to refer to an object as "the large red ball" rather than "the sphere") should be due in part to these priming effects. And in fact, as demonstrated by Goudbeek & Krahmer (2010) , speakers can in fact be influenced through priming to use attributes in their referring expressions that would otherwise have been dispreferred. These findings have motivated dual-route computational models of dialogue (Gatt et al., 2011; in which the properties used for referring expression selection are made on the basis of interaction between two parallel processes, each of which is periodically called upon to provide attributes of the target referent to be placed into a WM buffer that is consulted when properties are needed for RE generation (at which point selected properties are removed from that buffer). The first of these processes is a priming-based procedure in which incoming descriptions trigger changes in activation within a spreading activation model, and properties are selected if they are the highest-activation properties (above a certain threshold) for the target referent. The second of these processes is a preference-based procedure in which a set of properties is generated by a classic REG algorithm (cp. Gatt & Krahmer, 2018) such as the Incremental Algorithm, in which properties are incrementally selected according to a pre-established preference ordering designed or learned to reflect frequency of use, ease of cognitive or perceptual assessability, or some other informative metric (Dale & Reiter, 1995) . One advantage of this type of dual process model is that it accounts for audience design effects (in which speaker utterances at least appear to be intentionally crafted for ease-of-comprehension) within an egocentric framework, by demonstrating how priming influences on WM can themselves account for listener-friendly referring expressions. That is, if a concept is primed by an interlocutor's utterance, a speaker will be more likely to use that concept in their own references simply because it is in WM, with the side effect that that property will then be easy to process by the interlocutor responsible for its' inclusion in WM in the first place (Vogels et al., 2015) . Moreover, this phenomenon aligns well with evidence suggesting that despite the prevalence of lexical entrainment and alignment effects, people are actually slow to explicitly take others' perspectives into account (Bard et al., 2000; Fukumura & van Gompel, 2012; Gann & Barr, 2014) . Another advantage of this type of dual process model is its alignment with overarching dualprocess theories of cognition (e.g., Kahneman, 2011; Evans & Stanovich, 2013; Sloman, 1996; Oberauer, 2009 ): the first priming-driven process for populating WM, grounded in semantic activation, can be viewed as a reflexive System One process, whereas the second preference-driven process leveraging the Incremental Algorithm can be viewed as a deliberative System Two process. Of course in the model under discussion the two routes do not truly compete with each other or operate on different time courses, but are instead essentially sampled between; however, it is straightforward to imagine how the two processes used in this type of model could be instead deployed in parallel. One major disadvantage of this type of model, however, is that its focus with respect to WM is entirely on retrieval (i.e., how priming and preference-based considerations impact what information is retrieved from long-term memory into WM), and fails to satisfactorily account for maintenance within WM. Within Gatt et al. (2011) 's model, as soon as a property stored in WM is used in a description, it is removed from WM so that that space is available for another property to be considered. This behavior seems counter-intuitive, as it ensures that representations are removed from WM at just the time when it is established to be important and useful, which should be a cue to retain said representations rather than dispose of them. Moreover, this model is surprisingly organized from the perspective of models of WM such as Cowan (2001) A final complication for this model is its speaker-blind handling of priming. Specifically, within Gatt et al. (2011) 's model a speaker's utterances are only primed by their interlocutor's utterances, when in fact the choices a speaker makes should also impact the choices they themselves make in the future (Shintel & Keysar, 2007) , either due to Gricean notions of cooperativity (Grice, 1975) or, as we argue, because a speaker's decision to refer to a referent using a particular attribute should make that attribute more cognitively available to themselves in the immediate future. These concerns are addressed by our previously proposed model of robotic short-term memory (Williams et al., 2018b) , in which speakers rely on the contents of WM for initial attribute selection and, if their selected referring expression is not fully discriminating, select additional properties using a variant of the IA. While this does not align with dual-process models of cognition, it does account for both encoding and maintenance of WM, and provides a potentially more cognitively plausible account of REG with respect to WM dynamics. One shortcoming shared by both models, however, is neither the dual-process model of Gatt et al. (2011) nor our WM-focused model appropriately account for when and how information is removed from WM over time, or how this impacts REG. Different theories of WM "forgetting" necessarily lead to predicted differences in cognitive availability. Accordingly, these different models of forgetting should similarly predict cognitive availabilitybased differences in the properties selected during REG. To design effective intelligent agents, it is thus necessary to determine how different models of forgetting may be differentially effective at producing natural human-like referring expressions. In this work, we first computationalize two candidate models of WM forgetting within a robot cognitive architecture. Next, we propose a model of REG that is sensitive to the WM dynamics of encoding, retrieval, maintenance, and forgetting, discuss the particulars of deploying this type of model within an integrated robot architecture, where WM resources are divided by domain (i.e., people, locations, and objects) rather than by modality (i.e., visual vs. verbal). Finally, we provide a proof-of-concept demonstration of two parametrizations of our model into an integrated robot cognitive architecture, and demonstrate how these different parametrizations lead to cognitive availability-based differences in generated referring expressions. Models of forgetting in Working Memory are typically divided into two broad categories (Reitman, 1971; Jonides et al., 2008; Ricker et al., 2016) : decay-based models, and interference-based models. Decay-based models of WM (Brown, 1958; Atkinson & Shiffrin, 1968) , that time plays a causal role in WM forgetting, with a representation's inclusion in WM made on the basis of a level of activation that gradually decays over time if the represented information is not used or rehearsed. Accordingly, in such models, a piece of information is "forgotten" with respect to WM if it falls below some threshold of activation due to disuse. This model of forgetting is intuitively appealing due to the clear evidence that memory performance decreases over time (Brown, 1958; ?; Ricker et al., 2016) . Computational Advantages and Limitations: As with Gatt et al., spreading activation networks can be used to elegantly model how activation of representations impacts the rise and fall of activation of semantically related pieces of information. One disadvantage of this approach, however, is that activation levels need to be continuously re-computed for each knowledge representation in memory. While this may be an accurate representation of actual cognitive processing, artificial cognitive systems do not enjoy the massively parallelized architectures enjoyed by biological cognitive systems, meaning that this approach may face severe scaling limitations in practice. Computational Model: To allow for straightforward comparison with other models of forgetting, we define a simple model of decay that operates on individual representations outside the context of a semantic activation network. We begin by representing WM as a set W M = Q 0 , . . . , Q n , where Q i is a mental representation of a single entity, represented as a queue of properties currently activated for that entity. Next, we define an encoding procedure that specifies how the representations in WM are created and manipulated on the basis of referring expressions generated either by the agent or its' interlocutors. As shown in Alg. 1, this procedure operates by considering each property included in the referring expression, and updating the queue used to represent the entity being described, by placing that property in the back of the queue, or by moving the property to the back of the queue if it's already included in the representation. Note that this procedure can be used either after each utterance is heard (in which case the representation is updated based on all properties used to describe the entity) or incrementally (in which case the representation is updated after each property is heard). If used incrementally, then forgetting procedures may be interleaved with representation updates. Finally, we define a model of decay that operates on these representations. As shown in Alg. 2, this procedure operates by removing the property at the front of queue Q at fixed intervals defined by decay parameter δ. Algorithm 1 Per-entity encoding model In contrast, interference-based models (Waugh & Norman, 1965) argue that WM is a fixed-size buffer in which a piece of information is "forgotten" with respect to WM if it is removed (due to, e.g., being the least recently used representation in WM memory) to make room for some new representation. Interference-Based Models have been popular for nearly as long as decay-based models (Keppel & Underwood, 1962) due to observations that the evidence used as evidence for "decay" can just as easily be used as evidence for forgetting due to intra-representational interference, as longer periods of time directly correlate with higher frequencies of interfering events (Lewandowsky & Oberauer, 2015; Oberauer & Lewandowsky, 2008) , and because tasks with varying temporal lengths but consistent overall levels of interference have been shown to yield similar rates of forgetting, thus failing to support time-based decay (Oberauer & Lewandowsky, 2008) . Recent work has trended towards interference-based accounts of forgetting, with a number of further debates and competing models opening up within this broad theoretical ground. First, there is debate as to whether interference alone is sufficient to explain forgetting, or whether time-based decay still plays some role in conjunction with interference. Recent work suggests that in fact these two models may be differentially employed for different types of representations, with phonological representations forgotten due to interference and non-phonological representations forgotten due to a combination of interference and time-based decay (Ricker et al., 2016) . Second, within interference-based models, there exist competing models based on reasons for displacement. In particular, while theories of pure displacement (Waugh & Norman, 1965) posit that incoming representations replace maintained representations on the basis of frequency or recency of use, or on the basis of random chance (similar to caching strategies from computing systems research (Press et al., 2014) ), theories of retroactive interference instead posit that replacements are made on the basis of semantic similarity, with representations "forgotten" if they are too similar to incoming representations (Wickelgren, 1965; Lewis, 1996) . Third, within both varieties of interference-based models, there has been recent debate on the structure and organization of the capacity-limited medium of WM. Ma et al. (2014) contrasts four such models: (1) slot models, in which a fixed number of slots are available for storing coherent representations (Miller, 1956; Cowan, 2001) ; (2) resource models, in which a fixed amount of representational medium can be shared between an unbounded number of representations (with storage of additional features in one representation resulting in fewer feature-representing medium being available for other representations) (Wilken & Ma, 2004) ; (3) discrete-representation models, in which a fixed number of feature "quanta" are available to distribute across representations (Zhang & Luck, 2008) ; and (4) variable-precision models, in which working memory precision is statistically distributed (Fougnie et al., 2012) . Computational Advantages and Limitations: One advantage of interference-based models for artificial cognitive agents is decreased computational expense, as only a fixed number of entities or features must be maintained in WM, and WM need not be updated at continuous intervals if no new stimuli are processed. Rather, WM only needs to be updated when (1) new representations are encoded into WM, or (2) existing representations are manipulated. Another advantage of this approach is its conceptual alignment with the process of caching from computer science, which means that caching mechanisms from computing systems research, such as least-recently-used and least-frequently-used caching policies, can be straightforwardly leveraged, with prior work providing substantial information about their theoretical properties and guarantees. In fact, recent work has explored precisely how caching strategies from computer science can be used for this purpose (Press et al., 2014) ). Within the interference-based family of models, slot-based and discrete-representation models are likely the most easy to computationalize due to the ephemeral and undiscretized nature of "representational medium." Computational Model: To model interference-based forgetting, we use the same WM representation and encoding procedure as used to model decay-based forgetting, and propose a new model designed specifically for robotic knowledge representation. This model can be characterized as a per-entity discrete-representation displacement model. As shown in Alg. 3, this procedure operates by removing properties at the front of queue Q whenever the size of Q is greater than some size limitation imposed by parameter α. This model is characterized as discrete-representation because resource limitations are imposed at the level of discrete features rather than holistic representations. It is characterized as a displacement model because features are replaced on the basis of a Least-Recently Used (LRU) caching strategy (Knuth, 1997) rather than on the basis of semantic similarity, due to the pragmatic difficulty of assessing the similarity of different categories of properties without mandating the use of architectural features such as well-specified conceptual ontologies (cp. Tenorth & Beetz, 2009; Lemaignan et al., 2010) , which may not be available in all robot architectures. This model is characterized as per-entity because resource limitations are imposed locally (i.e., for each entity) rather than globally (i.e., shared across all entities). While this is obviously not a cognitively plausible characteristic, it was selected, as a starting point, to reduce the need for coordination across architectural components and to facilitate improved performance (as entity representations need not compete with each other). However, if desired, it would be straightforward to extend this approach to allow for imposition of global resource constraints. Algorithm 3 Per-entity discrete representation displacement model. procedure Displacement(Q, α) Q: the per-entity property queue α: maximum buffer size loop if |Q| > α then pop(Q) end if end loop end procedure Memory modeling has long been a core component of cognitive architectures, due to its central role in cognition (Baxter et al., 2011) . The relative attention paid to WM has, however, varied widely between cognitive architectures. ARCADIA, for example, has placed far more emphasis on attention than on memory (Bridewell & Bello, 2016) . While ARCADIA does include a visual short term memory component, it is treated as a "potentially infinite storehouse," with consideration of resource constraints left to future work (Bridewell & Bello, 2016) . ACT-R and Soar, in contrast, do place larger emphases on WM. ACT-R did not originally have explicit WM buffers, instead implicitly representing WM as the subset of LTM with activation above some particular level (Anderson et al., 1996) , with "forgetting" thus modeled through activation decay (cp. Cowan, 2001) . In more recent incarnations of ACT-R, a very small short-term buffer is maintained, with contents retrieved from LTM on the basis of both base-level activation (on the basis of recency and frequency) and informative cues. Similarly, Soar Laird (2012) has long emphasized the role of WM, due to its central focus on problem solving through continuous manipulation of WM (Rosenbloom et al., 1991) . And while Soar did not initially represent WM resource limitations (Young & Lewis, 1999) , it has also by now long included decay-based mechanisms on at least a subset of WM contents (Chong, 2003; Nuxoll et al., 2004) , as well as base-activation and cue-based retrieval methods such as those mentioned above Jones et al. (2016) . While the models may be intended primarily as models of human cognition, flaws and all, rather than as systems for enabling effective task performance through whatever means necessary regardless of cognitive plausibility, a significant body of work has well demonstrated the utility of these architectures for cognitive robotics (Laird et al., , 2017 Kurup & Lebiere, 2012) and human-robot interaction Trafton et al. (2013) . There has also been significant work in robotics seeking to use insights from psychological research on WM to better inform the design of select components of larger robot cognitive architectures that do not necessarily aspire towards cognitive plausibility. For example, a diverse body of researchers has collaborated on the development and use of the WM Toolkit (Phillips & Noelle, 2005; Gordon & Hall, 2006; Kawamura et al., 2008; Persiani et al., 2018) ; a software toolkit that maintains pointers to a fixed number of chunks containing arbitrary information. At each timestep, this toolkit proposes a new set of chunks, and then uses neural networks to select a subset of these chunks to retain. This model has been primarily used for enabling cognitive control, in which the link between robot perception and action is modulated by a learned memory management policy. Models of WM have also been leveraged within the field of Human-Robot Interaction. Broz et al. (2012) , for example, specifically model the episodic buffer sub-component of WM (Baddeley, 2000) . Baxter et al. (2013) leverages models of WM to better enable non-linguistic social interaction through alignment, similar to our approach in this work. Researchers have also leveraged models of WM to facilitate communication. For example, Hawes et al. (2007) leverage a model of WM within the CoSy architecture, with concessions made to accommodate the realistic needs of integrated robot architectures, in which specialized representations are stored and manipulated within distributed, parallelized, domain-specific components (see also . Similarly, in our own work within the DIARC architecture , we have demonstrated (as we will further discuss in this paper) the use of distributed WM buffers associated with such architectural components (Williams et al., 2018b) , as well as hierarchical models of common ground jointly inspired by models of WM (e.g. Cowan, 2001) and models of givenness from natural language pragmatics (e.g. Gundel et al., 1993) . In the next section we propose a new architecture that builds on this previous work of ours to allow for flexible selection (and comparison between) between different models of forgetting. Our forgetting models were integrated into the Distributed, Integrated, Affect, Reflection, Cognition (DIARC) Robot Cognitive architecture: a component-based architecture designed with a focus on robust spoken language understanding and generation . DIARC's Mnemonic and Linguistic components integrate via a consultant framework in which different architectural components (e.g., vision, mapping, social cognition) serve as heterogeneous knowledge sources that comprise a distributed model of Long Term Memory (Williams, 2017; . These consultants are used by DIARC's natural language pipeline for the purposes of reference resolution and REG. In recent work we have extended this framework to produce a new Short-Term Memory Augmented consultant framework in which consultants additionally maintain, for some subset of the entities for which they are responsible, a short term memory buffer of properties that have recently been used by interlocutors to refer to those entities. In this work, we build upon that STM (Short Term Memory)-Augmented Consultant Framework through the introduction of a new architectural component, the WM Manager, which is responsible for implementing the two forgetting strategies introduced in the previous section. Our model aligns with two key psychological insights. First, converging evidence from different communities suggests that humans have different resource limitations for different types of information (Wickens, 2008) , due to either decreased interference between disparate representations (Oberauer et al., 2012) or the use of independent domain-specific resource pools (Baddeley, 1992; Logie & Logie, 1995) . Our approach takes a robot-oriented perspective on this second hypothesis, with our use of the WM-Augmented consultant framework resulting in independent resource pools maintained for different types of entities (e.g., objects vs. locations vs. people) rather than for different modalities (e.g., visual vs. auditory) or different codes of processing (e.g., spatial vs. verbal). Second, while early models of WM suggested that WM resource limitations are bounded to a limited number of chunks (Miller, 1956) , more recent models instead suggest that the size of WM is affected by the complexity of those chunks (Mathy & Feldman, 2012) , and that maintaining multiple features of a single entity may detract from the total number of maintainable entities, and accordingly, the number of features maintainable for other entities (Alvarez & Cavanagh, 2004; Oberauer et al., 2016; Taylor et al., 2017) . Our approach, again, takes a robot-oriented perspective on these models, maintaining WM resources at the feature-level rather than entity-level, while enabling additional flexibility that may not be reflected in human cognition. Specifically, instead of enforcing global resource limits, we allow for flexible selection between decay-based and interference-based (i.e., resource-limited) memory management models, as well as for simultaneous employment of both models, in order to model joint impacts of interference and decay as discussed by Ricker et al. 2016) . Moreover, while we currently focus on local (per-entity) feature-based resource limitations, our system is designed to allow for global resource limitations (cp. Just & Carpenter, 1992; Ma et al., 2014) in future work, due to our use of a global WM Manager Component. Using this architecture, our different forgetting models can differentially affect REG without any direct interaction with the REG process. Rather, the WM Manager simply interfaces with the WM buffers maintained by each distributed consultant, which then implicitly impacts the properties available to the SD-PIA algorithm that we use for REG. This algorithm takes a lazy approach to REG in which the speaker initially attempts to rely only on the properties that are currently cached in WM, and only if this is insufficient to distinguish the target from distractors does the speaker retrieve additional properties from long-term memory. Incorporating the WM Manager into DIARC yields the configuration shown in Fig. 1 . When a teammate speaks to a robot running this configuration, text is recognized and then semantically parsed by the TLDL Parser , after which intention inference is performed by the Pragmatic Understanding Component (Williams et al., 2015) , whose results are provided to the Reference Resolution module of the Referential Executive (REX), which leverages the GROWLER algorithm (Williams et al., 2018a ) (see also ), which searches through Givenness Hierarchy (GH) (Gundel et al., 1993) informed data structures (Focus, Activated, Familiar) representing a hierarchically organized cache of relevant entity representations. This is important to highlight due to its relation to the WM buffers described in this work. While the WM buffers described in this work serve as a model of the robot's own WM, the Referential Executive's data structures can instead be viewed as either second-order theory-of-mind data structures or as a form of hierarchical common ground. When particular entities are mentioned by the robot's interlocutor or by the robot itself, pointers to the full representations for these entities (located in the robot's distributed long-term memory) are placed into the Referential Executive's GH-informed data structures, and the properties used to describe them are placed into the robot's distributed WM buffers. In addition, properties are placed into WM whenever the robot affirms that those properties hold through an LTM query. Critically, this can happen when considering other entities. For example, if property P is used during reference resolution, then P will be placed into WM for any distractors for which P is affirmed to hold, before being ruled out for other reasons. Similarly, if R is the target referent during REG, and if property P holds for R and is considered for inclusion in the generated RE, then for any distractors for which P also holds, P will be added to those distractors' WM buffers at the point where it is affirmed that P cannot be used to rule out those distractors because it holds for them as well. Once Reference Resolution is completed, if the robot decides to respond to the human utterance, it does so through a process that is largely the inverse of language understanding, including a dedicated REG module, which uses the properties cached in WM, along with properties retrievable from Long-Term Memory, to translate this intention into text (Williams & Scheutz, 2017; Williams et al., 2018b) . Human's Turn Robot's Turn Face Description Face Description Decay Interference Table 1 . 6 Face Case Study. Column contents denote properties used to describe target under each forgetting model, listed in the conssitent order below for easy comparison rather than in order selected. Properties: HL = light-hair(X), HD = dark-hair(X), HS = short-hair(X), GM = male(X), GF = female(X), CT = t-shirt(X), CL = lab-coat(X), CH = hoodie(X), GY = glasses(X), GN = no-glasses(X). We analyzed the proposed architecture by assessing two claims: (1) The proposed architecture demonstrates interactive alignment effects purely through WM dynamics; and (2) the proposed architecture demonstrates different referring behaviors when different models of forgetting are selected. These claims were assessed within the context of a "Guess Who"-style game in which partners take turns describing candidate referents (assigned from a set of 16 faces). On each player's turn, they are assigned a referent, and must describe that referent using a referring expression they believe will allow their interlocutor to successfully identify it (a process of REG). The other player must then process their interlocutor's referring expression and identify which candidate referent they believe to be their interlocutor's target (a process of Reference Resolution). Ideally, we would have assessed our claims in a setting in which a robot played this reference game with a naive human subject. This proved to be impossible due to the COVID-19 global pandemic. Instead, we present a case study in which a series of three six-round reference games are played between robot agents and a single human agent. All three games followed the same predetermined order of candidate referents and used the same pre-determined human utterances. The robot's responses were generated autonomously, with the robot in each of the three games using a different model of forgetting. In the first game, the robot uses our decay-based model of forgetting with δ = 10; in the second game, the robot uses our interference-based model of forgetting with α = 2; in the third game, the robot did retain any properties in short-term memory at all. The referring behavior under each model of forgetting is shown in Tab. 1. As shown in this table, the three examined models perform similarly in initial dialogue turns, but increasingly diverge over time. To help explain the observed differences in robot behavior, we examine specifically turn 6, in which the robot refers to Face 1 for the third time. This face could ostensibly be referred to using four properties: light-hair(X), short hair(X), lab-coat(X), and glasses(X). The architectural configuration that did not maintain representations in WM (No WM) operated according to the DIST-PIA algorithm (Williams & Scheutz, 2017) , which is a version of the Incremental Algorithm that is sensitive to uncertainty and that allows for distributed sources of knowledge. This algorithm first considers the highly preferred property light-hair(X), which is selected because it applies to Face 1 while ruling out distractors. Next, it considers short-hair(X), which it ignores because while it applies to Face 1, the faces with short hair are a subset of those with light hair, and thus short-hair(X) is not additionally discriminative. Next, the algorithm considers lab-coat(X), which it selects because it applies to Face 1 and rules out further distractors. Finally, to complete disambiguation, the algorithm considers and selects glasses(X). In contrast, the configuration that used the decay model had the following properties in WM: {lab-coat(X), short-hair(X), glasses(X)} (ordered from least-recently used to most-recently used 1 ). The algorithm starts by considering the properties stored in WM, beginning with labcoat(X), which is selected because it applies to Face 1 while ruling out distractors. Next, it considers short-hair(X), which is ignored because the set of entities with short hair is a subset of those wearing lab coats, and thus this is not additionally discriminative. Next, it considers glasses(X), which it selects because it applies to Face 1 and rules out distractors. In fact, {labcoat(X), glasses(X)} is fully discriminating for Face 1, so no further properties are needed. Finally, the configuration that used the interference model had the following properties in WM: {short-hair(X) glasses(X)}. This is easy to see as those properties were recently used in the Human's description of Face 6, and thus would have been considered for Face 1 when ruling it out during reference resolution. The algorithm thus starts by considering both of these properties, which are both selected because they apply to Face 1 and rule out distractors. However, because these are not sufficient for full disambiguation, the algorithm must also retrieve another property from LTM, i.e., lab-coat(X), which allows for completion of disambiguation. The differences in behavior demonstrated in this simple example validate both our claims. First, the proposed architecture's ability to demonstrate interactive alignment effects purely through WM dynamics is demonstrated by the systems' tendency to re-use properties originating from its interlocutor. Second, this example clearly demonstrates that the proposed architecture demonstrates different referring behaviors when different models of forgetting are selected. We have presented a flexible set of forgetting mechanisms for integrated cognitive architectures, and conducted a preliminary, proof-of-concept demonstration of these mechanisms, showing that they lead to different referring expressions being generated due to differences in cognitive availability between different properties. The next step of our work will be to fully explore the implications of different parametrizations of each of our presented mechanism, as well as the combined use of these mechanisms, on REG, and whether the referring expressions generated under different parametrizations are comparatively more or less natural, human-like, or effective, which would present obvious benefits for interactive, intelligent robots. In addition, the perspective taken in this paper may also yield insights and benefits for cognitive science more broadly. Specifically, we argue that our perspective may suggest alternative interpretations of the role of cognitive load on attribute choice. In 's work building on Gatt et al. (2011) 's model, they suggest that when speakers are under high cognitive load, they rely less on previously primed attributes, and are thus more likely to rely on their stable preference orderings. Their explanation for this finding 1. Future work should consider other algorithmic configurations, such as having properties within WM considered in the reverse order, or according to the preference ordering specified by the target referent's consultant. is that a decrease in available WM capacity leads to an inability to retrieve dialogue context into WM. We suggest an alternative explanation: cognitive load leads to decreased priming not because priming-activated representations cannot be retrieved into WM, but because those representations are less likely to be in WM in the first place. An additional promising direction for future work will thus be to compare the ability of the model presented in this paper to those presented by with respect to modeling of REG under cognitive load. The capacity of visual short-term memory is set both by visual information load and by number of objects Working memory: Activation limitations on retrieval Human memory: A proposed system and its control processes Working memory Controlling the intelligibility of referring expressions in dialogue Referential form, word duration, and modeling the listener in spoken dialogue. Approaches to studying world-situated language use Cognitive architecture for human-robot interaction: Towards behavioural alignment Memory-centred architectures: Perspectives on human-level cognitive competencies Tracking the time course of multidimensional stimulus discrimination: Analyses of viewing patterns and processing times during "same"-"different "decisions Conceptual pacts and lexical choice in conversation A theory of attention for cognitive systems Some tests of the decay theory of immediate memory Interaction histories and short term memory: Enactive development of turn-taking behaviors in a childlike humanoid robot The addition of an activation and decay mechanism to the soar architecture Referring as a collaborative process The magical number 4 in short-term memory: A reconsideration of mental storage capacity Cooking up referring expressions Generating referring expressions: Constructing descriptions in a domain of objects and processes Generating referring expressions involving relations. Fifth Conference of the European Chapter Computational interpretations of the gricean maxims in the generation of referring expressions Do speakers and listeners observe the gricean maxim of quantity? Dual-process theories of higher cognition: Advancing the debate Variability in the quality of visual working memory Producing pronouns and definite noun phrases: Do speakers use the addressee's discourse model? Speaking from experience: Audience design as expert performance. Language Attribute preference and priming in reference production: Experimental evidence and computational modeling Survey of the state of the art in natural language generation: Core tasks, applications and evaluation Evaluating algorithms for the generation of referring expressions using a balanced corpus System integration with working memory management for robotic behavior learning Preferences versus adaptation during referring expression generation Referring under load: Disentangling preference-based and alignment-based content selection processes in referring expression generation Alignment in interactive reference production: Content planning, modifier ordering, and referential overspecification Logic and conversation The oxford handbook of reference Cognitive status and the form of referring expressions in discourse Towards an integrated robot with multiple cognitive functions Learning lexical alignment policies for generating referring expressions for spoken dialogue systems Efficient computation of spreading activation using lazy evaluation The mind and brain of short-term memory A capacity theory of comprehension: individual differences in working memory Thinking, fast and slow Implementation of cognitive control for a humanoid robot Proactive inhibition in short-term retention of single items The egocentric basis of language use: Insights from a processing approach. Current directions in psychological science The art of computer programming Learning preferences for referring expression generation: effects of domain, language and algorithm What can cognitive architectures do for robotics? Biologically Inspired Cognitive Architectures The soar cognitive architecture Cognitive robotics using the Soar cognitive architecture. Workshops at the twenty-sixth AAAI conference on artificial intelligence A standard model of the mind: Toward a common computational framework across artificial intelligence, cognitive science, neuroscience, and robotics Oro, a knowledge management platform for cognitive architectures in robotics Rehearsal in serial recall: An unworkable solution to the nonexistent problem of decay Interference in short-term memory: The magical number two (or three) in sentence processing Visuo-spatial working memory Changing concepts of working memory What's magic about magic numbers? chunking and data compression in short-term memory A review of verbal and non-verbal human-robot interactive communication Generating anaphoric expressions: pronoun or definite description? The Relation of Discourse/Dialogue Structure and Reference The magical number seven, plus or minus two: Some limits on our capacity for processing information Comprehensive working memory activation in Soar. International Conference on Cognitive Modeling Psychology of learning and motivation Forgetting in immediate serial recall: Decay, temporal distinctiveness, or interference? Psychological review Modeling working memory: An interference model of complex span Givenness hierarchy theoretic cognitive status filtering. Proc. CogSci A working memory model improves cognitive control in agents and robots A biologically inspired working memory framework for robots Toward a mechanistic psychology of dialogue Centering: A parametric theory and its instantiations. Computational linguistics Caching algorithms and rational models of memory Building applied natural language generation systems Mechanisms of forgetting in short-term memory Decay theory of immediate memory: From brown (1958) to today A preliminary analysis of the soar architecture as a basis for general intelligence Spoken instruction-based one-shot object and action learning in a cognitive robotic architecture An overview of the distributed integrated cognition affect and reflection DIARC architecture You said it before and you'll say it again: Expectations of consistency in communication Does working memory have a single capacity limit? Robots that use language Knowrob-knowledge processing for autonomous personal robots ACT-R/E: An embodied cognitive architecture for human-robot interaction Computational models of referring: a study in cognitive science How cognitive load influences speakers' choice of referring expressions Acoustic similarity and retroactive interference in short-term memory Multiple resources and mental workload A detection theory account of change detection A consultant framework for natural language processing in integrated robot architectures Situated open world reference resolution for human-robot dialogue Going beyond command-based instructions: Extending robotic natural language interaction capabilities Towards givenness and relevancetheoretic open world reference resolution A framework for resolving open-world referential expressions in distributed heterogeneous knowledge bases Referring expression generation under uncertainty: Algorithm and evaluation framework Reference in robotics: A givenness hierarchy theoretic approach Augmenting robot knowledge consultants with distributed short term memory The Soar cognitive architecture and human working memory. Models of working memory: Mechanisms of active maintenance and executive control Discrete fixed-resolution representations in visual working memory