Elsevier

Information Sciences

Volume 373, 10 December 2016, Pages 79-94
Information Sciences

A competence-performance based model to develop a syntactic language for artificial agents

https://doi.org/10.1016/j.ins.2016.08.088Get rights and content

Abstract

The hypothesis of language use is an attractive theory in order to explain how natural languages evolve and develop in social populations. In this paper we present a model partially based on the idea of language games, so that a group of artificial agents are able to produce and share a symbolic language with syntactic structure. Grammatical structure is induced by grammatical evolution of stochastic regular grammars with learning capabilities, while language development is refined by means of language games where the agents apply on-line probabilistic reinforcement learning. Within this framework, the model adapts the concepts of competence and performance in language, as they have been proposed in some linguistic theories. The first experiments in this article have been organized around the linguistic description of visual scenes with the possibility of changing the referential situations. A second and more complicated experimental setting is also analyzed, where linguistic descriptions are enforced to keep word order constraints.

Introduction

Language is one of the fundamental cognitive skills needed for the development of advanced multi-agent or multi-robot systems, as it allows communication and cooperation among group members. This necessity has recently motivated research where computational models are applied for developing artificial languages with increasing complexity. Computational simulations, and even experiments with real robots, have also been used for understanding properties of natural languages [11], [38]. The present article belongs to the former approach: we propose a new model that combines evolutionary algorithms and reinforcement learning so that a team of agents can construct a shared artificial language, with the important property of having syntactic structure.

The initial approaches for communication within teams of agents have focused on creating a basic vocabulary or lexicon that is shared by all members of the group. The vocabulary is generally a simple mapping between symbols and meanings. This problem is not trivial because of the Symbol Grounding Problem: how to effectively connect a symbol with its meaning [7]. Humans do not seem to be limited by this issue, but artificial organisms have serious difficulties. While some works address this problem [12], [26], [28], [32], [33], [34], [37], other proposals bypass it and focus on the higher level, while the aspects related to the physical association between symbols and meanings are assumed to have been solved or are not considered [8], [17], [39].

Although a shared vocabulary is essential and it has to be the first step, syntactic competence is vital for an agent to efficiently describe reality in a symbolic way. In this work, we approach the syntactic alignment of agents in a team by means of language games applying on-line reinforcement learning algorithms and grammatical evolution. This process allows the agents to evolve and adapt their language to the current linguistic situation. Syntactic alignment refers here to the process that allows a team of agents to develop a common syntactic language, by pairwise interactions or language games, without being told by a centralized source. The concept of language games is partly inspired by the ideas of Wittgenstein [41] and it has been applied in most of Steel’s works (see [35] for a recent review) and related authors such as [3], [16]. The importance of syntactic competence is currently so relevant that other approaches based on different ideas have been proposed in [2], [4], [13], [14], [19], [24], [25], [29].

Most of the cited works, except [12], [17], [26], are focused on explaining properties of natural languages by means of computational simulations. The above three exceptions are more oriented towards artificial language development, but they do not address syntax in depth. A more recent work deals with syntax [18], by allowing team members to tune the probabilities of stochastic regular grammars by means of language games and reinforcement learning. However, the grammar production rules themselves cannot change, as they are encoded in the agents’ controller from the beginning. In this work, we follow their approach based on reinforcement learning and language games [18], but we study how the agents can evolve the grammar itself. This process allows the agents to look for the grammar that best describes the current linguistic situation which is a very important skill if the agents need to talk about different contexts. To solve this problem we propose a model that takes into account the concepts of Competence and Performance as they were initially defined by Chomsky [5]. Competence is the knowledge that a speaker has about his/her language and performance stands for the specific use of the language in a specific context. Competence is developed by means of an evolutionary algorithm in our model and it is essentially a search process in the space of possible grammars that describe the current linguistic situation. On the other hand, performance is achieved by means of a reinforcement learning mechanism, which rewards the most used structures.

The rest of the paper is organized as follows. Section 2 describes the two tasks used in the computational experiments. Section 3 studies the process of producing and evolving the symbolic artificial language by means of grammatical evolution. Section 4 analyzes how the language can be aligned by means of on-line reinforcement learning algorithms. Section 5 summarizes the whole model. Section 6 enhances the initial model to enforce syntactic order constraints. In Section 7 both the setting and the experimental results are presented and discussed. The last section draws the main conclusions.

Section snippets

Describing dynamic linguistic situations in two different tasks

The first task to be solved by the agents in the team consists of evolving a symbolic language to describe two types of linguistic situations. The agents are situated in an environment that can be perceived with their specialized sensors. As in this paper we are mainly interested in the high-level syntactic alignment process, we assume that agents are able to analyze the images they watch and build internal representations of them. Figs. 1 and 2 show the type of visual scenes that the agents

Evolving grammatical structures

As we explained in the Introduction section, the model that we propose is based on two strategies: evolution and learning. Evolution allows the agents to produce, evolve and adapt their grammatical structures (competence). Learning drives the process of achieving linguistic consensus (performance). In this section we will define in depth the evolutionary process as a competence developing process.

Evolutionary algorithms can be an effective way to study artificial language emergent phenomena,

Stochastic learning-grammars

In order to allow the agents to develop a suitable language for the description of the type of visual scenes displayed in Figs. 1 and 2, which are composed of objects, colors, or spatial relations, we propose to use an stochastic version of the solution grammars that we described in previous sections. In a stochastic solution grammar, each production rule has an associated probability (i.e. the probability of using the rule). These probabilities can be tuned by the agents by means of

The whole model

We summarise the whole model in order to provide a complete view of our work. Essentially, the model performs in two stages. First, an evolutionary algorithm is executed. In this evolutionary process each agent in the team tries to develop a private suitable solution grammar which can express the agent communicative intention. A communicative intention matrix helps to measure this intention. Each agent manages a population of chromosomes (binary strings) that are transformed into solution

Imposing order constraints

The previous model presents two potential problems when it has to scale to more complicated situations. On one hand, evolutionary processes usually need a long time and resources to find solutions, specially if the number of states in the search space increases. This situation is likely in a referential scene combining colors, objects, and spatial relations. In this case, there is a significant number of possible combinations of words associated to each item. On the other hand, the evolutionary

Suppositions and setting

It is important to comment here on the focus of our model and the experiments that have been carried out to validate it. We are interested in analyzing how a language with syntactic properties can emerge and be shared within a team of artificial agents. This process is high level and our approach is to focus on the symbolic aspects, which are more relevant for language development and spread. Nevertheless, higher procedures are supported by other kinds of processes such as perception and motor

Conclusions

In this work, we have proposed a model that combines evolution and learning, in order to study how a shared artificial language with syntactic structure, can emerge in a team of agents. The model considers concepts about natural language, such as competence and performance, and the possibility of innate grammatical knowledge expressed as a Universal Grammar. Competence is simulated by means of evolutionary algorithms that, for every agent, evolve a grammar suitable for the current linguistic

Acknowledgment

The second author has been supported by the Spanish Ministry of Science under contract ENE2014-56126-C2-2-R (AOPRIN-SOL).

References (42)

  • E.A.P. Hemberg

    An exploration of grammars in grammatical evolution

    (2010)
  • H. Jaeger et al.

    What can mathematical, computational and robotic models tell us about the origins of syntax?

    Biological Foundations and Origin of Syntax

    (2009)
  • D. Jung et al.

    Grounded symbolic communication between heterogeneous cooperating robots

    Auton. Robots

    (2000)
  • S. Kirby

    Spontaneous evolution of linguistic structure: an iterated learning model of the emergence of regularity and irregularity

    IEEE Trans. Evol. Comput.

    (2001)
  • S. Kirby

    Learning, bottlenecks and the evolution of recursive syntax

  • S.C. Levinson et al.

    The Background to the study of the language of space

  • J.V. Looveren

    Design and performance of pre-grammatical language games

    (2005)
  • D. Maravall et al.

    Self-emergence of lexicon consensus in a population of autonomous agents by means of binary-strings evolution strategies

  • D. Maravall et al.

    Alignment in vision-based syntactic language games for teams of robots using stochastic regular grammars and reinforcement learning: The fully autonomous case and the human supervised case

    Robot. Auton. Syst.

    (2013)
  • M.R. McClain

    Semantic based learning of syntax in an autonomous robot

    (2006)
  • R. Murphy

    Introduction to AI Robotics

    (2000)
  • Cited by (3)

    • Pixel convolutional neural network for multi-focus image fusion

      2018, Information Sciences
      Citation Excerpt :

      In 2012, Krizhevsky et al. [14] rekindled interest in CNNs by showing substantially higher image classification accuracy on the ImageNet Large Scale Visual Recognition Challenge (ILSVRC). Hereafter, CNN and its variations have been widely used in image recognition [27], human pose estimation [12], object localization [35], video analysis [16], natural language processing [28], medical image analysis [34] and other applications. As introduced in Section 1, although many FMs had been proposed, the FM in spatial domain-based multi-focus fusion is still a crucial part.

    View full text