N ew and E mergent W orld models T hrough I ndividual, E volutionary and S ocial Learning Main...

1
New and Emergent World models Through Individual, Evolutionary and Social Learning http://www.newties.org Main goal : to realize an evolving artificial society capable of exploring the environment and developing its own image of this environment through cooperation and interaction. Long-term target: to learn how to design agents that are able to adapt autonomously to, and then operate effectively in, environments whose features are not known in advance. Technical objectives: 1.To develop an artificial society with an emergent culture. 2.To create a powerful “emergence engine” as a combination of individual learning, evolutionary learning, and social learning. 3.To develop social learning mechanisms that allow sharing knowledge with other agents (incl. language evolution). System: scaleable, p2p infrastructure for large experiments: Agent complexity Population size World size and complexity Virtual time Logo IST LogFET Vrije Universiteit van Amsterdam NAPIER UNIVERSITY EDINBURGH Vrije Universiteit (NL) (Coordinator, agents, evolution): Prof.Dr. A.E. Eiben R. Griffioen C. Tzolov University of Surrey (UK) (Environments and challenges): Prof. N. Gilbert, S. Schuster Napier University (UK) (P2P infrastructure) Prof. B. Paechter Dr. B.G.W. Craenen Eötvös Loránd University (HU) (Individual learning, mining emergence) Dr. A. Lorincz A. Bontovics Dr. Gy. Hévizi Tilburg University (NL) (language, communication and co- operation) Dr. P. Vogt, Dr. F. Divina Learning mechanisms Evolutionary learning: • crossover and mutation on DQ-trees and • crossover and mutation on attitude genes Individual learning: reinforcement learning Social learning: • sharing knowledge bits directly (telepathy) • sharing knowledge bits via evolved language • learning / evolving language together Each time step every agent receives from the environment the objects it perceives. Agent may use this knowledge from the environment for decision- making and learning. At the end of each time step an agent produces an output to the environment, which is the action that it wants to perform. Agent environment interaction feedback ME MO RY PERCEPTION DECISION MAKING INDIVIDUAL / SOCIAL LEARNING GENETIC INFORMATION initial rules and values EVOLUTIONARY LEARNING Brain Genes ENVIRONMENT AGENT action visual, auditory information uses / creates Inside the Agent physical (e.g., sex) vs. mental properties (e.g., controller) inheritable vs. learnable properties inheritable – subject to evolution • attitude genes • basic controller structure (decision tree) learnable – subject to individual and social learning • refined controller structure • controller parameters, e.g., learned bias for action A • concepts used in controller, e.g., “enemy” Legend R EPR O D U C TIO N : AGENT M ALE T TURN LEFT MOVE TURN RIG HT A M ATE A VISUAL: AGENT FEM ALE T TURN LEFT MOVE TURN RIG HT A M ATE A SELF: M ALE T RANDOM 0.8 0.2 B NO YES Legend B T A Bias Test Action Decision 0.2 G enetic bias YES Boolean choice Legend R EPR O D U C TIO N : AGENT M ALE T NO YES TURN LEFT MOVE TURN RIG HT A TURN LEFT MOVE TURN RIG HT A 0.6 0.2 0.2 M ATE M ATE 1.0 A VISUAL: AGENT FEM ALE T NO YES TURN LEFT MOVE TURN RIG HT A TURN LEFT MOVE TURN RIG HT A 0.6 0.2 0.2 M ATE M ATE 1.0 A SELF: M ALE T RANDOM Controller = decision Q-tree Agent decisions by DQ-tree, also effected by attitude genes Agent Plant Token Road Environment Discrete time and space (grid world) with plants (energy) and tokens, locations with varying properties (e.g. road / obstacle). Results so far: Evolutionary learning: • learning food seeking behaviour • learning to distinguish poisonous/edible food Social learning: • developing shared lexicon for built-in concepts System: • research platform downloadable (single machine version) • distributed, p2p version beta FP6 FET Open
  • date post

    19-Dec-2015
  • Category

    Documents

  • view

    214
  • download

    0

Transcript of N ew and E mergent W orld models T hrough I ndividual, E volutionary and S ocial Learning Main...

Page 1: N ew and E mergent W orld models T hrough I ndividual, E volutionary and S ocial Learning  Main goal: to realize an evolving artificial.

New and Emergent World models Through Individual, Evolutionary and Social Learning

http://www.newties.org

Main goal: to realize an evolving artificial society capable of exploring the environment and developing its own image of this environment through cooperation and interaction.

Long-term target: to learn how to design agents that are able to adapt autonomously to, and then operate effectively in, environments whose features are not known in advance.

Technical objectives:

1. To develop an artificial society with an emergent culture.

2. To create a powerful “emergence engine” as a combination of individual learning, evolutionary learning, and social learning.

3. To develop social learning mechanisms that allow sharing knowledge with other agents (incl. language evolution).

System: scaleable, p2p infrastructure for large experiments:• Agent complexity• Population size• World size and complexity• Virtual time

Logo ISTLogFET

Vrije Universiteit van Amsterdam

NAPIER UNIVERSITYEDINBURGH

Vrije Universiteit (NL)(Coordinator, agents, evolution):Prof.Dr. A.E. Eiben R. GriffioenC. Tzolov

University of Surrey (UK)(Environments and challenges): Prof. N. Gilbert,S. Schuster

Napier University (UK)(P2P infrastructure) Prof. B. PaechterDr. B.G.W. Craenen

Eötvös Loránd University (HU) (Individual learning, mining emergence) Dr. A. Lorincz A. Bontovics

Dr. Gy. Hévizi

Tilburg University (NL)(language, communication and co-operation)Dr. P. Vogt, Dr. F. Divina

Learning mechanisms Evolutionary learning:

• crossover and mutation on DQ-trees and • crossover and mutation on attitude genes

Individual learning: reinforcement learning Social learning:

• sharing knowledge bits directly (telepathy)• sharing knowledge bits via evolved language• learning / evolving language together

Each time step every agent receives from the environment the objects it perceives. Agent may use this knowledge from the environment for decision-making and learning. At the end of each time step an agent produces an output to the environment, which is the action that it wants to perform.

Agent environment interaction

feedback

M E M O R Y

PERCEPTION

DECISION MAKING

INDIVIDUAL / SOCIAL LEARNING

GENETIC INFORMATION

initial rules and values

EVOLUTIONARY LEARNING

Brain Genes

ENVIRONMENT

AGENT

actionvisual, auditory information

uses / creates

Inside the Agent physical (e.g., sex) vs. mental properties (e.g., controller) inheritable vs. learnable properties inheritable – subject to evolution

• attitude genes• basic controller structure (decision tree)

learnable – subject to individual and social learning• refined controller structure• controller parameters, e.g., learned bias for action A• concepts used in controller, e.g., “enemy”

0.8 0.2

B

NO YES

B T ABias Test Action Decision 0.2 Genetic bias YES Boolean choice

Legend

REPRODUCTION:AGENTMALE

T

NO YES

TURNLEFT

MOVE TURNRIGHT

A

0.6 0.2 0.2

MATE

1.0

A

VISUAL:AGENTFEMALE

T

NO YES

TURNLEFT

MOVE TURNRIGHT

A

0.6 0.2 0.2

MATE

1.0

A

SELF:MALET

RANDOM

0.8 0.2

B

NO YES

B T ABias Test Action Decision 0.2 Genetic bias YES Boolean choice

Legend

B T ABias Test Action Decision 0.2 Genetic bias YES Boolean choice

Legend

REPRODUCTION:AGENTMALE

T

NO YES

TURNLEFT

MOVE TURNRIGHT

A

0.6 0.2 0.2

TURNLEFT

MOVE TURNRIGHT

A

0.6 0.2 0.2

MATE

1.0

MATE

1.0

A

VISUAL:AGENTFEMALE

T

NO YES

TURNLEFT

MOVE TURNRIGHT

A

0.6 0.2 0.2

TURNLEFT

MOVE TURNRIGHT

A

0.6 0.2 0.2

MATE

1.0

MATE

1.0

A

SELF:MALET

RANDOM

Controller = decision Q-tree

Agent decisions by DQ-tree, also effected by attitude genes

Agent

Plant

Token

Road

Environment

Discrete time and space (grid world) with plants (energy) and tokens, locations with varying properties (e.g. road / obstacle).

Results so far: Evolutionary learning:

• learning food seeking behaviour • learning to distinguish poisonous/edible food

Social learning:• developing shared lexicon for built-in concepts

System: • research platform downloadable (single machine version)• distributed, p2p version beta

FP6 FET Open