Abstract
Attempts to optimize simulated behaviors have typically relied on heuristics. A static set of if-then-else rules is derived and applied to the problem at hand. This approach, while mimicking the previously discovered decisions of humans, does not allow for true, dynamic learning. In contrast, evolutionary programming can be used to optimize the behavior of simulated forces which learn tactical courses of action adaptively. Actions of Computer-Generated Forces are created on-the-fly by iterative evolution through the state space topography. Tactical plans, in the form of a temporally linked set of task frames, are evolved independently for each entity in the simulation. Prospective courses of action at each time step in the scenario are scored with respect to the assigned mission (expressed as a Valuated State Space and normalizing function). Evolutionary updates of the plans incorporate dynamic changes in the developing situation and the sensed environment. This method can operate at any specified level of intelligence.
Preview
Unable to display preview. Download preview PDF.
References
Adamson, J., and Joshi, K. (1996) “Genetic Algorithms and Force Simulation,” Proceedings of the Sixth Conference on Computer Generated Forces and Behavioral Representation, D.E. Mullally (ed.), Orlando, FL, pp. 237–242.
Fogel, D.B. (1995) Evolutionary Computation, Toward a New Philosophy of Machine Intelligence, Piscataway, NJ: IEEE Press.
Fogel, L.J. (1995) “The Valuated State Space Approach and Evolutionary Computation for Problem Solving,” Computational Intelligence: A Dynamic System Perspective, M. Palaniswami, Y. Attikiouzel, R.J. Marks, D. Fogel, and T. Fukuda, IEEE Press, NY, pp. 129–136.
Fogel, L.J., Owens, A.J., and Walsh, M.J. (1966) Artificial Intelligence through Simulated Evolution, New York, NY: John Wiley.
Goldman, S.R. (1996) “Knowledge Acquisition and Delivery: Constructing Intelligent Software Command Entities,” Proceedings of the Sixth Conference on Computer Generated Forces and Behavioral Representation, D.E. Mullally (ed.), Orlando, FL, pp. 31–36.
Hieb, M.R. (1996) Training Instructable Agents Through Plausible Version Space Learning, Ph.D. dissertation, George Mason University, Fairfax VA.
Rajput, S., and Karr, C. (1996) “A New Mechanism for Cooperative Behavior in ModSAF,” Proceedings of the Sixth Conference on Computer Generated Forces and Behavioral Representation, D.E. Mullally (ed.), Orlando, FL, pp. 189–199.
Tambe, M., Johnson, W.L., Jones, R.M., Koss, F., Laird, J.E., Rosenbloom, P.S., and Schwamb, K. (Spring 1995) “Intelligent Agents for Interactive Simulation Environments,”, AI Magazine, vol. 16(1), pp. 15–40.
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1997 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Porto, V.W., Fogel, L.J. (1997). Evolution of intelligently interactive behaviors for simulated forces. In: Angeline, P.J., Reynolds, R.G., McDonnell, J.R., Eberhart, R. (eds) Evolutionary Programming VI. EP 1997. Lecture Notes in Computer Science, vol 1213. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0014830
Download citation
DOI: https://doi.org/10.1007/BFb0014830
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-62788-3
Online ISBN: 978-3-540-68518-0
eBook Packages: Springer Book Archive