Dimensionality Effects on the Markov Property in Shape Memory Alloy Hysteretic Environment
- Additional Document Info
- View All
Shape Memory Alloy actuators can be used for morphing, or shape change, by controlling their temperature, which is effectively done by applying a voltage difference across their length. Control of these actuators requires determination of the relationship between voltage and strain so that an input-output map can be developed. To determine this policy and map the hysteretic region, a Reinforcement Learning algorithm called Sarsa was used. Proper use of Reinforcement Learning requires that the learning environment have the Markov Property. However, hysteresis spaces are commonly referenced as non-Markovian due to the fact that state history is needed to properly predict future states and rewards. This paper reveals that this formerly non-Markovian learning environment of Shape Memory Alloy hysteresis can become Markovian by means of increasing the dimensionality of the measured states. The paper compares learning attempts in both versions of the environment and will show that Reinforcement Learning is successful in the modified learning environment by learning a near-optimal policy for controlling the length of a Shape Memory Alloy wire. This is then validated by using the modified Reinforcement Learning agent to learn a near-optimal control policy in an experimental setting. ©2009 IEEE.
author list (cited authors)
Kirkpatrick, K., & Valasek, J.