The Nervous System Appears to Minimize a Weighted Sum of Kinematic Error, Force, and Change in Force when Adapting to Viscous Environments during Reaching and Stepping D.J. Reinkensmeyer, J. Liu, J.L. Emken, and J.E Bobrow, University of California at Irvine The evolution of kinematic performance during reaching and walking in a viscous force field is well captured by a difference equation that relates the current and previous force to kinematic errors: (1) xi +1 = a0 xi + b1 Fi + b0 Fi +1 + c0 th where xi is a scalar measure of kinematic performance on the i movement (e.g. the average deviation from the normative trajectory), and Fi is a scalar measure of the force generated by the viscous force field (e.g. the average lateral force applied) (Fig. 1) [1-4]. These dynamics are consistent with the formation of an internal model of the force field [1, 2]. Here, we prove that these same performance dynamics minimize a cost function J that is a weighted sum of kinematic error, force, and change in force: λ 1 (2) J = ( x i +1 − x d ) 2 + ( ui +1 − αui ) 2 2 2 where ui = the force from the arm on the ith movement, and xd = the desired kinematic performance. The parameter λ weights the cost of the kinematic error versus the force terms. The parameter α weights the cost of the force magnitude term versus the force change term, with α = 0 weighting only magnitude, and α = 1 weighting only change. The minimum of this cost function occurs when: ∂x ∂J (3) = ( x i +1 − x d ) i +1 + λ (ui +1 − αui ) = 0 ∂ui +1 ∂ui +1 The robot/limb dynamics are spring dynamics for the class of viscous force field presumed here (Fig. 2): 1 (4) xi − x d = ( Fi + u i ) K

where K = limb stiffness. The controller that minimizes the cost function is thus: 1 (5) ui +1 = αui − ( x i +1 − x d ) λK The next kinematic performance xi+1 can be estimated using a Taylor’s series expansion of (4): ∂x ∂x 1 (6) x = x + i ∆u + i ∆ F = x + ( u − u ) i +1

i

∂ui

∂Fi

i

K

i +1

i

where we have assumed that the force field does not change, i.e. ∆F =0. Combining (5) and (6) gives: λαK 2 + 1 K (7) (8) ui +1 = fui − g ( xi − xd ) f = g= λK 2 + 1

λK 2 + 1

The error-based learning controller (7) increments the motor command in proportion to the previous error, in the direction that reduces error. This controller can be viewed as implementing an inverse model D-1 that estimates the arm force ui+1 required for the desired kinematic performance xd, i.e. ui+1 =D-1(xd), where D-1 depends on experience (ui and xi). If f < 1 the controller increments a decremented version of the previous motor command in a “forgetting” process. The condition f < 1 corresponds to α < 1 in the cost function. Thus, forgetting corresponds to minimizing force. This learning law is similar to recently proposed learning models [2, 5], except for the critical difference of the forgetting factor. Combining the controller (7) with the plant dynamics (4) gives: xi +1 = ( f −

g f 1 g ) xi − Fi + Fi +1 + (1 − ( f − )) x d (9) i.e. the same dynamics as (1) K K K K

Thus, the controller that creates the learning dynamics (1) is an error-based learning law that constructs an internal model and incorporates a forgetting factor. This controller minimizes a weighted of sum of error, force, and change in force. The learning law parameters can be identified by multiple linear regression of (9) on experimental data. The cost function weights that best explain the data variance can be inferred from the learning law parameters using (8). For stepping in a viscous force field (Fig. 1), the identified forgetting factor was significantly less than one, (p < 0.001, mean = 0.77 ± 0.1 SD, 10 subjects) verifying that the CNS minimizes force per (2). A useful prediction based on these learning dynamics is that internal model formation can be accelerated by transiently amplifying the force field strength (Fig. 3) [4].

[1] R. A. Scheidt, J. B. Dingwell, and F. A. Mussa-Ivaldi, "Learning to move amid uncertainty," J Neurophys, vol. 86, pp. 971-85, 2001. [2] K. A. Thoroughman and R. Shadmehr, "Learning of action through adaptive combination of motor primitives.," Nature, 407: 742-7, 2000. [3] J. Liu and D. J. Reinkensmeyer, "Motor adaptation as an optimal combination of computational strategy," Proceedings of the 2004 IEEE Engineering in Medicine and Biology Society, 2004. [4] J. Emken and D. J. Reinkensmeyer, "Accelerating motor adaptation by influencing neural computations," Proceedings of the 2004 IEEE Engineering in Medicine and Biology Society, 2004. [5] O. Donchin, J. T. Francis, and R. Shadmehr, "Quantifying generalization from trial-by-trial behavior of adaptive systems that learn with basis functions: theory and experiments in human motor control," Journal of Neuroscience, vol. 23, pp. 9032-9045, 2003. A

−20

y (cm)

−30 −40 −50

Step Height Error (cm)

0

20 x (cm)

40

60

B

Data Model

5 0 −5 0

5

10

15

20 25 Step Number

30

35

40

45

14

20

Baseline Fast Learn

12 Slope = 0.5823

R2 = 0.944 p= 0.000

15

Step Height (cm)

Direct Effect − Step Height Error (cm)

Figure 1 Left: Diagram of robotic device used to measure and perturb stepping. The robot makes use of a linear motor with two forcer coils and a V-shaped linkage to drive motion of its apex in the parasagittal plane. The apex is attached through a padded cuff and revolute joint to the subject’s lower shank. We used the robot to apply an upward force field that had a magnitude proportional to the forward velocity of the leg during swing. Right: Sample step data from a single subject for a single exposure to the force field. A: Step trajectories of the lower shank during exposure to the force field. Shown are the normal stepping trajectory in the null field (squares); the "direct effect", which is the first step in force field (triangles); a step produced after adaptation to field exposure (circles); and the "after effect," which is the first step in the null field following adaptation (diamonds). B: Step height error, referenced to the normal stepping height, before, during, and after application of the force field. The force field was turned "on" at step 10 and "off" at step 35. The symbols correspond to the trajectories in A. The dashed line shows the best fit of Equation 1. For this subject, the fit produced R2 = 0.86.

10 8 6

10

5

4 0 2

2 25

30

35

40 Force (N)

45

50

55

Figure 2 Spring-like dynamics of the leg in response to the viscous force field. Each data point is the change in step height due to unexpected application of the field. The field strength was varied; the abscissa shows the peak force applied. Each point is the average across ten subjects.

4

6

8 Steps

10

12

14

Figure 3 Example of how the learning dynamics can be exploited to accelerate learning. The “Baseline” data shows adaptation to the viscous force field during stepping for one subject, for ten exposures to the field (exposures overlaid, field turned on at step 5). By transiently amplifying the force field on Step 5 with a magnitude predictable using Equation (1), adaptation can be accelerated (“Fast Learn” experiment).

Optimization of Cost and Effort, Feedback Control using ...

The cost function weights that best explain the data variance can be inferred ... had a magnitude proportional to the forward velocity of the leg during swing.

82KB Sizes 1 Downloads 262 Views

Recommend Documents

Search, Effort, and Locus of Control Andrew McGee ...
University of Arkansas. February 2016 ... Address: Department of Economics, University of Arkansas, 402 Business. Building, Fayetteville ..... beliefs, other characteristics, and a linear time trend.7 In all of the tables discussed in this section, w

Optimization of Ship steering control system using ...
Tankers are full form ships and are thus generally difficult to maneuver, ... 3. The controller changes the heading of the vessel by manipulating the rudder.

Search, Effort, and Locus of Control Andrew McGee ...
imaginativeness or intellect (Goldberg, 1992), the Rosenberg Self-Esteem Scale (Rosenberg, 1965), and a ..... services), the number of job applications submitted (Caliendo et al. 2015), the ..... Princeton: Princeton University Press. Rotter ...

Using token leaky bucket with feedback control for ...
Oct 28, 2002 - Consider a host computer connected to the network through a single-server queuing system with constant service rate as depicted in fig. 1.

Feedback Control Tutorial
Design a phase lead compensator to achieve a phase margin of at least 45º and a .... Both passive component variations are specified in terms of parametric ...

Schaum's Outline of Feedback and Control System.pdf
Page 3 of 523. Schaum's Outline of Feedback and Control System.pdf. Schaum's Outline of Feedback and Control System.pdf. Open. Extract. Open with. Sign In.

Reducing Annotation Effort using Generalized ...
Nov 30, 2007 - between model predicted class distributions on unlabeled data and class priors. .... can choose the instance that has the highest expected utility according to .... an oracle that can reveal the label of each unlabeled instance.

Wage and effort dispersion
choose how much capital to purchase. While they address the ... the paper.1 A worker exerts a continuous effort e, which yields one of two levels of output. With .... it will get at least as many workers in expectation if not more, and will have larg

OPTIMIZATION OF ORBITAL TRAJECTORIES USING ...
a big number of times, reducing the exploration of the search space; at the end of .... (1971) directly calculate the velocity vector in the extreme points of the orbital arc ..... Algorithms in data analysis, Printed by Universiteitsdrukkerij Gronin