Robust Metropolis-Hastings Algorithm for Safe Reversible Markov Chain Synthesis Mahmoud El Chamie Abstract— This paper presents a new method to synthesize safe reversible Markov chains via classical Metropolis-Hastings (M-H) algorithm. Classical M-H algorithm does not impose safety constraints on the probability vector for the resulting Markov chain. This paper presents a new M-H algorithm for Markov chain synthesis that handles safety constraints while ensuring reversibility and a desired stationary (steadystate) distribution. Specifically, we provide a convex synthesis method that incorporates the safety constraints via proper choice of the proposal matrix for the M-H algorithm. The resulting proposal matrix is a stochastic matrix that ensures safety for a nominal stationary distribution. Then it is shown that the M-H algorithm with this proposal matrix, robust M-H algorithm, also ensures safety for a well-characterized convex set of stationary distributions, which also includes the nominal stationary distribution. We also present a convex synthesis method for the proposal matrix to maximize the size of this resulting set of feasible stationary distributions for the robust M-H algorithm. Simulation results are also provided to demonstrate that there is no tradeoff between the speed of convergence and the robustness.

I. I NTRODUCTION The Metropolis-Hasting (M-H) algorithm [1], [2], [3] is a method for obtaining random samples from a probability distribution when direct sampling is difficult. The M-H algorithm builds on the theory of Markov processes, MarkovChain-Monte-Carlo sampling methods [4], [5], [6], [7], [8], [9], graph theory [10], [11], and Lyapunov stability analysis. It is also aided by research in designing fast mixing Markov chains that incorporate constraints on transition probabilities [12], [13], [14]. Given a matrix K (called proposal matrix), this algorithm can be used to design the stochastic transition matrix M of a Markov chain to satisfy some system specifications (e.g., a prescribed stationary distribution). The matrix M inherits the key properties of the proposal matrix K, such as speed of convergence, while satisfying the system specifications. Thus the M-H is very useful for applications when online Markov chain synthesis is needed because it can be implemented easily and executed very efficiently on real-time processors. However, currently, the M-H algorithm cannot impose safety constraints on the probability distribution vector, i.e., upper bound constraints on the probability distribution. Such constraints appear naturally in many applications, e.g., to bound/minimize the probability of conflicts/collisions among agents in randomized motion planning of swarms [15], which may arise due to clustering during The authors are with the University of Texas at Austin, department of Aerospace Engineering and Engineering Mechanics, 210 E. 24th St., Austin, TX 78712 USA. Emails: [email protected] and [email protected]

Behc¸et Ac¸ıkmes¸e transition to the desired steady-state distribution. These hard constraints on the probability vector of the Markov chain are often referred to as safety constraints [16], [17]. Markov chain propagates the probability distribution (discrete) vector x(t) ∈ Rm as follows, x(t + 1) = M (t)x(t)

t = 0, 1, 2, . . .

(1)

where M (t) is a column stochastic matrix for all time; hence x(t) stays normalized as 1T x(t) = 1 for all t ≥ 0 where 1 is the vector of all ones. In many applications, it is desired to design the column stochastic matrix M to satisfy some specifications. For example, in randomized motion planning (RMP) [18], [19], [15], x(t) describes the probability of an agent (e.g., vehicle) to be in a certain state and M (t) determines the probability distributions for possible transitions. In gossiping and wireless sensor networks [20], [21], (1) is the dynamics for the evolution of an estimate for a relevant physical quantity as temperature, pressure, etc. In voting models [22], x(t) determines the preference of a group of people towards a given object of interest (e.g., application, leader, etc.). In belief propagation [23], (1) is a model for information dissemination in social networks. In consensus protocols [24], [25], the transition matrix (called weight matrix in the context of consensus protocols) is designed for fast convergence among networked agents for the value of a quantity of interest. When x(t) is a discrete probability vector for an underlying Markov chain governed by equation (1) dynamics, x(t) would satisfy some constraints naturally due to column stochasticity of M (t) such as x(t) ≥ 0 and 1T x(t) = 1 for all t = 0, 1, .... There can also be additional constraints characterized by hard safety upper bounds on the probability vector, i.e., x(t) ≤ d for all t ≥ 0, (2) where d ∈ Rm is a constant non-negative vector. The purpose of this paper is to expand the M-H algorithm for the generation of a transition matrix M (t) that is both convergent and incorporates safety constraints for time-varying system specifications (i.e., time-varying desired stationary distributions). In summary, our main contributions are: (i) Providing a new robust LP-based method to synthesize the proposal matrix of the M-H algorithm to satisfy safety constraints of the form given by (2); (ii) Providing an efficient M-H algorithm, generating safe Markov chains, that can handle time-varying desired stationary probability distributions; (iii) Studying the speed of convergence of the resulting transition matrix constructed by the robust

M-H algorithm in the presence of safety constraints. It is important to note that our earlier results [17] can synthesize constant Markov chains with all the constraints considered in this paper via convex optimization methods. The main contribution of this paper is obtaining feasible synthesis via the M-H algorithm, where the proposal matrix is designed such that the resulting Markov chain satisfies the safety constraints for a well-characterized set of stationary distributions. This allows us to quickly adapt to time-varying stationary distribution specifications without recomputing the proposal matrix, which is a very useful property for a number of applications where this adaptation must happen in realtime, see [26] for an example application. II. F ORMULATION OF R EVERSIBLE M ARKOV C HAIN S YNTHESIS WITH S AFETY C ONSTRAINTS A. Notation In this paper, small bold letters in general signifies a vector (e.g., v whose elements are indicated as v1 , v2 , . . . ), capital letters are in general matrices (e.g., M whose i-th row jth column element is denoted by Mij ). 0 and 1 are the matrices/vectors of conformable dimensions with all zeros and ones respectively; ei is a vector with its ith entry +1 and others zero; I is the identity matrix; xi = eTi x and Xij = eTi Xej ; Q = QT  ()0 is a symmetric positive (semi-)definite matrix; R > (≥)H implies that Rij > (≥)Hij for all i, j; v is a probability vector if v ≥ 0 and 1T v = 1; Pm is the set of probability m × 1 vectors, i.e., v ≥ 0 and 1T v = 1 for all v ∈ Pm . denotes the point-wise, Hadamard, product; An undirected graph with self-loops included is denoted by G = (V, E) where V = {1, . . . , m} is a set of vertices and E = {1, . . . , q} if the set of edges. The adjacency matrix A of G = (V, E) is: Aij = 1 if (i, j) ∈ E and Aij = 0 otherwise. Let CG ⊆ Rm×m be the set of matrices following the graph G defined as follows: X ∈ CG implies that Xij = 0 if the link (i, j) ∈ / E and Xij > 0 otherwise. B. Markov Chain Specifications We consider a Markov chain describing the evolution of a discrete probability vector x(t) ∈ Pm in time given by (1), where M (t) is a column stochastic matrix for all time; hence the probability vector x(t) ≥ 0 stays normalized as 1T x(t) = 1 for all t ≥ 0. Motivated by Markov chain terminology, M (t) is referred to as the transition matrix where Mij (t) is the probability of transition from a state j to state i at time t. We first consider the case with a constant transition matrix M (t) = M for all t. We will discuss the time-varying case after introducing the M-H algorithm to handle timevarying stationary distributions. Specifically our objective is to synthesize an m × m transition matrix M such that the resulting Markov chain via (1) has the following properties: ˆ ∈ Pm : lim x(t) = v ˆ ∀x(0) ∈ 1) Desired steady state: v t→∞ m P . 2) Reversibility: vˆi Mji = vˆj Mij , i, j = 1, . . . , m. 3) Transition constraints: Mij = 0 when (ij) ∈ / E.

4) Safety constraints: x(t) ≤ d for all t ≥ 0, and for a given vector 0 ≤ d ≤ 1. ˆ defines a probability distribuThe desired steady state v tion over the discrete set of states. In consensus protocols, 1 ˆ = m 1 is the uniform distribution. In RMP, it is the v desired density distribution of autonomous agents over a given operational area. Transition constraints are given by an adjacency matrix describing the set of feasible state transitions.1 Safety constraints bound the probability distribution during both: the transient and the stationary dynamics. For example, in RMP for multi-agent systems, this constraint can prevent overcrowding of agents in subregions. C. Convex Characterization of the Specifications This section summarizes our results on convex formulations of the four system specification properties for a timeinvariant Markov matrix M . These convex representations allow us to formulate Linear Matrix Inequality (LMI) problems for the synthesis of reversible Markov chains for given steady-state distributions [17]. Stochasticity, Transition, and Reversibility Constraints: First of all, the Markov matrix must satisfy the column stochasticity constraints described as M ≥ 0 and 1T M = 1T ,

(3)

which are linear equality and inequality constraints on M . The transition constraints can be expressed via the adjacency matrix A of the graph G that represents all feasible transitions M ∈ CG , ( 0 if(i, j) ∈ /E (11T − A) M = 0 , Mij = ⇐⇒ ∃  > 0 s.t. M ≥ A > 0 if(i, j) ∈ E (4) which impose linear constraints on M . Here we can choose  a sufficiently small positive scalar. Reversible chains have favorable properties (e.g., detailed balance condition) and are commonly utilized in applications of Markov Chain Monte Carlo (MCMC) methods, e.g., in birth death processes, M/M/1 queues, and symmetric random walks on graphs. A Markov matrix M is reversible with a stationary distribution v if and only if [27] M diag(v) = diag(v)M T .

(5)

The above equation implies that v is a steady-state distribution for the resulting Markov chain, which can be obtained by simply multiplying it by 1, M v = v. The set of all admissible Markov matrices with a given steady-state distribution, v, and adjacency matrix, A, is: MG (v) := {Matrices satisfying (3), (4), and (5)}.

(6)

The results in this paper hold (with minor modifications) when v has some zero entries. Without loss of generality, we assume in what follows that v > 0. 1 In consensus protocols, they represent possible communication links between agents. In RMP, transition constraints represent feasible transitions for an agent as a function of its current state.

We can parameterize all reversible Matrices having v as the steady-state distribution. Let ∆ be an m × m matrix defined by the stationary distribution v as follows: ( vi /vj if i < j (7) ∆ij (v) := 1 else. Next we give a useful parameterization of all reversible Markov matrices for a given v. Lemma 1. M ∈ MG (v) if and only if there exists an m×m matrix Y such that  M = ∆(v) Y + I − diag 1T (∆(v) Y ) , (8) M ≥ I, Y = Y T , Y ∈ CG where  > 0 is a positive scalar. Proof. Let Y be a matrix such that (8) is satisfied, then 1T M = 1T (∆(v) Y ) + 1T − 1T (∆(v) Y ) = 1T , and adding that M ≥ I ≥ 0 implies (3) is satisfied. For i < j, we have Mij = ∆ij (v)Yij = (vi /vj )Yij , and for i > j, we have Mij = ∆ij (v)Yij = Yij . Then, adding that Y is a symmetric matrix, we get Mij = (vi /vj )Yij = (vi /vj )Yji = (vi /vj )Mji , which implies that (5) is satisfied. Since Y ∈ CG and Mii ≥ , then M ∈ CG and (4) is satisfied. Therefore M ∈ MG (v). To show the other direction of the lemma, now suppose that M ∈ MG (v), it is sufficient to construct Y from M such that Y satisfies equations (8). It is indeed the case by constructing Y as follows: Yij = Mij if i > j, Yij = Yji if i < j, Yij =  if i = j. Lemma 1 shows that any any matrix M that generates a reversible chain can be parameterized by a set of linear constraints. Convergence to Steady-State Distribution: The following well known result (see for example [18]) shows that asymptotic convergence to v is ensured on matrix M when the spectral radius condition is satisfied, ρ(M − v1T ) < 1.

(9)

Lemma 2. Consider a column stochastic matrix, M , such that M v = v. Consider the dynamical system (1). Then, limt→∞ x(t) = v holds for any initial probability vector x(0) if and only if (9) is satisfied. The following lemma gives a known characterization for the convergence of the system using graph G properties Lemma 3. If G = (V, E) is ρ(M − v1T ) < 1 for all M ∈ MG (v).

connected,

then

Proof. When G is connected, then for any pair of vertices i and j, we can find a path iu1 u2 . . . ul−1 j, and thus (M l )ij ≥ Miu1 Mu1 u2 . . . Mul−1 j > 0 (the last inequality is due to (4)). Since l ≤ m, then M is irreducible as for every pair i, j of its index set, there exists a positive integer l ≡ l(i, j) such that (M l )ij > 0 [28, p. 18]. Moreover, since self-loops belong to E, then Mii > 0 for all i and thus M is primitive having a

unique stationary distribution v, so limt→∞ (M −v1T )t = 0 and the lemma follows. We will assume that the graph G is connected thereafter. Note that MG (v) is a convex set because (3), (4) and (5) are linear (in)equalities, which facilitate convex synthesis. Safety: We consider the following safety constraints [16], [29], [17]. x(t) ≤ d, t = 1, 2, . . . ,

if

x(0) ≤ d,

(10)

which impose that the probability vector (density distribution) is bounded by the vector d. These constraints can be useful in RMP for collision/conflict mitigation by limiting the densities, hence eliminating crowding during transitions. We have the following recent result [17] on Markov chain synthesis. Theorem 1. Consider the Markov chain dynamics given by (1) with a constant M matrix. Given a scalar γ ∈ [0, 1], for any x(0) ≤ d, the following holds x(t) ≤ (1 − γ) d

∀t = 1, 2, ...

if and only if there exist two variables S ∈ Rm×m and y ∈ Rm such that: M + S + y1T ≥ 0,  M + S + y1T d ≤ y + (1 − γ)d. S ≥ 0,

(11)

Proof. This theorem was proved in [17] for γ = 0. When γ ∈ [0, 1], the proof can be refined from [17] by simply having a factor (1 − γ) in the derived equations. III. ROBUST M ETROPOLIS -H ASTINGS (M-H) A LGORITHM This section presents a new M-H synthesis method for the M (t) matrix in (1) as a function of a time-varying desired distribution, v(t). If the desired distribution does not change with time, then the algorithm will generate a constant Markov matrix. However, if v is time-varying, the Markov matrix will be time-varying as well. Our main result adapts the wellknown M-H algorithm to handle the safety constraints for a time-varying stationary distribution. In particular, we show that M-H algorithm can be applied for a well-characterized class of stationary distributions by designing the proposal matrix to ensure convergent and safe Markov chains. Beyond having a new M-H result to generate safe Markov chains, the resulting algorithm presents a computationally inexpensive method, which makes it easily adaptable for online computations once a proper proposal matrix is generated offline via using convex optimization techniques. In summary, this section presents a method to synthesize the proposal matrix based on the results of the previous section, and will also characterize the set of stationary distributions for which MH algorithms generates safe reversible Markov chains.

A. M-H Algorithm The M-H algorithm [4], [2] is a Markov Chain Monte Carlo (MCMC) method for obtaining a sequence of random samples from a distribution. Definition 1. [M-H Algorithm for a given v] The M-H algorithm produces an M matrix given by: ( Kij Fij if i 6= j Mij = (12) P Kjj + k6=j (1 − Fkj )Kkj if i = j where K is a proposal matrix that satisfies K ≥ 0 and 1T K = 1T ; v > 0; and F is an acceptance matrix, which satisfies for i 6= j,   vi Kji i, j = 1, . . . , m. (13) Fij = min 1, vj Kij Note that if Kij = 0, then Mij = Mji = 0. Similarly, having Kij > 0 and Kji > 0 implies that Mij > 0 and Mji > 0. Consequently, when v > 0 and the matrix F is chosen by (13), we can impose transition constraints, given by (4), on the proposal matrix, K, to guarantee that M also satisfies the transition constraints. Lemma 4. Consider the M-H algorithm, for some v > 0, given by (12). If the proposal matrix K satisfies K ∈ MG (u) for some probability vector u, then the resulting Markov matrix M ∈ MG (v). Proof. The matrix M is by the algorithm’s construction a column stochastic matrix satisfying the reversibility assumption. Since K ∈ MG (u), then K ∈ CG , and thus M ∈ CG . As a result, M ∈ MG (v) and this ends the proof. Remark: The M-H can transform any proper proposal matrix into a Markov matrix with a desired stationary probability distribution v. Incorporating safety constraints into the resulting transition matrix M for a time-varying distribution v(t) is not straight forward. In the example of multi-agent system motion planning, even if the agents have sufficient computational capabilities (being able to solve optimization problems online), the safety might not be guaranteed. For example, two agents operating via two different safe Markov matrices might lead to an unsafe combined behavior. Therefore, it becomes natural to ask under what conditions, the M-H procedure can be safe? It turns out that if the proposal ˆ is safe and have some matrix for a nominal distribution v robustness properties, then the M-H procedure can also be ˆ as we will safe as long as v is in the neighborhood of v discuss next. B. Robust Proposal for Safe M-H Algorithm By construction, M-H algorithm can ensure reversible Markov chains with the desired steady-state distribution and transition constraints. However, the safety specification is not necessarily guaranteed. In this section, we study the effect of the proposal matrix K on the safety of M-H algorithm. For this purpose, we introduce the following definitions.

Definition 2 (S-Safe). M-H algorithm with a given proposal matrix K is called S-Safe if the resulting matrix M leads to Markov chains satisfying the safety constraints (10) for all steady-state distributions v ∈ S. Definition 3. Given a matrix K ≥ 0 and γ ∈ [0, 1], Vγ (K) ⊆ Pm is defined as follows: Vγ (K) = {v ∈ Pm : v satisfies (14)}, m X

max{0, vi Kki −vk Kik } ≤ γvi , for i = 1, . . . , m. (14)

k=1

Note that the set Vγ (K) is parameterized by γ, which can be considered as a robustness parameter as discussed next: The larger γ is, the larger is the set for which the M-H algorithm is safe. Proposition 1. Consider a proposal matrix K ∈ MG (ˆ v) used in M-H algorithm that results in a reversible Markov chain. Then Vγ (K) is a nonempty convex set for any γ ∈ [0, 1]. ˆ , the Proof. Vγ (K) is nonempty because by considering v stationary distribution of K, the following equations hold vˆi Kki = vˆk Kik for all i and k. Therefore the left hand side of (14) is zero and the equations are satisfied for any γ ∈ [0, 1]. ˆ is the unique value of v that In particular, if γ = 0, then v satisfies the inequalities and in this case Vγ is a singleton. If γ = 1,P then Vγ is the set of all probabilityP vectors (Vγ = Pm ) m m since k=1 max{0, vi Kki − vk Kik } ≤ k=1 vi Kki = vi . Vγ is a convex set of probability vectors v because of the following argument: given a matrix K, then vi Kki − vk Kik is a convex function of v (a linear function of v), max{0, vi Kki −Pvk Kik } is convex (the maximum of m convex functions), k=1 max{0, vi Kki − vk Kik } − γvi is convex (sum of convex functions), the set defined by Pm k=1 max{0, vi Kki − vk Kik } − γvi ≤ 0 is a convex set (sub-level set of a convex function), and finally Vγ is a convex set (intersection of convex sets). Moreover, Vγ ˆ because for any γ > 0, a defines a neighborhood around v ˆ would satisfy equations small enough perturbation around v (14). We can now give the main technical result of this paper. Theorem 2. Suppose that there exist variables Y ∈ Rm×m , S ∈ Rm×m ,  > 0, and y ∈ Rm such that the proposal matrix K satisfies the following conditions  K = ∆(ˆ v) Y + I − diag 1T (∆(ˆ v) Y ) , K ≥ I, Y = Y T , Y ∈ CG (K + S + y1T )d ≤ y + (1 − γ)d,

(15)

S ≥ 0, K + S + y1T ≥ 0, where ∆(ˆ v) is given by (7) and γ ∈ [0, 1]. Then the M-H algorithm with the proposal matrix K is Vγ -Safe. Proof. The proof is provided in the Appendix. It is important to note that, the M-H algorithm is safe for any desired distribution if γ = 1. Also note that

0.055

90%

Spectral Gap of Proposal Matrix K

Size of the neighborhood V as compared to the overall feasible ditributions

100%

80% 70% 60% 50% 40% 30% 20% 10% 0 0.2

0.3

0.4

0.5

0.6

0.7

0.8

Spectral Gap (1 − |λ2 |)

0.05

0.045

0.04

0

Robustness parameter γ

Fig. 1. The size of the neighborhood as function of the parameter γ. (One instance of a connected random network (RGG) of 10 nodes, each element in the safety vector is chosen uniformly at random from [0.4, 1], nominal stationary distribution is chosen uniformly at random from distributions satisfying the bound.)

Theorem 2 results in a Markov matrix K with reversible ˆ (via chains and the set Vγ is a convex set containing v Proposition 1). Fig. 1 illustrates the size of the set Vγ as a function of the robustness parameter γ.2 It demonstrates that, as γ approaches 1, the size of the set becomes larger and approaches the sample space, i.e., the simplex defining the probability distributions. This implies that maximizing γ maximizes the size of the set Vγ , which is the basis of the following LP-based synthesis for the proposal matrix K. maximize

γ

subject to

(15).

K,γ,S,Y,y

(16)

Remark: The robustness parameter γ and the matrix K are not independent, i.e., fixing one of them and solving for the other can render the problem infeasible. Notwithstanding this, (16) provides the maximum possible value for γ for which a feasible proposal matrix K can be found. C. Robustness and Convergence Speed Note that, in the optimization problem (16), the speed at which the proposal matrix K converges to its stationary ˆ is not taken into account by the constraints. distribution v Though the convergence is guaranteed (due to Lemma 3), the speed at which the system converges can be arbitrarily slow. Therefore, it is desirable to generate proposal matrices having both fast speed of convergence and robustness properties. It is well-known that the speed of convergence of a Markov matrix to its stationary distribution is governed by the second largest eigenvalue in magnitude |λ2 |. Since K is searched over Markov matrices for reversible chains, the fastest converging proposal matrix can be computed via solving the 2 The figure is generated using Monte Carlo simulations where random feasible distribution vectors are generated. The size of Vγ represents the percentage of simulation runs where the random generated vector v satisfies (14).

(1/5)γ max

(2/5)γ max

(3/5)γ max

(4/5)γ max

γ max

Fig. 2. The spectral gap of the proposal matrix is not sensitive to the robustness parameter γ. Therefore, we can have matrices that are both: robust and have fast convergence. (Random Geometric Graphs (RGG), n = 20 nodes, m = 53 links, connectivity constant c = 0.7, γmax = 0.4, uniformly random safety bound where di ∈ [0.4, 1], random nominal ˆ .) distribution v

following Semi-Definite Programming (SDP) problem [30] [19] for a prescribed robustness measure γ: minimize

λ

subject to

− λI  Q−1 KQ − qqT  λI

K,λ,S,Y,y

(17)

Eq. (15) ˆ 1/2 is the element-wise square root of v ˆ and where q = v Q = diag(q). Note that γ is not a variable in this formulation but a given value. The above SDP has a feasible solution for [0, γmax ] where γmax is computed via (16). Then a proportional measure of convergence speed is the spectral gap, defined as sg = 1 − λ∗ where λ∗ is the minimal value of the SDP (17). Since any feasible solution from (15) for a given γ1 is also a feasible solution for γ2 where γ2 ≤ γ1 , we expect the convergence speed to be a non-increasing function of γ when γ varies from 0 to γmax . It turns out that in fact, the speed of convergence has low sensitivity to the robustness parameter. In other words, we can have both fast convergence and robustness at the same time. Fig. 2 demonstrates this observation via simulation on an example with 20 states, i.e., x(t) ∈ P20 . IV. C ONCLUSIONS A robust M-H algorithm is introduced that ensures the satisfaction of safety upper-bound constraints for the synthesized Markov chains. The proposal matrix of the M-H algorithm is designed offline using a linear programming formulation to maximize the set of robust desired stationary distributions that the M-H algorithm can handle safely. This solution method is further improved to achieve fast convergence using a semidefinite programming formulation. Our results show that both, speed and robustness, can be achieved simultaneously.

to show that ui ≤ γ, X vk (Kki − Kik ) ui = vi 1

A PPENDIX P ROOF OF T HEOREM 2

k∈Ni

The first two conditions in Equation (15) ensure that K ∈ M(ˆ v) by Lemma 1. The last two conditions in Equation (15) ensure that K satisfy the conditions of Theorem 1 and thus K is safe. Moreover, using Lemma 4, the resulting transition matrix M from the M-H algorithm belongs to MG (v). It remains to prove that M is safe when v ∈ Vγ . Note that the last two lines are equivalent to the following expression (Theorem 1), T

Kx ≤ (1 − γ)d, for all 0 ≤ x ≤ d, x 1 = 1.

(18)

To show that M is safe, we need to show that for any 0 ≤ x ≤ d, xT 1 = 1, we have that M x ≤ d, or equivalently X

Mik xk ≤ di , for i = 1, . . . , m.

Notice that from the M-H algorithm (depending on the ratio Kik ) either (a) C ≤ 1, Mik = Kik , Mki = vvki Kik or C := vvki K ki (b) C ≥ 1, Mik = vvki Kki , Mki = Kki . Thus we can divide the neighbors of i in G, Ni , into two mutually exclusive sets: Ni1 = {k ∈ Ni ; condition (a) is satisfied}, and Ni2 = {k ∈ Ni ; condition (b) is satisfied} satisfying Ni1 ∪ Ni2 = Ni and Ni1 ∩ Ni2 = φ. Therefore we have: X

Mki

k∈Ni

=1−

X vk X Kik − Kki . vi 1 2

k∈Ni

k∈Ni

As a result, for i = 1, . . . , m X

Mik xk = Mii xi +

k

X

Mik xk

k∈Ni



 X vk X Kik − = 1 − Kki  xi vi 1 2 k∈Ni k∈Ni X X vi Kki xk + Kik xk + vk k∈Ni1 k∈Ni2   X X v k = Kik xk +  (Kki − Kik ) xi |{z} v i 1 k ≤d | {z } | k∈Ni {z } i ≤(1−γ)di :=ui  X  vi (19) + Kki − Kik xk vk k∈Ni2 | {z } ≤0

≤ (1 − γ)di + ui di .

=

1 X (vi Kik − vk Kik ) vi 1 1 vi

k∈Ni m X

max{0, vi Kik − vk Kik }

(21)

k=1

1 γvi vi = γ, ≤

(22) (23)

where (21) follows from the definition of Ni1 and the inequality (22) follows from the definition of Vγ . Combining (20) and (23) shows that M is safe (i.e., M x ≤ d for all x ≤ d), which concludes the proof. R EFERENCES

k

Mii = 1 −

=

(20)

Note that, the first inequality in (19) is due to K satisfying (18), the last inequality follows directly from the M-H algorithm (i.e., because for any k ∈ Ni2 , C ≥ 1). It remains

[1] P. Diaconis and L. Saloff-Coste, “What do we know about the Metropolis algorithm?,” Jrnl. of Computer and System Sciences, vol. 57, pp. 20–36, 1998. [2] L. J. Billera and P. Diaconis, “A geometric interpretation of the Metropolis-Hastings algorithm,” Statistical Science, vol. 16, no. 4, pp. 335–339, 2001. [3] J. S. Liu, “Metropolized independent sampling with comparisons to rejection sampling and importance sampling,” Statistics and Computing, vol. 6, pp. 113–119, 1996. [4] N. Metropolis and S. Ulam, “The Monte-Carlo method,” American Statistical Association, vol. 44, pp. 335–341, 1949. [5] P. Diaconis, “The Markov chain Monte Carlo revolution,” Bulletin of the American Mathematical Society, vol. 46, pp. 179–205, 2009. [6] W. Hastings, “Monte Carlo sampling methods using Markov chains and their applications,” Biometrika, vol. 57, pp. 97–109, 1970. [7] J. S. Liu, Monte Carlo strategies in scientific computing. SpringerVerlag, 2001. [8] G. Fishman, Monte Carlo, Concepts, algorithms and applications. Springer, New York, 1996. [9] D. Bayard and A. Schumitzky, “Implicit dual control based on particle filtering and forward dynamic programming,” International Journal of Adaptive Control and Signal Processing, vol. 24, no. 3, pp. 155–177, 2010. [10] N. Deo, Graph Theory with Applications to Engineering and Computer Science. Prentice-Hall, 1974. [11] M. Mesbahi and M. Egerstedt, Graph Theoretic Methods in Multiagent Networks. Princeton University Press, 2010. [12] P. Diaconis and D. Stroock, “Geometric bounds for eigenvalues of Markov Chains,” The Annals of Applied Probability, vol. 1, no. 1, pp. 36–61, 1991. [13] S. Boyd, P. Diaconis, and L. Xiao, “Fastest mixing Markov Chain on a graph,” SIAM Review, vol. 46, pp. 667–689, 2004. [14] S. Boyd, P. Diaconis, P. Parillo, and L. Xiao, “Fastest mixing Markov Chain on graphs with symmetries,” SIAM Jrnl. of Optimization, vol. 20, no. 2, pp. 792–819, 2009. [15] N. Demir, U. Eren, and B. Ac¸ıkmes¸e, “Decentralized probabilistic density control of autonomous swarms with safety constraints,” Autonomous Robots, In press, 2015. [16] A. Arapostathis, R. Kumar, and S. Tangirala, “Controlled markov chains with safety upper bound,” IEEE Trans. on Automatic Control, vol. 48, no. 7, pp. 1230–1234, 2003. [17] B. Ac¸ıkmes¸e, N. Demir, and M. Harris, “Convex necessary and sufficient conditions for density safety constraints in markov chain synthesis,” In press, IEEE Trans. on Automatic Control, 2015. [18] B. Ac¸ıkmes¸e and D. S. Bayard, “A Markov chain approach to probabilistic swarm guidance,” American Control Conference, Montreal, Canada, pp. 6300–6307, 2012. [19] B. Ac¸ıkmes¸e and D. S. Bayard, “Markov chain approach to probabilistic guidance for swarms of autonomous agents,” In Press, Asian Journal of Control, 2014.

[20] S. Boyd, A. Ghosh, B. Prabhakar, and D. Shah, “Randomized gossip algorithms,” IEEE/ACM Trans. Netw., vol. 14, pp. 2508–2530, June 2006. [21] K. Avrachenkov, M. El Chamie, and G. Neglia, “A local average consensus algorithm for wireless sensor networks,” in 2011 International Conference on Distributed Computing in Sensor Systems and Workshops (DCOSS), pp. 1–6, June 2011. [22] A. E. Sokhey and S. D. McClurg, “Social networks and correct voting,” The Journal of Politics, vol. 74, pp. 751–764, 7 2012. [23] L. Severini, M. El Chamie, and G. Neglia, “Topology versus link strengsth for information dissemination in networks,” in ALGOTEL 2013 (Pornic, Loire-Atlantique, France, May 28-31), p. 4, May 2013. [24] L. Xiao and S. Boyd, “Fast linear iterations for distributed averaging,” Systems and Control Letters, vol. 53, pp. 65–78, 2004. [25] M. El Chamie, G. Neglia, and K. Avrachenkov, “Distributed weight selection in consensus protocols by schatten norm minimization,” Automatic Control, IEEE Transactions on, vol. 60, pp. 1350–1355, May 2015. [26] B. Ac¸ıkmes¸e and D. Bayard, “Probabilistic swarm guidance for collaborative autonomous agents,” in American Control Conference (ACC), 2014, pp. 477–482, June 2014. [27] D. Aldous and J. A. Fill, “Reversible markov chains and random walks on graphs,” 2002. Unfinished monograph, recompiled 2014. [28] E. Seneta, Non-negative Matrices and Markov Chains. Springer Series in Statistics, Springer, 2006. [29] N. Demir, B. Ac¸ıkmes¸e, and M. Harris, “Convex optimization formulation of density upper bound constraints in markov chain synthesis,” in American Control Conference, June, 2014 2014. [30] S. Boyd, P. Diaconis, and L. Xiao, “Fastest mixing markov chain on a graph,” SIAM Rev., vol. 46, pp. 667–689, Apr. 2004.

Robust Metropolis-Hastings Algorithm for Safe ...

that the M-H algorithm with this proposal matrix, robust. M-H algorithm, also .... is a symmetric positive. (semi-)definite matrix; R>(≥)H implies that Rij >(≥)Hij.

277KB Sizes 0 Downloads 132 Views

Recommend Documents

A Robust Algorithm for Local Obstacle Avoidance
Jun 3, 2010 - Engineering, India. Index Terms— Gaussian ... Library along with Microsoft Visual Studio for programming. The robot is equipped with ...

A Robust Algorithm for Characterizing Anisotropic Local ...
755 College Road East, Princeton, NJ 08540, USA. 2. INRIA Rhône- ...... lung walls and near rib structures. All the 14 ... Academic Press, San Diego, 1990. 5.

A Robust Color Image Quantization Algorithm Based on ...
Clustering Ensemble. Yuchou Chang1, Dah-Jye Lee1, Yi Hong2, James Archibald1, and Dong Liang3. 1Department of Electrical and Computer Engineering, ...

A Robust Color Image Quantization Algorithm Based on ...
2Department of Computer Science, City University of Hong Kong, Kowloon, Hong Kong ...... Ph.D. degrees in computer science from the University of.

Safe Metropolis-Hastings Algorithm and Its Application ...
Oct 28, 2017 - agents in and out of the swarm and disaster inci- dents that cause a group of agents in a region to fail. ...... 1355, May 2015. [25] D. Aldous and J. A. Fill. Reversible markov chains and random walks on graphs, 2002. Un- finished mon

Polynomial algorithm for graphs isomorphism's
Polynomial algorithm for graphs isomorphism's i. Author: Mohamed MIMOUNI. 20 Street kadissia Oujda 60000 Morocco. Email1 : mimouni.mohamed@gmail.

Towards Robust Indexing for Ranked Queries ∗
Department of Computer Science. University of Illinois at Urbana-Champaign. Urbana, IL ... Database system should be able to process the ranked queries.

Robust Mechanisms for Risk-Averse Sellers - CiteSeerX
at least 1/2, which implies that we get at most 2Ç«2 utility for urisk-averse, compared to Ç«(1 − Ç«) at the price Ç«/(1 − Ç«). 2.4 Results and Techniques. We first show ...

Robust Confidence Regions for Incomplete ... - Semantic Scholar
Kyoungwon Seo. January 4, 2016. Abstract .... After implementing the simulation procedure above, one can evaluate the test statistic: Tn (θ) = max. (x,j)∈X×{1,...

RRED: Robust RED Algorithm to Counter Low-Rate Denial-of-Service ...
IN the past decades, quite a few Active Queue Management. (AQM) algorithms such as Random Early Detection (RED). [1] and its variants have been proposed to handle congestion and to improve the TCP performance ([1], [2], [3], [4]). Although these AQM

Architecture patterns for safe design
We have been inspired by computer science studies where design patterns have been introduced to ease software development process by allowing the reuse ...

ROBUST DECISIONS FOR INCOMPLETE MODELS OF STRATEGIC ...
Jun 10, 2011 - Page 1 .... parameters by only imposing best response or other stability ... when the decision maker has a prior for the parameter of interest, but ...

Best-Buddies Similarity for Robust Template ... - People.csail.mit.edu
1 MIT CSAIL. 2 Tel Aviv University ... ponent in a variety of computer vision applications such as ...... dation grant 1556/10, National Science Foundation Robust ... using accelerated proximal gradient approach. ... Online object tracking: A.

ROBUST DECISIONS FOR INCOMPLETE MODELS OF STRATEGIC ...
Jun 10, 2011 - Clearly, this particular problem is fairly basic, and there is no compelling ..... largest value when L(θ, a) is above that cutoff - see figure 1 for a graphical illustration. 4. ..... of each draw relative to the instrumental prior Ï

Robust Nonparametric Confidence Intervals for ...
results are readily available in R and STATA using our companion software ..... mance in finite samples by accounting for the added variability introduced by.

Nonlinear Spectral Transformations for Robust ... - Semantic Scholar
resents the angle between the vectors xo and xk in. N di- mensional space. Phase AutoCorrelation (PAC) coefficients, P[k] , are de- rived from the autocorrelation ...

TRAINABLE FRONTEND FOR ROBUST AND ... - Research
Google, Mountain View, USA. {yxwang,getreuer,thadh,dicklyon,rif}@google.com .... smoother M, PCEN can be easily expressed as matrix operations,.

BINAURAL PROCESSING FOR ROBUST RECOGNITION OF ...
ing techniques mentioned above, this leads to significant im- provements in binaural speech recognition. Index Terms—. Binaural speech, auditory processing, ...

An Evolutionary Algorithm for Homogeneous ...
fitness and the similarity between heterogeneous formed groups that is called .... the second way that is named as heterogeneous, students with different ...

Monotonic iterative algorithm for minimum-entropy autofocus
m,n. |zmn|2 ln |zmn|2 + ln Ez. (3) where the minimum-entropy phase estimate is defined as. ˆφ = arg min .... aircraft with a nose-mounted phased-array antenna.

CMII3 - Compensation Algorithm for Deterministic ...
Novel dispersive devices, such as chirped fiber Bragg gratings (CFBGs), can be used to temporally process broadband optical signals. Unlike optical fiber, these ...

Data Structure and Algorithm for Big Database
recommendation for further exploration and some reading lists with some ... There is a natural tendency for companies to store data of all sorts: financial data, ...

An Adaptive Fusion Algorithm for Spam Detection
An email spam is defined as an unsolicited ... to filter harmful information, for example, false information in email .... with the champion solutions of the cor-.

Focusing on Interactions for Composition for Robust ...
of composite services to model business-to-business ... service to represent a collaboration of web services ... briefly initial ideas of how the interactions for.