Application of a Novel Parallel Particle Swarm Optimization to the Design of Electromagnetic Absorbers Suomin Cui* and Daniel S. Weile Dept. of Electrical & Computer Engineering, University of Delaware, Newark, DE 19711 Email: [email protected], [email protected] 1. Introduction In 1995, Kennedy and Eberhart introduced particle swarm optimization (PSO) [1]. PSO is a new population-based stochastic optimization technique based on the movement of swarms and inspired by social behavior of bird flocking or fish schooling. In past several years, PSO has been successfully applied in many different application areas due to its robustness and simplicity. In comparison with other stochastic optimization techniques like genetic algorithms (GAs) [2] or simulated annealing (SA) [3] PSO has fewer complicated operations and fewer defining parameters, and can be coded in just a few lines. PSO has received increasing attention in EM community in recent years [4-5]. In this work, a novel PSO algorithm is applied to the design of electromagnetic absorbers. Most previous studies on PSO applications focus on novel methods of velocity calculation for each particle in the swarm in an otherwise standard (asynchronous) PSO scheme. Standard PSO schemes are called “asynchronous” here because the global optimum value (used in the velocity updates) is updated after each objective function evaluation. Indeed, because this rule ensures that the best information at hand is always immediately used, it is expected to lead to better results than a synchronous approach. The drawback of the asynchronous approach is that the behavior of each particle depends on all of the particles evaluated previously; thus, the inherent parallelism typical of population-based optimization techniques is destroyed. This disadvantage severely restricts the practical applications of PSO in electromagnetic engineering optimization, because the problems in this area are often expensive to analyze, and require parallel computing to be successfully optimized by a stochastic technique. Therefore, it is of importance to develop an easily implemented parallel PSO. This is described in the next section. The PSO description is followed by applications of the algorithm to absorber problems, and the conclusions of the work. 2. Description of the asynchronous PSO As mentioned above, PSO was inspired by flocks of birds. Suppose that a group of birds is randomly searching food in an area where only one piece of bread exists. Although none of the birds knows the exact location of the bread, if they each have some idea of their proximity to the bread, they can search the area cooperatively by communicating their positions and proximity to one another. The birds may fly in different ways to search in the area. In the asynchronous scheme, the first bird flies a certain distance and in a certain direction based on his own experience and the known location of the best position found so far. When he is done flying, if his location is closer to the bread than the previous globally optimal position, he communicates this information to the second bird. (Otherwise, the global optimal location remains the same.) The second bird then uses this information to update his speed and position, and so on. Since in this method of searching an area the birds all update their speeds and directions at different times, it will be called asynchronous.

A different method is obtained if all birds stop after flying in the same time interval, and communicate each other to get the current global best position. Once this communication is completed, the birds can update their directions and speeds based on their own history and the globally best information available, and fly again. This scheme will be called synchronous, as all of the birds update their information at the same time. Like the analogous birds’ flying schemes, PSO can be classified into asynchronous PSO and synchronous PSO. Standard PSO is asynchronous and is not described here; the interested reader is referred to the references [1,3-6]. In the synchronous PSO, because no relationship exists among all particles in the same iteration, we can update the positions of particles and use parallel computers to evaluate fitness values at the same time. Fig.1 shows the synchronous PSO proposed in this study. The most important factor for success in PSO is the selection of the velocity updating rule for each direction i of each particle j . The rule used here is the standard PSO update rule, given by (1) vij = w * vij + c1*rand(0,1)*( p ij − xij ) + c2 *rand(0,1) * ( g i − xij ) In Eq. (1), rand(x, y ) indicates a random number chosen from a uniform distribution between x and y, c1 and c2 are two acceleration constants (often chosen to be 2 from past experience), p ij is the local best position for each particle, g i is the global best position, v ij and x ij are the velocity and position for each particle, and w is the weight (i.e., inertia) factor. The time-dependent linear weight factor [6] often outperforms a fixed factor and is used in this study. The velocity is subject to a limiting process in which the components of the velocity are restricted to a value less than some predetermined value. The parallel PSO described above has the tendency to gravitate towards the boundaries, namely, x ij = 0 or x ij = 1 , even when these points are suboptimal. To overcome this problem, the component of the vector on the boundary is artificially reset to either a random number or the corresponding component of the current global optimum. While this approach could potentially decrease the efficiency of the search when the optimum is actually located on the boundary, such cases are likely rare in practice. It is worth noting this difficulty has not been reported for asynchronous PSO. Initialize population with random position and velocity vectors Next generation

Evaluate fitness functions for all particles

Search for local best position of each particle & Search for the global position

Update & clip position for each particle

Update velocity for each particle

Reached goal ?

Stop

Fig.1 The flowchart of the PSO

Fig.2 The performance of the PSO and GA

We used several toy problems in the literature to check the performance of the PSO, and found that: (1) PSO incorporating a velocity limit is a slightly better than the PSO without it, and a reasonable limit is around 0.2; (2) PSO is competitive with GAs in finding the optimal solution (i.e., PSO’s accuracy is comparable to that of GAs); and (3) PSO can be more efficient than GAs. To show how these conclusions were reached,

results for the optimization of Griewank function are presented. The Griewank function is given by x 1 10 2 10 cos( i ) + 1 xi ∈ [−10,10] . (2) f = xi − 4000 i =1 i i =1 Fig.2 displays the variation of the mean best objective functions for 20 trials on the generations for the PSO and binary GA with population sizes of 5, 15, and 40. From Fig.1, PSO with population sizes of 5, 15, and 40, and the binary GA with population sizes of 15 and 40 converge to the almost same optimum value at generation 120. (The GA with population size of 5 is much worse than the others.) These results demonstrate that PSO runs small population sizes can get results obtained by GAs only with larger population sizes. In short, for some problems, PSO appears as effective and more efficient than GAs. This result is further confirmed for absorber design as described in the next section.





3. Coating absorber designs In this section, the parallel PSO described above is applied to design a four-layer coating absorber backed by a perfect electric conductor (PEC) as shown in Fig. 2. Each layer of the coating absorber is occupied by a homogenous material whose thickness, permittivity, and permeability are denoted by ti , ε i = ε i′ − jε i′′ , and µi = µi′ − j µi′′ ( i = 1, 2,3, 4 ), respectively, and these 20 design variables make up the design space. The optimization goal is the minimization of the maximum the reflected power over a given frequency band and incident angle range for both polarizations. The design variables are limited as follows: ti ≤ 0.5 cm , ε i′ ∈ [1,10] , ε i′′∈ [0, 10] , µi′ ∈ [1,10] and µi′′∈ [0,10] . PSO, binary GA (BGA), real GA (RGA) and simulated annealing (SA) are used to optimize this problem. (Both GAs use the elitism procedure and niche technique). For PSO and both GAs the population size and the maximum generation number are set to 80 and 300 respectively; for SA, the stop criterion is that maximum number of objective function evaluations is 80 × 300 = 24,000 . The results obtained by these optimizers are shown in Table I. From this table, we can see all these optimizers return almost the same results. Table I: Performance of four different algorithms on a four layer absorber design problem Problem 20 − 25 GHz 0 − 10D 20 − 25 GHz 0 − 30D 20 − 40 GHz 0 − 10D 20 − 40 GHz 0 − 30D

PSO -42.68 dB -22.89 dB -42.38 dB -22.90 dB

RGA -42.42 dB -22.93 dB -42.32 dB -22.88 dB

BGA -42.66 dB -23.09 dB -42.34 dB -22.91 dB

SA -42.33 dB -22.89 dB -42.33 dB -22.89 dB

Fig. 3 shows the average best objective function value profile for 20 trials for the PSO and RGA with different population sizes for the frequency band 20 − 40 GHz and incident angle range 0 − 10D . For the population sizes 5, 10, and 40, shown in Figs. 3(a), (b), and (c), PSO clearly converges faster than the RGA. In these cases, PSO converges by iteration 50 to a result better than that achieved by RGA at generation 150. For a large population size, the convergence rates of both algorithms are almost same, and the results are very close as shown in Fig. 3(d), which had a population size of 80. This example shows that the PSO can get good optimum values by using very small population size, and confirms the high efficiency of the PSO. Indeed, the PSO results for 80 population members are not too different from those for 5 population members from the engineering point of view. More numerical results for polygonal absorbers will be presented in the talk to further demonstrate the advantages of the PSO approach.

4. Conclusion In this study, we presented the synchronous PSO, which is inherently (indeed, embarrassingly) parallel. The parallelism of the synchronous PSO makes practical the design of complex absorbers on clusters of computers. Nonetheless, it is found that parallel PSO must incorporate a special operator not found in standard PSO to keep the population from converging suboptimally to the boundary. This synchronous PSO was applied to optimize multilayer coatings and polygonal absorbers for wide band frequency and/or wide incident range. The apparent advantages of the PSO are summarized below: ● Simplicity: PSO it is extremely simple to code, even relative to GAs. ● Effectiveness: numerical simulations for both toy and practical problems in this study show that the PSO is competitive to other relatively complicated algorithms. ● Efficiency: numerical results presented in this paper show that for at least some classes of problems, PSO may be more efficient than GAs. Further extensions of this work include the development of mixed-parameter parallel PSO algorithms. References 1. J. Kennedy and R.C.Eberhart, “Particle swarm optimization,” Proceedings of IEEE International Conference on Neural Networks, Piscataway, NJ. pp. 1942-1948, 1995. 2. D.E. Goldberg, Genetic Algorithms in Search, Optimization & Machine Learning, Addison Wesley, 1989. 3. S. Kirkpatrick, J. C. D. Gelatt, and M. P. Vechi, “Optimization by simulated annealing,” Science, vol. 220, pp. 671-680, 1983. 4. J.Robinson andY.Rahmat-Samii, “Particle swarm optimization in electromagnetics,” IEEE Trans. Antennas and Propagation, vol. 52, No.2, pp. 397-407, 2004. 5. D.W.Boeringer and D.H.Werner, “Particle swarm optimization versus genetic algorithms for phased array synthesis,” IEEE Trans. Antennas and Propagation, vol. 52, No.3, pp.771-779 , 2004. 6. Y.Shi et al “ Parameter selection in particle swarm optimization,” In Evolutionary Programming VII: Proc. EP98, New York: Springer-Verlag, pp. 591-600,1998.

Fig.3 Comparison of the performance of the PSO and RGA for different population sizes

Application of a Novel Parallel Particle Swarm ...

Dept. of Electrical & Computer Engineering, University of Delaware, Newark, DE 19711. Email: [email protected], [email protected]. 1. Introduction. In 1995 ...

227KB Sizes 0 Downloads 254 Views

Recommend Documents

Application of a Parallel Particle Swarm Optimization ...
Application of a Parallel Particle Swarm Optimization. Scheme to the Design of Electromagnetic Absorbers. Suomin Cui, Senior Member, IEEE, and Daniel S.

Control a Novel Discrete Chaotic System through Particle Swarm ...
Control a Novel Discrete Chaotic System through. Particle Swarm Optimization. *. Fei Gao and Hengqing Tong. Department of mathematics. Wuhan University of ...

A Comparative Study of Differential Evolution, Particle Swarm ...
BiRC - Bioinformatics Research Center. University of Aarhus, Ny .... arPSO was shown to be more robust than the basic PSO on problems with many optima [9].

Srinivasan, Seow, Particle Swarm Inspired Evolutionary Algorithm ...
Tbe fifth and last test function is Schwefel function. given by: d=l. Page 3 of 6. Srinivasan, Seow, Particle Swarm Inspired Evolutionar ... (PS-EA) for Multiobjective ...

A Modified Binary Particle Swarm Optimization ... - IEEE Xplore
Aug 22, 2007 - All particles are initialized as random binary vectors, and the Smallest Position. Value (SPV) rule is used to construct a mapping from binary.

EJOR-A discrete particle swarm optimization method_ Unler and ...
Page 3 of 12. EJOR-A discrete particle swarm optimization method_ Unler and Murat_2010 (1).pdf. EJOR-A discrete particle swarm optimization method_ Unler ...

A Modified Binary Particle Swarm Optimization ...
Traditional Methods. ○ PSO methods. ○ Proposed Methods. ○ Basic Modification: the Smallest Position Value Rule. ○ Further Modification: new set of Update ...

An Improved Particle Swarm Optimization for Prediction Model of ...
An Improved Particle Swarm Optimization for Prediction Model of. Macromolecular Structure. Fuli RONG, Yang YI,Yang HU. Information Science School ...

An Interactive Particle Swarm Optimisation for selecting a product ...
Abstract: A platform-based product development with mixed market-modular strategy ... applied to supply chain, product development and electrical economic ...

Chang, Particle Swarm Optimization and Ant Colony Optimization, A ...
Chang, Particle Swarm Optimization and Ant Colony Optimization, A Gentle Introduction.pdf. Chang, Particle Swarm Optimization and Ant Colony Optimization, ...

Quantum Evolutionary Algorithm Based on Particle Swarm Theory in ...
Md. Kowsar Hossain, Md. Amjad Hossain, M.M.A. Hashem, Md. Mohsin Ali. Dept. of ... Khulna University of Engineering & Technology, ... Proceedings of 13th International Conference on Computer and Information Technology (ICCIT 2010).

particle swarm optimization pdf ebook download
File: Particle swarm optimization pdf. ebook download. Download now. Click here if your download doesn't start automatically. Page 1 of 1. particle swarm ...

Particle Swarm Optimization for Clustering Short-Text ... | Google Sites
Text Mining (Question Answering etc.) ... clustering of short-text corpora problems? Which are .... Data Sets. We select 3 short-text collection to test our approach:.

Quantum Evolutionary Algorithm Based on Particle Swarm Theory in ...
hardware/software systems design [1], determination ... is found by swarms following the best particle. It is ..... “Applying an Analytical Approach to Shop-Floor.

Phishing Website Detection Using Random Forest with Particle Swarm ...
Phishing Website Detection Using Random Forest with Particle Swarm Optimization.pdf. Phishing Website Detection Using Random Forest with Particle Swarm ...

Particle Swarm Optimization: An Efficient Method for Tracing Periodic ...
[email protected] e [email protected] ..... http://www.adaptiveview.com/articles/ipsop1.html, 2003. [10] J. F. Schutte ... email:[email protected].

Particle Swarm Optimization: An Efficient Method for Tracing Periodic ...
trinsic chaotic phenomena and fractal characters [1, 2, 3]. Most local chaos control ..... http://www.adaptiveview.com/articles/ipsop1.html, 2003. [10] J. F. Schutte ...

Entropy based Binary Particle Swarm Optimization and ... - GitHub
We found that non-ear segments have lesser 2-bit entropy values ...... CMU PIE Database: 〈http://www.ri.cmu.edu/research_project_detail.html?project_.

a novel parallel clustering algorithm implementation ... - Varun Jewalikar
calculations. In addition to the 3D hardware, today's GPUs include basic 2D acceleration ... handling 2D graphics from Adobe Flash or low stress 3D graphics.

a novel parallel clustering algorithm implementation ...
parallel computing course which flattened the learning curve for us. We would ...... handling 2D graphics from Adobe Flash or low stress 3D graphics. However ...

A Novel Storage Scheme for Parallel Turbo Decoder
We do this by restricting the whole number of colors seen by a SISO processor when it decode the two component codes. If p χ can be restricted, the total tri-state buffer consumption will decrease. The resultant “reordered first fit” algorithm i

A Novel Parallel Architecture with Fault-Tolerance for ...
paper we provide a novel parallel architecture named Dual-. Assembly-Pipeline(DAP) with fault-tolerance, in which we join bi-directional data streams by considering the processing nodes' failures. Especially, virtual machines in a ... distributed in

Application Specific Parallel Architectures
an evaluation system that can help the designer in the ... A system for ASA development has to be ...... two neighbors and the host, the 2N memory loads and.

a novel parallel clustering algorithm implementation ...
In the process of intelligent grouping of the files and websites, clustering may be used to ..... CUDA uses a recursion-free, function-pointer-free subset of the C language ..... To allow for unlimited dimensions the process of loading and ... GPU, s