VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

Post-processing of lines obtained by raster-to-vector conversion E.Bodansky, A.Gribov, M.Pilouk Environmental Systems Research Institute, Inc – ESRI 380 New York St., Redlands, CA 92373-8100 USA (ebodansky, agribov, mpilouk}@esri.com

Summary: The main goal of this paper is to analyze different methods of postprocessing lines that have resulted from raster-to-vector conversion of black and white line drawing, mostly the borders of connected components of maps. These methods are implemented with generalization, compression, and smoothing algorithms. The paper shows that smoothing algorithms can enhance precision and that using smoothing and compression algorithms in succession gives a more accurate result than using only a compression algorithm. The paper also shows that a map in vector format may require more memory than a map in raster format. The influence of hardware and software on choosing post-processing methods is discussed. The Appendix contains a description of the smoothing method (continuous local weighted averaging) used in the paper. Keywords: raster-to-vector conversion, border, centerline, post-processing, smoothing, compression, accuracy. 1. Introduction The result of raster-to-vector conversion of a black and white line drawing can be represented as borders of connected components and as centerlines. Some definitions for the term centerline can be find in [1]. The simplest vectorization consists in building piecewise linear curves or polylines that are borders of connected components. It has a unique solution, which can be obtained without any control parameters. There are two routine methods of building borders of connected components [2, 3]. The first method results in borders as orthogonal polylines [2], lines that consist only of horizontal and vertical segments. The second method results in a set of digital lines that consist of border pixels. In this case, borders in the form of polylines can be obtained with special approximation algorithms. The description of one of the appropriate algorithms can be found in [3, chapter 6]. Post-processing borders should increase accuracy, defragment straight lines, arcs, and circles, compress data, and generalize. Very often it is impossible to separate these tasks. For example, increasing accuracy while suppressing a high frequency noise gives the effect of data compression. The same effect can be achieved with defragmentation or generalization, and data compression can be used for defragmentation and generalization. It is necessary to explain what we mean by increasing accuracy. Figure 1 shows a piece of a contour map that was converted into a raster format with a resolution of 300 dpi. Any differences between contours on a source map and appropriate contours of the raster image of this map will be called errors of digital description, errors that appear because of scanning, binarization, and discretization.

1

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

For error correction we can use some a priori information about depicted objects. It is known that isolines (in our case contours) cannot intersect. Hence any intersections on Figure 1 are errors. These are most likely caused by scanning and binarization. Only an operator can correct these errors. To ascertain this it is enough to see Figure 2a. Should it be two contours (Figure 2b) or only one (Figure 2c)? Another type of error can be corrected automatically. It is known that contours are smooth. The main reason of failure to carry out this condition is a discretization error. This error can be corrected with smoothing algorithm after choosing a value of a control parameter. In this paper we will analyze only the second type of error. The importance of the pre-processing goals listed above change as times goes by. Not long ago it was very important to compress data. We can read in the literature that this was one of the main goals of raster-to-vector conversion [3] and [4, chapter 8]. But now the situation is changed. Progress in computer technology has resulted in huge increases of available computer memory and processing speed. That is why data compression is not as critical as before. Progress in computer technology has also given one more very important result: we can now use very sophisticated algorithms for smoothing and generalization that were considered unrealizable not long ago.

2. Post-processing algorithms All algorithms that can be used for post-processing can be fall into three groups: a) Deletion of some specific features of curves (for example, arms and narrow peninsulas from a coast line). b) Deletion of some vertices of polylines. c) Approximation of lines with some mathematical functions. For the rigorous solution of generalization tasks, the best algorithms are those of the first group. Because these algorithms are essentially recognizing algorithms, they are very complex. In [5], the development of such algorithms is characterized as a “fantastically complex problem.” Besides they have a very large computational complexity. Generalization of different objects may require different algorithms. It is unlikely that the algorithm assigned for cost line generalization can be used to generalize roads, contours or building silhouettes. But there is some progress in developing such algorithms [6]. At the beginning of the sixties, difficulties of development and implementing such algorithms seemed so insuperable that the second group algorithms were used not only for compression but also for generalization. Obviously algorithms of the second group have to be used for data compression and straight line defragmentation. But till now they were also used for generalization (in cartography) and smoothing [7]. Many algorithms of this group [5, 8-11] were developed. These algorithms result in a new polyline with vertices (critical points) that are a subset of vertices of the source polyline. Critical points are the points that determine the shape of the processed polyline. The distances of other vertices (noncritical points) from the resulting polyline have to be less than some threshold h, chosen by an operator. The algorithms of this group differ from each other in inessentials. The Douglas-Peucker algorithm does not depend on the direction the polyline is processed [5]. The algorithms described in [8-11] require in some cases less processing time than the previous one. In

2

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

[11], a new version of the Douglas-Peucker algorithm is suggested. A worst-case running time of this version is proportional to n (log n) whereas for the other algorithms it is n*n. The author of [12] wrote that the Douglas-Peucker method is “mathematically superior” and the author of [13] that “the generalizations produced by the Douglas algorithm were overwhelmingly – by 86 percent of all sample subjects – deemed the best perceptual representations of the original lines.” But it is necessary to note that all these algorithms, including the Douglas-Peucker algorithm, have a common shortcoming: errors of vertices of resulting polylines are equal to errors of critical points of the source polylines, and therefore they cannot be reduced with decreasing the control parameter h. Because of this, resulting polylines are uneven, orientation of segments of resulting polylines depends on errors of the critical points’ coordinates, and these algorithms cannot be used to filter random noise The lower the resolution of the image the more noticeable the errors are. These shortcomings cause undesirable consequences when these algorithms are used for map generalization, defragmentation of straight lines, and noise suppression. Figures 3a and 3b show a connected component and the border after processing with the DouglasPeucker algorithm, with values of the threshold h = 1 and h = 0.75, respectively. We have already said that contours have to be smooth. The source image (Figure 1) contains a large amount of binarization noise. Comparing borders before and after compression, it is easy to conclude that it is easier to smooth source borders manually than to smooth compressed borders manually. The same conclusion may be drawn for straight lines. To define the correct direction of straight segments after compression is more difficult than to define direction of these segments before compression. So we can conclude that compression algorithms result in irreversibly losing information about a noise. 3. Smoothing and accuracy The most appropriate algorithms for suppressing a discretization noise (at least for contours) are those of the third group, which use mathematical functions to approximate the original polylines. They include algorithms that define straight lines and polynomials with the least square method and algorithms of local smoothing, which are especially important for cartography. An approximation can be done for a whole polyline (global approximation), or, sequentially, for small pieces of polyline (the local approximation). In [5], several such methods are described and their shortcomings are listed: a. The resulting curves are smoother than real objects. b. These methods look for an average behavior of polylines and consequently corrupt the critical points that define a polyline shape. c. Processed polylines contain more points than source polylines. We have objections to all these conclusions: a. Methods of this group use a control parameter, which defines a power of smoothing. An operator can choose a value for this parameter that gives the required power of smoothing. Figures 4a and 4b show polylines obtained with the same algorithm of smoothing but with different values of the smoothing parameter (d = 3 and d = 5). Developed by A.Gribov, this algorithm is described in the Appendix.

3

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

b. There are different methods for looking for critical points [13, 14], based on evaluation and analysis of a curvature and the first derivation of a curvature. If critical points are found, it is possible to split a curve at critical points. Each of the new curves can be smoothed with fixed end points. c. First let us try to understand why this algorithm can produce too many points. These algorithms can calculate smoothed curve points with any given step. Usually this step is constant, i.e. distances between any two consecutive points are almost equal. The appropriate number of points required for a good approximation of the curve with a polyline depends on the nature of the processing curve and is defined by the operator. Extra points can appear if the chosen step was too small or if curves contain long segments with a very small curvature. It is possible to delete extra points by increasing a step or by using compression algorithms. We explain in the next section why using smoothing and compression algorithms consecutively gives a better result than simply using the same compression algorithm. 4. Using different post-processing algorithms consecutively

Consider one essential difference between the algorithms of the second (compression) and the third (approximation) groups. We have already noticed that all vertices of the processed polyline are a subset of the vertices of the source polyline. That means that along with the vertices we inherit their errors so that these algorithms are unsuitable for noise suppression. Methods of the third group produce polylines with vertices that do not coincide with vertices of the source polyline; they can, therefore, be used for accuracy enhancement. Assume that a smoothing algorithm produced points located on the original smooth curve without any errors. Let us plot a polyline using these points as vertices. If the vertices were calculated with a small step, the polyline would be a good approximation of the smooth curve. If we process this polyline with a compression algorithm, the vertices of the compressed polyline would be free of error. The maximum distance between the source and resulting polylines would be less than the threshold h, i.e. we could build the approximation with any given accuracy. Of course, smoothing algorithms produce the result with some error. But if the control parameter of smoothing is selected correctly, errors of the smoothed polylines will be much less than errors of the original polyline. Processing this smoothed polyline using a compression algorithm with small enough threshold h will not increase the error enough to matter. Figures 5a and 5b show borders of connected components after smoothing with the parameter of smoothing d = 5, followed by processing with the Douglas-Peucker algorithm with the threshold h equals 0.25 and 0. 5, respectively. Comparing the polylines shown in Figures 5 and 3 visually, it is easy to arrive at the conclusion that accuracy increases if you use smoothing before compression. The polylines shown on Figure 5b have only 5% more vertices than the polylines shown on Figure 3b.

4

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

Figures 6a, 6b, and 6c show the result of post-processing a circle. A fragment of the circle and its border obtained after discretization are shown in Figure 6a. The same fragment and its borders obtained after compression and after smoothing with compression are shown in Figures 6b and 6c, respectively. Table 1 contains mean root square errors of the source and the processed borders of the circle, the parameters of processing, and the number of vertices. The smoothed and compressed border has the lowest root mean square error and the lowest number of vertices. This is an empirical corroboration of the fact that our predictions were correct. 5. Different formats and requirement for memory Fairly often you can read that vector representation of maps, engineering drawings, and other line drawings is much more economical than raster representation, and that this is one of the main incentives for raster-to-vector conversion [3, 4]. Maybe this is right for drawings consisting mainly of straight lines and circles, such as engineering drawings and electrical schematics, but it is not true for maps and other documents consisting mainly of irregular curved lines. To show this, we evaluated the memory required for storing the piece of a contour map shown in Figure 1 in both raster and vector formats. The raster image of this fragment consists of 293 rows and 879 columns of 257547 pixels. Because raster-to-vector conversion systems usually process black and white images, only 1 bit is required for each pixel. So for storing the image in uncompressed raster format, approximately 32 Kbytes will be required. If the image will be converted into the run-length encoding (RLE) format, the required memory will decrease to 24 Kbytes. The exact borders of all connected components shown on Figure 1 include 14169 points. To store information about each point will require 16 bytes (2 coordinates in the double format). So for the exact borders in the vector format, more than 220 Kbytes will be required, or almost 7 times more than for the raster format. A vector database saves not only coordinates but also other information about the geometry and topology of image features: area of polygons, length of perimeters of the polygons, number of holes, and so on. After smoothing (for the algorithm of smoothing see the Appendix) with the power of smoothing d = 5 and d = 3, the number of points increased to 26921 and 48174, respectively. If, after smoothing with d = 5, the resulting image is processed with the DouglasPeucker algorithm with parameter h = 0.25, then the number of points will reduce to 5872, but all the same a vector representation of the map fragment will require much more memory (about 92 Kbytes) than the raster representation. If the image consists of linear elements (as Figure 1), it can be represented in a vector format as a set of centerlines with their width. That will require almost two times less memory in comparison with borders. But it will still need more than the raster image. The analyzed example allows one to conclude that memory compression cannot be the main reason for raster-to-vector conversion of line drawings (at least for maps). Obviously this transformation is performed for other, more important goals. That is why the importance of this problem was not reduced by the enhancement of computers and the tremendous increase in available computer memory.

5

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

6. Conclusion The raw results of raster-to-vector conversion of linear drawings often require post-processing, usually by generalization, compression, and smoothing. Smoothing can be used to increase accuracy by depressing discretization noise. Increasing accuracy is especially important for processing images with a small resolution. Smoothing can be implemented with the algorithm of local averaging. The continuous weighted local averaging algorithm used in this paper is given in the Appendix. This algorithm has some advantages; particularly that it does not require preliminary densification and that it attains a worst-case running time proportional to O(n), where n is a number of vertices of a source polyline. Compression algorithms do not increase the accuracy of the vectorization, but they can be used after smoothing for reducing the required memory with little diminution of accuracy. A vector format of map representation may require more memory than a raster one. But raster-to-vector conversion is performed not for saving memory but for other more important goals. That is why the significance of the vectorization was not reduced by enhancement of computers and the tremendous increase of available computer memory. Appendix. Smoothing with continuous local averaging. We analyze curves that are approximated with piecewise linear polylines. The smoothing algorithm is based on an approximation of a source curve with a mathematical function. A local averaging algorithm calculates points of a smoothed curve, taking into account only few neighboring points of the source curve. In [5], the three such algorithms were mentioned. The straight segments of a smoothed curve are calculated with the least-squares method, using a constant number n of neighboring vertices of the source polyline [14]. Then the procedure repeats for a set of points shifted forward one vertex and so on. Two other algorithms calculate the coordinates of linear segment ends as an average value of coordinates of a fixed number of neighboring vertices of the source polyline. In [15], each vertex of a resulting curve is an average of the neighboring vertices, and in [16] it is a weighted average. If lengths of segments of a source polyline differ from each other a lot, it is necessary to densify them before smoothing. In [17] and [18], coordinates of points of the smoothed polyline are calculated as the convolution of the coordinates of points of the source polyline and the Gaussian. Our smoothing method differs from them by the convolution kernel. Instead of −

t −τ

the Gaussian we used function e d , d > 0 , where ( τ ) is the length of a path along source polyline. A local approximation is performed with the second order polynomial uτ (t ) = α 0u (τ ) + α1u (τ ) ⋅ (t − τ ) + α 2u (τ ) ⋅ (t − τ ) , 2

where u changes by turn with x and y. Values of coordinates x(τ ) and y (τ ) are calculated as 6

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

x(τ ) = α 0x (τ ) ,

y (τ ) = α 0y (τ ) ,

where α 0u (τ ) is the solution of the next equation 2

∑ c (τ )α (τ ) = b (τ ), j = 0, N , where u i

i+ j

i =0

L

ck (τ ) = ∫ e



u j

t −τ d

(t − τ )k dt , and

t −τ d

(t − τ )k u(t )dt

0

L

b (τ ) = ∫ e u k



0

A calculated curve has smooth first derivations at calculated points. To simplify the expressions, we assign d = 1 , and the influence of this coefficient on the power of smoothing takes into account the value τ . L

I i (τ ) = ∫ e

− t −τ

(t − τ )i u (t )dt =

0

τ

L

= ∫ et −τ (t − τ ) u (t )dt + ∫ eτ −t (t − τ ) u (t )dt = i

i

τ

0

= I iL (τ ) + I iR (τ ) I iL (τ 0 ) = I iL (0) = 0

I iR (τ n ) = I iR (L ) = 0 L

I iL (τ k +1 ) =

∫e

(t − τ k +1 )i u (t )dt =

t −τ k +1

τ k +1 τk

= ∫ et −τ k +1 (t − τ k +1 ) u (t )dt + i

0

τk

= ∫e

t −τ k +τ k −τ k +1

τk

∫e

τ k −τ k +1

τ k −τ k +1

τ k −τ k +1

=e

t −τ k +1

(t − τ k +1 )i u (t )dt =

τk

τ k +1

(t − τ k + τ k − τ k +1 ) u (t )dt + ∫ et −τ (t − τ k +1 )i u(t )dt = i

(t −τ k )

k +1

∑ {C (t − τ ) (τ i

i

k

− τ k +1 )

i− j

j

j

j =0

0

=e

∫e

τk

0

=e

τ k +1

k

τ k +1

}u (t )dt + ∫ e

t −τ k +1

(t − τ k +1 )i u (t )dt =

τk

 j  τ k +1 t −τ i− j j i t −τ k C i (τ k − τ k +1 ) ∫ e (t − τ k ) u (t )dt  + ∫ e k +1 (t − τ k +1 ) u (t )dt = ∑  τ k j =0  0  τk

i

∑ {C (τ i

j =0

i

− τ k +1 )

i− j

j

k

τ k +1

} ∫e

I (τ k ) + L j

t −τ k +1

(t − τ k +1 )i u (t )dt

τk

7

……………….(1)

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

I iR (τ k −1 ) =

L

∫e

τ k −1 −t

(t − τ k −1 )i u (t )dt =

τ k −1 τk

L

τ −t τ −t ∫ e k−1 (t − τ k −1 ) u (t )dt + ∫ e k−1 (t − τ k −1 ) u (t )dt =

=

i

τ k −1

i

τk

τk

L

τ −t τ −τ +τ − t ∫ e k−1 (t − τ k −1 ) u(t )dt + ∫ e k −1 k k (t − τ k + τ k − τ k −1 ) u(t )dt =

=

i

τ k −1

i

τk

τk

L

i

τ k −1

=

i

{

τ −t τ −τ τ −t j ∫ e k−1 (t − τ k −1 ) u(t )dt + e k−1 k ∫ e k ∑ Ci (t − τ k ) (τ k − τ k −1 )

=

j =0

τk

i− j

j

}u(t )dt =

L i    j i i− j j τ k −1 − t τ k −1 −τ k τ k −t ( ) ( ) ( ) ( ) ( ) e t − τ u t dt + e C τ − τ e t − τ u t dt  = ∑ k −1 i k k −1 k ∫ ∫  j =0  τ k −1 τ k  τk

τk

=

i

{

}

τ −t τ −τ j R ∫ e k−1 (t − τ k −1 ) u(t )dt + e k−1 k ∑ Ci (τ k − τ k −1 ) I j (τ k ) ………………(2) i

j =0

τ k −1

i− j

τ k +1

If

at

first

you

calculate

the

integrals

∫e

t −τ k +1

(t − τ k +1 )i u (t )dt

and

τk

τk

∫e

τ k −1 −t

(t − τ k −1 )i u(t )dt ,

the integrals I iL (τ k ) ? I iR (τ k ) , where τ k is a parameter

τ k −1

corresponding to the vertex of the source polyline, can be calculated with the recursive equations (1) and (2). This algorithm attains a worst-case running time proportional to O(n ) . This algorithm gives a good result for smoothing centerlines, too. Reference. 1. E.Bodansky, M.Pilouk. “Using Local Deviations of Vectorization to Enhance the Performance of Raster-to-Vector Conversion Systems.” International Journal on Document Analysis and Recognition, Vol. 3, #2, December 2000, pp. 67-72. 2. F.Quek. “An algorithm for the rapid computation of boundaries of run-length encoded regions.” Pattern Recognition, 33, 2000, pp. 1637-1649. 3. D.Lisitzkii. The main principle of map-making regions. (Osnovniie prinzipi cartografirovaniia mestnosti). Nedra, 1988 (in Russian). 4. S.Ablameyko, T.Pridmore. Machine Interpretation of Line Drawing Images. (Technical Drawings, Maps, and Diagrams.) Springer, 2000.

8

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

5. D.Douglas, Th. Peucker. “Algorithm for the reduction of the number of points required to represent a digitized line or its caricature.” The Canadian Cartographer, Vol.10, #2, December 1973, pp. 112-122. 6. Z.Wang, J-C.Muller. “Line Generalization Based on Analysis of Shape Characteristics.” Cartography and Geographic Information Systems, Vol.25, #1, 1998, pp. 3-15. 7. D.Dori, W.Liu. “Sparse pixel vectorization: An Algorithm and its performance evaluation.” IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol.21, #3, 1999, pp. 202-215. 8. U.Ramer. “Extraction of Line Structures from Photographs of Curved Objects.” Computer Graphics and Image Processing. Vol.4, 1975, pp. 81 – 103. 9. J.Sklansky, V. Gonzalez. “Fast Polygonal Approximation of Digitized Curves.” Pattern Recognition, Vol.12, 1980, pp. 327-331. 10.Y.Curozumi, W.Davis. “Polygonal approximation by minimax method.” Computer Graphics and Image Processing, Vol.19, 1982, pp.248-264. 11.J.Snoeyink, J. Hershberger.”Speeding up the Douglas-Peucker Line-Simplification Algorithm.” http://www.cs.ubc.ca/spider/snoeyink/papers/papers.html 12. R.B.McMaster Automated line generalization. Cartographica, Vol. 24, #2, 1987, pp. 74-111. 13. E.R.White. ”Assessment of line-generalization algorithms using characteristic points.” American Cartographer, Vol. 12, #1, 1985, pp. 17-27. 14. A.Boyle. “Computer Aided Compilation.” Hydrographic Conference. Ottawa, January, 1970. 15. C.Coeman, F.Vander Weiden. “The Application of Computation and Automatic Drawing Instruments to Structural Generalization”, Cartographic Journal, Vol.7, #1, 1970, pp.47-49. 16. J.Juncaitis, J.Junkins. “Mathematical Techniques for Cartography”, Final Contract Report for U.S.Army Engineers Topographic Laboratory, Fort Belvoir, Virginia, Contract No. DAAK02-72-C-0256, February 1973, pp.15-20. 17. H.Asada, M.Brady. “The curvature primal sketch.” IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol.8, #1, 1986, pp. 2-14.

9

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

18. F.Mokhtarian, A.Mackworth. “Scale-Based Description and Recognition of Planar Curves and Two-dimensional Shapes.” IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol.8, #1, 1986, pp. 34-43.

Figure 1. A piece of a contour map scanned with a resolution 300 dpi.

a)

b)

c)

Figure 2. A fragment of Fig.1. a) An error b) Two contours c) One contour

Figure 3. A connected component and the compressed border. a) h = 1.0 b) h = 0.75

a)

b)

10

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

Figure 4. A connected component and the smoothed border. a) d = 3 b) d = 5

a)

b)

Figure 5. A connected component and the border smoothed and compressed. a) d = 5, h = 0.25 b) d = 5, h = 0.5

a)

b)

a)

11

VISION (Machine Vision Association of SME), VOL. 18, NO. 1, http://www.sme.org/mva, First Quarter 2002

b)

c) Figure 6. A fragment of a circle and its border obtained with discretization (a), compression (b), and smoothing with compression (c). Table 1. Result of discretization and post-processing a circle Case

Parameter of smoothing (d) Source border (fig 6a) ------------Compressed border (fig. 6b) ------------Smoothed and compressed 20.0 border (fig. 6c)

12

Compression threshold (h)

Root mean Number square error of verices

-------------0.9 0.9

0.01321 0.01365 0.01227

316 136 31

Post-processing of lines obtained by raster-to-vector ...

VISION (Machine Vision Association of SME), VOL. 18, NO. .... algorithm, the vertices of the compressed polyline would be free of error. The maximum distance ...

171KB Sizes 0 Downloads 103 Views

Recommend Documents

Detection of permeabilisation obtained by micropulses ... - IEEE Xplore
electric pulses is compared. ... Data show that very low repetition frequencies can be much ... treatment: the combination of electric permeabilisation and.

2016 Confidentiality of Personally Identifiable Information Obtained ...
2016 Confidentiality of Personally Identifiable Information Obtained Through Child Find Activities.pdf. 2016 Confidentiality of Personally Identifiable Information ...

Recognition of qualification obtained through Distance Education.PDF ...
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Recognition of ...

Recognition of qualification obtained through Distance Education.PDF
Recognition of qualification obtained through Distance Education.PDF. Recognition of qualification obtained through Distance Education.PDF. Open. Extract.

Art. 12 Particularities of hollow-core briquettes obtained out of ...
Page 1 of 18. Maderas-Cienc Tecnol 20(1):2018. Ahead of Print: Accepted Authors Version. 1. 1 DOI:10.4067/S0718-221X2018005000012. 2 PARTICULARITIES OF HOLLOW-CORE BRIQUETTES OBTAINED OUT OF. 3 SPRUCE AND OAK WOODEN WASTE. 4. Cosmin Spirchez1. , Aure

Lines by SRAP Markers
Taq DNA polymerase (Fermentas). Amplifications were carried out in a MJ Research PTC-100 thermocycler programmed for 32 cycles of 1 min at 940C, 1 min at 470C, 1 min at 720C, and ending with 5 min at 720C. PCR products were separated by electrophores

Efficient Decoding of Permutation Codes Obtained from ...
Index Terms—Permutation codes, Distance preserving maps ... have efficient decoding, are known to achieve this upper bound. (see [1], [2]). ... denote this mapping. The example below illustrates the map of b = (1, 1, 0, 1) to the permutation vector

pdf-12105\the-nursery-rhymes-of-england-obtained-principally-from ...
... the apps below to open or edit this item. pdf-12105\the-nursery-rhymes-of-england-obtained-principally-from-oral-tradition-1-by-j-o-halliwell-phillipps.pdf.

pdf-1312\the-nursery-rhymes-of-england-obtained-principally-from ...
... the apps below to open or edit this item. pdf-1312\the-nursery-rhymes-of-england-obtained-princi ... on-with-additions-by-james-orchard-halliwell-after.pdf.

EVALUATION OF CANDIDATE LINES AGAINST WHEAT RUSTS.pdf ...
These lines will remain in NUWYT 2005-06. for further ... even long distances ( Singh et al., 2005). ... results of these trials, the candidate wheat .... NRL-2017 AMSEL/TUI CM107503-12Y-020Y-010M-3Y- 010M-1Y-0M-0AP NIFA, Peshawar.

TOTAL NUMBER OF DEGREES OBTAINED IN ...
12th-No Diploma HS Grad Some College Associates Bachelor's Master's Professional Doctoral Assoc+. 15 to 19 ...... PERCENT BACHELOR'S DEGREE OR HIGHER FOR POPULATION OVER 25. 2005. 2006 .... The difference in educational attainment between Arizona and

Efficient Decoding of Permutation Codes Obtained from ...
N. Thus it is of interest to consider other means of obtaining permutation codes, for .... the transmitted symbol corresponding to bi = 0 is different from the received ...

Concurrency of Four Euler Lines
Apr 9, 2001 - F−. D+. D−. O. Figure 2. The Parry reflection point. This is the reflection of the circumcenter in the focus of the Kiepert parabola.

Reconstruction of Orthogonal Polygonal Lines
algorithm has a low computational complexity and can be used for restoration of orthogonal polygonal lines with many vertices. It was developed for a raster- to-vector conversion system ArcScan for ArcGIS and can be used for interactive vectorization

Points and Lines Axis-parallel Lines Hyperplanes ... - Semantic Scholar
segments in the tour (or the path) are axis-parallel. Four Optimization Problems. • Minimum-Link Covering Tour. • Minimum-Link Rectilinear Covering Tour.

Comparing plumage colour measurements obtained ...
2) on a dark velvet surface (reflectance. 0%), trying to imitate the plumage surface of the bird. (Bennett et al. 1997). Afterwards we measured the samples with the ...

parental lines of hybrid rice
Climatic factors play a major role in altering the floral biology .... 017-0LSL yy„OI X SZ'y-£IO'0-„-0 I X Lşşſ-00Z 0ſuſuɔAaog \sug og SK2GI (e ț7ț7068 SV 9.

Architecture of Product Lines
23 Sep 2009 - Families. “We consider a set of programs to constitute a family whenever it is worthwhile to study programs from the set by first studying the common properties of the ... We call a family of products designed to take advantage .....

Test Pair of lines
2) Find the joint equation of the following pair of lines: 3) passing through (2, 3) and parallel ... 17)AOAB is formed by the lines x2 – 4xy + y2. = 0 and the line AB.

eCounty Lines -
LEADING OFF: Federal Spending Bill Fully Funds PILT, Reauthorizes SRS. There is good news for Colorado counties in the $1.3 trillion federal spending bill Congress passed last week. In addition to funding the federal government through September 30,

pdf-1835\behind-enemy-lines-by-gwynne-bounds.pdf
Sign in. Loading… Whoops! There was a problem loading more pages. Retrying... Whoops! There was a problem previewing this document. Retrying.