OPTIMISING FIGURE OF MERIT FOR PHONETIC SPOKEN TERM DETECTION Roy Wallace, Robbie Vogt, Brendan Baker, Sridha Sridharan Speech and Audio Research Laboratory, Queensland University of Technology (QUT), Brisbane, Australia {rg.wallace,r.vogt,bj.baker,s.sridharan}@qut.edu.au

ABSTRACT This paper introduces a novel technique to directly optimise the Figure of Merit (FOM) for phonetic spoken term detection. The FOM is a popular measure of STD accuracy, making it an ideal candidate for use as an objective function. A simple linear model is introduced to transform the phone log-posterior probabilities output by a phone classifier to produce enhanced log-posterior features that are more suitable for the STD task. Direct optimisation of the FOM is then performed by training the parameters of this model using a nonlinear gradient descent algorithm. Substantial FOM improvements of 11% relative are achieved on held-out evaluation data, demonstrating the generalisability of the approach. Index Terms— spoken term detection, speech processing, speech recognition, information retrieval 1. INTRODUCTION The emergence of vast collections of recorded speech is driving an urgent need for technologies to enable access to the information in these collections. Consequently, there has been recent interest in Spoken Term Detection (STD), the task of detecting occurrences of search terms rapidly and accurately in audio archives [1]. STD generally consists of the two distinct phases of indexing and searching. The role of indexing is to process the raw audio data into a form that is suitable for rapidly searching for query terms. Indexing is performed once in an off-line process, while many searches are then performed on this index. The performance of an STD system is characterised by both search accuracy and search speed. Accuracy, particularly, relates to the usefulness of the results produced by a search. The Figure of Merit (FOM) is a well-established evaluation metric of STD accuracy [2] based on the expected rate of detected search term occurrences over the low false alarm rate operating region. This paper introduces a novel technique for improving the accuracy of a phonetic-based STD system by directly maximising the FOM. For the system presented in this study, a phonetic posteriorfeature matrix is generated during indexing and searched with a fast Viterbi decoding pass. The Figure of Merit is directly optimised through its use as an objective function to train a transformation of the posterior-feature matrix. This direct optimisation of FOM is a form of discriminative training, as the FOM can be formulated in terms of the separation of scores attributed to true search term occurrences and false alarms. Discriminative methods have been previously proposed in the context of STD, however, often these approaches do not seek to directly maximise the STD metric. In [3], for example, a Minimum Classification Error (MCE) criterion is used to improve the word error rate (WER) of the initial word transcript, with no assurance that optimising MCE will lead to optimal STD accuracy.

Maximum FOM training has been applied in other detection problems such as language [4] and topic [5] identification. Few studies have, however, aimed at the direct maximisation of FOM for STD [6, 7]. The technique used in [7] trains importance weights for a small number of feature functions without much justification of their selection or quantification of their contributions to overall performance. The task in [7] is also best described as utterance retrieval, that is, the detection of utterances containing the term, as opposed to the detection of individual term occurrences as in STD. In [8], the concept of enhanced phone posteriors is introduced where a posterior-feature matrix is used for phone and word recognition, but trained to optimise per-frame phone classification. In the work presented in this paper, the phone posteriors are once again adjusted but with FOM as the objective function (rather than phone error rate) in both training and evaluation. In Section 2, the Figure of Merit is defined, followed by details of the baseline STD system in section 3. Section 4 describes the technique used to optimise the Figure of Merit, followed by experimental results in Section 5 and conclusions in Section 6. 2. FIGURE OF MERIT In general, STD accuracy is measured in terms of both detected term occurrences (hits) and false alarms. The Figure of Merit (FOM) metric measures the percentage of correctly detected term occurrences averaged over all operating points between 0 and 10 false alarms per term per hour [2]. This work uses the term-weighted FOM, obtained by averaging the detection rate at each operating point across a set of evaluation search terms, to avoid introducing a bias towards frequently occurring terms [1]. The FOM can be formally defined in terms of a set of STD results. Given a set of query terms, q ∈ Q, search is first performed on T hours of data, resulting in a set of events, e ∈ E, where e is either a hit or false alarm. Each event e has the attributes qe , le , se where qe is the query term to which the event refers, the label le = 1 if the event is a hit or 0 for a false alarm and se is the score of the event. Also, for each term, q ∈ Q, define the set of hits as E + q = {e ∈ E : le = 1 ∧ qe = q} and the set of false alarms as E − q = {e ∈ E : le = 0 ∧ qe = q}. The FOM can then be defined as FOM =

  X 1 X hek max 0, A − (1 − H (ek , ej )) A + − ek ∈E

ej ∈E

(1) where A = 10T |Q|, he =

1 |Q||E + qe |

and

( 1 sek > sej H (ek , ej ) = 0 otherwise.

In this formulation, each hit, ek , contributes a value of between 0 and hek , depending on the number of false alarms which outscore it. The value is 0 when the event is outscored by A false alarms or more. Such events have no effect on FOM, so it is possible to re-define − − the FOM in terms of truncated results sets, − R ⊂ E ,+containing the top scoring false alarms with R ≈ A, and R ⊂ E + ,  containing the top scoring hits which outscore all e ∈ E − − R− . By summing over these subsets, (1) can thus be re-written as FOM =

X 1 X he k H (ek , ej ) . A + − ek ∈R

(2)

ej ∈R

Equation (2) can be interpreted as the weighted proportion of correctly ranked pairs of hits and false alarms, similar to the definition of the AUC (area under the ROC curve) in [7]1 . 3. PHONE POSTERIOR-FEATURE MATRIX STD SYSTEM The indexing and search approach adopted for the STD system is based on that described in [9]. The indexing phase produces a posterior-feature matrix, as described below in Section 3.1. Search is then performed in this matrix, as described in Section 3.2. The use of separate indexing and searching phases allows search to be performed at 50 times faster than real-time, and the system is also suitable for use as a post-processor for candidates selected by other (faster) STD techniques, if necessary [9].

of phones constituting the search term’s pronunciation(s). The background model is an open loop of all phones. Classical Viterbi decoding is then carried out, with the emission probability of a phone p in the network at frame t being directly provided by the corresponding log-posterior probability, xt,p . A phone insertion penalty, tuned on development data, is used to counteract a tendency to favour a short phone duration. At each frame, a score is computed for each search term. This is calculated as the log likelihood ratio between the score of the token exiting the term model and the score of the token in the background model. For an event beginning at frame b and having a length of n frames, this score is s=

b+n−1 X t=b

(pTt xt − g Tt xt ) =

b+n−1 X

(pt − g t )T xt

(3)

t=b

where pt is a mask vector representing the state occupancy at time t of the hypothesised Viterbi alignment for the search term. That is, the elements of pt are ( 1 i is the index of the current phone pt,i = 0 otherwise. Similarly, g t represents the hypothesised alignment for the background model at time t. Whenever the score for a term at a particular time is greater than a threshold (and greater than potential overlapping candidates), the event is output as a putative term occurrence.

3.1. Indexing Indexing involves the generation of a posterior-feature matrix, as follows. As in [9], the LCRC FeatureNet phoneme classifier [10] is first used to produce phone posterior probabilities for each frame of audio. These output posterior probabilities are then logarithmised2 . The resulting log-posteriors form the contents of the posterior-feature matrix, X = [x1 , x2 , . . . , xU ], with xt = [xt,1 , xt,2 , . . . , xt,N ]T , and xt,p referring to the logposterior of phone p at frame t, in an utterance of U frames. These log-posteriors provide the emission probabilities for each state in the decoding network used in search (see Section 3.2). In contrast to [9], phones are modelled with a single state only, to reduce index size and the number of parameters to be trained. To give an indication of the quality of the phone classifier, using the log-posteriors directly for open-loop phone recognition gives 46% accuracy on the evaluation set described in Section 5.1. 3.2. Search Once the index has been constructed, the system can accept search terms in the form of a word or phrase. A pronunciation lexicon is used to convert the search term into a sequence of phones. Letter-tosound rules may be used for out-of-vocabulary terms. Search then consists of a modified Viterbi decoding on the posterior-feature matrix. As in [9], a network is constructed from context-independent phones, with two connected parts; term model and background model. The term model consists of the sequence(s) 1 In this work, however, we aim to detect individual term occurrences, rather than classify whole utterances. The AUC characterises average performance over all utterance retrieval operating points, whereas the FOM refers to average performance over STD operating points between 0 and 10 false alarms per term per hour. 2 The alternative transformations suggested in [9] did not improve STD accuracy in empirical trials.

4. OPTIMISING THE FIGURE OF MERIT The contribution of this paper is a novel method for direct optimisation of FOM for an STD system. This is achieved by introducing an extra layer of modelling that provides a mechanism for transformation of the posterior-feature matrix. In the remainder of this section, this mechanism is first described, followed by a description of the optimisation algorithm used to maximise FOM on the training set. 4.1. Enhanced posterior-feature linear model A simple linear model is introduced to transform the phone logposterior probabilities output by the phone classifier, X, to produce enhanced log-posterior features, X 0 , that are more suitable for the STD task. The linear transform is decomposed into a decorrelating transform, V , and an enhancement transform W , giving x0t = W V xt .

(4)

The decorrelating transform, V , is obtained through principal component analysis (PCA) of the log-posterior features. Performing PCA additionally provides the opportunity for dimensionality reduction by retaining only the M ≤ N directions of highest variability. Dimensionality reduction reduces susceptibility to over-fitting of the proposed model by reducing the number of free parameters to train in W , and can also have the benefit of suppressing low-energy directions that may be dominated by noise. The weighting matrix, W , is an N ×M transform that produces a set of enhanced posterior-features from the decorrelated features. The goal of the novel training algorithm is to optimise the weights in W that maximise FOM directly. While the original posteriorfeatures, X, were optimised for phone classification, it is hypothesised that a discriminative algorithm optimising FOM directly will place additional emphasis on differentiating phones that provide the

most useful information in an STD task. The algorithm for optimising W is described in Section 4.2. Using the enhanced posteriors directly in the searching, the score for an event is given by (pt − g t )T W V xt =

t=b

b+n−1 X

(5)

The use of (5) instead of (3) actually requires no change to the searching phase. Compared to the baseline approach, this is implemented simply as a transformation of the posterior-feature matrix index. Alternatively, rather than storing the N -dimensional features x0t = W V xt in the index, the M -dimensional decorrelated features V xt could instead be stored, with final multiplication by the weighting matrix W performed at search time. In the case that M < N , this approach provides for index compression, which is desirable in some applications. 4.2. Optimisation algorithm This section details the method for training the weights, W . The goal of the optimisation algorithm is to maximise FOM, however, the FOM is not a continuously differentiable function. Therefore, this section introduces a continuously differentiable objective function, f , that is a close approximation to −FOM. The function f is defined by replacing the step function, H (ek , ej ) in (2) with a sigmoid, ς (ek , ej ). That is, X 1 X ς (ek , ej ) (6) he k f =− A − + ς (ek , ej ) =

ej ∈R

1 . 1 + exp −α(sek − sej )

0.600 0.575

Train Eval (audio & terms) Train audio/Eval terms Eval audio/Train terms

0.550 0.525 0.500 0

10

20

30

40

50

Iteration

Fig. 1: FOM achieved after each gradient descent iteration on training data (Train). Also shown is the FOM achieved when the trained weights are used to search for held-out (Eval) terms and/or audio. will increase f if such a change generally causes the scores of hits to increase relative to the scores of false alarms. For computation of the Newton-Raphson step size, the Hessian is approximated with the diagonal of the Hessian, and is forced to be positive-definite by adding a multiple of the identity matrix when necessary [13]. The diagonal of the Hessian is found to be X ∂2f α2 X ς (ek , ej ) (1 − ς (ek , ej )) he k = − A ∂W 2 − + ek ∈R

ej ∈R

(1 − 2ς (ek , ej )) D (ek , ej )

(10)

Di,j (ek , ej ) = d2i,j (ek , ej ). 5. EXPERIMENTAL RESULTS 5.1. Training and evaluation data

(7)

The data used for training and evaluation is conversational telephone speech selected from the Fisher corpus [14]. Selected conversations are annotated as having high signal and conversation quality, from American English speakers and not made via speaker-phone. Training of the LCRC FeatureNet phoneme classifier uses 100 hours of speech with force-aligned 10ms frame phone labels, with each frame modelled with 15 log mel-filterbank channel outputs. The phone set consists of 43 phones, including a pause/silence model. The training set for the gradient descent algorithm consists of 10 hours of speech and 400 eight-phone search terms (referred to as training terms) with 1041 occurrences. The evaluation set consists of 8.7 hours of speech and 400 eight-phone search terms (evaluation terms). Search terms are selected randomly from those with at least one true occurrence in the reference transcript. Speakers and search terms do not overlap across training/evaluation sets.

(8)

5.2. FOM optimisation results

ej ∈R

∂sej ∂sek d (ek , ej ) = − ∂W ∂W b+n−1 X ∂s = (pt − g t ) (V xt )T ∂W

0.625

with the elements of D(ek , ej ) given by

The parameter α is a tunable constant controlling the slope of the sigmoid. A value of α = 1 was found to be reasonable in preliminary experiments, and is used in this work. The optimisation approach taken here is that of gradient descent, similar to [11, 12], whereby the weights, W , are found that correspond to a minimum of f . Specifically, the algorithm used is the Polak-Ribière variant of nonlinear conjugate gradients (CG) with the Newton-Raphson method and backtracking line search [13]. Prior to gradient descent, the weights are initialised such that W = V T , that is, the inverse of the decorrelating PCA transform. In the case that M = N , this initialisation ensures that (5) is equivalent to (3) before optimisation. In this way, a reasonable starting point is assured, and any adjustment of the weights away from this point during gradient descent constitutes an adaptation away from the baseline configuration. Given (5) and (6), the derivative of f with respect to the weights to be trained, W , is found to be X ∂f α X =− hek ς (ek , ej ) (1 − ς (ek , ej )) d (ek , ej ) ∂W A + − ek ∈R

0.675 0.650

(pt − g t )T x0t .

t=b

ek ∈R

0.700

FOM

s=

b+n−1 X

0.725

(9)

t=b

Intuitively, (9) shows that the change in an event’s score is related to the difference in alignments between the term model, and the background model. Furthermore, (8) and (7) show that a change in W

This section presents the improvement in FOM achieved on evaluation data by using the enhanced posterior-features with the weights W trained according to the algorithm in Section 4.2. Figure 1 shows the FOM achieved on training and evaluation sets using the weights obtained after each of the first fifty CG iterations. Continued but small improvements were observed on further iterations.

M

Training set (FOM) Initial Max

Eval. set (FOM) Initial Max

43 (M = N ) 40 35 25

0.569 0.566 0.566 0.437

0.547 0.534 0.522 0.390

0.703 (+24%) 0.699 (+24%) 0.686 (+21%) 0.678 (+55%)

0.606 (+11%) 0.597 (+12%) 0.586 (+12%) 0.572 (+47%)

Table 1: FOM achieved before FOM optimisation (Initial FOM, with W = V T ) and after training (Max FOM), and relative improvement compared to initial FOM, for different values of M , the number of dimensions retained after PCA. Training terms (FOM) Eval. terms (FOM) Initial Max Initial Max Training audio 0.569 0.703 (+24%) 0.523 0.591 (+13%) Eval. audio 0.597 0.654 (+10%) 0.547 0.606 (+11%) Table 2: FOM achieved before and after training, and relative improvement compared to baseline, evaluated when searching using various combinations of the audio and search term sets. Training terms Terms Occurrences Training audio Eval. audio

400 60

1041 512

Eval. terms Terms Occurrences 212 400

1089 1267

Table 3: Number of search terms occurring at least once and number of term occurrences in the training and evaluation sets. Before FOM optimisation (iteration 0), search is effectively performed directly on the phone log-posterior probabilities as in [9]. This baseline approach results in an FOM of 0.547 on the evaluation set. Table 1 shows that the FOM optimisation approach results in substantially improved FOM for both the training and evaluation sets in all situations. Figure 1 illustrates the optimisation algorithm for the first fifty iterations when all dimensions of PCA (M = N ) are retained, corresponding to the first row of Table 1. In this case, the FOM on the evaluation set continues to improve along with the training set, up to 0.606 (+11%), without over-training. Performance on the training terms and data, not surprisingly, enjoys a higher relative gain (+24%). As mentioned in Section 4.1, the number of feature dimensions retained after PCA, M , is a tunable parameter that influences the dimensions of W (an N × M matrix), and consequently the number of free parameters of the data. Table 1 shows that it is advantageous to retain all dimensions, that is, to use M = N = 43. For example, using M = 25 initially severely degrades FOM (from 0.547 to 0.390). However, it can be seen that FOM optimisation provides the highest relative gain on the evaluation data in this case (+47%). In fact, by using M = 25, an index compression factor of about 0.6 may be achieved with a relative decrease in FOM of less than 6% (from 0.606 to 0.572). Depending on the application, this may be a desirable compromise. To further explore the generalisation characteristics, the dependence on terms and/or audio was investigated by evaluating the FOM achieved in two further situations — that is, searching for the evaluation terms in the training audio and secondly, searching for the training terms in the evaluation audio. Table 3 indicates the number of terms from each term list that occur in each block of audio. Note that the FOM optimisation algorithm uses only terms and audio from the training set. These further results, included in Figure 1 and Table 2, illustrate that FOM is substantially improved for all combinations.

It is evident that the algorithm is not overly tuned to the training search terms because searching for these terms in the evaluation audio does not provide a greater relative gain (+10% c.f. +11%). Rather, there is an indication of slight dependence on the training audio as, for both term lists, the relative gain achieved on the training audio is greater than that achieved on the evaluation audio (+13% c.f. +11% for evaluation terms). Overall, the technique appears to generalise well to search terms and audio not used during training. Future work could involve improving the algorithm for training the parameters of the linear model. Also, incorporating the parameters of the neural network-based phone classifier in the optimisation process is expected to yield further improvements. 6. CONCLUSIONS This paper proposed a novel technique for direct optimisation of the Figure of Merit for phonetic STD. A gradient descent algorithm was shown to be effective for this purpose. By modelling vectors of log-posterior probabilities and training weights to exploit this information to optimise an objective function that is a close approximation to the FOM, the resulting system offers substantial improvements over the baseline that uses log-posterior probabilities directly. 7. REFERENCES [1] National Institute of Standards and Technology, “The Spoken Term Detection (STD) 2006 evaluation plan,” September 2006. [2] J.R. Rohlicek, W. Russell, S. Roukos, and H. Gish, “Continuous hidden Markov modeling for speaker-independent word spotting,” in Proc. ICASSP, 1989, pp. 627–630 vol.1. [3] U. Chaudhari, Hong-Kwang Jeff Kuo, and B. Kingsbury, “Discriminative graph training for ultra-fast low-footprint speech indexing,” in Proc. Interspeech, 2008, pp. 2175–2178. [4] Donglai Zhu, Haizhou Li, Bin Ma, and Chin-Hui Lee, “Discriminative learning for optimizing detection performance in spoken language recognition,” in ICASSP 2008, 2008, pp. 4161–4164. [5] Sheng Gao, Wen Wu, Chin-Hui Lee, and Tat-Seng Chua, “A MFoM learning approach to robust multiclass multi-label text categorization,” in ICML ’04, New York, NY, USA, 2004, p. 42, ACM. [6] E. Chang, Improving wordspotting performance with limited training data, Ph.D. thesis, Massachusetts Institute of Technology, 1995. [7] D. Grangier, J. Keshet, and S. Bengio, Automatic Speech and Speaker Recognition: Large Margin and Kernel Methods, chapter Discriminative Keyword Spotting, pp. 173–194, Wiley, 2009. [8] H. Ketabdar and H. Bourlard, “Hierarchical integration of phonetic and lexical knowledge in phone posterior estimation,” in Proc. ICASSP, 2008, pp. 4065–4068. [9] I. Szoke, P. Schwarz, P. Matejka, L. Burget, M. Karafiat, and J. Cernocky, “Phoneme based acoustics keyword spotting in informal continuous speech,” Lect Notes Comput Sc, vol. 3658, pp. 302, 2005. [10] P. Schwarz, Phoneme Recognition Based on Long Temporal Context, Ph.D. thesis, Brno University of Technology, 2008. [11] C. Burges, T. Shaked, E. Renshaw, M. Deeds, N. Hamilton, and G. Hullender, “Learning to rank using gradient descent,” in ICML, 2005. [12] V.C. Raykar, R. Duraiswami, and B. Krishnapuram, “A fast algorithm for learning large scale preference relations,” in AISTATS, 2007, vol. 2, pp. 388–395. [13] J. Nocedal and S.J. Wright, Numerical optimization, Springer-Verlag, 1999. [14] C. Cieri, D. Graff, O. Kimball, D. Miller, and K. Walker, “Fisher English training speech and transcripts,” December 2004.

OPTIMISING FIGURE OF MERIT FOR PHONETIC ...

urgent need for technologies to enable access to the information in .... 2The alternative transformations suggested in [9] did not improve STD accuracy in empirical ... in W , and can also have the benefit of suppressing low-energy di- rections ...

224KB Sizes 1 Downloads 144 Views

Recommend Documents

Thermoelectric figure of merit calculations for semiconducting nanowires
May 4, 2011 - A model for the thermoelectric properties of nanowires was used to demonstrate the contrasting influences of quantization and degeneracy on ...

Phonetic Symbols.pdf
Phonetic Symbols. for Old English through Modern English. Consonants. bilabial labiodental dental alveolar palatoalveolar palatal velar glottal. nasal m. me. n.

Optimising Typed Programs
Typed intermediate languages of optimising compilers are becoming increasingly recognised, mainly for two reasons. First, several type based optimisations and analyses have been suggested that cannot be done in an untyped setting. Such analyses and o

A Bangla Phonetic Encoding for Better Spelling ... - Semantic Scholar
Encode the input word using phonetic coding rules;. 2. Look up a phonetically ..... V. L. Levenshtein, “Binary codes capable of correcting deletions, insertions ...

Figure 16
Liang Jia is with Consumer Hardware Engineering, Google Inc, Mountain View, CA 94043, USA ..... Through the closed feedback loop control, Ve will be increased to Ve'. .... And the results match the simulation closely shown in Figure 9. In the.

Phonetic Realization of Contrastive Focus in Korean
following domain. .... free environment, the prompt question, the discourse ... 100. 150. 200. 250. 300. F. 0 (H z). Always. Only. Non-FP. [FOC]. [FOC]. [FOC] ...

phonetic encoding for bangla and its application to ...
These transformations provide a certain degree of context for the phonetic ...... BHA. \u09AD. “b” x\u09CD \u09AE... Not Coded @ the beginning sরণ /ʃɔroɳ/.

A phonetic study of voiced, voiceless and alternating ...
The Newsletter of the Center for Research in Language, University of California, San Diego, La Jolla CA 92093-0526 .... between voiced, voiceless and alternating stops is necessary to account for all of the data. * I would ..... Indiana University.

Infants attempting to learn the phonetic categories of ...
Statistical learning, cross-constraints and the acquisition of speech categories: a computational approach. Joseph Toscano. Bob McMurray [email protected] [email protected]. Dept. of Psychology. Dept. of Psychology. University of Iowa. Un

The Effect of Language Models on Phonetic Decoding ...
EVALUATION PROCEDURE. 3.1 Metrics and data. STD accuracy is measured in terms of simultaneously maximising the percentage of detected occurrences (detec- tion rate) and ... to a “standard” large vocabulary speech recognition config- uration, usin

Merit List of M.A. Education.pdf
Loading… Page 1. Whoops! There was a problem loading more pages. Retrying... Merit List of M.A. Education.pdf. Merit List of M.A. Education.pdf. Open. Extract.

Presentation - Optimising the guidance on significant benefit ...
Apr 25, 2017 - Industry stakeholder platform on research and development support. Presented by Matthias Hofer on 25 ... authorisation application. Page 2. Orphan environment after 16 years of EU orphan legislation. Recent developments ...

COMMON MERIT ...
Page 1 of 10. All india. NEET-UG. 2017 Rank. Common. Rank. MP. State. Rank. NEET Roll. No. Candidate Name Father's Name Gender MP. Domicile. Category Eligible. Category. Percentile Score. (out of. 720). 25297 1 0 901903014. ASHNEET KAUR. ANAND. KAVIN

Second Merit List of B.Ed.pdf
64 24641028 PG17160606 45.00 UR none HEMANSHU AGARWAL SHARDA PRASAD. AGARWAL. Page 3 of 45. Second Merit List of B.Ed.pdf. Second Merit ...

Phonetic Sound Cards - With Pictures.pdf
Page 2 of 42. S.G.SIVAKUMAR D.T.Ed., M.A., B.Ed. P.U.P.SCHOOL,. PERAMBAKKAM. KADAMBATHUR BLOCK. THIRUVALLUR. DISTRICT. www.asiriyar.com. Page 2 of 42. Page 3 of 42. S.G.SIVAKUMAR D.T.Ed., M.A., B.Ed. P.U.P.SCHOOL,. PERAMBAKKAM. KADAMBATHUR BLOCK. THI

Merit List.pdf
006927 Jibran Ali Waqar Ahmed. 006936 Kamran Ali shah Ayaz Ali shah. 007943 Kaneez Fatima Wali Muhammad. 008280 Khadija Farooq Azam Farooq Azam.

Supplemental Figure S5 – Related to Figure 5 - Broad Institute ...
(A) Correlation of mRNA expression and clinical features in 519 samples of glioblastoma multiforme (doi:10.7908/C1445K69). (B) Correlation of mRNA-Seq expression data from. 480 samples from kidney renal clear cell carcinoma with donor age. (doi:10.79

Merit List of M.Ed..pdf
80 1510639791 1510639791 27.7 UR none VIKRAM SHARMA SUBHASH CHAND. 81 25347074 PG17132547 27.5 UR none Rashmi sharma Ramotar ...

Figure 7 Global distribution of haemoglobin disorders, in terms of ...
The designations employed and the presentation of material on this map do not imply the expression of any opinion whatsoever on the part of the World Health.