Using Sub-sequence Information with kNN for Classification of Sequential Data Pradeep Kumar1,2, M. Venkateswara Rao1,2, P. Radha Krishna1, and Raju S. Bapi2 1

Institute for Development and Research in Banking Technology IDRBT, Castle Hills, Masab Tank, Hyderabad, India-500057 Ph No: 91-40-23534981, Fax No: 91-40-23535157 2 University of Hyderabad, Gachibowli, Hyderabad, India-500046 {pradeepkumar, prkrishna}@idrbt.ac.in, [email protected], [email protected]

Abstract. With the enormous growth of data, which exhibit sequentiality, it has become important to investigate the impact of embedded sequential information within the data. Sequential data are growing enormously, hence an efficient classification of sequential data is needed. k-Nearest Neighbor (kNN) has been used and proved to be an efficient classification technique for two-class problems. This paper uses sliding window approach to extract sub-sequences of various lengths and classification using kNN. We conducted experiments on DARPA 98 IDS dataset using various distance/similarity measures such as Jaccard similarity, Cosine similarity, Euclidian distance and Binary Weighted Cosine (BWC) measure. Our results demonstrate that sub-sequence information enhances kNN classification accuracy for sequential data, irrespective of the distance/similarity metric used. Keywords: Sequence mining, k-Nearest Neighbor Classification, Similarity/Distance metric, Intrusion detection.

1 Introduction Data are very vital for a commercial organization. These data are sequential or nonsequential in nature. Sequence mining helps us in discovering formal relations in sequence data. Sequence pattern mining is the mining of frequently occurring patterns related to time or other sequences [7, 15]. An example of the rule that sequence mining algorithm would discover is -- “A user who has visited rediff website is likely to visit yahoo website within next five page visits.” Sequence mining plays a vital role in domains such as telecommunication records, protein classification, signal processing and intrusion detection. It is important to note that datasets in these problems need not necessarily have inherent temporality [7, 15]. Studies on sequential pattern mining mostly concentrate on symbolic patterns [1, 10, 17]. As in symbolic patterns, numerical curve patterns usually belong to the scope of trend analysis and prediction in statistical time series analysis. Many other parameters also influence the results of sequential pattern mining. These parameters include duration of time sequence (T), event folding window (w) and time interval between two events (int). If we assign w as the whole duration T, we get time independent G. Chakraborty (Ed.): ICDCIT 2005, LNCS 3816, pp. 536 – 546, 2005. © Springer-Verlag Berlin Heidelberg 2005

Using Sub-sequence Information with kNN for Classification of Sequential Data

537

frequent patterns. An example of such a rule is “ In 1999, customers who bought PCs also bought digital cameras”. If w is set to be 1, that is, no event sequence folding occurs, then all events are considered to be discrete time events. The rule of the type “Customers who bought hard disk and then memory chip are likely to buy CD-Writer later on” is example of such a case. If w were set to be something between 1 and T, events occurring between sliding windows of specified length would be considered. An example rule is “Sale of PC in the month of April 1999 is maximum”. Sequential data are growing at a rapid pace. A pre-defined collection of historical data with their observed nature helps in determining the nature of newly arriving data stream and hence will be useful in classification of the new data stream. In data mining, classification algorithms are popularly used for exploring the relationships among various object features at various conditions. Sequence data sets are similar in nature except that they have an additional temporal dimension [22]. Classification algorithms help in predicting future trends as well as extracting a model of important data classes. Many classification algorithms have been proposed by researchers in machine learning [21], expert systems [20], statistics [8]. Classification algorithms have been successfully applied to the problems, where the dependent variable (class variable) depends on non-sequential independent (explanatory) variables [3]. Typical classification algorithms are Support Vector Machines, Decision Trees, Bayesian Classification, Neural Networks, k-Nearest Neighbor (kNN) and Association Classification. To deal with the sequential information, sequential data are transformed into non-sequential variables. This leads to a loss of sequential information of the data. Although traditional classification is robust and efficient for modeling non-sequential data, they fail to capture sequential information of the dataset. Intrusion detection is the process of monitoring and analyzing the events occurring in a computer system in order to detect signs of security problems [2]. Computer security can be achieved by maintaining audit data. Cryptographic techniques, authentication means and firewalls have gained importance with the advent of new technologies. With the ever-increasing size of audit data logs, it becomes crucial for network administrators and security analysts to use some efficient Intrusion Detection System (IDS), to reduce the monitoring activity. Data mining techniques are useful in providing important contributions to the field of intrusion detection. IDSs based on examining sequences of system calls often define normal behavior of an application by sliding a window of fixed size across a sequence of traces of system calls. System call traces are normally produced with programs like strace on Linux systems and truss on Solaris systems. Several methods have been proposed for storing system calls traces’ information and to use these for detecting anomalies in an IDS. Forrest et al. [5, 9] stored normal behavior by sliding a window of fixed size L across sequence of system call traces and recorded which system call followed the system call in position 0 at offsets 1 through L-1. Liao et al. [12] applied kNN classifier with Cosine similarity measure considering frequencies of system calls with sliding window size w =1. A similar work with modified similarity measure using a combination of Cosine as well Jaccard has also been carried out in [18]. The central theme of this paper is to investigate that vital information stored in subsequences, plays any role in building a classifier. In this paper, we combine sequence analysis problem with kNN classification algorithm, to design an efficient classifier

538

P. Kumar et al.

for sequential data. Sequence analysis can be categorized into two types, depending on the nature of the treatment. Either we can consider the whole sequence as one or sub-sequences of different sizes. Our hypothesis is that sequence or order of information plays a role in sequence classification. We extracted sequence information from sub-sequences and used this information for building various distance/similarity metrics. With the appropriate distance/similarity metric, a new session is classified using kNN classifier. In order to evaluate the efficiency and behavior of the classifier with the encoded vector measures, Receiver Operating Characteristics (ROC) curve is used. Experiments are conducted on DARPA 98 IDS [13] dataset to show the viability of our model. Like other classification algorithms, kNN classification algorithm does not make a classifier in advance. Hence, it is suitable for classification of data streams. Whenever a new data stream comes, kNN finds the k near neighbors to new data stream from training data set using some distance/similarity metric [4, 6]. kNN is the best choice for making a good classifier, when simplicity and accuracy is important issues [11]. The rest of the paper is organized as follows - Section 2 gives a brief description of the nearest neighbor classification algorithm. In section 3, we briefly discuss about the distance/similarity measures used in the experiments. In section 4, we outline our proposed approach. The Section 5 provides the experimental results on DARPA 98 IDS dataset. Finally, we conclude in section 6.

2 Nearest Neighbor Classification kNN classifier are based on learning by analogy. KNN classification algorithm assumes that all instances correspond to points in an n-dimensional space. Nearest neighbors of an instance are described by a distance/similarity measure. When a new sample comes, a kNN classifier searches the training dataset for the k closest sample to the new sample using distance/similarity measure for determining the nature of new sample. These k samples are known as the k nearest neighbors of the new sample. The new sample is assigned the most common class of its k nearest neighbors. Nearest neighbor algorithm can be summarized as follows: Begin Training Construct Training sample T from the given dataset D. Classification Given a new sample s to be classified, Let I1… Ik denote the k instances from T that are nearest to new sample s Return the class from k nearest neighbor samples. Returned class is the class of new sample. End In the nearest neighbor model, choice of a suitable distance function and the value of the members of nearest neighbors (k) are very crucial. The k represents the complexity of nearest neighbor model. The model is less adaptive with higher k values [7].

Using Sub-sequence Information with kNN for Classification of Sequential Data

539

3 Distance/Similarity Measures Distance/similarity measure plays an important role in classifying or grouping observations in homogeneous groups. In other words, a distance/similarity measure establishes the relationship between the rows of the data matrix. Preliminary information for identifying homogeneous groups is provided by the distance/similarity measure. Between any pair of observations xi and xj function of the corresponding row vector in the data matrix is given by: Dij = f (xi , xj ) where i,j = 1, 2, 3,…,n For an accurate classifier, it is important to formulate a metric to determine whether an event is deemed normal or anomalous. In this section, we briefly discuss various measures such as Jaccard similarity measure, Cosine similarity measure, Euclidian distance measure and BWC measure. We used sub-sequence information with these different measures in kNN classifier for cross comparison purpose. 3.1 Jaccard Similarity Function Jaccard similarity function is used for measuring similarity between binary values [19]. It is defined as the degree of commonality between two sets. It is measured as a ratio of number of common attributes of X AND Y to the number of elements possessed by X OR Y. If X and Y are two distinct sets then the similarity between X and Y is: S(X,Y) =

| X ∩Y | | X ∪Y |

Consider two sets X =〈 M, N, P, Q, R, M, S, Q〉 and Y = 〈P, M, N, Q, M, P, P〉. X ∩ Y is given as 〈M, N, P, Q〉 and X ∪ Y is 〈M, N, P, Q, R, S〉. Thus, the similarity between X and Y is 0.66. 3.2 Cosine Similarity Cosine similarity is a common vector based similarity measure. Cosine similarity measure is commonly used in text databases [16]. Cosine similarity metric calculates the angle of difference in direction of two vectors, irrespective of their lengths. Cosine similarity between two vectors X and Y is given by: S(X,Y) =

X •Y | X || Y |

Direct application of Cosine similarity measure is not possible across sets. Sets are first converted into n-dimensional vector space. Over these transformed vectors Cosine similarity measure is applied to find the angular similarity. For two sets, X = 〈M, N, P, Q, R, M, S, Q〉 and Y = 〈P, M, N, Q, M, P, P〉 the equivalent transformed frequency vector is Xv = < 2,1,1,2,1,1> and Yv = < 2,1,3,1,0,0 >. The Cosine similarity of the transformed vector is 0.745.

540

P. Kumar et al.

3.3 Euclidean Distance Euclidean distance is a widely used distance measure for vector spaces [16]. For two vectors X and Y in an n- dimensional Euclidean space, it is defined as the square root of the sum of difference of the corresponding dimensions of the vector. Mathematically, it is given as

D(X,Y) =

⎡n 2⎤ ( X Y ) − ∑ ⎢⎣ s =1 ⎥⎦ s

1/ 2

s

Similar, to the Cosine similarity metric, application of Euclidean measure on sets is not possible. Similar approach as used in Cosine similarity measure to transform sets into vector is applicable here also. For two sets, X = 〈 M, N, P, Q, R, M, S, Q〉 and Y = 〈P, M, N, Q, M, P, P〉 the equivalent transformed frequency vector is Xv = < 2,1,1,2,1,1> and Yv = < 2,1,3,1,0,0 >. The Euclidean measure of the transformed vector is 2.64. 3.4 Binary Weighted Cosine (BWC) Metric Rawat et.al.[18] proposed BWC similarity measure for measuring similarity across sequences of system calls. They showed the effectiveness of the proposed measure on IDS. They applied kNN classification algorithm with BWC metric measure to enhance the capability of the classifier. BWC similarity measure considers both the number of shared elements between two sets as well as frequencies of those elements in traces. The similarity measure between two sequences X and Y is given by S (X, Y)=

X •Y | X ∩Y | * | X || Y | | X ∪ Y |

BWC measure is derived from Cosine similarity as well as Jaccard similarity measure. Since the Cosine similarity measure is a contributing component in a BWC similarity measure hence, BWC similarity measure is also a vector based similarity measure. The transformation step is same as carried out in Cosine similarity measure or Euclidean measure for sets. For two sets, X =〈M, N, P, Q, R, M, S, Q〉 and Y = 〈P, M, N, Q, M, P, P〉 the Cosine similarity is given as 0.745 and Jaccard similarity as 0.66. Hence, the computed BWC similarity measure comes out to be 0.49.

4 Proposed Methodology This section illustrates the methodology for extracting sequential information from the sets, thus making it applicable to be used by various vector based distance/similarity metrics. We considered sub-sequences of fixed sizes: 1,2,3… This fixed size subsequence is called window. This window is slided over the traces of system calls to find the unique sub-sequences of fixed length s over the whole dataset. A frequency count of each sub-sequence is recorded. Consider a sequence, which consists of traces of system calls.

Using Sub-sequence Information with kNN for Classification of Sequential Data

541

execve open mmap open mmap mmap mmap mmap mmap open mmap exit execve open mmap open mmap mmap mmap mmap mmap open mmap exit Sliding window of size 3 Total length of sequence is 12 with the sliding window size w (=3) we will have total sub-sequences of size 3 as 12 –3 + 1= 10. These 10 sub-sequences of size 3 are execve open mmap mmap mmap mmap mmap open mmap

open mmap open mmap mmap mmap open mmap exit

mmap open mmap mmap mmap mmap

open mmap mmap mmap mmap open

From among these 10 generated sliding window-sized sub-sequences unique subsequences with their frequencies are as follows: execve open mmap open mmap open open mmap mmap mmap mmap mmap

1 1 1 3

mmap open mmap mmap mmap open open mmap exit

2 1 1

With these encoded frequencies for sub-sequences, we can apply any vector based distance/similarity measure, thus incorporating the sequential information with vector space. The traditional classification algorithm – the kNN classification algorithm [4, 7] with suitable distance/similarity metric can be used to build an efficient classifier. Our proposed methodology consists of two phases namely training and testing phase. Dataset D consists of m sessions. Each session is of variable length. Initially in training phase, all the unique sub-sequences of size s are extracted from the whole dataset. Let n be the number of unique sub-sequences of size w, generated from the dataset D. A matrix C of size m × n is constructed where Cij is given by count of jth unique sub-sequence in the ith session. A distance/similarity metric is constructed by applying distance/similarity measure over the C matrix. The model is trained with the dataset consisting of normal sessions. In testing phase, whenever a new process P comes to the classifier, it looks for the presence of any new sub-sequence of size s. If a new sub-sequence is found, the new process is marked as abnormal. When there is no new sub-sequence in new process P, calculate the similarity of new process with all the sessions. If similarity between any session in training set and new process is equal to 1, mark it as normal. In other case, pick the k highest values of similarity between new process P and training dataset. From this k maximum values, calculate the average similarity for k-nearest neighbors. If the average similarity value is greater than user defined threshold value ( τ ) mark the new process P as normal, else mark P as abnormal.

5 Experimental Results Experiments were conducted using k-Nearest Neighbor classifier with Jaccard similarity function, Cosine similarity measure, Euclidean distance and BWC metric.

542

P. Kumar et al.

Each distance/similarity metric was individually experimented with kNN classifier on DARPA 98 IDS dataset. DARPA 98 IDS dataset consists of TCPDUMP and BSM audit data. The network traffic of an Air Force Local Area Network was simulated to collect TCPDUMP and BSM audit data [13]. The audit logs contain seven weeks of training data and two weeks of testing data. There were 38 types of network-based attacks and several realistic intrusion scenarios conducted in the midst of normal background data. Detailed discussion of DARPA dataset is given at [12]. For experimental purpose, 605 unique processes were used as a training dataset, which were free from all types of attacks. Testing was conducted on 5285 normal processes. In order to test the detection capability of proposed approach, we incorporate 55 intrusive sessions into our test data. For kNN classification experiments, k=5 was considered. With various discussed distance/similarity measures in the above section (Jaccard similarity measure, Cosine similarity measure, Euclidean distance measure and BWC similarity measure) at different sub-sequence lengths (sliding window size) L=1,3,5 experiments were carried out. Here, L=1 means that no sequential information is captured whereas, for L > 1 some amount of order information across elements of the data is preserved. sub-seq L=1

sub-seq L =3

sub-seq L =5

1.2

Detection Rate

1 0.8 0.6 0.4 0.2 0 0

0.005 0.01 False Positive Rate

0.015

Fig. 1. ROC curve for Jaccard similarity metric using kNN classification for k =5

To analyze the efficiency of classifier, ROC curve is used. The ROC curve is an interesting tool to analyze two-class problems [14]. ROC curve is very useful where situations detection of rarely occurring event is done. ROC curve depicts the relationship between False Positive Rate (FPR) and Detection Rate (DR) at various threshold values. DR is the ratio of the number of intrusive sessions (abnormal) detected correctly to the total number of intrusive sessions. The FPR is defined as the number of normal processes detected as abnormal, divided by the total number of normal processes. ROC curve gives an idea of the trade off between FPR and DR achieved by classifier. An ideal ROC curve would be parallel to FPR axis at DR equal to 1.

Using Sub-sequence Information with kNN for Classification of Sequential Data

Detection Rate

Sub-seq L=1

Sub-seq L=3

543

Sub-seq L=5

1.2 1 0.8 0.6 0.4 0.2 0 0

0.1

0.2

0.3

0.4

False Positive Rate Fig. 2. ROC curve for Cosine similarity metric using kNN classification for k =5 sub-seq L=1

sub-seq L=3

Sub-seq L=5

1.2

Detection Rate

1 0.8 0.6 0.4 0.2 0 0

0.05

0.1

0.15

False Positive Rate

Fig. 3. ROC curve for Euclidian distance metric using kNN classification for k =5

Corresponding ROC curves for Jaccard similarity measure, Cosine similarity measure, Euclidean distance measure and BWC measure are shown in fig 1, 2, 3 and 4 respectively. It can be observed from fig 1,2,3 and 4 that as the sliding window size increases from L =1 to L = 5, high DR (close to ideal value of 1) is observed with all the distance/similarity metrics. Rate of increase in false positive is less for Jaccard similarity measure (0.0050.015) as compared to different distance/similarity metrics such as Cosine similarity (0.1-0.4), Euclidian distance (0.05-0.15) and BWC similarity (0.1-0.7). Table 1 depicts the factor (FPR or Threshold value) that was traded off in order to achieve high DR. For example, in the case of Jaccard similarity measure, FPR was traded off for threshold values (highlighted in bold face) in order to achieve high DR.

544

P. Kumar et al.

Sub-seq L=1

Sub-seq L=5

Sub-seq L=3

1.2 Detection Rate

1 0.8 0.6 0.4 0.2 0 0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

False Positive Rate

Fig. 4. ROC curve for BWC similarity metric using kNN classification for k =5 Table 1. Results for different distance/similarity metric Jaccard similarity measure

τ

L =1 L =3 L =5

0.94 0.95 0.89

Cosine similarity measure

FPR 0.0056 0.011 0.0105

τ

0.99 0.99 0.75

Euclidian distance measure

FPR 0.29 0.12 0.03

τ

0.99 0.99 0.99

BWC similarity measure

FPR 0.12 0.07 0.06

τ 0.89 0.7 0.65

FPR 0.096 0.28 0.30

Thus, our results support the hypothesis that classification accuracy of sequential data can be improved by incorporating the order information embedded in sequences. We also performed experiments with different k values for nearest neighbor classifier with all the four measures. Table 2. False positive rate at maximum attained detection rate for different sub-sequence length for different distance/similarity measure at k =7

Jaccard similarity Euclidian distance Cosine distance BWC measure

L=1 0.0058 0.94 0.3286 0.0885

L=3 0.0102 0.0047 0.1799 0.0783

L =5 0.0105 0.0085 0.0387 0.0787

We present the false positive rate at maximum attained detection rate for different sub-sequence lengths L = 1, 3, 5 with all the distance/similarity measures in table 2 for k =7. It can be observed that, as per the trend, the FPR is increasing with the increasing sub-sequence lengths for all the four measures. We also performed experiments with k =10 and the trend is also found to be consistent (Results are not included here).

Using Sub-sequence Information with kNN for Classification of Sequential Data

545

6 Conclusion Using Intrusion Detection as an example domain, we demonstrated in this paper the usefulness of utilizing sub-sequence information for kNN classification of sequential data. We presented results on DARPA 98 IDS dataset wherein we systematically varied the length of the sliding window from 1 to 5 and used various distance /similarity measures such as Jaccard similarity, Cosine similarity, Euclidian distance and BWC similarity measure. As the sub-sequence information is increased, the high DR is achieved with all the four measures. Our results show that if order information is made available, a traditional classifier such as kNN can be adapted for sequence classification problem. We are currently working on design of new similarity measure, for capturing complete sequential information. Although the current paper presented results in the domain of information security, we feel this methodology can be adopted for the domains such as web mining, text mining and bio-informatics.

References 1. Agrawal, R., Faloutsos, C. and Swami, A.: Efficient similarity search in sequence databases. In proceedings of the 4th Int'l Conference on Foundations of Data Organization and Algorithms. Chicago, IL, 1993. pp 69-84. 2. Bace, R.: Intrusion Detection. Macmillan Technical Publishing, 2000. 3. Buckinx, W., Moons, E., Van den Poel, D. and Wets, G: Customer-Adapted Coupon Targeting Using Feature Selection, Expert Systems with Applications 26, No. 4 2004, 509-518. 4. Dasarathy, B.V.: Nearest-Neighbor Classification Techniques, IEEE Computer Society Press, Los Alomitos, CA, 1991. 5. Forrest S, Hofmeyr S A, Somayaji A and Longstaff T.A.: A Sense of self for UNIX process. In Proceedings of the IEEE Symposium on Security and Privacy, pages 120-128, Los Alamitos, CA, 1996. IEEE Comuputer Socity Press. 6. Gludici, P: Applied Data Mining , Statistical methods for business and industry, Wiely publication, 2003. 7. Han, Jiawei., Kamber, Micheline.: Data Mining , Concepts and Techniques, Morgan Kaufmann Publishers, 2001. 8. Hastie, T., Tibshirani, R. and Friedman, J. H.: The Elements of Statistical Learning, Data Mining, Inference, and Prediction, Springer, 2001. 9. Hofmeyr S A, Forrest S, and Somayaji A.: Intrusion Detection Using Sequences of System calls. Journal of Computer Security, 1998, 6:151-180. 10. Keogh, E., Chakrabarti, K., Pazzani, M. and Mehrotra, S.: Locally adaptive dimensionality reduction for indexing large time series databases. In proceedings of ACM SIGMOD Conference on Management of Data. Santa Barbara, CA, 2003. pp 151-162. 11. Khan, M., Ding, Q. and Perrizo, W.: k-Nearest Neighbor Classification on Spatial Data Streams Using P-Trees, In the Proceedings of the 6th Pacific-Asia Conference on Advances in Knowledge Discovery and Data Mining, 2002. 12. Liao, Y., Rao Vemuri, V.: Using Text Categorization Techniques for Intrusion Detection. USENIX Security Symposium 2002: 51-59. 13. MIT Lincoln Laboratory, http://www.ll.mit.edu/IST/ideval/.

546

P. Kumar et al.

14. Marques de sa, J.P: Pattern recognition: concepts, methods and applications, SpringerVerlag 2001. 15. Pujari, A.K.: Data Mining Techniques, Universities Press INDIA, 2001. 16. Qian, G, Sural, S., Gu, Y., Pramanik, S.: Similarity between Euclidean and cosine angle distance for nearest neighbor queries. SAC 2004: 1232-1237 17. Ratanamahatana, C. A. and Keogh. E..: Making Time-series Classification More Accurate Using Learned Constraints. In proceedings of SIAM International Conference on Data Mining (SDM '04), Lake Buena Vista, Florida, 2004. pp. 11-22. 18. Rawat, S. Pujari, A.K., Gulati, V.P.,and Vemuri, V. Rao.: Intrusion Detection using Text Processing Techniques with a Binary-Weighted Cosine Metric. International Journal of Information Security, Springer-Verlag, Submitted 2004. 19. Sams String Metrics, http://www.dcs.shef.ac.uk/~sam/stringmetrics.html 20. Sholom M. Weiss and Casimir A. Kulikowski: Computer Systems That Learn: Classification and Prediction Methods from Statistics, Neural Nets, Machine Learning, and Expert Systems (Machine Learning Series), Morgan Kaufmann Publishers Inc. San Francisco, CA, USA , 1991. 21. Tom M. Mitchell.: Machine learning, Mc Graw Hill 1997. 22. Wang, Jason T.L.; Zaki, Mohammed J.; Toivonen, Hannu T.T.; Shasha, Dennis: Data mining in bioinformatics, Springer-Verlag 2005

Using Sub-sequence Information with kNN for ...

in a computer system in order to detect signs of security problems [2]. ..... Rate of increase in false positive is less for Jaccard similarity measure (0.005-.

213KB Sizes 0 Downloads 214 Views

Recommend Documents

Granger Causality Driven AHP for Feature Weighted kNN
Jan 17, 2017 - The kNN algorithm [1,2] remains a popular choice for pattern ...... [32] K. Fukunaga, Introduction to Statistical Pattern Recognition, Computer Science .... obtained the Ph.D. degree from the University of Jadavpur, Kolkata, India in .

Fuzzy-KNN-Prediksii.pdf
of Secondary Structure of Proteins (DSSP) [17] where the secondary structure is. classified as eight states. By grouping these eights states into three classes Coil.

SVM-KNN: Discriminative Nearest Neighbor ...
Michael Maire. Jitendra Malik. Computer Science Division, EECS Department. Univ. of California, Berkeley, CA 94720 ... yields excellent results in practice. The basic idea is to find close neighbors to a query sample and ... rather by similarity to p

BOOTSTRAPPING PERCEPTION USING INFORMATION ... - eSMCs
Homeokinesis: A new principle to back up evolution with learning (IOS Press, 1999). [6] Edgington, M., Kassahun, Y., and Kirchner, F., Using joint probability ...

Using Syntactic Information for Improving why-Question ...
Improving the QA system. – Better retrieval technique (search engine). – Better ranking system. • Syntactic knowledge between question and answer helps?

Using lexico-semantic information for query expansion ...
retrieval engine using Apache Lucene (Jakarta,. 2004). Documents have been .... method (1.2K vs 1.4K, as can be seen in 1). The proximity-based method ...

Using lexico-semantic information for query expansion ...
back loop that feeds lexico-semantic alternations .... in the top-k passages returned by the system. The ..... http://lucene.apache.org/java/docs/index.html. Kaisser ...

Using Syntactic Information for Improving why-Question ...
Improving the QA system. – Better retrieval technique (search engine). – Better ranking system. • Syntactic knowledge between question and answer helps?

Using lexico-semantic information for query expansion ...
Using lexico-semantic information for query expansion in passage retrieval for question answering. Lonneke van der Plas. LATL ... Information retrieval (IR) is used in most QA sys- tems to filter out relevant passages from large doc- ..... hoofdstad

System and method for obtaining and using location specific information
Sep 1, 2010 - supports the coordinate entry or linked to an existing Web ..... positions to any GPS receiver that is within the communica tion path and is tuned ...

System and method for obtaining and using location specific information
(73) Assignee: Apple Inc., Cupertino, CA (US). (21) App1.No.: 12/874,155. (22) Filed: Sep. 1, 2010. Related US. Patent Documents. Reissue of: (64) Patent No.:.

Longest Common Subsequence (LCS) Retrieval in ...
ABSTRACT. Alignment of DNA sequences is one of the routines in bioinformatics. The longest common subsequence (LCS) retrieval problem can be treated as ...

Using sycamore education information - anytime ... -
and click the Login bu on. Sycamore Education is web-based and works in all modern web browsers including Internet Explorer and. Google Chrome. Mobile apps are also available for iOS and Android users. Visit your mobile webstore to download the laste

Information security when using the EudraVigilance system
[PDF]Information security when using the EudraVigilance systemfeedproxy.google.com/~r/EmaUpdates/~3/aIYj0klfpUE/open_document.jspCachedApr 18, 2017 - The Agency is committed to ensuring the confidentiality, integrity and availability of its informati

Mobile Information Agent using
using the internet as the network infrastructure to make available heterogeneous museum information .... Finally, .net framework deals with standards efficiently.

Finding and Using Internet Information
Sep 21, 2006 - Concerning the Internet, students coming to me for their third ... scheduled a computer-related assignment for an entire class or has borrowed a ...

USING COLLECTIVE INFORMATION IN SEMI ...
The application of the semi-supervised learning tech- niques to the ASR is not ... duced to selecting a best transcription from the lattice. The existing algorithms ...

Using NET-SNMP with A91750
Creating the document. Michael Arp ... most likely have net-snmp in their application repository, so it's very easy to install. On Windows you ... For windows, please look through the official documentation here: ​http://www.net-snmp.org/wiki/​.

Using E-Learning for Helping Children with Diabetes
into E-Learning systems and how it can help people with diabetes, especially when it comes to children who are largely unaware and poorly informed about the menace of the disease. This research addresses children' needs expectations, and proposes a d