Abstract
We present a study of sera derived from the malaria medical analysis of 189 subjects. The feature space is 18-dimensional and each serum is represented by a binary number. The subjects are divided into three different groups: no malaria, clinical malaria and asymptomatic subjects. We studied the main characteristics of the data and we selected 7 out of the 18 antigens as the most important for group discrimination. We propose a novel representation of the data in the so-called relational space, where the coded data of pairs of patients are plotted. We are able to separate the groups with 58% accuracy, about 15% points better than several conventional methods with which we compare our results.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Malaria is one of the most common infectious diseases and an enormous public health problem. According to the World Health Organisation, “Half of the world’s population is at risk of malaria, and an estimated 247 million cases led to nearly 881 000 deaths in 2006” (from the World malaria report 2008 [20]). Some researchers evaluated that deaths are up to 50% higher than those reported by the World Health Organisation [15]. Snow et al. estimated that there were approximately 515 (range 300–660) million cases of Plasmodium falciparum malaria in 2002, while the number of deaths was between 700,000 and 2.7 million [15]. This represents at least one death every 30 s.
The vast majority of cases occur in children under the age of 5 and especially in Africa where 90% of malaria fatalities occur [2]. Despite efforts to reduce its transmission and increase treatment, there has been little change in those areas which are at risk by this disease since 1992 [10]. Indeed, if the prevalence of malaria stays on its present upwards course, the death rate could double in the next 20 years [2]. Precise statistics are unknown, because many cases occur in rural areas where people do not have access to hospitals or the means to afford health care. Consequently, the majority of cases are undocumented.
Malaria is caused by protozoan parasites of the genus Plasmodium [19]. Only four types of the plasmodium parasite can infect humans [19]. The most serious forms of the disease are caused by Plasmodium falciparum and Plasmodium vivax, but other related species (Plasmodium ovale, Plasmodium malariae) can also affect humans [19]. These human-pathogenic plasmodium species are usually referred to as malaria parasites and they are transmitted by female anopheles mosquitoes [19]. The parasites multiply within the red blood cells, causing symptoms that include symptoms of anaemia (light headedness, shortness of breath, tachycardia, etc.) as well as other general symptoms such as fever, chills, nausea, flu-like illness, and in severe cases, coma and death [19].
No vaccine is currently available for malaria; preventative drugs must be taken continuously to reduce the risk of infection (http://www.malariavaccine.org/). These prophylactic drug treatments are often very expensive for most people living in endemic areas [20]. Malaria infections are treated by the use of antimalarial drugs, such as quinine or artemisinin derivatives, although drug resistance is increasingly common [19].
Research on malaria vaccine is mainly focused on finding differences between affected people, healthy people and people that are protected against it. A considerable role is played by new technologies such as microarray techniques and bio-informatics. Algorithms of pattern recognition and signal processing are speeding up the malaria vaccine development. Consider only that the completed genome sequences of bacteria and parasites have disclosed the presence of ∼1,000−5,000 candidate proteins (and variants thereof) for each microorganism [8].
In this work, we analyse and cluster sera derived from the medical analysis of 189 subjects. The data are the result of a study using microarray technology to characterise the serum reactivity profiles of 189 children, 3–9 years old, living in Gambia. In this study, Gray et al. [8] assessed the antibody responses against 18 recombinant proteins derived from 4 leading blood-stage vaccine candidates for Plasmodium falciparum malaria.
The sera are divided into three different groups: no malaria (group A) including children having no evidence of infection throughout the study period; clinical malaria (group B), children having at least 1 episode of fever (temperature more than 37.5°C) and parasitemia ≥5,000 μl; and asymptomatic (group C), children with parasitemia or acquired splenomegaly but with no evidence of fever or other symptoms of clinical malaria [8].
To analyse the data and cluster them, Gray et al. [8] used the k-means clustering technique. The data were clustered into three clusters. Their results are reported in Table 1, where we present the number of subjects from each group that were classified in each of the three clusters created.
We note from these results that it is not possible to associate the three identified clusters with the three classes of subjects we have, without committing a very significant error. For example, cluster 2 could easily be considered to represent group A or group B. If we consider cluster 2 to represent group B, cluster 1 to represent group C and cluster 3 to represent group A, 108 out of the 189 subjects will be wrongly classified, producing a classification accuracy of 43% only.
In this paper, first, in Sect. 2, we analyse the data to work out which antigens have the most discriminatory power between the groups, and then examine what the minimum possible classification error one can achieve is. In Sect. 3, we present novel methodology to approach this ideally minimum error, by considering the context of the data expressed by their pairwise relations. Our results are presented in Sect. 4 and our conclusions in Sect. 5.
2 Data characteristics
Figure 1 shows the data divided into three sets: group A, group B and group C. In this figure, each matrix column represents the subjects’ immunology response, while the ith row shows the positive (“1”) or negative (“0”) response of all subjects to the ith antibody. The three groups are composed, respectively, of 55 subjects (group A), 81 subjects (group B) and 53 subjects (group C). Each subject is represented by a binary 18-dimensional vector: a component is 1 if the child’s serum has a positive response and 0 if it has negative response to the corresponding antibody.
To discover whether there are proteins which permit us to distinguish better one group from the others, we show in Fig. 2 the mean value for each matrix row. We observe that some antigens have the same mean value (i.e. antigens 2 or 6), while for antigens 1, 10, 11, 13, 14, 15, 18, the mean values are very different. We select the most discriminative antigens by considering the mean value μA, μB and μC for sets A, B and C and the mean μ for the whole population. We choose those antigens that have a mean value, which, for at least one set, is considerably different from the mean value of the whole population. Defining \(\Upsigma^*\) vector as
we select the ith antigen if the ith component of \(\Upsigma^*\) is considerably large.Footnote 1 Figure 3 shows the value of \(\Upsigma^*\) for each antigen. There is a natural gap in the values of these antigens that allows us to divide them into two categories. However, one may also define a threshold value following Otsu’s method [11], where the threshold is selected to make the two populations as compact as possible. It turns out that a threshold in the range [0.05–0.07] minimises the total intraclass spread. Using such a threshold allows us to select the most meaningful antigens. The selected antigens are those that lie above the line in Fig. 3. We call these antigens principal antigens.
2.1 Confidence intervals and significance of the principal antigens
To quantify how significant the difference in mean value is for the principal antigens selected, we considered the value of each antigen as a random variable with a binomial probability distribution.
Assuming that the outcome of a random experiment is either 0 or 1, the probability of getting k 1s in n trials is given by
where p is the probability of getting 1 in a single trial. Let us denote by k A i the total number of subjects of group A that are positive to the ith antigen (they have “1” in their ith component). In a similar way, we define k B i and k C i . According to these definitions, the probability of finding a person that has “1” at the ith component is
where N A = 55, N B = 81 and N C = 53. Therefore, the probability of getting k i people positive with respect to the ith antigen is given by
where N = 189. We also introduce the probability of finding a person that is positive with respect to the ith antigen in groups A, B and C asFootnote 2
Now we want to check whether p A i , p B i and p C i are significantly different from p i . For example, if the probability to get k i positive subjects in the whole population is the same as that of getting k A i positive subjects in set A, then
For this purpose, we defined the confidence intervals for the binomial distribution as well as for the normal distribution.Footnote 3 Considering Fig. 4 and set A, for each antigen we defined two sets:
Then
According to the above definitions, we define the confidence interval of at least \(\epsilon{\%}\) for group A, with respect to the ith antigen, as the set \(\mathcal{K}_{i,{\text{A}}}\) where threshold V th is chosen so that the sum of the impulses with amplitude greater than V th is at least \(\frac{\epsilon}{100},\) while the sum of the impulses of the probability distribution with amplitude less than V th is less than \((1-\frac{\epsilon}{100}){:}\)
In the above formulae, we considered group A and probability p i in place of p A i . For fixed \(\epsilon,\) if k A i belongs to \(\mathcal{K}_{i,{\text{A}}}\) for small values of \(\epsilon\) (big values of V th) it means that p A i is close to p i , while if k A i belongs to \(\overline{\mathcal{K}}_{i,{\text{A}}}\) for big values of \(\epsilon\) (small values of V th) it means that p A i is significantly different from p i .
Changing the level of the confidence interval, we can analyse the importance of the antigens for the three groups. For confidence ranging from 50 to 99%, Table 2 presents the antigens that are out of the corresponding interval.
Let us focus our attention to antigen 15. The mean value for antigen 15 in group A is outside the 92% confidence interval. This means that if we consider the whole population and a subset of it composed only of group A, this subset does not represent the population for antigen 15 in 92% of the cases. According to that, we can say that the “significance” of antigen 15 is 92%.
From Table 2, we can see that all seven principal antigens have a significance value between 92 and 55%. Moreover, some antigens are significant for more than one group. However, note that there is no antigen that characterises in any significant way group B. This is as expected, because group B is the biggest group and forms a great part of the population (43%). This means that the value of μ is close to μB and there are no antigens that characterise the difference between group B and the whole population. We do not include antigen 9 to the principal antigens, because the ninth value of \(\Upsigma^*\) is considerably lower than its first and tenth components (see Fig. 3).
2.2 The minimum possible error
When each subject is represented by a small number of antigens (in our case 18 or 7), expressed as binary patterns, it is possible that different subjects have identical representations. If subjects with identical representations belong to the same group, this is a good characteristic of the data. If, however, they belong to different groups, then one has to realise that it is not possible to distinguish these two subjects using only the information that is available. In order to work out the minimum possible error allowed by the information that is available, we introduce here the concept of the ideal classifier: the ideal classifier can classify the subjects by committing the minimum possible error. This becomes clear with an example. Let us assume that three subjects have identical representation when we use only 18 genes, although two of them belong to group A and one of them to group B. The ideal classifier will put all three to the same class, which will be identified with group A, so that only one out of the three subjects will be wrongly classified. A less than ideal classifier will place all three in the same class, which may be identified with group B (two of the subjects will be wrongly classified) or with group C (all three subjects will be wrongly classified). The ideal classifier, of course, is expected to classify with the minimum error subjects with identical data and with no error subjects with discriminating data.
The ideal classifier will have an 82.54% accuracy in the 18-dimensional space and a 62.96% accuracy in the 7-dimensional space. Note that these minimum possible errors may not be achievable in practice, as the non-ambiguous patterns may be very diverse and incoherent, even when the subjects belong to the same group, and thus, no real classifier will be able to group together subjects of the same group. The classifiers we shall propose in this paper will attempt to approach this minimum possible error.
3 Proposed methodology
Figure 5 shows schematically the essence of our idea: let us consider that each subject is represented by two features, and let us assume that we have two classes of data that have a strong overlap. If we try to classify a new pattern using these two features, any pattern that happens to fall in the overlapping area of the training patterns will be classified with the same accuracy as pure chance. Let us consider now the pairwise relationships in the training data, between patterns that belong to the same group. It is as if we replace the individual points in this feature space by a structure that connects the members of each group with each other. Let us then introduce the new pattern to the structure of each class. Obviously, the structure of each class will be perturbed when a new pattern has to be incorporated into it. We shall classify the new pattern to the class with the least perturbed structure.
To use this idea in practice, we have to decide two things:
-
(a)
how to construct features from a pattern that describes a subject and
-
(b)
how to describe the structure with the pairwise relations.
These two issues will be discussed in the subsections that follow.
3.1 Feature creation: binary code
As we stated earlier, the value of each component can be “0” or “1”. This characteristic suggested to us the idea of encoding the subjects with a binary number. Let us consider, e.g., one element P A,1 of set A:
Interpreting this pattern as a binary number, we have:
The order of “0” and “1” is not immunologically important. In fact, what is important is the positive or negative response against antigens and not the order in which we consider the antigens. Changing the antigen order we can codify the same subject in different ways. Indeed, we have as many as 18! = 6402373705728000 possible permutations.
To deal with smaller numbers, we decided to consider just the seven principal antigens. Table 3 lists the frequencies of occurrence of these seven antigens. The antigens are ordered in each row from the least frequent (L.F.) to the most frequent (M.F.) occurrence in each group and in all groups together.
Using this table, we may select convenient permutations that may allow us to distinguish the groups.
Let us consider, e.g., permutation {13, 1, 11, 10, 18, 14, 15}. We called it α-code. In this permutation, antigen 13 represents the most significant bit (MSB), while antigen 15 is the least significant bit (LSB). In this way, antigens {18, 14, 15} have little weight, because they do not permit us to separate the groups. The positions of antigens 13 and 1 (MSB) allow us to maximise the distance between groups C and {A, B}. To maximise the distance between groups B and C, we consider antigen 11 as the third significant bit. The fourth significant bit then has to be antigen 10.
Another possible choice is considering the global occurrence. Let us call β-code the antigen order given by the last row in Table 3: {11, 10, 1, 13, 18, 14, 15}.
Considering different permutations we can create mono- and multi-dimensional feature spaces where along each axis we represent a different code.
3.2 Pairwise structure: the relational space
The pairwise relations we consider are represented in the feature space by lines that join the corresponding points, i.e. edges that connect corresponding vertices (see Fig. 5). The relational space is a space where each edge is represented by a point, with coordinates that measure something which characterises the vertices it connects. For our data, we define this relational space as follows. We consider each group separately and encode subjects according to different codes. Let us consider the α-code and β-code and then create two vectors for each group containing the codes of the subjects. Let us call these vectors α-vector and β-vector. Before drawing the graph in the relational space, we sort the elements of the α- and β-vectors in increasing order and pair the ith element of the sorted α-vector with the ith element of the sorted β-vector. Then, we plot these new pairs in the relational space. Obviously, each point now does not represent a single person but a pair of subjects belonging to the same group.
Let us consider a simple example. Consider three subjects P 1, P 2 and P 3 who belong to group A with data shown in Table 4. The sorted α-vector of group A is (41, 67, 118) and the β-vector is (11, 62, 81) (see Table 5). In the relational space, we plot points (41, 11), (67, 62) and (118, 81), which correspond to subject pairs: (P α3 , P β1 ), (P α1 ,P β2 ) and (P α2 , P β3 ), respectively. Let us consider the meaning of pair (41, 11). It corresponds to patients (P 3, P 1). Patient P 3 is one who does not have the antigen that most discriminates between groups C and {A, B} activated. (That is why this patient’s α-code value is minimal.) Patient P 1 does not have the antigen that most discriminates between groups C and B also activated. So, these patients have something in common, and we may justify their pairing on this basis. Consider also the meaning of pair (118, 81). It corresponds to patients (P 2, P 3). Patient P 2 has activated the antigens that discriminate between groups C and {A, B}, while patient P 3 has activated the antigen that discriminates between groups C and B. Again, these patients have something in common, and it is reasonable to pair them.
3.3 Classifier
In the relational space, which we created in the previous section, every point represents a pair of subjects that are known to belong to the same group. Assume now that the data of a new subject arrive. We wish to classify this subject to one of the three groups. We propose the following algorithm.
We assign the test pattern to each group in turn, and re-classify the points of this group in the relational space to assess which group shows most improvement by the incorporation of the test pattern. We assign the test pattern to the group that shows the most improvement. To use this approach, we have to specify:
-
(a)
the classifier used to classify the points in the relational space;
-
(b)
what is meant by “showed most improvement”.
As a classifier we use the k nearest neighbour classifier (k-nn). In the experimental section, we shall show that this classifier produced better results than the support vector machine (SVM), the naive Bayes and the C4.5 tree classifier.
To assess which class assignment to the test pattern improves the coherence of the clusters in the relational space most, we use the following procedure. We use the leave-one-out method to classify all points in the relational space of the group we disturb. First we consider the group as it is, before we introduce the test pattern. We consider one point of this group (representing a pair of subjects) at a time and classify it using all remaining points. This way we work out what fraction of the group would be classified correctly when we have the training data only. Then we introduce the new pattern in this group, create the pairs of patients and classify them using all training patterns in the relational space, and, thus, identify the fraction of the points that are correctly classified now. We assign the test pattern to the group for which we observed the largest improvement in classification when we incorporated it into the group. In other words, we assign the test pattern to the group that becomes most coherent in its relational space representation when we incorporate the new pattern into it.
4 Experimental results
First, we have to check whether the results produced by straightforward k-means reported in Table 1 could be improved if we used a more sophisticated algorithm, like, e.g., a SVM. Then we examine whether the results improve if we use various combinations of features constructed using the binary code. Finally, we shall show that the best results are obtained by the proposed method that uses the relational feature space.
4.1 Classification of the data in the original space
Let us consider first the method of principal component analysis (PCA) and linear discriminant analysis (LDA). These techniques are usually employed to reduce multi-dimensional data sets to a lower dimensionality feature space where the classes may be better separated. In Fig. 6a, we project the 18D feature space into a 2D plane where the first and second principal components are considered, while in Fig. 6b we depict the LDA projection into a 2D space, where the classes are assumed to be maximally separable [16, 18]. We see that the groups are overlapped and it is hard to distinguish them clearly.
We present in Table 6 the classification results using SVMs [3, 4, 7, 9] and k-nn [1, 5, 17] classifiers in the 18-dimensional and 7-dimensional feature space using the leave-one-out method. Since our data are “0” or “1” we did not need to scale them.Footnote 4 We tried k = 1, 2, 3, 4, 5 and we report here only the results with k = 3 because these turned out to be the best. In Table 6, k is the number of nearest neighbours, C is the Tikhonov constant of the SVM problem and p is the parameter of the SVM with the RBF kernel. All parameters in this table were chosen according to the recommendations in [6]. The authors suggest to consider a “grid-search” on C and p using cross-validation, and to use exponentially growing sequences of C and p. That is what we did and the combination of parameter values with the best cross-validation accuracy was picked.
4.2 Classification of the data using their binary codes
Besides the α-code and β-code, we considered some other antigen permutations. Let us define the γ-code = {1, 10, 11, 13, 18, 14, 15} and the δ-code = {10, 11, 1, 13, 18, 14, 15}. In the γ-code, we use antigen 1 as the MSB to help us separate group C from groups A and B, while in the δ-code we consider antigens 10 and 11 as the most significant ones to separate group A from group B. Table 7 lists all the codes we considered.
Using these binary codes, we created new feature spaces and analysed them with the k-nn classifier with k = 3, testing the method with the leave-one-out cross-validation. We report the results of the various feature spaces we considered in Table 8.
In order to get an idea which class is confused with which, we give in Table 9 the confusion matrices of the classifications we obtained in three of these feature spaces.
4.3 Classification of the data in the relational space
Figure 7 shows the data plotted in the relational space of the α- and β-codes. Let us call this space (αr, βr). In this figure, we can distinguish easily all groups: they are overlapped just in the corner near zero where we have several subjects with all or most genes not expressed. In order to appreciate the importance of sorting the patients before pairing them, we show in Fig. 8 the relational space obtained by randomly pairing the patients and by pairing each patient with himself.
Table 10 reports the results of the k-nn classifier with the leave-one-out cross-validation method. The reported results correspond to the k = 3 classifier that produced the best results compared with classifiers with k = 1, 2, 4, 5. The confusion matrices of the best results, obtained in the (αr, βr) and the (δr, αr, βr) spaces,Footnote 5 are shown in Table 9.
The (αr, βr) space was also analysed by replacing the k-nn classifier with an SVM with different kernels. Table 11 reports these results.
We also used the naive Bayes classifier in the relational space, obtained from [12]. The naive Bayes classifier estimates a class-conditional distribution using a histogram of feature values and assuming feature independence. The number of histogram bins may be treated as a free parameter. The algorithm counts the number of training examples for each of the classes in each of the bins, and classifies the object to the class that gives maximum posterior probability. The prior probabilities in our case were set up according to:
Figure 9 reports the results of this classifier for different number of bins. In all cases, they are worse than those of the k-nn classifier.
Finally, we explored the C4.5 decision tree as a classifier in the original 18-dimensional and 7-dimensional spaces, and in the relational space [13, 14]. At each node of the tree, C4.5 chooses one attribute of the data that most effectively splits the set of samples into subsets enriched in one class or the other. Its criterion is the normalised information gain (difference in entropy) that results from choosing an attribute for splitting the data. The attribute with the highest normalised information gain is chosen to make the decision. In Table 12, we report the results of our analysis, where we have considered different values for the parameters “confidence-factor” and “trials”. We used a grid-search, considering account pairs of (“confidence-factor”, “trials”), with “confidence-factor” taking values 0.1, 0.2, …, 1.0, and “trials” being an integer between 1 and 10. For the other parameters, default values were used. The accuracies were computed with the leave-one-out technique. The best accuracy identified was lower than that of the 3-nn classifier.
The summary of the best results obtained by each classifier is shown in Table 13. When we assign the test pattern to each group in turn for the classification, we modify the placement of the points in the relational space. While the SVM, C4.5 and naive Bayes classifier are more vulnerable to this modification, the 3-nn method is more robust and produces higher accuracy.
4.4 A systematic way to create codes
Due to the large number of antigen possible permutations, one may device several codes and relational spaces. We estimated that the CPU time required to explore all possible permutations of 7 antigens is about 12 years! In this section, we propose the following algorithm that allows one to generate good codings and relational spaces.
-
1.
Consider the two most important antigens.
-
2.
Analyse all possible combinations.
-
3.
Select the relational space that gives the highest accuracy. If two different relational spaces have the same accuracy, explore all possible solutions.
-
4.
Add to the best relational space one antigen at every possible position with respect to the existing antigens, but without changing the relative order of the other antigens.
-
5.
Go to Step 3.
The algorithm finishes when the accuracy starts to decrease. The accuracy is computed using our classifier and leave-one-out testing.
The results of this algorithm are reported in Table 14.
Obviously, this algorithm does not allow us to find the best combination, but permits us to construct reasonably good relational spaces in finite computing time.
4.5 Computational complexity
Let us call \(\mathcal{C}^{\text{train}}\) the computational complexity for training a classifier and \(\mathcal{C}^{\text{class}}\) its computational complexity for classifying an unknown pattern. \(\mathcal{C}^{\text{train}}\) will be a function of the training patterns we have, \(\mathcal{N}_{\text{train}},\) and both will be functions of the dimensionality of the space we use d. When a new pattern has to be classified, we combine it with the training patterns we have and classify the pairs we create. If we have \(\mathcal{N}_{\text{class}}\) classes, we do that by assigning the new pattern to each class in turn. The overall complexity of the algorithm then is:
5 Discussion and conclusions
In this paper, we introduced the concept of the relational space, where instead of representing individual patterns, we represent pairs of patterns. In this space, a subject is classified not simply according to the data that describe him/her, but according to the relationship these data have with the data of other subjects. Working in the relational space allowed us to improve the classification accuracy of the malaria subjects by about 15% above the accuracy obtained by conventional methods.
We also introduced in this paper the concept of principal antigens, defined as those that have a statistically significant specificity for a group of subjects as well as the concept of the ideal classifier, used to indicate what the possible minimum error is with which the available data could be classified. Because of a large number of subjects characterised by the same pattern of nil expression of the selected antigens, this maximum classification accuracy was found to be 63% for the case of using the 7 principal antigens. The classification accuracy we obtained using the relational space was 58%, i.e. it approached this ideal accuracy much closer than any of the other methods considered (which had accuracies of the order of 40%). Although the ideal accuracy, when using all 18 antigens, is as high as 83%, this accuracy is not really attainable, as the extra genes that allow the differentiation when we are talking in idealistic terms do not actually have any specificity for any particular group of subjects. Such accuracies are obviously not high enough for a real practical application. However, the methodology improves significantly the original raw data classification accuracy. It is possible that the improvement in classification accuracy at levels appropriate for a practical system will come from the inclusion of further, more discriminatory raw measurements. In any case, our results demonstrate that the use of a relational space may improve on the classification results of individual feature spaces.
The methodology we presented in constructing the relational space is just an example. Alternative ways of defining relationships between individuals may be devised, which may prove to lead to better classifiers. However, the concepts we introduced here, namely that of relational space, binary coding for feature construction and principal antigen are generic and they may be used for analysing other data sets and developing different classification schemes.
Notes
The mean values μ = [μ1 μ2 … μ18]T, μA = [μ A1 μ A2 … μ A18 ]T, μB = [μ B1 μ B2 … μ B18 ]T and μC = [μ C1 μ C2 … μ C18 ]T are 18-dimensional vectors where each component is the mean value of the ith antigen.
Note that μ = [μ1 μ2 … μ18]T and p = [p 1 p 2 … p 18]T have the same values, in the same way as μ A i = p A i , μ B i = p B i and μ C i = p C i for i = 1, …, 18; we changed the notation here in order to underline the probabilistic meaning of these quantities.
For the normal distribution, N(μ, σ), about 68% of the values are within [μ − σ, μ + σ], about 95% of the values are within [μ − 2σ, μ + 2σ] and about 99.7% lie within [μ − 3σ, μ + 3σ].
Also note that for binary data, the Euclidean metric, the Hamming distance and the L1 norm coincide.
The 3D relational space (δr, αr, βr) was obtained in the same way as the (αr, βr) space, considering triplets of patients of the same group.
References
Alippi C, Roveri M (2007) Reducing computational complexity in k-NN based adaptive classifiers. In: Proceedings of the IEEE international conference on computational intelligence for measurement systems and applications (CIMSA 2007), Ostuni, Italy, 27–29 June 2007
Breman J (2001) The ears of the hippopotamus: manifestations, determinants, and estimates of the malaria burden. Am J Trop Med Hygiene 64(1–2 Suppl):1–11
Burges CJC (1998) A tutorial on support vector machines for pattern recognition. Data Min Knowl Discov 2(2):121–167
Boser BE, Guyon I, Vapnik V (1992) A training algorithm for optimal margin classifiers. In: Proceedings of the fifth annual workshop on computational learning theory. ACM Press, pp 144–152
Chakrabarti A, Chazelle B, Gum B, Lvov A (1999) A lower bound on the complexity of approximate nearest-neighbor searching on the hamming cube. In: Proceedings of the 31st annual ACM symposium on theory of computing (STOC99), pp 305–311
Chang CC, Lin CJ (2001) LIBSVM: a library for support vector machines. Software available at http://www.csie.ntu.edu.tw/cjlin/libsvm
Cortes C, Vapnik V (1995) Support-vector networks. Mach Learn 20(3):273–297
Gray JC, Corran PH, Mangia E, Gaunt MW, Li Q, Tetteh KKA, Polley SD, Conway DJ, Holder AA, Bacarese-Hamilton T, Riley EM, Crisanti A (2007) Profiling the antibody immune response against blood stage malaria vaccine candidates. Proteomics Protein Markers Clin Chem 53(7):1244–1253
Gunn SR (1998) Support vector machines for classification and regression. Technical Report. Faculty of Engineering, Science and Mathematics School of Electronics and Computer Science, University of Southampton
Hay S, Guerra C, Tatem A, Noor A, Snow R (2004) The global distribution and population at risk of malaria: past, present, and future. Lancet Infect Dis 4(6):327–336
Otsu N (1979) A threshold selection method from grey level histograms. IEEE Trans Syst Man Cybern 9:62–66
PRTools, version 4.0.14, 4 Mar 2005, http://prtools.org/
Quinlan JR (1993) C4.5: programs for machine learning. Morgan Kaufmann Publishers, San Francisco
Quinlan JR. C4.5 Release 8. http://www.rulequest.com/Personal/. Accessed 8 June 2011
Snow RW, Guerra CA, Noor AM, Myint HY, Hay SI (2005) The global distribution of clinical episodes of Plasmodium falciparum malaria. Nature 434(7030):214–217
Theodoridis S, Koutroumbas K (2009) Pattern recognition. Academic Press, London
Vaidya PM (1989) An O(n log n) algorithm for the all-nearest-neighbors. Problem Discrete Comput Geom 4:101–115
Welling M. Fisher linear discriminant analysis. Department of Computer Science, University of Toronto, Canada. http://www.ics.uci.edu/~welling/classnotes/papers_class/Fisher-LDA.pdf. Accessed 8 June 2011
WHO (2010) Guidelines for the treatment of malaria. ISBN: 9789241547925. http://www.who.int/malaria/publications/atoz/9789241547925/en/index.html. Accessed 8 June 2011
WHO (2008) World malaria report. ISBN: 9789241563697. WHO reference number: WHO/HTM/GMP/2008.1. http://www.who.int/malaria/publications/atoz/9789241563697/en/index.html. Accessed 8 June 2011
Acknowledgments
The authors are grateful to Julian Gray, Giorgio Mazzoleni and Andrea Crisanti of the Department of Biological Sciences of Imperial College London, for supplying the data. Paolo Pintus is grateful to Fabio Roli and Luca Citi for useful advise. This work was also partly supported by the EPSRC portfolio grant ‘‘Integrated Electronics’’.
Author information
Authors and Affiliations
Corresponding author
Additional information
This work was done when he was visiting Imperial College as a student of the Leonardo da Vinci programme, ‘‘Bioinformatics and Bionanotechnology (BIO-NANO)’’, European project coordinated by Consorzio TUCEP, Perugia, Italy.
Rights and permissions
About this article
Cite this article
Pintus, P., Petrou, M. Relational space classification for malaria diagnosis. Pattern Anal Applic 14, 261–272 (2011). https://doi.org/10.1007/s10044-011-0224-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10044-011-0224-z