Data mining is the process of finding interesting patterns in large quantities of data. While implementing clustering algorithms, it is important to be able to quantify the proximity of objects to one another. Proximity measures are mainly mathematical techniques that calculate the similarity/dissimilarity of data points. Usually, proximity is measured in terms of similarity or dissimilarity i.e., how alike objects are to one another.
Real-Life Example Use-case : Predicting COVID-19 patients on the basis of their symptoms
With the rise of COVID-19 cases, many people are not being able to seek proper medical advice due to the shortage of both human and infrastructure resources. As a result, we as engineers can contribute our bit to solve this problem by providing a basic diagnosis to help in identifying the people suffering from COVID-19. To help us we can make use of Machine Learning algorithms to ease out this task, among which clustering algorithms come in handy to use.
For this, we make two clusters based on the symptoms of the patients who are COVID-19 positive or negative and then predict whether a new incoming patient is suffering from COVID-19 or not by measuring the similarity/dissimilarity of the observed symptoms (features) with that of the infected person’s symptoms.
–Numerical measure of how alike two data objects are.
– Is higher when objects are more alike.
– Often falls in the range [0,1].
–Numerical measure of how different two data objects are.
– Lower when objects are more alike.
– Minimum dissimilarity is often 0.
– Upper limit varies.
Dissimilarity matrix is a matrix of pairwise dissimilarity among the data points. It is often desirable to keep only lower triangle or upper triangle of a dissimilarity matrix to reduce the space and time complexity.
1. It’s square and symmetric(AT= A for a square matrix A, where AT represents its transpose).
2. The diagonals members are zero, meaning that zero is the measure of dissimilarity between an element and itself.
Nominal attributes can have two or more different states e.g. an attribute ‘color’ can have values like ‘Red’, ‘Green’, ‘Yellow’, ‘Blue’, etc. Dissimilarity for nominal attributes is calculated as the ratio of total number of mismatches between two data points to the total number of attributes.
Nominal means “relating to names.” The values of a nominal attribute are
symbols or names of things. Each value represents some kind of category, code,
or state and so nominal attributes are also referred to as categorical.
Examples: ID numbers, eye color, zip codes.
Let M be the total number of states of a nominal attribute. Then the states can be numbered from 1 to M. However, the numbering does not denote any kind of ordering and can not be used for any mathematical operations.
Let m be total number of matches between two-point attributes and p be total number of attributes, then the dissimilarity can be calculated as,
d(i, j)=(p-m)/p
We can calculate similarity as,
s(i, j)=1-d(i, j)
EXAMPLE,
Roll No | Marks | Grades |
1 | 96 | A |
2 | 87 | B |
3 | 83 | B |
4 | 96 | A |
In this example we have four objects as Roll No from 1 to 4.
Now, we apply the formula(described above) for finding the proximity of nominal attributes:
– d(1,1)= (p-m)/p = (2-2)/2 = 0 – d(2,2)= (p-m)/p = (2-2)/2 = 0
– d(2,1)= (p-m)/p = (2-0)/2 = 1 – d(3,2)= (p-m)/p = (2-1)/2 = 0.5
– d(3,1)= (p-m)/p = (2-2)/2 = 1 – d(4,2)= (p-m)/p = (2-0)/2 = 1
– d(4,1)= (p-m)/p = (2-2)/2 = 0 – d(3,3)= (p-m)/p = (2-2)/2 = 0
– d(4,3)= (p-m)/p = (2-0)/2 = 1 – d(4,4)= (p-m)/p = (2-2)/2 = 0
– As seen from the calculation, we observe that the similarity between an object with itself is 1, which seems intuitively correct.
An ordinal attribute is an attribute whose possible values have a meaningful order or ranking among them, but the magnitude between successive values is not known. However, to do so, it is important to convert the states to numbers where each state of an ordinal attribute is assigned a number corresponding to the order of attribute values.
Examples: rankings (e.g., taste of potato chips on a scale from 1-10), grades, height {tall, medium, short}.
Since a number of states can be different for different ordinal attributes, it is therefore required to scale the values to a common range, e.g [0,1]. This can be done using the given formula,
zif=(rif−1)/(Mf−1)
where M is a maximum number assigned to states and r is the rank(numeric value) of a particular object.
The similarity can be calculated as:
s(i, j)=1-d(i, j)
EXAMPLE,
|
|
1 | High |
2 | Low |
3 | Medium |
4 | High |
In this example, we have four objects having ID from 1 to 4.
Here for encoding our attribute column, we consider High=1, Medium=2, and Low=3. And, the value of Mf=3(since there are three states available)
Now, we normalize the ranking in the range of 0 to 1 using the above formula.
So, High=(1-1)/(3-1)=0, Medium=(2-1)/(3-1)=0.5, Low=(3-1)/(3-1)=1.
Finally, we are able to calculate the dissimilarity based on difference in normalized values corresponding to that attribute.
– d(1,1)= 0-0 = 0 – d(2,2)= 3-3 = 0
– d(2,1)= 1-0= 1 – d(3,2)= 0.5-0 = 0.5
– d(3,1)= 0.5-0 = 0.5 – d(4,2)= 1-0 = 1
– d(4,1)= 0-0 =0 – d(3,3)= 0.5-0.5 = 0
– d(4,3)= 0.5-0=0 – d(4,4)= 0-0 = 0
Thanks for reading! 😊
This brings us to the end of our article on proximity measures for nominal and ordinal attributes. I hope you liked my article. Now, as far as proximity measures for binary and numeric attributes are concerned, Well, that’s another blog post for another time.
If you liked this and want to know more, go visit my other articles on Data Science and Machine Learning by clicking on the Link
Please feel free to contact me on Linkedin, Email.
Something not mentioned or want to share your thoughts? Feel free to comment below And I’ll get back to you. 🙂
Till then Stay Home, Stay Safe to prevent the spread of COVID-19 and Keep Learning!
Currently, I pursuing my Bachelor of Technology (B.Tech) in Computer Science and Engineering from Indian Institute of Technology Jodhpur(IITJ). I am very enthusiastic about Machine learning, Deep Learning, and Artificial Intelligence.
The media shown in this article are not owned by Analytics Vidhya and is used at the Author’s discretion.