MarketJaccard index
Company Profile

Jaccard index

The Jaccard index is a statistic used for gauging the similarity and diversity of sample sets. It is defined in general taking the ratio of two sizes, the intersection size divided by the union size, also called intersection over union (IoU).

Overview
The Jaccard index measures similarity between finite non-empty sample sets and is defined as the size of the intersection divided by the size of the union of the sample sets: : J(A, B) = \frac = \frac. The Jaccard index can be interpreted as a normalized measure of overlap between two sets, where the intersection represents shared elements and the union represents the total set of distinct elements. By definition, 0 \le J(A, B) \le 1. If the sets A and B have no elements in common, their intersection is empty, so |A \cap B|=0 and therefore J(A,B)=0. The other extreme is that the two sets are equal. In that case A\cap B=A\cup B=A=B, so then J(A,B)=1. The Jaccard index is widely used in computer science, ecology, genomics and other sciences where binary or binarized data are used. Both the exact solution and approximation methods are available for hypothesis testing with the Jaccard index. Both the exact solution and approximation methods are available for hypothesis testing with the Jaccard index. Jaccard similarity also applies to bags, i.e., multisets. This has a similar formula, but the symbols used represent bag intersection and bag sum (not union). The maximum value is 1/2. : J(A, B) = \frac = \frac. The Jaccard distance, which measures dissimilarity between sample sets, is complementary to the Jaccard index and is obtained by subtracting the Jaccard index from 1 or, equivalently, by dividing the difference of the sizes of the union and the intersection of two sets by the size of the union: : d_J(A, B) = 1 - J(A, B) = \frac. An alternative interpretation of the Jaccard distance is as the ratio of the size of the symmetric difference A \mathbin\triangle B = (A \cup B) - (A \cap B) to the union. Jaccard distance is commonly used to calculate an matrix for clustering and multidimensional scaling of n sample sets. These distance measures are commonly used in cluster analysis for grouping similar observations. This distance is a metric on the collection of all finite sets. There is also a version of the Jaccard distance for measures, including probability measures. If \mu is a measure on a measurable space X, then we define the Jaccard index by : J_\mu(A, B) = \frac{\mu(A \cap B)}{\mu(A \cup B)}, and the Jaccard distance by : d_\mu(A, B) = 1 - J_\mu(A,B) = \frac{\mu(A \mathbin\triangle B)}{\mu(A \cup B)}. The definition is not well-defined when \mu(A \cup B) = 0 or \mu(A \cup B) = \infty. The MinHash min-wise independent permutations locality sensitive hashing scheme may be used to efficiently compute an accurate estimate of the Jaccard similarity index of pairs of sets, where each set is represented by a constant-sized signature derived from the minimum values of a hash function. The Jaccard index is particularly useful for analyzing large-scale and sparse datasets in modern data mining applications. == Similarity of asymmetric binary attributes ==
Similarity of asymmetric binary attributes
Given two objects, A and B, each with n binary attributes, the Jaccard index is a useful measure of the overlap that A and B share with their attributes. Each attribute of A and B can either be 0 or 1. The total number of each combination of attributes for both A and B are specified as follows: :M_{11} represents the total number of attributes where A and B both have a value of 1. :M_{01} represents the total number of attributes where the attribute of A is 0 and the attribute of B is 1. :M_{10} represents the total number of attributes where the attribute of A is 1 and the attribute of B is 0. :M_{00} represents the total number of attributes where A and B both have a value of 0. Each attribute must fall into one of these four categories, meaning that :M_{11} + M_{01} + M_{10} + M_{00} = n. The Jaccard similarity index, J, is given as :J = {M_{11} \over M_{01} + M_{10} + M_{11}}. The Jaccard distance, dJ, is given as :d_J = {M_{01} + M_{10} \over M_{01} + M_{10} + M_{11}} = 1 - J. Statistical inference can be made based on the Jaccard similarity index, and consequently related metrics. The distinction between similarity measures that include or exclude joint absences has been widely discussed in numerical taxonomy and classification literature. The Jaccard index is defined as J(A,B)=\frac{M_{11}}{M_{11}+M_{10}+M_{01}} where M_{11} is the number of attributes where both values are 1, M_{10} where only A is 1, and M_{01} where only B is 1. == Weighted Jaccard similarity and distance ==
Weighted Jaccard similarity and distance
If \mathbf{x} = (x_1, x_2, \ldots, x_n) and \mathbf{y} = (y_1, y_2, \ldots, y_n) are two vectors with all real x_i, y_i \geq 0, then their Jaccard similarity index (also known then as Ruzicka similarity) is defined as :J_\mathcal{W}(\mathbf{x}, \mathbf{y}) = \frac{\sum_i \min(x_i, y_i)}{\sum_i \max(x_i, y_i)}, and Jaccard distance (also known then as Soergel distance) :d_{J\mathcal{W}}(\mathbf{x}, \mathbf{y}) = 1 - J_\mathcal{W}(\mathbf{x}, \mathbf{y}). With even more generality, if f and g are two non-negative measurable functions on a measurable space X with measure \mu, then we can define :J_\mathcal{W}(f, g) = \frac{\int\min(f, g) d\mu}{\int \max(f, g) d\mu}, where \max and \min are pointwise operators. Then Jaccard distance is :d_{J\mathcal{W}}(f, g) = 1 - J_\mathcal{W}(f, g). Then, for example, for two measurable sets A, B \subseteq X, we have J_\mu(A,B) = J(\chi_A, \chi_B), where \chi_A and \chi_B are the characteristic functions of the corresponding set. Calculating the weighted Jaccard similarity index for two vectors typically requires a single pass over the data, resulting in a linear computational complexity of O(n), where n represents the number of dimensions. In data science applications, this is often optimized for sparse vectors by iterating only over non-zero elements, which significantly reduces processing time for high-dimensional datasets. This optimization shifts the complexity to O(k), where k is the number of non-zero attributes.# Psuedocode for O(k) complexity function jaccardIndex(vector_A, vector_B): intersection_sum = 0 union_sum = 0 # Get total unique keys all_keys = unique_keys(vector_A.keys() + vector_B.keys()) for key in all_keys: val_a = vector_A.get(key, 0) val_b = vector_B.get(key, 0) intersection_sum += min(val_a, val_b) union_sum += max(val_a, val_b) return intersection_sum / union_sum To further scale processing time, techniques such as MinHashing and locality sensitive hashing are used to approximate the index using compact signatures. These algorithmic enhancements ensure that similarity measures remain scalable and efficient, despite an increase in data volume and dimensionality. == Probability Jaccard similarity and distance ==
Probability Jaccard similarity and distance
The weighted Jaccard similarity described above generalizes the Jaccard Index to positive vectors, where a set corresponds to a binary vector given by the indicator function, i.e. x_i \in \{0,1\}. However, it does not generalize the Jaccard Index to probability distributions, where a set corresponds to a uniform probability distribution, i.e. :x_i = \begin{cases} \frac{1} & i \in X \\ 0 & \text{otherwise} \end{cases} It is always less if the sets differ in size. If |X| > |Y|, and x_i = \mathbf{1}_X(i)/|X|, y_i = \mathbf{1}_Y(i)/|Y| then :J_\mathcal{W}(x,y) = \frac Instead, a generalization that is continuous between probability distributions and their corresponding support sets is :J_\mathcal{P}(x,y) = \sum_{x_i\neq 0, y_i \neq 0} \frac{1}{\sum_{j} \max\left(\frac{x_j}{x_i}, \frac{y_j}{y_i}\right)} which is called the "Probability" Jaccard. It has the following bounds against the Weighted Jaccard on probability vectors. :J_\mathcal{W}(x,y) \leq J_\mathcal{P}(x,y) \leq \frac{2J_\mathcal{W}(x,y)}{1+J_\mathcal{W}(x,y)} Here the upper bound is the (weighted) Sørensen–Dice coefficient. The corresponding distance, 1 - J_\mathcal{P}(x,y), is a metric over probability distributions, and a pseudo-metric over non-negative vectors. The Probability Jaccard Index has a geometric interpretation as the area of an intersection of simplices. Every point on a unit k-simplex corresponds to a probability distribution on k+1 elements, because the unit k-simplex is the set of points in k+1 dimensions that sum to 1. To derive the Probability Jaccard Index geometrically, represent a probability distribution as the unit simplex divided into sub simplices according to the mass of each item. If you overlay two distributions represented in this way on top of each other, and intersect the simplices corresponding to each item, the area that remains is equal to the Probability Jaccard Index of the distributions. Optimality of the Probability Jaccard Index Consider the problem of constructing random variables such that they collide with each other as much as possible. That is, if X\sim x and Y\sim y, we would like to construct X and Y to maximize \Pr[X=Y]. If we look at just two distributions x,y in isolation, the highest \Pr[X=Y] we can achieve is given by 1 - \text{TV}(x,y) where \text{TV} is the Total Variation distance. However, suppose we weren't just concerned with maximizing that particular pair, suppose we would like to maximize the collision probability of any arbitrary pair. One could construct an infinite number of random variables one for each distribution x, and seek to maximize \Pr[X=Y] for all pairs x,y. In a fairly strong sense described below, the Probability Jaccard Index is an optimal way to align these random variables. For any sampling method G and discrete distributions x,y, if \Pr[G(x) = G(y)] > J_\mathcal{P}(x,y) then for some z where J_\mathcal{P}(x,z)>J_\mathcal{P}(x,y) and J_\mathcal{P}(y,z)>J_\mathcal{P}(x,y), either \Pr[G(x) = G(z)] or \Pr[G(y) = G(z)] . That is, no sampling method can achieve more collisions than J_\mathcal{P} on one pair without achieving fewer collisions than J_\mathcal{P} on another pair, where the reduced pair is more similar under J_\mathcal{P} than the increased pair. This theorem is true for the Jaccard Index of sets (if interpreted as uniform distributions) and the probability Jaccard, but not of the weighted Jaccard. (The theorem uses the word "sampling method" to describe a joint distribution over all distributions on a space, because it derives from the use of weighted minhashing algorithms that achieve this as their collision probability.) This theorem has a visual proof on three element distributions using the simplex representation. == Tanimoto similarity and distance ==
Tanimoto similarity and distance
Various forms of functions described as Tanimoto similarity and Tanimoto distance occur in the literature and on the Internet. Most of these are synonyms for Jaccard similarity and Jaccard distance, but some are mathematically different. Many sources cite an IBM Technical Report as the seminal reference. In "A Computer Program for Classifying Plants", published in October 1960, a method of classification based on a similarity ratio, and a derived distance function, is given. It seems that this is the most authoritative source for the meaning of the terms "Tanimoto similarity" and "Tanimoto Distance". The similarity ratio is equivalent to Jaccard similarity, but the distance function is not the same as Jaccard distance. Tanimoto's definitions of similarity and distance In that paper, a "similarity ratio" is given over bitmaps, where each bit of a fixed-size array represents the presence or absence of a characteristic in the plant being modelled. The definition of the ratio is the number of common bits, divided by the number of bits set (i.e. nonzero) in either sample. Presented in mathematical terms, if samples X and Y are bitmaps, X_i is the ith bit of X, and \land , \lor are bitwise and, or operators respectively, then the similarity ratio T_s is : T_s(X,Y) = \frac{\sum_i ( X_i \land Y_i)}{\sum_i ( X_i \lor Y_i)} If each sample is modelled instead as a set of attributes, this value is equal to the Jaccard index of the two sets. Jaccard is not cited in the paper, and it seems likely that the authors were not aware of it. Tanimoto goes on to define a "distance" based on this ratio, defined for bitmaps with non-zero similarity: : T_d(X,Y) = -\log_2 ( T_s(X,Y) ) This coefficient is, deliberately, not a distance metric. It is chosen to allow the possibility of two specimens, which are quite different from each other, to both be similar to a third. It is easy to construct an example which disproves the property of triangle inequality. Other definitions of Tanimoto distance Tanimoto distance is often referred to as a synonym for Jaccard distance 1-T_s. This function is a proper distance metric. In application, Tanimoto distance can be harmfully confused with Jaccard distance as a proper distance metric. If Jaccard or Tanimoto similarity is expressed over a bit vector, then it can be written as : f(A,B) =\frac{ A \cdot B}{\|A\|^2 +\|B\|^2 - A \cdot B} where the same calculation is expressed in terms of vector scalar product and magnitude. This representation relies on the fact that, for a bit vector (where the value of each dimension is either 0 or 1) then :A \cdot B = \sum_i A_iB_i = \sum_i ( A_i \land B_i) and :\|A\|^2 = \sum_i A_i^2 = \sum_i A_i. This is a potentially confusing representation, because the function as expressed over vectors is more general, unless its domain is explicitly restricted. Properties of T_s do not necessarily extend to f. In particular, the difference function 1-f does not preserve triangle inequality, and is not therefore a proper distance metric, whereas 1 - T_s is. There is a real danger that the combination of "Tanimoto Distance" being defined using this formula, along with the statement "Tanimoto Distance is a proper distance metric" will lead to the false conclusion that the function 1-f is in fact a distance metric over vectors or multisets in general, whereas its use in similarity search or clustering algorithms may fail to produce correct results. Lipkus uses a definition of Tanimoto similarity which is equivalent to f, and refers to Tanimoto distance as the function 1-f. It is, however, made clear within the paper that the context is restricted by the use of a (positive) weighting vector W such that, for any vector A being considered, A_i \in \{0,W_i\}. Under these circumstances, the function is a proper distance metric, and so a set of vectors governed by such a weighting vector forms a metric space under this function. == Jaccard index in binary classification confusion matrices ==
Jaccard index in binary classification confusion matrices
In confusion matrices employed for binary classification, the Jaccard index can be framed in the following formula: :\text{Jaccard index} = \frac{TP}{TP + FP + FN} where TP are the true positives, FP the false positives and FN the false negatives. The binary classification formula of the Jaccard Index measures how the predicted positive results match with the actual positive results. It shows how much overlap of what a model predicts are true positive cases to real positive cases. When the Jaccard index is higher it means that the prediction is closer to the actual result. The Jaccard index is an overlap metric that can be defined with the elements in a confusion matrix. The true positives represent the cases that are predicted positive and are actually positive, creating an overlap of the two groups. The false positives are cases that were predicted positive but are actually not. Similarly, false negatives are cases that were predicted negative but in reality are actually positive. The Jaccard index compares the true positives to the set of relevant information. Jaccard index is useful when trying to compare predicted results against actual results. Using binary classification alongside with Jaccard index helps explain how reliable a prediction model is. It is used in many fields, and commonly used in medicine for lab testing.{{Cite journal == Application to computer science and graph theory ==
Application to computer science and graph theory
In computer science, the Jaccard Index can be used to measure similarity between vertices in a graph by comparing their adjacency sets. Given two vertices, their similarity is calculated by taking the size of the intersection of their neighborhoods divided by the size of their union. This measure is widely applied in Link prediction, community detection, and graph classification, where it serves as a guide to estimate the likelihood of an edge forming between two nodes in networks. The computation of the Jaccard similarity between two vertices can be expressed using their adjacency sets, as shown below. // Javascript Code: function jaccardSimilarity(graph1, graph2){ const numNodes1 = graph1.length; const numNodes2 = graph2.length; let similarity = 0; // Iterate over all pairs of nodes in both graphs for(let i = 0; i b.includes(value)); } // Helper function to compute the union of two arrays function union(a, b){ return [...new Set([...a, ...b])]; } Above shows an implementation based on set operations over adjacency lists. In practice, graph representations like adjacency lists are used to improve the efficiency of intersection and union math. For large graphs, computing similarity across all pairs of vertices may be computationally expensive. Programmers may use approximations to avoid taking on the entire cost of this computation. == See also ==
tickerdossier.comtickerdossier.substack.com