HOME

TheInfoList



OR:

In
applied mathematics Applied mathematics is the application of mathematical methods by different fields such as physics, engineering, medicine, biology, finance, business, computer science, and industry. Thus, applied mathematics is a combination of mathemati ...
, topological based data analysis (TDA) is an approach to the analysis of datasets using techniques from
topology In mathematics, topology (from the Greek words , and ) is concerned with the properties of a geometric object that are preserved under continuous deformations, such as stretching, twisting, crumpling, and bending; that is, without closing ...
. Extraction of information from datasets that are high-dimensional, incomplete and noisy is generally challenging. TDA provides a general framework to analyze such data in a manner that is insensitive to the particular
metric Metric or metrical may refer to: * Metric system, an internationally adopted decimal system of measurement * An adjective indicating relation to measurement in general, or a noun describing a specific type of measurement Mathematics In mathe ...
chosen and provides
dimensionality reduction Dimensionality reduction, or dimension reduction, is the transformation of data from a high-dimensional space into a low-dimensional space so that the low-dimensional representation retains some meaningful properties of the original data, ideally ...
and robustness to noise. Beyond this, it inherits
functor In mathematics, specifically category theory, a functor is a mapping between categories. Functors were first considered in algebraic topology, where algebraic objects (such as the fundamental group) are associated to topological spaces, and m ...
iality, a fundamental concept of modern mathematics, from its topological nature, which allows it to adapt to new mathematical tools. The initial motivation is to study the shape of data. TDA has combined
algebraic topology Algebraic topology is a branch of mathematics that uses tools from abstract algebra to study topological spaces. The basic goal is to find algebraic invariants that classify topological spaces up to homeomorphism, though usually most classify ...
and other tools from pure mathematics to allow mathematically rigorous study of "shape". The main tool is persistent homology, an adaptation of homology to
point cloud Point or points may refer to: Places * Point, Lewis, a peninsula in the Outer Hebrides, Scotland * Point, Texas, a city in Rains County, Texas, United States * Point, the NE tip and a ferry terminal of Lismore, Inner Hebrides, Scotland * Poin ...
data. Persistent homology has been applied to many types of data across many fields. Moreover, its mathematical foundation is also of theoretical importance. The unique features of TDA make it a promising bridge between topology and geometry.


Basic theory


Intuition

TDA is premised on the idea that the shape of data sets contains relevant information. Real high-dimensional data is typically sparse, and tends to have relevant low dimensional features. One task of TDA is to provide a precise characterization of this fact. For example, the trajectory of a simple predator-prey system governed by the
Lotka–Volterra equations The Lotka–Volterra equations, also known as the predator–prey equations, are a pair of first-order nonlinear differential equations, frequently used to describe the dynamics of biological systems in which two species interact, one as a pre ...
forms a closed circle in state space. TDA provides tools to detect and quantify such recurrent motion. Many algorithms for data analysis, including those used in TDA, require setting various parameters. Without prior domain knowledge, the correct collection of parameters for a data set is difficult to choose. The main insight of persistent homology is to use the information obtained from all parameter values by encoding this huge amount of information into an understandable and easy-to-represent form. With TDA, there is a mathematical interpretation when the information is a
homology group In mathematics, homology is a general way of associating a sequence of algebraic objects, such as abelian groups or modules, with other mathematical objects such as topological spaces. Homology groups were originally defined in algebraic topolog ...
. In general, the assumption is that features that persist for a wide range of parameters are "true" features. Features persisting for only a narrow range of parameters are presumed to be noise, although the theoretical justification for this is unclear.


Early history

Precursors to the full concept of persistent homology appeared gradually over time. In 1990, Patrizio Frosini introduced the size function, which is equivalent to the 0th persistent homology. Nearly a decade later,
Vanessa Robins Vanessa Robins is an Australian applied mathematician whose research interests include computational topology, image processing, and the structure of granular materials. She is a fellow in the departments of applied mathematics and theoretical phy ...
studied the images of homomorphisms induced by inclusion. Finally, shortly thereafter, Edelsbrunner et al. introduced the concept of persistent homology together with an efficient algorithm and its visualization as a persistence diagram. Carlsson et al. reformulated the initial definition and gave an equivalent visualization method called persistence barcodes, interpreting persistence in the language of commutative algebra. In algebraic topology the persistent homology has emerged through the work of Sergey Barannikov on Morse theory. The set of critical values of smooth Morse function was canonically partitioned into pairs "birth-death", filtered complexes were classified, their invariants, equivalent to persistence diagram and persistence barcodes, together with the efficient algorithm for their calculation, were described under the name of canonical forms in 1994 by Barannikov.


Concepts

Some widely used concepts are introduced below. Note that some definitions may vary from author to author. A
point cloud Point or points may refer to: Places * Point, Lewis, a peninsula in the Outer Hebrides, Scotland * Point, Texas, a city in Rains County, Texas, United States * Point, the NE tip and a ferry terminal of Lismore, Inner Hebrides, Scotland * Poin ...
is often defined as a finite set of points in some Euclidean space, but may be taken to be any finite metric space. The Čech complex of a point cloud is the ''nerve'' of the ''cover'' of balls of a fixed radius around each point in the cloud. A persistence module \mathbb indexed by \mathbb is a vector space U_t for each t \in \mathbb, and a linear map u_t^s: U_s \to U_t whenever s \leq t, such that u_t^t=1 for all t and u_t^su_s^r=u_t^r whenever r \leq s \leq t. An equivalent definition is a functor from \mathbb considered as a partially ordered set to the category of vector spaces. The persistent homology group PH of a point cloud is the persistence module defined as PH_k(X)= \prod H_k(X_r), where X_r is the Čech complex of radius r of the point cloud X and H_k is the homology group. A persistence barcode is a
multiset In mathematics, a multiset (or bag, or mset) is a modification of the concept of a set that, unlike a set, allows for multiple instances for each of its elements. The number of instances given for each element is called the multiplicity of that e ...
of intervals in \mathbb, and a persistence diagram is a multiset of points in \Delta(:= \). The Wasserstein distance between two persistence diagrams X and Y is defined as W_p _qX,Y):= \inf_ \left \sum_ (\Vert x-\varphi(x)\Vert _q)^p\rightwhere 1 \leq p,q \leq \infty and \varphi ranges over bijections between X and Y. Please refer to figure 3.1 in Munch for illustration. The bottleneck distance between X and Y is W_ _qX,Y):= \inf_ \sup_ \Vert x-\varphi(x)\Vert _q. This is a special case of Wasserstein distance, letting p=\infty.


Basic property


Structure theorem

The first classification theorem for persistent homology appeared in 1994 via Barannikov's canonical forms. The classification theorem interpreting persistence in the language of commutative algebra appeared in 2005: for a finitely generated persistence module C with field F coefficients, H(C;F) \simeq \bigoplus_i x^ \cdot F \oplus \left(\bigoplus_j x^ \cdot (F (x^\cdot F )\right). Intuitively, the free parts correspond to the homology generators that appear at filtration level t_i and never disappear, while the torsion parts correspond to those that appear at filtration level r_j and last for s_j steps of the filtration (or equivalently, disappear at filtration level s_j+r_j). Persistent homology is visualized through a barcode or persistence diagram. The barcode has its root in abstract mathematics. Namely, the category of finite filtered complexes over a field is semi-simple. Any filtered complex is isomorphic to its canonical form, a direct sum of one- and two-dimensional simple filtered complexes.


Stability

Stability is desirable because it provides robustness against noise. If X is any space which is homeomorphic to a simplicial complex, and f,g:X\to \mathbb are continuous tame functions, then the persistence vector spaces \ and \ are finitely presented, and W_(D(f),D(g)) \leq \lVert f-g \rVert_, where W_ refers to the bottleneck distance and D is the map taking a continuous tame function to the persistence diagram of its k-th homology.


Workflow

The basic workflow in TDA is: # If X is a point cloud, replace X with a nested family of
simplicial complex In mathematics, a simplicial complex is a set composed of points, line segments, triangles, and their ''n''-dimensional counterparts (see illustration). Simplicial complexes should not be confused with the more abstract notion of a simplicial ...
es X_ (such as the Čech or Vietoris-Rips complex). This process converts the point cloud into a filtration of simplicial complexes. Taking the homology of each complex in this filtration gives a persistence module H_i(X_)\to H_i(X_) \to H_i(X_) \to \cdots # Apply the structure theorem to provide a parameterized version of
Betti number In algebraic topology, the Betti numbers are used to distinguish topological spaces based on the connectivity of ''n''-dimensional simplicial complexes. For the most reasonable finite-dimensional spaces (such as compact manifolds, finite simplici ...
, persistence diagram, or equivalently, barcode. Graphically speaking,


Computation

The first algorithm over all fields for persistent homology in algebraic topology setting was described by Barannikov through reduction to the canonical form by upper-triangular matrices. The first algorithm for persistent homology over F_2 was given by Edelsbrunner et al. Zomorodian and Carlsson gave the first practical algorithm to compute persistent homology over all fields. Edelsbrunner and Harer's book gives general guidance on computational topology. One issue that arises in computation is the choice of complex. The Čech complex and
Vietoris–Rips complex In topology, the Vietoris–Rips complex, also called the Vietoris complex or Rips complex, is a way of forming a topological space from distances in a set of points. It is an abstract simplicial complex that can be defined from any metric space ...
are most natural at first glance; however, their size grows rapidly with the number of data points. The Vietoris–Rips complex is preferred over Čech complex because its definition is simpler and the Čech complex requires extra effort to define in a general finite metric space. Efficient ways to lower the computational cost of homology have been studied. For example, the α-complex and witness complex are used to reduce the dimension and size of complexes. Recently,
Discrete Morse theory Discrete Morse theory is a combinatorial adaptation of Morse theory developed by Robin Forman. The theory has various practical applications in diverse fields of applied mathematics and computer science, such as configuration spaces, homology com ...
has shown promise for computational homology because it can reduce a given simplicial complex to a much smaller cellular complex which is homotopic to the original one. This reduction can in fact be performed as the complex is constructed by using
matroid theory In combinatorics, a branch of mathematics, a matroid is a structure that abstracts and generalizes the notion of linear independence in vector spaces. There are many equivalent ways to define a matroid axiomatically, the most significant being ...
, leading to further performance increases. Another recent algorithm saves time by ignoring the homology classes with low persistence. Various software packages are available, such a
javaPlexDionysusPHATDIPHAGUDHIRipser
an
TDAstats
A comparison between these tools is done by Otter et al
Giotto-tda
is a Python package dedicated to integrating TDA in the machine learning workflow by means of a
scikit-learn scikit-learn (formerly scikits.learn and also known as sklearn) is a free software machine learning library for the Python programming language. It features various classification, regression and clustering algorithms including support-vector ...
br>
API. An R packag
TDA
is capable of calculating recently invented concepts like landscape and the kernel distance estimator. Th
Topology ToolKit
is specialized for continuous data defined on manifolds of low dimension (1, 2 or 3), as typically found in scientific visualization. Another R package
TDAstats
implements the Ripser library to calculate persistent homology.


Visualization

High-dimensional data is impossible to visualize directly. Many methods have been invented to extract a low-dimensional structure from the data set, such as
principal component analysis Principal component analysis (PCA) is a popular technique for analyzing large datasets containing a high number of dimensions/features per observation, increasing the interpretability of data while preserving the maximum amount of information, and ...
and
multidimensional scaling Multidimensional scaling (MDS) is a means of visualizing the level of similarity of individual cases of a dataset. MDS is used to translate "information about the pairwise 'distances' among a set of n objects or individuals" into a configurati ...
. However, it is important to note that the problem itself is ill-posed, since many different topological features can be found in the same data set. Thus, the study of visualization of high-dimensional spaces is of central importance to TDA, although it does not necessarily involve the use of persistent homology. However, recent attempts have been made to use persistent homology in data visualization. Carlsson et al. have proposed a general method called MAPPER. It inherits the idea of Serre that a covering preserves homotopy. A generalized formulation of MAPPER is as follows: Let X and Z be topological spaces and let f:X\to Z be a continuous map. Let \mathbb = \_ be a finite open covering of Z. The output of MAPPER is the nerve of the pullback cover M(\mathbb,f):=N(f^(\mathbb)), where each preimage is split into its connected components. This is a very general concept, of which the Reeb graph and merge trees are special cases. This is not quite the original definition. Carlsson et al. choose Z to be \mathbb or \mathbb^2, and cover it with open sets such that at most two intersect. This restriction means that the output is in the form of a
complex network In the context of network theory, a complex network is a graph (network) with non-trivial topological features—features that do not occur in simple networks such as lattices or random graphs but often occur in networks representing real ...
. Because the topology of a finite point cloud is trivial, clustering methods (such as single linkage) are used to produce the analogue of connected sets in the preimage f^(U) when MAPPER is applied to actual data. Mathematically speaking, MAPPER is a variation of the
Reeb graph A Reeb graphY. Shinagawa, T.L. Kunii, and Y.L. Kergosien, 1991. Surface coding based on Morse theory. IEEE Computer Graphics and Applications, 11(5), pp.66-78 (named after Georges Reeb by René Thom) is a mathematical object reflecting the evolutio ...
. If the M(\mathbb,f) is at most one dimensional, then for each i \geq 0, H_i(X) \simeq H_0(N(\mathbb);\hat_i) \oplus H_1(N(\mathbb);\hat_). The added flexibility also has disadvantages. One problem is instability, in that some change of the choice of the cover can lead to major change of the output of the algorithm. Work has been done to overcome this problem. Three successful applications of MAPPER can be found in Carlsson et al. A comment on the applications in this paper by J. Curry is that "a common feature of interest in applications is the presence of flares or tendrils." A free implementation of MAPPER is availabl
online
written by Daniel Müllner and Aravindakshan Babu. MAPPER also forms the basis of Ayasdi's AI platform.


Multidimensional persistence

Multidimensional persistence is important to TDA. The concept arises in both theory and practice. The first investigation of multidimensional persistence was early in the development of TDA,. Carlsson-Zomorodian introduced the theory of multidimensional persistence in and in collaboration with Singh introduced the use of tools from symbolic algebra (Grobner basis methods) to compute MPH modules. Their definition presents multidimensional persistence with n parameters as a \mathbb^n graded module over a polynomial ring in n variables. Tools from commutative and homological algebra are applied to the study of multidimensional persistence in work of Harrington-Otter-Schenck-Tillman. The first application to appear in the literature is a method for shape comparison, similar to the invention of TDA. The definition of an ''n''-dimensional persistence module in \mathbb^n is * vector space V_s is assigned to each point in s=(s_1,...,s_n) * map \rho_s^t:V_s \to V_t is assigned if s\leq t(s_i \leq t_i, i=1,...,n) * maps satisfy \rho_r^t=\rho_s^t \circ \rho_r^s for all r \leq s \leq t It might be worth noting that there are controversies on the definition of multidimensional persistence. One of the advantages of one-dimensional persistence is its representability by a diagram or barcode. However, discrete complete invariants of multidimensional persistence modules do not exist. The main reason for this is that the structure of the collection of indecomposables is extremely complicated by Gabriel's theorem in the theory of quiver representations, although a finitely generated n-dim persistence module can be uniquely decomposed into a direct sum of indecomposables due to the Krull-Schmidt theorem. Nonetheless, many results have been established. Carlsson and Zomorodian introduced the rank invariant \rho_M(u,v), defined as the \rho_M(u,v)=\mathrm(x^:M_u\to M_v), in which M is a finitely generated n-graded module. In one dimension, it is equivalent to the barcode. In the literature, the rank invariant is often referred as the persistent Betti numbers (PBNs). In many theoretical works, authors have used a more restricted definition, an analogue from sublevel set persistence. Specifically, the persistence Betti numbers of a function f:X\to \mathbb^k are given by the function \beta_f: \Delta^ \to \mathrm, taking each (u,v) \in \Delta^ to \beta_f(u,v):= \mathrm (H(X(f\leq u)\to H(X(f\leq v))), where \Delta^ := \ and X(f\leq u):=\. Some basic properties include monotonicity and diagonal jump. Persistent Betti numbers will be finite if X is a compact and locally contractible subspace of \mathbb^n. Using a foliation method, the k-dim PBNs can be decomposed into a family of 1-dim PBNs by dimensionality deduction. This method has also led to a proof that multi-dim PBNs are stable. The discontinuities of PBNs only occur at points (u,v) (u\leq v) where either u is a discontinuous point of \rho_M (\star,v) or v is a discontinuous point of \rho (u,\star) under the assumption that f\in C^0(X,\mathbb^k) and X is a compact, triangulable topological space. Persistent space, a generalization of persistent diagram, is defined as the multiset of all points with multiplicity larger than 0 and the diagonal. It provides a stable and complete representation of PBNs. An ongoing work by Carlsson et al. is trying to give geometric interpretation of persistent homology, which might provide insights on how to combine machine learning theory with topological data analysis. The first practical algorithm to compute multidimensional persistence was invented very early. After then, many other algorithms have been proposed, based on such concepts as discrete morse theory and finite sample estimating.


Other persistences

The standard paradigm in TDA is often referred as sublevel persistence. Apart from multidimensional persistence, many works have been done to extend this special case.


Zigzag persistence

The nonzero maps in persistence module are restricted by the preorder relationship in the category. However, mathematicians have found that the unanimity of direction is not essential to many results. "The philosophical point is that the decomposition theory of graph representations is somewhat independent of the orientation of the graph edges". Zigzag persistence is important to the theoretical side. The examples given in Carlsson's review paper to illustrate the importance of functorality all share some of its features.


Extended persistence and levelset persistence

Some attempts is to lose the stricter restriction of the function. Please refer to the Categorification and cosheaves and Impact on mathematics sections for more information. It's natural to extend persistence homology to other basic concepts in algebraic topology, such as cohomology and relative homology/cohomology. An interesting application is the computation of circular coordinates for a data set via the first persistent cohomology group.


Circular persistence

Normal persistence homology studies real-valued functions. The circle-valued map might be useful, "persistence theory for circle-valued maps promises to play the role for some vector fields as does the standard persistence theory for scalar fields", as commented in Dan Burghelea et al. The main difference is that Jordan cells (very similar in format to the
Jordan block In the mathematical discipline of matrix theory, a Jordan matrix, named after Camille Jordan, is a block diagonal matrix over a ring (whose identities are the zero 0 and one 1), where each block along the diagonal, called a Jordan block, has th ...
s in linear algebra) are nontrivial in circle-valued functions, which would be zero in real-valued case, and combining with barcodes give the invariants of a tame map, under moderate conditions. Two techniques they use are Morse-Novikov theory and graph representation theory. More recent results can be found in D. Burghelea et al. For example, the tameness requirement can be replaced by the much weaker condition, continuous.


Persistence with torsion

The proof of the structure theorem relies on the base domain being field, so not many attempts have been made on persistence homology with torsion. Frosini defined a pseudometric on this specific module and proved its stability. One of its novelty is that it doesn't depend on some classification theory to define the metric.


Categorification and cosheaves

One advantage of
category theory Category theory is a general theory of mathematical structures and their relations that was introduced by Samuel Eilenberg and Saunders Mac Lane in the middle of the 20th century in their foundational work on algebraic topology. Nowadays, ca ...
is its ability to lift concrete results to a higher level, showing relationships between seemingly unconnected objects. Bubenik et al. offers a short introduction of category theory fitted for TDA. Category theory is the language of modern algebra, and has been widely used in the study of algebraic geometry and topology. It has been noted that "the key observation of is that the persistence diagram produced by depends only on the algebraic structure carried by this diagram." The use of category theory in TDA has proved to be fruitful. Following the notations made in Bubenik et al., the indexing category P is any preordered set (not necessarily \mathbb or \mathbb), the target category D is any category (instead of the commonly used \mathrm_), and
functor In mathematics, specifically category theory, a functor is a mapping between categories. Functors were first considered in algebraic topology, where algebraic objects (such as the fundamental group) are associated to topological spaces, and m ...
s P \to D are called generalized persistence modules in D, over P. One advantage of using category theory in TDA is a clearer understanding of concepts and the discovery of new relationships between proofs. Take two examples for illustration. The understanding of the correspondence between interleaving and matching is of huge importance, since matching has been the method used in the beginning (modified from Morse theory). A summary of works can be found in Vin de Silva et al. Many theorems can be proved much more easily in a more intuitive setting. Another example is the relationship between the construction of different complexes from point clouds. It has long been noticed that Čech and Vietoris-Rips complexes are related. Specifically, V_r(X) \subset C_(X) \subset V_(X). The essential relationship between Cech and Rips complexes can be seen much more clearly in categorical language. The language of category theory also helps cast results in terms recognizable to the broader mathematical community. Bottleneck distance is widely used in TDA because of the results on stability with respect to the bottleneck distance. In fact, the interleaving distance is the
terminal object In category theory, a branch of mathematics, an initial object of a category is an object in such that for every object in , there exists precisely one morphism . The dual notion is that of a terminal object (also called terminal element): ...
in a poset category of stable metrics on multidimensional persistence modules in a
prime field In mathematics, the characteristic of a ring , often denoted , is defined to be the smallest number of times one must use the ring's multiplicative identity (1) in a sum to get the additive identity (0). If this sum never reaches the additive ide ...
. Sheaves, a central concept in modern
algebraic geometry Algebraic geometry is a branch of mathematics, classically studying zeros of multivariate polynomials. Modern algebraic geometry is based on the use of abstract algebraic techniques, mainly from commutative algebra, for solving geometrical ...
, are intrinsically related to category theory. Roughly speaking, sheaves are the mathematical tool for understanding how local information determines global information. Justin Curry regards level set persistence as the study of
fibers Fiber or fibre (from la, fibra, links=no) is a natural or artificial substance that is significantly longer than it is wide. Fibers are often used in the manufacture of other materials. The strongest engineering materials often incorporate ...
of continuous functions. The objects that he studies are very similar to those by MAPPER, but with sheaf theory as the theoretical foundation. Although no breakthrough in the theory of TDA has yet used sheaf theory, it is promising since there are many beautiful theorems in algebraic geometry relating to sheaf theory. For example, a natural theoretical question is whether different filtration methods result in the same output.


Stability

Stability is of central importance to data analysis, since real data carry noises. By usage of category theory, Bubenik et al. have distinguished between soft and hard stability theorems, and proved that soft cases are formal. Specifically, general workflow of TDA is The soft stability theorem asserts that HF is
Lipschitz continuous In mathematical analysis, Lipschitz continuity, named after German mathematician Rudolf Lipschitz, is a strong form of uniform continuity for functions. Intuitively, a Lipschitz continuous function is limited in how fast it can change: there e ...
, and the hard stability theorem asserts that J is Lipschitz continuous. Bottleneck distance is widely used in TDA. The isometry theorem asserts that the interleaving distance d_I is equal to the bottleneck distance. Bubenik et al. have abstracted the definition to that between functors F,G: P\to D when P is equipped with a sublinear projection or superlinear family, in which still remains a pseudometric. Considering the magnificent characters of interleaving distance, here we introduce the general definition of interleaving distance(instead of the first introduced one): Let \Gamma, K \in \mathrm (a function from P to P which is monotone and satisfies x \leq \Gamma(x) for all x\in P). A (\Gamma, K)-interleaving between F and G consists of natural transformations \varphi\colon F \Rightarrow G\Gamma and \psi: G \Rightarrow FK, such that (\psi\Gamma)=\varphi F\eta_ and (\varphi\Gamma)=\psi G\eta_. The two main results are * Let P be a preordered set with a sublinear projection or superlinear family. Let H:D \to E be a functor between arbitrary categories D,E. Then for any two functors F,G:P\to D, we have d_I(HF,HG) \leq d_I(F,G). * Let P be a poset of a metric space Y , X be a topological space. And letf,g:X\to Y (not necessarily continuous) be functions, and F,G to be the corresponding persistence diagram. Then d_I(F,G) \leq d_(f,g):=\sup_d_Y(f(x),g(x)). These two results summarize many results on stability of different models of persistence. For the stability theorem of multidimensional persistence, please refer to the subsection of persistence.


Structure theorem

The structure theorem is of central importance to TDA; as commented by G. Carlsson, "what makes homology useful as a discriminator between topological spaces is the fact that there is a classification theorem for finitely generated abelian groups." (see the
fundamental theorem of finitely generated abelian groups In abstract algebra, an abelian group (G,+) is called finitely generated if there exist finitely many elements x_1,\dots,x_s in G such that every x in G can be written in the form x = n_1x_1 + n_2x_2 + \cdots + n_sx_s for some integers n_1,\dots, ...
). The main argument used in the proof of the original structure theorem is the standard
structure theorem for finitely generated modules over a principal ideal domain In mathematics, in the field of abstract algebra, the structure theorem for finitely generated modules over a principal ideal domain is a generalization of the fundamental theorem of finitely generated abelian groups and roughly states that finite ...
. However, this argument fails if the indexing set is (\mathbb,\leq). In general, not every persistence module can be decomposed into intervals. Many attempts have been made at relaxing the restrictions of the original structure theorem. The case for pointwise finite-dimensional persistence modules indexed by a locally finite subset of \mathbb is solved based on the work of Webb. The most notable result is done by Crawley-Boevey, which solved the case of \mathbb. Crawley-Boevey's theorem states that any pointwise finite-dimensional persistence module is a direct sum of interval modules. To understand the definition of his theorem, some concepts need introducing. An interval in (\R,\leq) is defined as a subset I \subset \mathbb having the property that if r, t \in I and if there is an s \in \R such that r \leq s \leq t, then s\in I as well. An interval module k_I assigns to each element s\in I the vector space k and assigns the zero vector space to elements in \R \setminus I. All maps \rho_s^t are the zero map, unless s,t \in I and s\leq t, in which case \rho_s^t is the identity map. Interval modules are indecomposable. Although the result of Crawley-Boevey is a very powerful theorem, it still doesn't extend to the q-tame case. A persistence module is q-tame if the rank of \rho_s^t is finite for all s< t. There are examples of q-tame persistence modules that fail to be pointwise finite. However, it turns out that a similar structure theorem still holds if the features that exist only at one index value are removed. This holds because the infinite dimensional parts at each index value do not persist, due to the finite-rank condition. Formally, the observable category \mathrm is defined as \mathrm/\mathrm, in which \mathrm denotes the full subcategory of \mathrm whose objects are the ephemeral modules (\rho^t_s=0 whenever s < t). Note that the extended results listed here do not apply to zigzag persistence, since the analogue of a zigzag persistence module over \mathbb R is not immediately obvious.


Statistics

Real data is always finite, and so its study requires us to take stochasticity into account. Statistical analysis gives us the ability to separate true features of the data from artifacts introduced by random noise. Persistent homology has no inherent mechanism to distinguish between low-probability features and high-probability features. One way to apply statistics to topological data analysis is to study the statistical properties of topological features of point clouds. The study of random simplicial complexes offers some insight into statistical topology. K. Turner et al. offers a summary of work in this vein. A second way is to study probability distributions on the persistence space. The persistence space B_\infty is \coprod_n B_/ , where B_n is the space of all barcodes containing exactly n intervals and the equivalences are \ \backsim \ if x_n = y_n. This space is fairly complicated; for example, it is not complete under the bottleneck metric. The first attempt made to study it is by Y. Mileyko et al. The space of persistence diagrams D_p in their paper is defined as D_p := \left\where \Delta is the diagonal line in \mathbb^2. A nice property is that D_p is complete and separable in the Wasserstein metric W_p(u,v)=\left(\inf_\int_ \rho^p(x,y) \, \mathrm\gamma(x,y)\right)^. Expectation, variance, and conditional probability can be defined in the Fréchet sense. This allows many statistical tools to be ported to TDA. Works on null hypothesis significance test,
confidence interval In frequentist statistics, a confidence interval (CI) is a range of estimates for an unknown parameter. A confidence interval is computed at a designated ''confidence level''; the 95% confidence level is most common, but other levels, such as 9 ...
s, and robust estimates are notable steps. A third way is to consider the cohomology of probabilistic space or statistical systems directly, called information structures and basically consisting in the triple (\Omega,\Pi,P), sample space, random variables and probability laws. Random variables are considered as partitions of the n atomic probabilities (seen as a probability (n-1)-simplex, , \Omega, =n) on the lattice of partitions (\Pi_n). The random variables or modules of measurable functions provide the cochain complexes while the coboundary is considered as the general homological algebra first discovered by Hochschild with a left action implementing the action of conditioning. The first cocycle condition corresponds to the chain rule of entropy, allowing to derive uniquely up to the multiplicative constant, Shannon entropy as the first cohomology class. The consideration of a deformed left-action generalises the framework to Tsallis entropies. The information cohomology is an example of ringed topos. Multivariate k-
Mutual information In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual dependence between the two variables. More specifically, it quantifies the " amount of information" (in units such ...
appear in coboundaries expressions, and their vanishing, related to cocycle condition, gives equivalent conditions for statistical independence. Minima of mutual-informations, also called synergy, give rise to interesting independence configurations analog to homotopical links. Because of its combinatorial complexity, only the simplicial subcase of the cohomology and of information structure has been investigated on data. Applied to data, those cohomological tools quantifies statistical dependences and independences, including
Markov chains A Markov chain or Markov process is a stochastic model describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event. Informally, this may be thought of as, "What happen ...
and
conditional independence In probability theory, conditional independence describes situations wherein an observation is irrelevant or redundant when evaluating the certainty of a hypothesis. Conditional independence is usually formulated in terms of conditional probabil ...
, in the multivariate case. Notably, mutual-informations generalize
correlation coefficient A correlation coefficient is a numerical measure of some type of correlation, meaning a statistical relationship between two variables. The variables may be two columns of a given data set of observations, often called a sample, or two components ...
and
covariance In probability theory and statistics, covariance is a measure of the joint variability of two random variables. If the greater values of one variable mainly correspond with the greater values of the other variable, and the same holds for the le ...
to non-linear statistical dependences. These approaches were developed independently and only indirectly related to persistence methods, but may be roughly understood in the simplicial case using Hu Kuo Tin Theorem that establishes one-to-one correspondence between mutual-informations functions and finite measurable function of a set with intersection operator, to construct the Čech complex skeleton. Information cohomology offers some direct interpretation and application in terms of neuroscience (neural assembly theory and qualitative cognition ), statistical physic, and deep neural network for which the structure and learning algorithm are imposed by the complex of random variables and the information chain rule. Persistence landscapes, introduced by Peter Bubenik, are a different way to represent barcodes, more amenable to statistical analysis. The persistence landscape of a persistent module M is defined as a function \lambda:\mathbb\times\mathbb\to \bar, \lambda(k,t):=\sup(m\geq 0\mid\beta^\geq k), where \bar denotes the
extended real line In mathematics, the affinely extended real number system is obtained from the real number system \R by adding two infinity elements: +\infty and -\infty, where the infinities are treated as actual numbers. It is useful in describing the algebra ...
and \beta^=\mathrm(\mathrm(M(a\leq b))). The space of persistence landscapes is very nice: it inherits all good properties of barcode representation (stability, easy representation, etc.), but statistical quantities can be readily defined, and some problems in Y. Mileyko et al.'s work, such as the non-uniqueness of expectations, can be overcome. Effective algorithms for computation with persistence landscapes are available. Another approach is to use revised persistence, which is image, kernel and cokernel persistence.


Applications


Classification of applications

More than one way exists to classify the applications of TDA. Perhaps the most natural way is by field. A very incomplete list of successful applications includes data skeletonization, shape study, graph reconstruction, image analysis, material, progression analysis of disease, sensor network, signal analysis, cosmic web, complex network, fractal geometry, viral evolution, propagation of contagions on networks , bacteria classification using molecular spectroscopy, hyperspectral imaging in physical-chemistry, remote sensing, feature selection, and early warning signs of financial crashes. Another way is by distinguishing the techniques by G. Carlsson,


Characteristics of TDA in applications

There are several notable interesting features of the recent applications of TDA: # Combining tools from several branches of mathematics. Besides the obvious need for algebra and topology, partial differential equations, algebraic geometry, representation theory, statistics, combinatorics, and Riemannian geometry have all found use in TDA. # Quantitative analysis. Topology is considered to be very soft since many concepts are invariant under homotopy. However, persistent topology is able to record the birth (appearance) and death (disappearance) of topological features, thus extra geometric information is embedded in it. One evidence in theory is a partially positive result on the uniqueness of reconstruction of curves; two in application are on the quantitative analysis of Fullerene stability and quantitative analysis of
self-similarity __NOTOC__ In mathematics, a self-similar object is exactly or approximately similar to a part of itself (i.e., the whole has the same shape as one or more of the parts). Many objects in the real world, such as coastlines, are statistically se ...
, separately. # The role of short persistence. Short persistence has also been found to be useful, despite the common belief that noise is the cause of the phenomena. This is interesting to the mathematical theory. One of the main fields of data analysis today is
machine learning Machine learning (ML) is a field of inquiry devoted to understanding and building methods that 'learn', that is, methods that leverage data to improve performance on some set of tasks. It is seen as a part of artificial intelligence. Machine ...
. Some examples of machine learning in TDA can be found in Adcock et al.
conference
is dedicated to the link between TDA and machine learning. In order to apply tools from machine learning, the information obtained from TDA should be represented in vector form. An ongoing and promising attempt is the persistence landscape discussed above. Another attempt uses the concept of persistence images. However, one problem of this method is the loss of stability, since the hard stability theorem depends on the barcode representation.


Impact on mathematics

Topological data analysis and persistent homology have had impacts on
Morse theory In mathematics, specifically in differential topology, Morse theory enables one to analyze the topology of a manifold by studying differentiable functions on that manifold. According to the basic insights of Marston Morse, a typical differentiab ...
. Morse theory has played a very important role in the theory of TDA, including on computation. Some work in persistent homology has extended results about Morse functions to tame functions or, even to continuous functions. A forgotten result of R. Deheuvels long before the invention of persistent homology extends Morse theory to all continuous functions. One recent result is that the category of
Reeb graph A Reeb graphY. Shinagawa, T.L. Kunii, and Y.L. Kergosien, 1991. Surface coding based on Morse theory. IEEE Computer Graphics and Applications, 11(5), pp.66-78 (named after Georges Reeb by René Thom) is a mathematical object reflecting the evolutio ...
s is equivalent to a particular class of cosheaf. This is motivated by theoretical work in TDA, since the Reeb graph is related to Morse theory and MAPPER is derived from it. The proof of this theorem relies on the interleaving distance. Persistent homology is closely related to
spectral sequence In homological algebra and algebraic topology, a spectral sequence is a means of computing homology groups by taking successive approximations. Spectral sequences are a generalization of exact sequences, and since their introduction by , they h ...
s. In particular the algorithm bringing a filtered complex to its canonical form permits much faster calculation of spectral sequences than the standard procedure of calculating E^r_ groups page by page. Zigzag persistence may turn out to be of theoretical importance to spectral sequences.


See also

*
Dimensionality reduction Dimensionality reduction, or dimension reduction, is the transformation of data from a high-dimensional space into a low-dimensional space so that the low-dimensional representation retains some meaningful properties of the original data, ideally ...
* Data mining *
Computer vision Computer vision is an interdisciplinary scientific field that deals with how computers can gain high-level understanding from digital images or videos. From the perspective of engineering, it seeks to understand and automate tasks that the human ...
*
Computational topology Algorithmic topology, or computational topology, is a subfield of topology with an overlap with areas of computer science, in particular, computational geometry and computational complexity theory. A primary concern of algorithmic topology, as it ...
*
Discrete Morse theory Discrete Morse theory is a combinatorial adaptation of Morse theory developed by Robin Forman. The theory has various practical applications in diverse fields of applied mathematics and computer science, such as configuration spaces, homology com ...
*
Shape analysis (digital geometry) This article describes shape analysis to analyze and process geometric shapes. Description ''Shape analysis'' is the (mostly) automatic analysis of geometric shapes, for example using a computer to detect similarly shaped objects in a database or ...
* Size theory *
Algebraic topology Algebraic topology is a branch of mathematics that uses tools from abstract algebra to study topological spaces. The basic goal is to find algebraic invariants that classify topological spaces up to homeomorphism, though usually most classify ...


References


Further reading

Brief Introduction *
Source Material for Topological Data Analysis
by Mikael Vejdemo-Johansson Monograph * Video Lecture
Introduction to Persistent Homology
an
Topology for Data Analysis
by Matthew Wright
The Shape of Data
by Gunnar Carlsson Textbook on Topology * Available for Download *

by Robert Ghrist Other Resources of TDA
Applied Topology
by Stanford
Applied algebraic topology research network
, by the Institute for Mathematics and its Applications *Topological Kernel Learning: Discrete Morse Theory is used to connect kernel machine learning with topological data analysis. https://www.researchgate.net/publication/327427685_Topological_Kernel_Learning {{DEFAULTSORT:Topological Data Analysis Computational topology Data analysis Homology theory Applied mathematics Articles with example R code