like PCA - Principal Component Analysis - are widely used in Machine Learning for a variety of tasks. But besides the well-known standard methods there are a lot more tools available, especially in the of Manifold Learning. We will interactively explore these tools and present applications for Data Visualization and Feature Engineering using scikit-learn.
|
Manifolds are a complex topic, but they are very important in mathematics and physics. By understanding manifolds, we can better understand the world around us. A manifold is a curved surface that looks like a flat plane when you zoom in. For example, the surface of a sphere is a manifold. It looks curved from a distance, but if you look at a small enough patch of the sphere, it looks flat. Manifolds are important in mathematics and physics because they allow us to describe curved surfaces using flat geometry. This is useful because flat geometry is much simpler to understand than curved geometry.
For example, we can use manifolds to describe the space around us. Even though the space around us is curved, we can describe it using flat geometry by breaking it up into a bunch of small patches. Each patch is a manifold, and we can use flat geometry to describe each patch.
Manifolds are also important in relativity. Relativity is a theory of physics that describes space and time. In relativity, space and time are combined into a four-dimensional manifold.
This manifold is curved, but we can use flat geometry to describe it by breaking it up into a bunch of small patches. Each patch is a manifold, and we can use flat geometry to describe each patch. Here is a simple analogy to help you understand manifolds:
Imagine you have a map of the world. The map is flat, but the Earth is a sphere. This means that the map is distorted. For example, Greenland appears to be larger than Africa on a map, but in reality Africa is much larger.
Manifold learning is an approach to non-linear dimensionality reduction . Algorithms for this task are based on the idea that the dimensionality of many data sets is only artificially high. Manifold learning | SciKitLearn
Manifold learning is based on the assumption that many high-dimensional datasets lie on a low-dimensional manifold , which is a curved surface that is embedded in a higher-dimensional space . It is particularly useful for datasets that lie on a low-dimensional manifold , even if the manifold is non-linear. Manifold learning algorithms aim to find a low-dimensional embedding of the data that preserves the intrinsic geometry of the manifold. This can be useful for visualization, data analysis, and machine learning tasks. Some popular manifold learning algorithms include:
Manifold learning algorithms have been used in a wide variety of applications, including:
Imagine you have two different datasets, one of images of cats and one of images of dogs. Both datasets are high-dimensional, meaning that each image is represented by a long list of numbers.
Manifold alignment is a technique that can be used to find a common representation of these two datasets, even though they are different. It does this by assuming that the two datasets lie on a common manifold.
A manifold is a curved surface that locally resembles a flat plane. For example, the Earth's surface is a manifold. It is curved, but if you look at a small enough patch, it looks flat.
Manifold alignment works by finding a projection from each dataset to the manifold. This projection maps each image to a point on the manifold. The goal is to find projections that preserve the relationships between the images in each dataset.
Once the projections have been found, the two datasets can be represented in the same space. This makes it possible to compare the images in the two datasets directly.
For example, manifold alignment could be used to develop a system that can identify cats and dogs in images, even if the images are taken from different angles or in different lighting conditions.
Here is a simple analogy to help you understand manifold alignment:
Manifold alignment is like finding a way to unfold the map so that it accurately represents the surface of the Earth. This would allow you to compare the distances between different places on the Earth more accurately.
Manifold alignment is a powerful technique that can be used to solve a variety of problems in machine learning. It is often used in applications such as image recognition, natural language processing, and data visualization.
An adversarial example is a specially crafted input that is designed to fool a machine learning model. Adversarial examples are often created by adding small perturbations to normal inputs. These perturbations are imperceptible to humans, but they can cause the model to make incorrect predictions.
The manifold of normal examples is a mathematical object that represents the set of all normal inputs. It is a curved surface, and each normal input corresponds to a point on the manifold.
The reformer network is a type of neural network that can be used to defend against adversarial attacks. It works by moving adversarial examples towards the manifold of normal examples. This makes it more likely that the model will classify the adversarial examples correctly.
To understand how the reformer network works, it is helpful to think about an analogy. Imagine that the manifold of normal examples is a mountain range. The reformer network is like a force that pushes adversarial examples uphill, towards the highest peaks of the mountain range. The higher up the mountain an adversarial example is, the more likely it is to be classified correctly.
The reformer network is effective for correctly classifying adversarial examples with small perturbation because it moves the adversarial examples towards the manifold of normal examples. This makes it more likely that the model will classify the adversarial examples correctly.
Here is a concrete example of how the reformer network could be used to defend against an adversarial attack on an image classification model:
Suppose an attacker creates an adversarial image of a cat that is classified as a dog by the model. The reformer network could be used to move the adversarial image towards the manifold of normal images of cats. This would make it more likely that the model will classify the adversarial image correctly, as a cat.
The reformer network is a promising new defense against adversarial attacks. It is still under development, but it has shown promising results in experiments.
Einstein manifolds are special types of curved surfaces that are important in the theory of general relativity. General relativity is a theory of gravity that describes space and time as a curved four-dimensional surface. Einstein manifolds are surfaces that satisfy the Einstein field equations, which are the equations that govern gravity.
One way to think about Einstein manifolds is to imagine a trampoline with a bowling ball in the middle. The bowling ball will cause the trampoline to bend. This is similar to how gravity causes spacetime to bend.
If you roll a marble across the trampoline, it will follow a curved path. This is because the marble is following the curvature of the trampoline. In the same way, objects in spacetime follow curved paths because of the curvature of spacetime.
Einstein manifolds are mathematical models of bent spacetime. By studying Einstein manifolds, we can learn more about the nature of gravity and the structure of the universe.
Here is a simpler analogy:
Imagine you have a piece of paper. You can draw a straight line on the paper. This is like a flat surface.
Now, imagine you crumple the piece of paper into a ball. You can still draw a line on the paper, but it will be curved. This is like a curved surface.
Einstein manifolds are like curved surfaces, but they are four-dimensional. This means that they have four directions, instead of just two.
Einstein manifolds are important because they are models of the spacetime in which we live. Spacetime is the fabric of the universe, and it is curved because of gravity. By studying Einstein manifolds, we can learn more about the nature of gravity and how it works.
Personal tools.
Deep learning systems have been used extensively in several fields in recent years, but deep neural network (DNN) can also make incorrect decisions and lead to significant losses. Testing tasks based on DNN systems often require annotation of test data to obtain oracle information. However, collecting and annotating large amounts of disparate data from application scenarios is very expensive and time consuming. Thus, we propose DeepSense, an effective neural network test prioritization technique, to select more test inputs that reveal neural network faults as early as possible in the unlabeled test dataset. DeepSense considers the full range of model fault detection capabilities, including faults near the boundary and near the centroid of class. To be specific, we first designed multiple mutation features and extracted them from the neural network model and input samples based on different mutation operators. Then, the sample embedding features are extracted to construct an undirected weighted graph, and a random walk is performed to calculate the distance similarity and manifold similarity, and then spatial nearest neighbor features are designed and extracted. Finally, MLP is used to combine multiple mutation features and spatial nearest neighbor features to predict the input sample fault revealing ability and set priorities accordingly. We evaluate DeepSense on four popular image datasets, and results show that DeepSense significantly outperforms existing test input prioritization techniques.
This is a preview of subscription content, log in via an institution to check access.
Subscribe and save.
Price includes VAT (Russian Federation)
Instant access to the full article PDF.
Rent this article via DeepDyve
Institutional subscriptions
Availability of data and materials.
Data images supporting the experiment training are publicly available in the MINIST, Fashion-MINIST, CIFAR-10 and SVHN repository as part of this record: MINIST: http://yann.lecun.com/exdb/mnist/ Fashion-MINIST: https://yann.lecun.com/exdb/mnist/ CIFAR-10: http://www.cs.toronto.edu/ \(\sim\) kriz/cifar.html SVHN: http://ufldl.stanford.edu/housenumbers/
Han SC, Lin CH, Shen C, Wang Q, Guan XH (2023) Interpreting adversarial examples in deep learning: a review. ACM Comput Surv. https://doi.org/10.1145/3594869
Article Google Scholar
Cinà AE, Grosse K, Demontis A, Vascon S, Zellinger W (2023) Wild patterns reloaded: a survey of machine learning security against training data poisoning. ACM Comput Surv. https://doi.org/10.1145/3585385
Xiong P, Tegegn M, Sarin JS (2023) It is all about data: a survey on the effects of data on adversarial robustness. ACM Comput Surv. https://doi.org/10.48550/arXiv.2303.09767
Long T, Gao Q, Xu LL, Zhou ZB (2022) A survey on adversarial attacks in computer vision: taxonomy, visualization and future directions. Comput Secur 121:102847. https://doi.org/10.1016/j.cose.2022.102847
Sun JZ, Li J, Wen SL (2023) DeepMC: DNN test sample optimization method jointly guided by misclassification and coverage. Appl Intell 53(12):15787–15801
Al-Qadasi H, Wu CS, Falcone Y, Bensalem S (2022) DeepAbstraction: 2-level prioritization for unlabeled test inputs in deep neural networks. In: International conference on artificial intelligence testing, IEEE, pp 64–71
Feng Y, Shi QK, Gao XY (2020) DeepGini: prioritizing massive tests to enhance the robustness of deep neural networks. In: International symposium on software testing and analysis, ACM, pp 177–178
Li Y, Li M, Lai QX (2021) TestRank: bringing order into unlabeled test instances for deep learning tasks. Neural Inf Process Syst 34:20874–20886
Google Scholar
Byun T, Rayadurgam S, Heimdahl MPE (2021) Black-box testing of deep neural networks. In: International symposium on software reliability engineering. IEEE, pp 309–320
Gao XY, Feng Y, Yin YN (2022) Adaptive test selection for deep neural networks. In: International conference on software engineering. IEEE, pp 73–85
Zhang L, Sun XC, Li Y, Zhang ZY (2019) A noise-sensitivity-analysis-based test prioritization technique for deep neural networks. https://doi.org/10.48550/arXiv.1901.00054 . Accessed 1 Jan 2019
Ma W, Papadakis M, Tsakmails A (2021) Test selection for deep learning systems. ACM Trans Softw Eng Methodol 30(2):1–22
Kim J, Feldt R, Yoo S (2019) Guiding deep learning system testing using surprise adequacy. In: International conference on software engineering. IEEE, pp 1039–1049
Weiss M, Chakraborty R, Tonella P (2021) A review and refinement of surprise adequacy. In: International workshop on deep learning for testing and testing for deep learning. IEEE, pp 17–24
Wang ZY, Xu SH, Cai XR, Ji H (2020) Test input selection for deep neural networks. J Phys Conf Ser 1693(1):012017
Shen WJ, Li YH, Chen L (2020) Multiple-boundary clustering and prioritization to promote neural network retraining. In: Automated software engineering. IEEE, pp 410–422
Wang Z, You HM, Chen JJ (2021) Prioritizing test inputs for deep neural networks via mutation analysis. In: Software engineering. IEEE, pp 397–409
Cayton L (2005) Algorithms for manifold learning. Univ. of California at San Diego Tech. Rep.
Guan S, Loew M (2020) Analysis of generalizability of deep neural networks based on the complexity of decision boundary. In: Machine learning and applications. IEEE, pp 101–106
Petrović G, Ivanković M, Fraser G, Just R (2021) Does mutation testing improve testing practices?. In: Software engineering. IEEE, pp 910–921
Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. In: Annual conference on neural information processing systems, Wiley, p 9
Zhou DY, Weston J, Gretton A (2003) Ranking on data manifolds. In: neural information processing systems, MIT Press, pp 169–176
Iscen A, Tolias G, Avrithis Y (2017) Efficient diffusion on region manifolds: Recovering small objects with compact CNN representations. In: computer vision and pattern recognition. IEEE, pp 926–935
Page L, Brin S, Motwani Winograd T (1998) The PageRank citation ranking: bringing order to the web. In: ASIS, pp 161–172
Cho M, Lee KM (2012) Mode-seeking on graphs via random walks. In: Computer vision and pattern recognition. IEEE, pp 606–613
Zhang CN, Benz P, Imatiaz T, Kweon IS (2020) Understanding adversarial examples from the mutual influence of images and perturbations. In: Computer vision and pattern recognition, IEEE, pp 14509–14518
Carlini N, Wagner D (2017) Towards evaluating the robustness of neural networks. In: Security and privacy. IEEE, pp 39–57
Kurakin A, Goodfellow I, Bengio S (2016) Adversarial machine learning at scale. https://doi.org/10.48550/arXiv.1611.01236 . Accessed 11 Feb 2017
Papernot, N. McDaniel, P. Jha S (2016) The limitations of deep learning in adversarial settings. In: Security and privacy. IEEE, pp 372–387
Download references
Authors and affiliations.
Nanchang Hangkong University, Feng and South Avenue, Nanchang, 330063, China
FengYu Yang, YuAn Chen, Tong Chen & Ying Ma
Nanjing University of Aeronautics and Astronautics, Imperial street, Nanjing, 210016, China
FengYu Yang
Jiangxi Hongdu Aviation Industry (Group) Corporation Limited, Qingyunpu Road, Nanchang, China
You can also search for this author in PubMed Google Scholar
Correspondence to Tong Chen .
Conflict of interest.
The authors have no potential Conflict of interest to disclose. The authors have no relevant financial or non-financial interests to disclose.
This paper does not involve human or animal studies. There are no human or animal subjects in this paper and informed consent does not apply. The authors did not receive support from any organization for the submitted work.
Publisher's note.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
Reprints and permissions
Yang, F., Chen, Y., Chen, T. et al. DeepSense: test prioritization for neural network based on multiple mutation and manifold spatial distribution. Evol. Intel. (2024). https://doi.org/10.1007/s12065-024-00961-4
Download citation
Received : 12 December 2023
Revised : 19 June 2024
Accepted : 20 June 2024
Published : 30 July 2024
DOI : https://doi.org/10.1007/s12065-024-00961-4
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
You are accessing a machine-readable page. In order to be human-readable, please install an RSS reader.
All articles published by MDPI are made immediately available worldwide under an open access license. No special permission is required to reuse all or part of the article published by MDPI, including figures and tables. For articles published under an open access Creative Common CC BY license, any part of the article may be reused without permission provided that the original article is clearly cited. For more information, please refer to https://www.mdpi.com/openaccess .
Feature papers represent the most advanced research with significant potential for high impact in the field. A Feature Paper should be a substantial original Article that involves several techniques or approaches, provides an outlook for future research directions and describes possible research applications.
Feature papers are submitted upon individual invitation or recommendation by the scientific editors and must receive positive feedback from the reviewers.
Editor’s Choice articles are based on recommendations by the scientific editors of MDPI journals from around the world. Editors select a small number of articles recently published in the journal that they believe will be particularly interesting to readers, or important in the respective research area. The aim is to provide a snapshot of some of the most exciting work published in the various research areas of the journal.
Original Submission Date Received: .
Find support for a specific problem in the support section of our website.
Please let us know what you think of our products and services.
Visit our dedicated information section to learn more about MDPI.
Angle estimation using learning-based doppler deconvolution in beamspace with forward-looking radar.
2.2. angle estimation using ssl-fae, 2.3. training ssl-fae, 3. numerical simulations and results, 3.1. experiment 1, 3.1.1. case 1, 3.1.2. case 2, 3.1.3. case 3, 3.1.4. case 4, 3.1.5. case 5, 3.2. experiment 2, 3.3. experiment 3, 3.4. experiment 4, 4. conclusions, author contributions, data availability statement, acknowledgments, conflicts of interest, abbreviations.
FLR | Forward-looking radar |
SVR | Support vector regression |
SNR | Signal-to-noise ratio |
APD | Antenna pattern deconvolution |
SVM | Support vector machines |
ASP | Array signal processing |
SSL-FAE | Semi-supervised learning framework for FLR angle estimation |
MVDR | Minimum variance distortionless response |
CS | Compressive sensing |
RKHS | Reproducing kernel Hilbert space |
RBF | Radial basis function |
SCR | Signal-to-clutter ratio |
Click here to enlarge figure
Parameter | Symbol | Value |
---|---|---|
frequency | 77 GHz | |
bandwidth | 2 GHz | |
pulsewidth | T | 2 s |
pulse repetition frequency | PRF | 10 kHz |
altitude | H | 100 m |
velocity | v | 0–250 km/h |
Symbol | Value |
---|---|
0.1 | |
0.1 |
10 m/s | 40 m/s | 70 m/s | 100 m/s | 130 m/s | 160 m/s | |
---|---|---|---|---|---|---|
SVR | ||||||
SSL-FAE |
MVDR | Bayesian | Doppler–MVDR | Doppler–Bayesian | SVR | SSL-FAE | |
---|---|---|---|---|---|---|
Training | - | - | - | - | 0.3399 s | 1.5508 s |
Estimation | 0.0874 s | 0.0055 s | 0.4118 s | 0.0073 s | 0.0016 s | 0.0028 s |
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
Li, W.; Xu, X.; Xu, Y.; Luan, Y.; Tang, H.; Chen, L.; Zhang, F.; Liu, J.; Yu, J. Angle Estimation Using Learning-Based Doppler Deconvolution in Beamspace with Forward-Looking Radar. Remote Sens. 2024 , 16 , 2840. https://doi.org/10.3390/rs16152840
Li W, Xu X, Xu Y, Luan Y, Tang H, Chen L, Zhang F, Liu J, Yu J. Angle Estimation Using Learning-Based Doppler Deconvolution in Beamspace with Forward-Looking Radar. Remote Sensing . 2024; 16(15):2840. https://doi.org/10.3390/rs16152840
Li, Wenjie, Xinhao Xu, Yihao Xu, Yuchen Luan, Haibo Tang, Longyong Chen, Fubo Zhang, Jie Liu, and Junming Yu. 2024. "Angle Estimation Using Learning-Based Doppler Deconvolution in Beamspace with Forward-Looking Radar" Remote Sensing 16, no. 15: 2840. https://doi.org/10.3390/rs16152840
Further information, mdpi initiatives, follow mdpi.
Subscribe to receive issue release notifications and newsletters from MDPI journals
Grab your spot at the free arXiv Accessibility Forum
Help | Advanced Search
Title: transient anisotropic kernel for probabilistic learning on manifolds.
Abstract: PLoM (Probabilistic Learning on Manifolds) is a method introduced in 2016 for handling small training datasets by projecting an Itô equation from a stochastic dissipative Hamiltonian dynamical system, acting as the MCMC generator, for which the KDE-estimated probability measure with the training dataset is the invariant measure. PLoM performs a projection on a reduced-order vector basis related to the training dataset, using the diffusion maps (DMAPS) basis constructed with a time-independent isotropic kernel. In this paper, we propose a new ISDE projection vector basis built from a transient anisotropic kernel, providing an alternative to the DMAPS basis to improve statistical surrogates for stochastic manifolds with heterogeneous data. The construction ensures that for times near the initial time, the DMAPS basis coincides with the transient basis. For larger times, the differences between the two bases are characterized by the angle of their spanned vector subspaces. The optimal instant yielding the optimal transient basis is determined using an estimation of mutual information from Information Theory, which is normalized by the entropy estimation to account for the effects of the number of realizations used in the estimations. Consequently, this new vector basis better represents statistical dependencies in the learned probability measure for any dimension. Three applications with varying levels of statistical complexity and data heterogeneity validate the proposed theory, showing that the transient anisotropic kernel improves the learned probability measure.
Comments: | 44 pages, 14 figures |
Subjects: | Machine Learning (stat.ML); Machine Learning (cs.LG) |
classes: | 68Q32, 68T05, 62R30, 6 0J20 |
classes: | G.3 |
Cite as: | [stat.ML] |
(or [stat.ML] for this version) | |
Focus to learn more arXiv-issued DOI via DataCite |
Access paper:.
Code, data and media associated with this article, recommenders and search tools.
arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.
Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.
Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs .
IMAGES
VIDEO
COMMENTS
The manifold hypothesis is related to the effectiveness of nonlinear dimensionality reduction techniques in machine learning. Many techniques of dimensional reduction make the assumption that data lies along a low-dimensional submanifold, such as manifold sculpting, manifold alignment, and manifold regularization .
The Manifold Hypothesis explains ( heuristically) why machine learning techniques are able to find useful features and produce accurate predictions from datasets that have a potentially large number of dimensions ( variables). The fact that the actual data set of interest actually lives on in a space of low dimension, means that a given machine ...
Manifold learning is an approach to non-linear dimensionality reduction. Algorithms for this task are based on the idea that the dimensionality of many data sets is only artificially high. ... "Multidimensional scaling by optimizing goodness of fit to a nonmetric hypothesis ... L.J.P.; Journal of Machine Learning Research 15(Oct):3221-3245, 2014.
The manifold hypothesis. Chapter 1: Multidimensional Scaling. Classical, metric, and non-metric MDS algorithms. Example applications to quantitative psychology and social science. Chapter 2: ISOMAP. Geodesic distances and the isometric mapping algorithm. Implementation details and applications with facial images and coil-100 object images.
The manifold hypothesis is that natural data forms lower-dimensional manifolds in its embedding space. ... New layers, specifically motivated by the manifold perspective of machine learning, may be useful supplements. (This is a developing research project. It's posted as an experiment in doing research openly.
Manifold learning (ML), known also as non-linear dimension reduction, is a set of methods to find the low dimensional structure of data. Dimension reduction for large, high dimensional data is not merely a way to reduce the data; the new representations and descriptors obtained by ML reveal the geometric shape of high dimensional point clouds, and allow one to visualize, de-noise and interpret ...
In recent years there has been increased interest in understanding the interplay between deep generative models (DGMs) and the manifold hypothesis. Research in this area focuses on understanding the reasons why commonly-used DGMs succeed or fail at learning distributions supported on unknown low-dimensional manifolds, as well as developing new models explicitly designed to account for manifold ...
Manifold learning (ML), also known as nonlinear dimension reduction, is a set of methods to find the low-dimensional structure of data. Dimension reduction for large, high-dimensional data is not merely a way to reduce the data; the new representations and descriptors obtained by ML reveal the geometric shape of high-dimensional point clouds and allow one to visualize, denoise, and interpret them.
In its early days, the primary applications of manifold learning were to reduce the representation dimension, to beat the curse of dimensionality in machine learning applications such as face recognition (cf. []).Another parallel stream of work has been to leverage these methods for understanding structure of data by embedding into 2D or 3D maps.
TESTING THE MANIFOLD HYPOTHESIS CHARLESFEFFERMAN,SANJOYMITTER,ANDHARIHARANNARAYANAN Contents 1. Introduction 984 1.1. Definitions 988 1.2. Constants 988 1.3. d-planes 988 ... to the underlying hypothesis as the "manifold hypothesis." Manifold learning, in particular, fitting low dimensional nonlinear manifolds to sampled data points in
Manifold learning is a popular and quickly-growing subfield of machine learning based on the assumption that one's observed data lie on a low-dimensional manifold embedded in a higher-dimensional space. This thesis presents a mathematical perspective on manifold learning, delving into the intersection of kernel learning, spectral graph theory, and differential geometry. Emphasis is placed on ...
The Manifold Hypothesis explains (heuristically) why machine learning techniques are able to find useful features and produce accurate predictions from datasets that have a potentially large number of dimensions ( variables).The fact that the actual data set of interest actually lives on in a space of low dimension, means that a given machine learning model only needs to learn to focus on a ...
2. We obtain a minimax lower bound on the sample complexity of any rule for learning a manifold from Fin Theorem 6 showing that for a fixed error, the the dependence of the sample complexity on intrinsic dimension, curvature and volume must be at least exponen- tial, polynomial, and linear, respectively. 3.
ents are being projected on the data manifold [14,40]. In explainable machine learning, it has been shown that expla-nations can be manipulated by modifying the model outside of the image manifold, and that one can defend against such attacks by projecting the explanations back on the manifold [13]. The hypothesis that natural image data ...
Significance. Accurate simulation of fluids is important for many science and engineering problems but is very computationally demanding. In contrast, machine-learning models can approximate physics very quickly but at the cost of accuracy. Here we show that using machine learning inside traditional fluid simulations can improve both accuracy ...
The Manifold Hypothesis is a widely accepted tenet of Machine Learning which asserts that nominally high-dimensional data are in fact concentrated near a low-dimensional manifold, embedded in high-dimensional space. This phenomenon is observed empirically in many real world situations, has led to development of a wide range of statistical methods in the last few decades, and has been suggested ...
The present article gives an introduction to information geometry and surveys its applications in the area of machine learning, optimization and statistical inference. Information geometry is explained intuitively by using divergence functions introduced in a manifold of probability distributions and other general manifolds. They give a Riemannian structure together with a pair of dual ...
Physics-informed machine-learning (PIML) enables the integration of domain knowledge with machine learning (ML) algorithms, which results in higher data efficiency and more stable predictions. This provides opportunities for augmenting—and even replacing—high-fidelity numerical simulations of complex turbulent flows, which are often expensive due to the requirement of high temporal and ...
The paper presents a new geometrically motivated method for non-linear regression based on Manifold learning technique. The regression problem is to construct a predictive function which estimates an unknown smooth mapping f from q-dimensional inputs to m-dimensional outputs based on a training data set consisting of given 'input-output' pairs.
Unsupervised non-linear manifold learning (Izenman, Citation 2012; Fukami & Taira, Citation 2023) is a novel approach for capturing geometric features. It treats the data as distributions on low-dimensional manifolds, and uses the non-Euclidean distance to capture the intrinsic attributes of features.
Manifold hypothesis states that data points in high-dimensional space actually lie in close vicinity of a manifold of much lower dimension. In many cases this hypothesis was empirically verified and used to enhance unsupervised and semi-supervised learning. Here we present new approach to manifold hypothesis checking and underlying manifold dimension estimation. In order to do it we use two ...
Some form of adjustment to the model architecture or training algorithm is necessary, since we show that generalization of neural networks alone does not imply the alignment of model gradients with the data manifold. Subjects: Machine Learning (cs.LG); Computer Vision and Pattern Recognition (cs.CV) Cite as: arXiv:2206.07387 [cs.LG]
Deep learning systems have been used extensively in several fields in recent years, but deep neural network (DNN) can also make incorrect decisions and lead to significant losses. Testing tasks based on DNN systems often require annotation of test data to obtain oracle information. However, collecting and annotating large amounts of disparate data from application scenarios is very expensive ...
The measurement of the target azimuth angle using forward-looking radar (FLR) is widely applied in unmanned systems, such as obstacle avoidance and tracking applications. This paper proposes a semi-supervised support vector regression (SVR) method to solve the problem of small sample learning of the target angle with FLR. This method utilizes function approximation to solve the problem of ...
PLoM (Probabilistic Learning on Manifolds) is a method introduced in 2016 for handling small training datasets by projecting an Itô equation from a stochastic dissipative Hamiltonian dynamical system, acting as the MCMC generator, for which the KDE-estimated probability measure with the training dataset is the invariant measure. PLoM performs a projection on a reduced-order vector basis ...