MinBackProp – Backpropagating through Minimal Solvers

  • Abstract
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

We present an approach to backpropagating through minimal problem solvers in end-to-end neural network training. Traditional methods relying on manually constructed formulas, finite differences, and autograd are laborious, approximate, and unstable for complex minimal problem solvers. We show that using the Implicit function theorem to calculate derivatives to backpropagate through the solution of a minimal problem solver is simple, fast, and stable. We compare our approach to (i) using the standard autograd on minimal problem solvers and relate it to existing backpropagation formulas through SVD-based and Eig-based solvers and (ii) implementing the backprop with an existing PyTorch Deep Declarative Networks (DDN) framework. We demonstrate our technique on a toy example of training outlier-rejection weights for 3D point registration and on a real application of training an outlier-rejection and RANSAC sampling network in image matching. Our method provides 100% stability and is 10 times faster compared to autograd, which is unstable and slow, and compared to DDN, which is stable but also slow.

Similar Papers
  • Research Article
  • 10.30917/att-vk-1814-9588-2023-1-4
Diagnosis of dermatophytosis in cats using artificial neural networks
  • Feb 1, 2023
  • Veterinaria i kormlenie
  • А.А Bushmina + 2 more

The purpose of the research, the results of which are presented in this article, is to determine the possibility and evaluate the effectiveness of using a trained neural network in the diagnosis of ringworm. The article provides an analysis of the methods used for diagnosing dermatomycosis in veterinary practice. One of the actively developing areas at present is the use of artificial neural networks in the diagnosis of animal diseases. The authors have developed a method for diagnosing dermatophytosis using a trained neural network. To identify hair damaged by dermatophyte spores in cats, a trained artificial neural network YOLO v5 was used, based on the YOLO architecture (high-precision artificial neural network), which provides high accuracy and speed of object detection in images. Diagnostics was carried out in three stages. The first stage: the diagnosis of hair in cats damaged by dermatophyte spores was carried out using a trained artificial neural network. The second stage: microscopy by a veterinary specialist of the veterinary center. The third stage: comparison of the received data from the trained artificial neural network and veterinary specialists. Three comparative experiments were carried out on 20 depersonalized samples with different ratios from healthy and sick animals. As a result of testing the trichoscopy method using artificial neural networks for diagnosing spore-damaged hair dermatitis in cats, it was found that a trained artificial neural network of 60 studied samples diagnosed dermatophyte spore damage in 20 samples, a veterinarian - in 17. All positive results were confirmed by a mycological laboratory study. and identification of the pathogen. It has been established that the use of a trained artificial neural network increases the diagnostic efficiency by 15% and reduces the time to perform diagnostic microscopy by 60.3%. The application of the proposed method allows to reduce the time of microscopic examination, improve the accuracy of interpretation of the results, automate methods for identifying causative agents of ringworm in small animals and take timely measures to treat the animal.

  • Research Article
  • Cite Count Icon 15
  • 10.1016/j.vlsi.2020.05.002
Logarithm-approximate floating-point multiplier is applicable to power-efficient neural network training
  • May 14, 2020
  • Integration
  • Taiyu Cheng + 4 more

Logarithm-approximate floating-point multiplier is applicable to power-efficient neural network training

  • Dataset
  • Cite Count Icon 2
  • 10.15200/winn.156631.13064
The Essential Tools of Scientific Machine Learning (Scientific ML)
  • Aug 20, 2019
  • Christopher Rackauckas

Scientific machine learning is a burgeoning discipline which blends scientific computing and machine learning.Traditionally, scientific computing focuses on large-scale mechanistic models, usually differential equations, that are derived from scientific laws that simplified and explained phenomena.On the other hand, machine learning focuses on developing non-mechanistic data-driven models which require minimal knowledge and prior assumptions.The two sides have their pros and cons: differential equation models are great at extrapolating, the terms are explainable, and they can be fit with small data and few parameters.Machine learning models on the other hand require "big data" and lots of parameters but are not biased by the scientists ability to correctly identify valid laws and assumptions.However, the recent trend has been to merge the two disciplines, allowing explainable models that are data-driven, require less data than traditional machine learning, and utilize the knowledge encapsulated in centuries of scientific literature.The promise is to fuse a priori domain knowledge which doesn't fit into a "dataset", allow this knowledge to specify a general structure that prevents overfitting, reduces the number of parameters, and promotes extrapolatability, while still utilizing machine learning techniques to learn specific unknown terms in the model.This has started to be used for outcomes like automated hypothesis generation and accelerated scientific simulation.The purpose of this blog post is to introduce the reader to the tools of scientific machine learning, identify how they come together, and showcase the existing open source tools which can help one get started.We will be focusing on differentiable programming frameworks in the major languages for scientific machine learning: C++, Fortran, Julia, MATLAB, Python, and R.We will be comparing two important aspects: efficiency and composability.Efficiency will be taken in the context of scientific machine learning: by now most tools are well-optimized for the giant neural networks found in traditional machine learning, but, as will be discussed here, that does not necessarily make them efficient when deployed inside of differential equation solvers or when mixed with probabilistic programming tools.Additionally, composability is a key aspect of scientific machine learning since our toolkit is not ML in isolation.Our goal is not to do machine learning as seen in a machine learning conference (classification, NLP, etc.), and it's not to do traditional machine learning as applied to scientific data.Instead, we are putting ML models and techniques into the heart of scientific simulation tools to accelerate and enhance them.Our neural networks need to fully integrate with tools that simulate satellites and robotics simulators.They need to integrate with the packages that we use in our scientific work for verifying numerical accuracy, tracking units, estimating uncertainty, and much more.We need our neural networks to play nicely with existing packages for delay

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 13
  • 10.3390/a14040107
Identification of Intrinsically Disordered Protein Regions Based on Deep Neural Network-VGG16
  • Mar 28, 2021
  • Algorithms
  • Pengchang Xu + 2 more

The accurate of i identificationntrinsically disordered proteins or protein regions is of great importance, as they are involved in critical biological process and related to various human diseases. In this paper, we develop a deep neural network that is based on the well-known VGG16. Our deep neural network is then trained through using 1450 proteins from the dataset DIS1616 and the trained neural network is tested on the remaining 166 proteins. Our trained neural network is also tested on the blind test set R80 and MXD494 to further demonstrate the performance of our model. The MCC value of our trained deep neural network is 0.5132 on the test set DIS166, 0.5270 on the blind test set R80 and 0.4577 on the blind test set MXD494. All of these MCC values of our trained deep neural network exceed the corresponding values of existing prediction methods.

  • Conference Article
  • Cite Count Icon 24
  • 10.1109/3ict.2018.8855743
Training of Artificial Neural Network Using PSO With Novel Initialization Technique
  • Nov 1, 2018
  • Hafiz Tayyab Rauf + 3 more

Artificial neural networks (ANN) have been widely used in the field of data classification. Normally, training of neural network is applied with the traditional back propagation technique. As, this approach has various drawbacks, training of neural network is done with Particle Swarm Optimization (PSO). PSO has been widely used to solve the diverse kind of optimization problems. Population initialization performs a significant role in meta-heuristic algorithms. This paper describes a new initialization population approach Log Logistic termed as PSOLL-NN to create the initialization of the swarm. The proposed algorithm has been tested for weight optimization of feed forward neural network; and compared with back propagation Algorithm (BPA), standard PSO (PSO-NN), PSO initialized with Halton Sequence (PSOH-NN), Torus sequence (PSOT-NN) and Sobol sequence (PSOS-NN). The experimental results show that the proposed technique performed exceptionally better than the other traditional techniques. Moreover, the outcome of our work presents a foresight that how the proposed initialization technique can be used as an efficient alternative to standard training approaches for the data classification problems.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 2
  • 10.3390/math11010164
Forecasting the Efficiency of Innovative Industrial Systems Based on Neural Networks
  • Dec 28, 2022
  • Mathematics
  • Aleksey I Shinkevich + 2 more

Approaches presented today in the scientific literature suggest that there are no methodological solutions based on the training of artificial neural networks to predict the direction of industrial development, taking into account a set of factors—innovation, environmental friendliness, modernization and production growth. The aim of the study is to develop a predictive model of performance management of innovative industrial systems by building neural networks. The research methods were correlation analysis, training of neural networks (species—regression), extrapolation, and exponential smoothing. As a result of the research, the estimation efficiency technique of an innovative industrial system in a complex considering the criteria of technical modernization, development, innovative activity, and ecologization is developed; the prognostic neural network models allow to optimize the contribution of signs to the formation of target (set) values of indicators of efficiency for macro and micro-industrial systems that will allow to level a growth trajectory of industrial systems; the priority directions of their development are offered. The following conclusions: the efficiency of industrial systems is determined by the volume of sales of goods, innovative products and waste recycling, which allows to save resources; the results of forecasting depend significantly on the DataSet formulated. Although multilayer neural networks independently select important features, it is advisable to conduct a correlation analysis beforehand, which will provide a higher probability of building a high-quality predictive model. The novelty of the research lies in the development and testing of a unique methodology to assess the effectiveness of industrial systems: it is based on a multidimensional system approach (takes into account factors of innovation, environmental friendliness, modernization and production growth); it combines a number of methodological tools (correlation, ranking and weighting); it expands the method of effectiveness assessment in terms of the composition of variables (previously presented approaches are limited to the aspects considered).

  • Research Article
  • Cite Count Icon 2
  • 10.1007/s12239-019-0128-2
Minimization of Surface Deflection in Rectangular Embossing Using Automatic Training of Artificial Neural Network and Genetic Algorithm
  • Nov 1, 2019
  • International Journal of Automotive Technology
  • Sungmin Cho + 1 more

Surface deflection is a phenomenon that causes fine wrinkles on the outer surfaces of sheet metal and deteriorates product external appearance. It is quantitatively defined as the difference between the section curve of the sheet and the ideal curve. In this study, using neural networks, a prediction model for surface deflection according to material properties was constructed and combined with a genetic algorithm; the combination of the material properties was studied to predict the minimum surface deflection. Because of the limited number of simulation data, neural networks were developed using several sampling methods such as central composite design, Latin hypercube sampling, and random sampling. In the training of the neural networks, the optimal hyper-parameter of the neural network was found automatically using Latin hypercube sampling. In conclusion, for prediction of surface deflection in rectangular embossing, neural networks made by central composite design showed the best performance. In addition, it was confirmed that the procedure of combining automatic training of a neural network and the genetic algorithm accurately predicted the set of material properties that generates the minimum surface deflection. Also, the quantity of surface deflection predicted by the neural network was very close to that predicted by finite element analysis.

  • Dissertation
  • 10.15476/elte.2019.135
Affine Correspondences and their Applications for Model Estimation
  • Jan 1, 2019
  • Dániel Béla Baráth

This work aims to solve sub-problems of two major fields in computer vision: minimal problems in two- or multi-view geometric model estimation and robust model fitting. Minimal solvers, i.e. algorithms solving estimation problems from a minimal sample of data points, are involved in most of the vision pipelines as an the engine of the applied robust method, e.g. RANSAC [1] and its recent variants. Vision pipelines, including calibration, structure-from-motion, image matching and retrieval, benefits from efficient minimal solvers which improve their performance upon. Given a minimal point correspondence set in two views, state-of-the-art solvers, with a few exceptions, use them solely through their coordinates. Nevertheless, as it will be demonstrated in this thesis, exploiting affine correspondences which encode higher-order geometric information leads to methods superior to the state-of-the-art in terms of stability and the number of data points required. Methods will be proposed for surface normal, homography, epipolar geometry, and focal length estimation. The second major part of this work focuses on robust model fitting which is also a significant part of vision tasks. The base problem is to fit a single or more model instances, e.g. planes to a 3D point cloud or fundamental matrix to point correspondences, interpreting the input whilst it is contaminated by noise and contains outliers. We consider outliers as points not belonging to any desired model instance. First, a method is proposed to distinguish inliers and outliers in a set of correspondences without necessarily assuming an underlying model. Then a new local optimization step is proposed for locally optimized RANSAC (LO-RANSAC) outperforming its state-of-the-art variants. Finally, we focus on multi-homography, then general multi-class multi-instance, fitting – the problem of interpreting the input data as a mixture of noisy observations originating from multiple instances of multiple classes. The methods proposed in this work were validated both on synthesized and publicly available real world datasets.

  • Research Article
  • 10.5937/grmk1501003k
Estimation of concrete compressive strength using artificial neural network
  • Jan 1, 2015
  • Gradjevinski materijali i konstrukcije
  • Srdjan Kostic + 1 more

In present paper, concrete compressive strength is evaluated using back propagation feed-forward artificial neural network. Training of neural network is performed using Levenberg-Marquardt learning algorithm for four architectures of artificial neural networks, one, three, eight and twelve nodes in a hidden layer in order to avoid the occurrence of overfitting. Training, validation and testing of neural network is conducted for 75 concrete samples with distinct w/c ratio and amount of superplasticizer of melamine type. These specimens were exposed to different number of freeze/thaw cycles and their compressive strength was determined after 7, 20 and 32 days. The obtained results indicate that neural network with one hidden layer and twelve hidden nodes gives reasonable prediction accuracy in comparison to experimental results (R=0.965, MSE=0.005). These results of the performed analysis are further confirmed by calculating the standard statistical errors: the chosen architecture of neural network shows the smallest value of mean absolute percentage error (MAPE=, variance absolute relative error (VARE) and median absolute error (MEDAE), and the highest value of variance accounted for (VAF).

  • Conference Article
  • Cite Count Icon 25
  • 10.1109/isuma.1990.151303
Fuzzy neural expert system with automated extraction of fuzzy If-Then rules from a trained neural network
  • Dec 3, 1990
  • Y Hayashi + 1 more

Proposes a fuzzy neural expert system (FNES) which has a feedforward fuzzy neural network whose input layer consists of fuzzy cell groups and crisp (non-fuzzy) cell groups. Here, the truthfulness of fuzzy information and crisp information of training data is represented by fuzzy cell groups and crisp cell groups, respectively. The expert system has the following two functions: generalization of the information derived from the training data and embodiment of knowledge in the form of the fuzzy neural network; and extraction of fuzzy If-Then rules with linguistic relative importance of each proposition in an antecedent (If-part) from a trained fuzzy neural network. The paper also gives a method to extract automatically fuzzy If-Then rules from the trained neural network. To prove the effectiveness and validity of the proposed fuzzy neural expert system, a fuzzy neural expert system for medical diagnosis has been developed. >

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 1
  • 10.3389/frai.2024.1368569
Training neural networks with universal adiabatic quantum computing.
  • Jun 21, 2024
  • Frontiers in artificial intelligence
  • Steve Abel + 2 more

The training of neural networks (NNs) is a computationally intensive task requiring significant time and resources. This article presents a novel approach to NN training using adiabatic quantum computing (AQC), a paradigm that leverages the principles of adiabatic evolution to solve optimization problems. We propose a universal AQC method that can be implemented on gate quantum computers, allowing for a broad range of Hamiltonians and thus enabling the training of expressive neural networks. We apply this approach to various neural networks with continuous, discrete, and binary weights. The study results indicate that AQC can very efficiently evaluate the global minimum of the loss function, offering a promising alternative to classical training methods.

  • Conference Article
  • Cite Count Icon 3
  • 10.1109/nigercon.2017.8281897
Signal power loss prediction based on artificial neural networks in microcell environment
  • Nov 1, 2017
  • Virginia Chika Ebhota + 2 more

In a bid to predict the propagation loss of electromagnetic signals, different models based on empirical and deterministic formulas have been used. In this study, different artificial neural network models which are very effective for prediction were used for the prediction of signal power loss in a microcell environment, Obio-Akpor, Port Harcourt, Nigeria. The signal power loss of the area is studied based on three artificial neural network algorithms with nine training functions. For the training of the artificial neural network, the input data were the distance from the transmitter and the signal power loss. Training of neural network is a demanding task in the field of supervised learning research. This is because the main difficulty in adopting artificial neural network is in finding the most suitable combination of learning and training functions for the prediction task. We compared the performance of three training algorithms in feedforward back propagation multi layer perceptron neural network. Nine training functions under three training algorithms were selected: the Gradient descent based algorithms, the Conjugate gradient based algorithms and the Quasi-Newton based algorithms. The work compared the training algorithms on the basis of mean square error, mean absolute error, standard deviation, correlation coefficient, regression on training and validation and the rate of convergence. The general performance of the training functions demonstrates their effectiveness to yield accurate results in short time. The conclusion on the training functions is based on the simulation results using measurement data from the micro environment.

  • Research Article
  • Cite Count Icon 103
  • 10.1016/j.eswa.2010.09.028
Comparing performances of backpropagation and genetic algorithms in the data classification
  • Sep 18, 2010
  • Expert Systems with Applications
  • H Hasan Örkcü + 1 more

Comparing performances of backpropagation and genetic algorithms in the data classification

  • Research Article
  • Cite Count Icon 104
  • 10.1016/s0167-9236(00)00086-5
Reliable classification using neural networks: a genetic algorithm and backpropagation comparison
  • Nov 6, 2000
  • Decision Support Systems
  • Randall S Sexton + 1 more

Reliable classification using neural networks: a genetic algorithm and backpropagation comparison

  • Conference Article
  • Cite Count Icon 4
  • 10.1109/ijcnn.2003.1223849
Piecewise-linear modeling of analog circuits using trained feed-forward neural networks and adaptive clustering of hidden neurons
  • Jul 20, 2003
  • S Doboli + 2 more

This paper presents a new technique for automatically creating analog circuit models. The method extracts piecewise linear models from trained neural networks. A model is a set of linear dependencies between circuit performances and design parameters. The paper illustrates the technique for an OTA circuit - an amplifier circuit widely used in filters and A/D converters for which models for gain and bandwidth were automatically generated. As experiments show, the obtained models have simple form that accurately fits the sampled points and the behavior of the trained neural networks. These models are useful for fast simulation of systems with non-linear behavior and performances.

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.