Lecture: Introduction to Deep Learning Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab

Total Page:16

File Type:pdf, Size:1020Kb

Lecture: Introduction to Deep Learning Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture: Introduction to Deep Learning Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Slides adapted from Justin Johnson Stanford University Lecture 19 - 1 6 Dec 2016 So far this quarter Edge detection SIFT RANSAC K-Means Linear classifiers Image features PCA / Eigenfaces Mean-shift Stanford University Lecture 19 - 2 5 Dec6 Dec 2016 2016 Current Research ECCV 2016 (photo credit: Justin Johnson) Stanford University Lecture 19 - 3 6 Dec 2016 Current Research Convolutional Recurrent neural Autoencoders neural networks networks Fully convolutional networks Multilayer perceptrons ECCV 2016 (photo credit: Justin Johnson) What are these things? Why are they important? Stanford University Lecture 19 - 4 5 Dec6 Dec 2016 2016 Deep Learning ● Learning hierarchical representations from data ● End-to-end learning: raw inputs to predictions ● Can use a small set of simple tools to solve many problems ● Has led to rapid progress on many problems ● (Very loosely!) Inspired by the brain ● Today: 1 hour introduction to deep learning fundamentals ● To learn more take CS 231n in the Spring Stanford University Lecture 19 - 5 5 Dec6 Dec 2016 2016 Deep learning for many different problems Stanford University Lecture 19 - 6 6 Dec 2016 Stanford University Lecture 19 - 7 5 Dec6 Dec 2016 2016 Slide credit: CS 231n, Lecture 1, 2016 Stanford University Lecture 19 - 8 5 Dec6 Dec 2016 2016 Slide credit: Kaiming He, ICCV 2015 Stanford University Lecture 19 - 9 5 Dec6 Dec 2016 2016 ≈5.1 Human Human benchmark from Russakovsky et al, “ImageNet Large Scale Visual Recognition Challenge”, IJCV 2015 Slide credit: Kaiming He, ICCV 2015 Stanford University Lecture 19 - 10 5 Dec6 Dec 2016 2016 Slide credit: Kaiming He, ICCV 2015 Stanford University Lecture 19 - 11 5 Dec6 Dec 2016 2016 Deep Learning for Object Detection Slide credit: Ross Girshick, ICCV 2015 Stanford University Lecture 19 - 12 5 Dec6 Dec 2016 2016 Deep Learning for Object Detection Current SOTA: 85.6 He et al, “Deep Residual Learning for Image Recognition”, CVPR 2016 Slide credit: Ross Girshick, ICCV 2015 Stanford University Lecture 19 - 13 5 Dec6 Dec 2016 2016 Object segmentation Figure credit: Dai, He, and Sun, “Instance-aware Semantic Segmentation via Multi-task Network Cascades”, CVPR 2016 Stanford University Lecture 19 - 14 5 Dec6 Dec 2016 2016 Pose Estimation Figure credit: Cao et al, “Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields”, arXiv 2016 Stanford University Lecture 19 - 15 5 Dec6 Dec 2016 2016 Deep Learning for Image Captioning Figure credit: Karpathy and Fei-Fei, “Deep Visual-Semantic Alignments for Generating Image Descriptions”, CVPR 2015 Stanford University Lecture 19 - 16 5 Dec6 Dec 2016 2016 Dense Image Captioning Figure credit: Johnson*, Karpathy*, and Fei-Fei, “DenseCap: Fully Convolutional Localization Networks for Dense Captioning”, CVPR 2016 Stanford University Lecture 19 - 17 5 Dec6 Dec 2016 2016 Visual Question Answering Figure credit: Agrawal et al, “VQA: Visual Question Answering”, ICCV 2015 Figure credit: Zhu et al, “Visual7W: Grounded Question Answering in Images”, CVPR 2016 Stanford University Lecture 19 - 18 5 Dec6 Dec 2016 2016 Image Super-Resolution Figure credit: Ledig et al, “Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network”, arXiv 2016 Stanford University Lecture 19 - 19 5 Dec6 Dec 2016 2016 Generating Art Figure credit: Mordvintsev, Olah, and Tyka, “Inceptionism: Going Deeper into Neural Networks”, https://research.googleblog.com/2015/06/inceptionism-going-deeper-into-neural.html Figure credit: Gatys, Ecker, and Bethge, “Image Style Transfer using Convolutional Neural Networks”, CVPR 2016 Figure credit: Johnson, Alahi, and Fei-Fei: “Perceptual Losses for Real-Time Style Transfer and Super-Resolution”, ECCV 2016, https://github.com/jcjohnson/fast-neural-style Stanford University Lecture 19 - 20 5 Dec6 Dec 2016 2016 Outside Computer Vision Speech Recognition Machine Translation Figure credit: Bahdanau, Cho, and Bengio, “Neural Machine Translation by jointly learning to align and translate”, ICLR 2015 Figure credit: Amodei et al, “Deep Speech 2: End-to-End Speech Text Synthesis Recognition in English and Mandarin”, arXiv 2015 Speech Synthesis Figure credit: Sutskever, Martens, and Hinton, “Generating Text with Recurrent Neural Networks”, ICML 2011 Figure credit: var der Oord et al, “WaveNet: A Generative Model for Raw Audio”, arXiv 2016, https://deepmind.com/blog/wavenet-generative-model-raw-audio/ Stanford University Lecture 19 - 21 5 Dec6 Dec 2016 2016 Deep Reinforcement Learning Playing Atari games AlphaGo beats Lee Sedol Silver et al, “Mastering the game of Go with deep neural networks and tree search”, Nature 2016 Image credit: http://www.newyorker.com/tech/elements/alphago-lee-sedol-and-the-reassuring-future- of-humans-and-machines Mnih et al, “Human-level control through deep reinforcement learning”, Nature 2015 Stanford University Lecture 19 - 22 5 Dec6 Dec 2016 2016 Outline ● Applications ● Motivation ● Supervised learning ● Gradient descent ● Backpropagation ● Convolutional Neural Networks Stanford University Lecture 19 - 23 5 Dec6 Dec 2016 2016 Outline ● Applications ● Motivation ● Supervised learning ● Gradient descent ● Backpropagation ● Convolutional Neural Networks Stanford University Lecture 19 - 24 5 Dec6 Dec 2016 2016 Recall: Image Classification Write a function that maps images to labels (also called object recognition) Dataset: ETH-80, by B. Leibe; Slide credit: L. Lazebnik No obvious way to implement this function! Slide credit: CS 231n, Lecture 2, 2016 Stanford University Lecture 19 - 25 5 Dec6 Dec 2016 2016 Recall: Image Classification Example training set Slide credit: CS 231n, Lecture 2, 2016 Stanford University Lecture 19 - 26 5 Dec6 Dec 2016 2016 Recall: Image Classification Figure credit: D. Hoiem, L. Lazebnik Stanford University Lecture 19 - 27 5 Dec6 Dec 2016 2016 Recall: Image Classification Features remain fixed Classifier is learned from data Figure credit: D. Hoiem, L. Lazebnik Stanford University Lecture 19 - 28 5 Dec6 Dec 2016 2016 Recall: Image Classification Features remain fixed Classifier is learned from data Problem: How do we know which features to use? We may need different features for each problem! Figure credit: D. Hoiem, L. Lazebnik Stanford University Lecture 19 - 29 5 Dec6 Dec 2016 2016 Recall: Image Classification Features remain fixed Classifier is learned from data Problem: Solution: How do we know which features Learn the features jointly to use? We may need different with the classifier! features for each problem! Figure credit: D. Hoiem, L. Lazebnik Stanford University Lecture 19 - 30 5 Dec6 Dec 2016 2016 Image Classification: Feature Learning Training labels Training Learned Model Learned Image Classifier features features Learned Classifier Learned Model Learned features Prediction Learned Classifier Stanford University Lecture 19 - 31 5 Dec6 Dec 2016 2016 Image Classification: Deep Learning Training labels Learned Model Training Low-level features Low-level Classifier features Mid-level Model is features deep: it High-level has many Mid-level High-level layers features features features Classifier Models can learn hierarchical features Stanford University Lecture 19 - 32 5 Dec6 Dec 2016 2016 Image Classification: Deep Learning Training labels Learned Model Training Low-level features Low-level Classifier features Mid-level Model is features deep: it High-level has many Mid-level High-level layers features features features Classifier Models can learn hierarchical features Figure credit: Zeiler and Fergus, “Visualizing and Understanding Convolutional Networks”, ECCV 2014 Stanford University Lecture 19 - 33 5 Dec6 Dec 2016 2016 Outline ● Applications ● Motivation ● Supervised learning ● Gradient descent ● Backpropagation ● Convolutional Neural Networks Stanford University Lecture 19 - 34 5 Dec6 Dec 2016 2016 Supervised Learning in 3 easy steps How to learn models from data Step 1: Define Model Stanford University Lecture 19 - 35 5 Dec6 Dec 2016 2016 Supervised Learning in 3 easy steps How to learn models from data Step 1: Define Model Predicted output Model Input data Model (image label) structure (Image pixels) weights Stanford University Lecture 19 - 36 5 Dec6 Dec 2016 2016 Supervised learning b Stanford University Lecture 19 - 37 5 Dec 2016 Supervised Learning in 3 easy steps How to learn models from data Step 1: Step 2: Collect data Define Model Predicted output Model Input data Model (image label) structure (Image pixels) weights Training True input output Stanford University Lecture 19 - 38 5 Dec6 Dec 2016 2016 Supervised Learning in 3 easy steps How to learn models from data Step 1: Step 2: Collect data Define Model Predicted output Model Input data Model (image label) structure (Image pixels) weights Training True input output Step 3: Learn the model Stanford University Lecture 19 - 39 5 Dec6 Dec 2016 2016 Supervised Learning in 3 easy steps How to learn models from data Step 1: Step 2: Collect data Define Model Predicted output Model Input data Model (image label) structure (Image pixels) weights Training True input output Predicted Step 3: Learn the model output Loss function: Regularizer: Learned Minimize average Measures “badness” Penalizes complex weights loss over training set of prediction models Stanford University Lecture 19 - 40 5 Dec6 Dec 2016 2016 Supervised Learning: Linear regression Sometimes called “Ridge Regression” with regularizer Input and output are vectors Loss is Euclidean distance Linear Regression
Recommended publications
  • Mean Shift Paper by Comaniciu and Meer Presentation by Carlo Lopez-Tello What Is the Mean Shift Algorithm?
    Mean Shift Paper by Comaniciu and Meer Presentation by Carlo Lopez-Tello What is the Mean Shift Algorithm? ● A method of finding peaks (modes) in a probability distribution ● Works without assuming any underlying structure in the distribution ● Works on multimodal distributions ● Works without assuming the number of modes Why do we care about modes? ● Given a data set we can assume that it was sampled from some pdf ● Samples are most likely to be drawn from a region near a mode ● We can use the modes to cluster the data ● Clustering has many applications: filtering, segmentation, tracking, classification, and compression. Why do we care about modes? Why use mean shift for clustering? ● K-means needs to know how many clusters to use. Clusters data into voronoi cells. ● Histograms require bin size and number of bins ● Mixture models require information about pdf structure Intuition ● We have a set of data that represents discrete samples of a distribution ● Locally we can estimate the density of the distribution with a function ● Compute the gradient of this estimation function ● Use gradient ascent to find the peak of the distribution How does it work? ● We estimate the density using: ● Where h (bandwidth) is the region around x where we are trying to estimate the density and k is some kernel function ● Instead of using the gradient of f, we use the mean shift vector: How to find a mode? 1. Start at any point 2. Compute mean shift 3. if mean shift is zero: possible mode found 4. else move to where mean shift is pointing go to 2 ● To find multiple modes we need to try all points that are more than h distance apart ● Prune modes by perturbing them and checking for convergence ● Combine modes that are close together.
    [Show full text]
  • Identification Using Telematics
    risks Article Identification Using Telematics: An Algorithm to Identify Dwell Locations Christopher Grumiau 1, Mina Mostoufi 1,* , Solon Pavlioglou 1,* and Tim Verdonck 2,3 1 Allianz Benelux, 1000 Brussels, Belgium; [email protected] 2 Department of Mathematics (Faculty of Science), University of Antwerp, 2000 Antwerpen, Belgium; [email protected] 3 Department of Mathematics (Faculty of Science), Katholieke Universiteit Leuven, 3000 Leuven, Belgium * Correspondence: mina.mostoufi@allianz.be (M.M.); [email protected] (S.P.) Received: 16 June 2020; Accepted: 21 August 2020; Published: 1 September 2020 Abstract: In this work, a method is proposed for exploiting the predictive power of a geo-tagged dataset as a means of identification of user-relevant points of interest (POI). The proposed methodology is subsequently applied in an insurance context for the automatic identification of a driver’s residence address, solely based on his pattern of movements on the map. The analysis is performed on a real-life telematics dataset. We have anonymized the considered dataset for the purpose of this study to respect privacy regulations. The model performance is evaluated based on an independent batch of the dataset for which the address is known to be correct. The model is capable of predicting the residence postal code of the user with a high level of accuracy, with an f1 score of 0.83. A reliable result of the proposed method could generate benefits beyond the area of fraud, such as general data quality inspections, one-click quotations, and better-targeted marketing. Keywords: telematics; address identification; POI; machine learning; mean shift clustering; DBSCAN clustering; fraud detection 1.
    [Show full text]
  • DBSCAN++: Towards Fast and Scalable Density Clustering
    DBSCAN++: Towards fast and scalable density clustering Jennifer Jang 1 Heinrich Jiang 2 Abstract 2, it quickly starts to exhibit quadratic behavior in high di- mensions and/or when n becomes large. In fact, we show in DBSCAN is a classical density-based clustering Figure1 that even with a simple mixture of 3-dimensional procedure with tremendous practical relevance. Gaussians, DBSCAN already starts to show quadratic be- However, DBSCAN implicitly needs to compute havior. the empirical density for each sample point, lead- ing to a quadratic worst-case time complexity, The quadratic runtime for these density-based procedures which is too slow on large datasets. We propose can be seen from the fact that they implicitly must compute DBSCAN++, a simple modification of DBSCAN density estimates for each data point, which is linear time which only requires computing the densities for a in the worst case for each query. In the case of DBSCAN, chosen subset of points. We show empirically that, such queries are proximity-based. There has been much compared to traditional DBSCAN, DBSCAN++ work done in using space-partitioning data structures such can provide not only competitive performance but as KD-Trees (Bentley, 1975) and Cover Trees (Beygelzimer also added robustness in the bandwidth hyperpa- et al., 2006) to improve query times, but these structures are rameter while taking a fraction of the runtime. all still linear in the worst-case. Another line of work that We also present statistical consistency guarantees has had practical success is in approximate nearest neigh- showing the trade-off between computational cost bor methods (e.g.
    [Show full text]
  • Audio Event Classification Using Deep Learning in an End-To-End Approach
    Audio Event Classification using Deep Learning in an End-to-End Approach Master thesis Jose Luis Diez Antich Aalborg University Copenhagen A. C. Meyers Vænge 15 2450 Copenhagen SV Denmark Title: Abstract: Audio Event Classification using Deep Learning in an End-to-End Approach The goal of the master thesis is to study the task of Sound Event Classification Participant(s): using Deep Neural Networks in an end- Jose Luis Diez Antich to-end approach. Sound Event Classifi- cation it is a multi-label classification problem of sound sources originated Supervisor(s): from everyday environments. An auto- Hendrik Purwins matic system for it would many applica- tions, for example, it could help users of hearing devices to understand their sur- Page Numbers: 38 roundings or enhance robot navigation systems. The end-to-end approach con- Date of Completion: sists in systems that learn directly from June 16, 2017 data, not from features, and it has been recently applied to audio and its results are remarkable. Even though the re- sults do not show an improvement over standard approaches, the contribution of this thesis is an exploration of deep learning architectures which can be use- ful to understand how networks process audio. The content of this report is freely available, but publication (with reference) may only be pursued due to agreement with the author. Contents 1 Introduction1 1.1 Scope of this work.............................2 2 Deep Learning3 2.1 Overview..................................3 2.2 Multilayer Perceptron...........................4
    [Show full text]
  • Deep Mean-Shift Priors for Image Restoration
    Deep Mean-Shift Priors for Image Restoration Siavash A. Bigdeli Meiguang Jin Paolo Favaro University of Bern University of Bern University of Bern [email protected] [email protected] [email protected] Matthias Zwicker University of Bern, and University of Maryland, College Park [email protected] Abstract In this paper we introduce a natural image prior that directly represents a Gaussian- smoothed version of the natural image distribution. We include our prior in a formulation of image restoration as a Bayes estimator that also allows us to solve noise-blind image restoration problems. We show that the gradient of our prior corresponds to the mean-shift vector on the natural image distribution. In addition, we learn the mean-shift vector field using denoising autoencoders, and use it in a gradient descent approach to perform Bayes risk minimization. We demonstrate competitive results for noise-blind deblurring, super-resolution, and demosaicing. 1 Introduction Image restoration tasks, such as deblurring and denoising, are ill-posed problems, whose solution requires effective image priors. In the last decades, several natural image priors have been proposed, including total variation [29], gradient sparsity priors [12], models based on image patches [5], and Gaussian mixtures of local filters [25], just to name a few of the most successful ideas. See Figure 1 for a visual comparison of some popular priors. More recently, deep learning techniques have been used to construct generic image priors. Here, we propose an image prior that is directly based on an estimate of the natural image probability distribution. Although this seems like the most intuitive and straightforward idea to formulate a prior, only few previous techniques have taken this route [20].
    [Show full text]
  • 4 Nonlinear Regression and Multilayer Perceptron
    4 Nonlinear Regression and Multilayer Perceptron In nonlinear regression the output variable y is no longer a linear function of the regression parameters plus additive noise. This means that estimation of the parameters is harder. It does not reduce to minimizing a convex energy functions { unlike the methods we described earlier. The perceptron is an analogy to the neural networks in the brain (over-simplified). It Pd receives a set of inputs y = j=1 !jxj + !0, see Figure (3). Figure 3: Idealized neuron implementing a perceptron. It has a threshold function which can be hard or soft. The hard one is ζ(a) = 1; if a > 0, ζ(a) = 0; otherwise. The soft one is y = σ(~!T ~x) = 1=(1 + e~!T ~x), where σ(·) is the sigmoid function. There are a variety of different algorithms to train a perceptron from labeled examples. Example: The quadratic error: t t 1 t t 2 E(~!j~x ; y ) = 2 (y − ~! · ~x ) ; for which the update rule is ∆!t = −∆ @E = +∆(yt~! · ~xt)~xt. Introducing the sigmoid j @!j function rt = sigmoid(~!T ~xt), we have t t P t t t t E(~!j~x ; y ) = − i ri log yi + (1 − ri) log(1 − yi ) , and the update rule is t t t t ∆!j = −η(r − y )xj, where η is the learning factor. I.e, the update rule is the learning factor × (desired output { actual output)× input. 10 4.1 Multilayer Perceptrons Multilayer perceptrons were developed to address the limitations of perceptrons (introduced in subsection 2.1) { i.e.
    [Show full text]
  • Comparative Analysis of Clustering Techniques for Movie Recommendation
    MATEC Web of Conferences 225, 02004 (2018) https://doi.org/10.1051/matecconf/201822502004 UTP-UMP-VIT SES 2018 Comparative Analysis of Clustering Techniques for Movie Recommendation Aditya TS1, Karthik Rajaraman2, and M. Monica Subashini3,* 1School of Electronics Engineering, VIT University, Vellore, India. 2School of Computer Science Engineering, VIT University, Vellore, India 3School of Electrical Engineering, VIT University, Vellore, India Abstract. Movie recommendation is a subject with immense ambiguity. A person might like a movie but not a very similar movie. The present recommending systems focus more on just few parameters such as Director, cast and genre. A lot of Power intensive methods such as Deep Convolutional Neural Network (CNN) has been used which demands the use of Graphics processors that require more energy. We try to accomplish the same task using lesser Energy consuming algorithms such as clustering techniques. In this paper, we try to create a more generalized list of similar movies in order to provide the user with more variety of movies which he/she might like, using clustering algorithms. We will compare how choosing different parameters and number of features affect the cluster's content. Also, compare how different algorithms such as K-mean, Hierarchical, Birch and mean shift clustering algorithms give a varied result and conclude which method will suit for which scenarios of movie recommendations. We also conclude on which algorithm clusters stray data points more efficiently and how different algorithms provide different advantages and disadvantages. 1 Introduction A movie recommendation system using four different clustering algorithms is built on the same cleaned dataset with identical features.
    [Show full text]
  • The Variable Bandwidth Mean Shift and Data-Driven Scale Selection
    The Variable Bandwidth Mean Shift and Data-Driven Scale Selection Dorin Comaniciu Visvanathan Ramesh Peter Meer Imaging & Visualization Department Electrical & Computer Engineering Department Siemens Corp orate Research Rutgers University 755 College Road East, Princeton, NJ 08540 94 Brett Road, Piscataway, NJ 08855 Abstract where the d-dimensional vectors fx g representa i i=1:::n random sample from some unknown density f and the Wepresenttwo solutions for the scale selection prob- kernel, K , is taken to be a radially symmetric, non- lem in computer vision. The rst one is completely non- negative function centered at zero and integrating to parametric and is based on the the adaptive estimation one. The terminology xedbandwidth is due to the fact of the normalized density gradient. Employing the sam- d that h is held constant across x 2 R . As a result, the ple p oint estimator, we de ne the Variable Bandwidth xed bandwidth pro cedure (1) estimates the densityat Mean Shift, prove its convergence, and show its sup eri- eachpoint x by taking the average of identically scaled orityover the xed bandwidth pro cedure. The second kernels centered at each of the data p oints. technique has a semiparametric nature and imp oses a For p ointwise estimation, the classical measure of the lo cal structure on the data to extract reliable scale in- ^ closeness of the estimator f to its target value f is the formation. The lo cal scale of the underlying densityis mean squared error (MSE), equal to the sum of the taken as the bandwidth which maximizes the magni- variance and squared bias tude of the normalized mean shift vector.
    [Show full text]
  • A Comparison of Clustering Algorithms for Face Clustering
    University of Groningen Research internship A comparison of clustering algorithms for face clustering Supervisors: Author: A.Sobiecki A.F. Bijl (2581582) Dr M.H.F.Wilkinson Groningen, the Netherlands, July 24, 2018 1 Abstract Video surveillance methods become increasingly widespread and popular in many organizations, including law enforcement, traffic con- trol and residential applications. In particular, the police performs inves- tigations based on searching specific people in videos and in pictures. Because the number of such videos is increasing, manual examination of all frames becomes impossible. Some degree of automation is strongly needed. Face clustering is a method to group faces of people into clusters contain- ing images of one single person. In the current study several clustering algorithms are described and applied on different datasets. The five clustering algorithms are: k-means, threshold clustering, mean shift, DBSCAN and Approximate Rank-Order. In the first experiments these clustering techniques are applied on subsets and the whole Labeled Faces in the Wild (LFW) dataset. Also a dataset containing faces of people appearing in videos of ISIS is tested to evaluate the performance of these clustering algorithms. The main finding is that threshold clustering shows the best performance in terms of the f-measure and amount of false positives. Also DBSCAN has shown good performance during our experiments and is considered as a good algorithm for face clustering. In addition it is discouraged to use k-means and for large datsets
    [Show full text]
  • Generalized Mean Shift with Triangular Kernel Profile Arxiv:2001.02165V1 [Cs.LG] 7 Jan 2020
    Generalized mean shift with triangular kernel profile S. Razakarivony, A. Barrau .∗ January 8, 2020 Abstract The mean shift algorithm is a popular way to find modes of some probability density functions taking a specific kernel-based shape, used for clustering or visual tracking. Since its introduction, it underwent several practical improvements and generalizations, as well as deep theoretical analysis mainly focused on its convergence properties. In spite of encouraging results, this question has not received a clear general answer yet. In this paper we focus on a specific class of kernels, adapted in particular to the distributions clustering applications which motivated this work. We show that a novel Mean Shift variant adapted to them can be derived, and proved to converge after a finite number of iterations. In order to situate this new class of methods in the general picture of the Mean Shift theory, we alo give a synthetic exposure of existing results of this field. 1 Introduction The mean shift algorithm is a simple iterative algorithm introduced in [8], which became over years a cornerstone of data clustering. Technically, its purpose is to find local maxima of a function having the following shape: 2 ! 1 X jjx − xijj arXiv:2001.02165v1 [cs.LG] 7 Jan 2020 f(x) = k ; (1) Nhq h i q where (xi)1≤i≤N is a set of data points belonging to a vector space R , h is a scale factor, k(:) is a q convex and decreasing function from R≥0 to R≥0 and ||·|| denotes the Euclidean norm in R . This optimization is done by means of a sequence n ! x^n, which we will explicite in Sect.
    [Show full text]
  • Petroleum and Coal
    Petroleum and Coal Article Open Access IMPROVED ESTIMATION OF PERMEABILITY OF NATURALLY FRACTURED CARBONATE OIL RESER- VOIRS USING WAVENET APPROACH Mohammad M. Bahramian 1, Abbas Khaksar-Manshad 1*, Nader Fathianpour 2, Amir H. Mohammadi 3, Bengin Masih Awdel Herki 4, Jagar Abdulazez Ali 4 1 Department of Petroleum Engineering, Abadan Faculty of Petroleum Engineering, Petroleum University of Technology (PUT), Abadan, Iran 2 Department of Mining Engineering, Isfahan University of Technology, Isfahan, Iran 3 Discipline of Chemical Engineering, School of Engineering, University of KwaZulu-Natal, Howard College Campus, King George V Avenue, Durban 4041, South Africa 4 Faculty of Engineering, Soran University, Kurdistan-Iraq Received June 18, 2017; Accepted September 30, 2017 Abstract One of the most important parameters which is regarded in petroleum industry is permeability that its accurate knowledge allows petroleum engineers to have adequate tools to evaluate and minimize the risk and uncertainty in the exploration of oil and gas reservoirs. Different direct and indirect methods are used to measure this parameter most of which (e.g. core analysis) are very time -consuming and cost consuming. Hence, applying an efficient method that can model this important parameter has the highest importance. Most of the researches show that the capability (i.e. classification, optimization and data mining) of an Artificial Neural Network (ANN) is suitable for imperfections found in petroleum engineering problems considering its successful application. In this study, we have used a Wavenet Neural Network (WNN) model, which was constructed by combining neural network and wavelet theory to improve estimation of permeability. To achieve this goal, we have also developed MLP and RBF networks and then compared the results of the latter two networks with the results of WNN model to select the best estimator between them.
    [Show full text]
  • Learning Sabatelli, Matthia; Loupe, Gilles; Geurts, Pierre; Wiering, Marco
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by University of Groningen University of Groningen Deep Quality-Value (DQV) Learning Sabatelli, Matthia; Loupe, Gilles; Geurts, Pierre; Wiering, Marco Published in: ArXiv IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from it. Please check the document version below. Document Version Early version, also known as pre-print Publication date: 2018 Link to publication in University of Groningen/UMCG research database Citation for published version (APA): Sabatelli, M., Loupe, G., Geurts, P., & Wiering, M. (2018). Deep Quality-Value (DQV) Learning. ArXiv. Copyright Other than for strictly personal use, it is not permitted to download or to forward/distribute the text or part of it without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license (like Creative Commons). Take-down policy If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. Downloaded from the University of Groningen/UMCG research database (Pure): http://www.rug.nl/research/portal. For technical reasons the number of authors shown on this cover page is limited to 10 maximum. Download date: 19-05-2020 Deep Quality-Value (DQV) Learning Matthia Sabatelli Gilles Louppe Pierre Geurts Montefiore Institute Montefiore Institute Montefiore Institute Université de Liège, Belgium Université de Liège, Belgium Université de Liège, Belgium [email protected] [email protected] [email protected] Marco A.
    [Show full text]