{"id":3689,"date":"2018-02-25T12:53:47","date_gmt":"2018-02-25T12:53:47","guid":{"rendered":"https:\/\/datakeen.co\/3-deep-learning-architectures-explained-in-human-language\/"},"modified":"2021-08-27T10:31:22","modified_gmt":"2021-08-27T10:31:22","slug":"3-deep-learning-architectures-explained-in-human-language","status":"publish","type":"post","link":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/","title":{"rendered":"3 Deep Learning Architectures explained in Human Language"},"content":{"rendered":"<p>[et_pb_section fb_built=&#8221;1&#8243; _builder_version=&#8221;3.22&#8243; global_colors_info=&#8221;{}&#8221;][et_pb_row _builder_version=&#8221;3.25&#8243; background_size=&#8221;initial&#8221; background_position=&#8221;top_left&#8221; background_repeat=&#8221;repeat&#8221; global_colors_info=&#8221;{}&#8221;][et_pb_column type=&#8221;4_4&#8243; _builder_version=&#8221;3.25&#8243; custom_padding=&#8221;|||&#8221; global_colors_info=&#8221;{}&#8221; custom_padding__hover=&#8221;|||&#8221;][et_pb_text _builder_version=&#8221;4.10.4&#8243; background_size=&#8221;initial&#8221; background_position=&#8221;top_left&#8221; background_repeat=&#8221;repeat&#8221; hover_enabled=&#8221;0&#8243; global_colors_info=&#8221;{}&#8221; sticky_enabled=&#8221;0&#8243;]<\/p>\n<p><!--:--><!--:en--><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/3_Algorithms_Deep_Learning.png\"><img decoding=\"async\" class=\"aligncenter size-full wp-image-690\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/3_Algorithms_Deep_Learning.png\" alt=\"\" width=\"1920\" height=\"1080\" srcset=\"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141504\/3_Algorithms_Deep_Learning.png 1920w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141504\/3_Algorithms_Deep_Learning-300x169.png 300w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141504\/3_Algorithms_Deep_Learning-768x432.png 768w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141504\/3_Algorithms_Deep_Learning-1024x576.png 1024w\" sizes=\"(max-width: 1920px) 100vw, 1920px\" \/><\/a><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">\u201cDeep&#8221; Learning has attracted much attention during these past years. And for a good reason: this subset of machine learning has stood out impressively in several research fields: facial recognition, speech synthesis, machine translation, and many others.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">These research fields have in common to be perceptual problems related to our senses and our expression. They have long represented a real challenge for researchers because it is extremely difficult to model vision or voice by means of algorithms and mathematical formulas.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">As a result, the first models that have been implemented in these fields have been constructed with a good deal of business expertise (in speech recognition: decomposition into phonemes, in machine translation: application of grammatical and syntactic rules). Years of research have been dedicated to the exploitation and processing of these non-structured data in order to derive meaning.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">The problem is that these new representations of data invented by researchers have failed to generalize at full extent to any text, sound or image. If you used Google Translate before 2014, year when they switched to a 100% deep learning model, you will remember the obvious limitations at the time.<\/span><\/p>\n<p><span style=\"color: #000000;\">Deep learning places itself directly on top of raw data without distortion or pre-aggregation. Then, thanks to a very large number of parameters that self-adjust over learning, will learn from implicit links existing in the data.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">Before going into details of three different algorithms * used in deep learning for different use cases, let\u2019s start by simply defining the model at the heart of deep learning: the &#8220;neural network&#8221;.<\/span><\/p>\n<p><span style=\"color: #000000;\">* We also talk about different network architectures.<\/span><\/p>\n<h3>\u00a0<\/h3>\n<h3><b>1. Neural networks<\/b><\/h3>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;\">Let me begin by saying that neural networks have very little to do with the neural system and the brain. The analogy between a neuron and a one-neuron neural network is essentially graphic, insofar as there is a flow of information from one end to the other network.<\/span><\/p>\n<p><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_107.png\"><img decoding=\"async\" class=\"aligncenter size-full wp-image-683\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_107.png\" alt=\"neural network vs neuron\" width=\"927\" height=\"411\" srcset=\"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141506\/S%C3%A9lection_107.png 927w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141506\/S%C3%A9lection_107-300x133.png 300w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141506\/S%C3%A9lection_107-768x341.png 768w\" sizes=\"(max-width: 927px) 100vw, 927px\" \/><\/a><\/p>\n<p><span style=\"color: #000000;\">The first layer of a neural network is called the input layer. It is through this layer that your data will enter the network. Prior to &#8220;feeding&#8221; the network with your data you will need to change it to numbers if they are not already.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">We&#8217;ll take the example of sentiment analysis on textual data.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">Let\u2019s say you have 10,000 comments on your ecommerce website about products sold:<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">With your team you have labeled 1000 of them (we&#8217;ll see that you can also rely on pre-trained neural networks) into 3 classes (satisfied | neutral | dissatisfied). This number of 3 classes, often taken in the sense of analysis, is an example and you can actually set more.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">&#8211; &#8220;I loved it, very good taste&#8221;;<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">&#8211; &#8220;I didn\u2019t like the packaging that much&#8221;;<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">&#8211; &#8220;I thought it was pretty good&#8221;<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">The final layer, called output layer, will provide you with the classification &#8220;satisfied \/ neutral \/ dissatisfied&#8221;.<\/span><\/p>\n<p><span style=\"color: #000000;\">And all layers between the input and output layer, layers called &#8220;hidden&#8221; are all different representations of the data. A representation may be the number of words in a sentence, the number of punctuation (?!) in a sentence, etc. You will not have to specify the network these representations; if statistically they help to correctly classify the sentences the network will teach alone.<\/span><\/p>\n<p><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_100.png\"><img decoding=\"async\" class=\"size-full wp-image-688\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_100.png\" alt=\"simple neural network\" width=\"743\" height=\"443\" srcset=\"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141622\/S%C3%A9lection_100.png 743w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141622\/S%C3%A9lection_100-300x179.png 300w\" sizes=\"(max-width: 743px) 100vw, 743px\" \/><\/a><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;\">To illustrate these layers take another example: that of the estimated price of a home.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">As we can see here we take four input variables: the size of the house, number of bedrooms, the postal code and the degree of richness of the area. The output is not seeking to classify but to predict a number: the price of the house. This is a problem known as regression.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">The italicized words to examples of representations that the neural network will make the data after having seen many.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">The parameters of the network are updated thanks to a process called &#8220;backpropagation&#8221;. And the more hidden layers there are in the network the &#8220;deeper&#8221; it is, hence the name &#8220;Deep&#8221; Learning.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">Let us now see 3 different types of architectures of neural networks.<\/span><\/p>\n<h3>\u00a0<\/h3>\n<h3><span style=\"color: #000000;\"><b>2. Convolutional Neural Networks (CNN)<\/b><\/span><\/h3>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;\">These networks are used for all use cases around image or video which include face recognition or image classification.<\/span><\/p>\n<p><span style=\"color: #000000;\">For example Bai Du (the equivalent of Google in China) has set up portals powered by face recognition to let enter only employees of the company.<\/span><\/p>\n<p><span style=\"color: #000000;\">Snapchat and many mobile applications have leveraged the breakthroughs of deep learning and CNNs to increase their face recognition capacities in order to add extra layers on your face such as funny bunny ears and a pink nose.<\/span><\/p>\n<p><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_101.png\"><img decoding=\"async\" class=\"aligncenter size-full wp-image-687\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_101.png\" alt=\"convolutional neural network\" width=\"928\" height=\"389\" srcset=\"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141623\/S%C3%A9lection_101.png 928w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141623\/S%C3%A9lection_101-300x126.png 300w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141623\/S%C3%A9lection_101-768x322.png 768w\" sizes=\"(max-width: 928px) 100vw, 928px\" \/><\/a><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">The name \u201cconvolution\u201d comes from a mathematical operation: convolution between functions.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">Put simply, the convolution applies a filter to the input image, the filter parameters are learned through the learning. A learnt filter will be able of detecting features in an image, for example angles, and use them to classify at best the image.<\/span><\/p>\n<p><span style=\"color: #000000;\"><span style=\"font-weight: 400;\">The image is first decomposed into 3 channels (R, G, B) pixels per pixel, we obtain three matrices of size <\/span><b>n x n<\/b><span style=\"font-weight: 400;\"> (where <\/span><b>n<\/b><span style=\"font-weight: 400;\"> is the number of pixels).<\/span><\/span><\/p>\n<p><span style=\"color: #000000;\"><span style=\"font-weight: 400;\">Below is an example of a convolution with a <\/span><b>6 x 6<\/b><span style=\"font-weight: 400;\"> size matrix:<\/span><\/span><\/p>\n<blockquote>\n<p><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_102.png\"><img decoding=\"async\" class=\"aligncenter size-full wp-image-686\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_102.png\" alt=\"neural network convolution\" width=\"843\" height=\"409\" srcset=\"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141623\/S%C3%A9lection_102.png 843w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141623\/S%C3%A9lection_102-300x146.png 300w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141623\/S%C3%A9lection_102-768x373.png 768w\" sizes=\"(max-width: 843px) 100vw, 843px\" \/><\/a><\/p>\n<\/blockquote>\n<p><span style=\"font-weight: 400; color: #000000;\">It is important to note two important advantages inherent to convolutional networks:<\/span><\/p>\n<ul>\n<li><span style=\"font-weight: 400; color: #000000;\">the network can learn by steps to recognize characteristics in an image. To recognize a face for instance: it will learn to recognize first of eyelids and pupils, and then recognize eyes;<\/span><\/li>\n<li><span style=\"color: #000000;\">once an item to a learned image place the network will be able to recognize it anywhere else in the picture.<\/span><\/li>\n<\/ul>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<h3>3. Recurrent neural networks (RNN)<\/h3>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">Recurrent neural networks are at the heart of many substantial improvements in areas as diverse as speech recognition, automatic music composition, sentiment analysis, DNA sequence analysis, machine translation.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">The main difference with other neural networks is that they take into account a sequence of data, often a sequence evolving over time. For example in the case of analyzing temporal data (time series) the network will still have in memory a part or all of the observations previous to the data being analyzed.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">The pattern of this network is produced here:<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_105.png\"><img decoding=\"async\" class=\"aligncenter size-full wp-image-685\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_105.png\" alt=\"recurrent neural network\" width=\"827\" height=\"463\" srcset=\"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141624\/S%C3%A9lection_105.png 827w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141624\/S%C3%A9lection_105-300x168.png 300w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141624\/S%C3%A9lection_105-768x430.png 768w\" sizes=\"(max-width: 827px) 100vw, 827px\" \/><\/a><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">Instead of taking into account separately input data (in the way a CNN would analyse image per image) the recurrent network takes into account data previously processed.<\/span><\/p>\n<p><span style=\"color: #000000;\">Some architectures, called bidirectional, can also take into account future data. For instance when analyzing text to identify named entities (people, companies, countries, etc.) the network would need to see the words of the whole sentence.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">Example:<\/span><\/p>\n<ul>\n<li><span style=\"color: #000000;\"><span style=\"font-weight: 400;\">&#8220;<\/span><span style=\"font-weight: 400;\">I see<\/span><span style=\"font-weight: 400;\"> [Jean] Valjean still have escaped you, Javert!&#8221;<\/span><\/span><\/li>\n<li><span style=\"color: #000000;\">&#8220;<span style=\"font-weight: 400;\">I see <\/span><span style=\"font-weight: 400;\">[Jean] R. plays in this adaptation of \u2018Les Mis\u00e9rables\u2019\u201d.<\/span><\/span><\/li>\n<\/ul>\n<p><span style=\"font-weight: 400; color: #000000;\">The beginning of the sentence (underlined) is not enough to identify who is \u2018Jean\u2019.<\/span><\/p>\n<p>&nbsp;<\/p>\n<h3><span style=\"color: #000000;\"><b>4. Autoencoders<\/b><\/span><\/h3>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;\">Autoencoders are applied mainly to anomaly detection (for example to detect fraud in banking or to find faults in an industrial production line). They can also be used in dimensionality reduction (close to the objective of a Principal Component Analysis). Indeed the goal of autoencoders is to teach the machine what constitutes &#8220;normal&#8221; data.<\/span><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">The architecture of our network is the following:<\/span><\/p>\n<p><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_106.png\"><img decoding=\"async\" class=\"aligncenter size-full wp-image-684\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_106.png\" alt=\"autoencoder\" width=\"812\" height=\"474\" srcset=\"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141624\/S%C3%A9lection_106.png 812w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141624\/S%C3%A9lection_106-300x175.png 300w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141624\/S%C3%A9lection_106-768x448.png 768w\" sizes=\"(max-width: 812px) 100vw, 812px\" \/><\/a><\/p>\n<p><span style=\"color: #000000;\">The network will therefore represent data through one or more hidden layers so that the output will be as close as possible to the input data.<\/span><\/p>\n<p><span style=\"color: #000000;\">The objective to find the same data back as the output of the network is characteristic of autoencoders (analogous to the identity function f (x) = x).<\/span><\/p>\n<p><span style=\"color: #000000;\">The encoding and decoding stage it is not however specific to autoencoders. Indeed, they are found in machine translation in recurrent neural networks.<\/span><\/p>\n<p><span style=\"color: #000000;\">After training the network with enough data it will be possible to identify suspicious or anomalous observations when they exceed a certain threshold compared to the new &#8220;standard&#8221;.<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #000000;\"><b>Conclusion:<\/b><\/span><\/p>\n<p><span style=\"color: #000000;\">We saw 3 major types of neural networks:<\/span><\/p>\n<ul>\n<li><span style=\"color: #000000;\">Convolution networks with applications in facial recognition and image classification;<\/span><\/li>\n<li><span style=\"color: #000000;\">Recurrent networks with applications in the timeseries, text and voice analysis;<\/span><\/li>\n<li><span style=\"color: #000000;\">Autoencoders with applications to anomaly detection as well as dimensionality reduction.<\/span><\/li>\n<\/ul>\n<p><span style=\"color: #000000;\">Other architectures exist such as GANs, generative adversarial networks, which are composed of a model generating candidates for a given task, for example image creation, and another that evaluates the different outputs. Or Reinforcement Learning, a method used by Deepmind to train their Alpha Go and Alpha Go Zero models.<\/span><\/p>\n<p><span style=\"color: #000000;\">Obviously there are limits: for example it is possible to fool convolutional network by adding a particular sound to image undetectable to the human eye but can be fatal for a model that has not been sufficiently tested robustness. New architectures such as capsule networks have merged to face this particular problem.<\/span><\/p>\n<p><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_099.png\"><img decoding=\"async\" class=\"aligncenter size-full wp-image-689\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_099.png\" alt=\"\" width=\"863\" height=\"346\" srcset=\"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141622\/S%C3%A9lection_099.png 863w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141622\/S%C3%A9lection_099-300x120.png 300w, https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141622\/S%C3%A9lection_099-768x308.png 768w\" sizes=\"(max-width: 863px) 100vw, 863px\" \/><\/a><\/p>\n<p><span style=\"font-weight: 400; color: #000000;\">All in all it is certain that deep learning has a bright future with many business applications to come.<\/span><\/p>\n<p><span style=\"color: #000000;\"><span style=\"font-weight: 400;\">Ga\u00ebl Bonnardot,<\/span><span style=\"font-weight: 400;\"><br \/><\/span><em><span style=\"font-weight: 400;\"><br \/><\/span><span style=\"font-weight: 400;\">Cofounder and CTO at <strong>Datakeen<\/strong><\/span><\/em><span style=\"font-weight: 400;\"><br \/><\/span><span style=\"font-weight: 400;\"><br \/><\/span><span style=\"font-weight: 400;\">At Datakeen we seek to simplify the use and understanding of new machine learning paradigms by the business functions of all industries.<\/span><span style=\"font-weight: 400;\"><br \/><\/span><span style=\"font-weight: 400;\"><br \/><\/span><span style=\"font-weight: 400;\">Contact us for more information: contact@datakeen.co<\/span><\/span><!--:--><!--:de--><!--:--><\/p>\n<p>[\/et_pb_text][\/et_pb_column][\/et_pb_row][\/et_pb_section]<\/p>\n","protected":false},"excerpt":{"rendered":"<p>&nbsp; \u201cDeep&#8221; Learning has attracted much attention during these past years. And for a good reason: this subset of machine learning has stood out impressively in several research fields: facial recognition, speech synthesis, machine translation, and many others. These research fields have in common to be perceptual problems related to our senses and our expression. [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":3691,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_yoast_wpseo_focuskw":"","_yoast_wpseo_title":"","_yoast_wpseo_metadesc":"","_et_pb_use_builder":"on","_et_pb_old_content":"<!--:fr--><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/3_Algorithms_Deep_Learning.png\"><img class=\"aligncenter size-full wp-image-690\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/3_Algorithms_Deep_Learning.png\" alt=\"\" width=\"1920\" height=\"1080\" \/><\/a>\r\n\r\n<span style=\"color: #000000;\">L'apprentissage \u201cprofond\u201d ou \u201cdeep learning\u201d fait beaucoup parler de lui ces derni\u00e8res ann\u00e9es. Et pour cause, ce sous ensemble de l'apprentissage machine ('machine learning'') s'est impos\u00e9 de mani\u00e8re impressionnante dans plusieurs champs de recherche: reconnaissance faciale, synth\u00e8se vocale, traduction automatique, et bien d'autres.<\/span>\r\n\r\n<span style=\"color: #000000;\">Ces champs de recherche ont pour point commun d'\u00eatre des probl\u00e8mes perceptifs, li\u00e9s \u00e0 nos sens et \u00e0 notre expression. Ils ont ainsi represent\u00e9 pendant longtemps un veritable d\u00e9fi pour les chercheurs car il est extr\u00eamement difficile de traduire la vue ou la voix aux moyens d'algorithmes et de formules math\u00e9matiques.<\/span>\r\n\r\n<span style=\"color: #000000;\">Il en r\u00e9sulte que les premiers mod\u00e8les qui ont \u00e9t\u00e9 mis en place dans ces domaines ont \u00e9t\u00e9 construits \u00e0 partir d'une certaine expertise m\u00e9tier (pour la reconnaissance vocale: la d\u00e9composition en phon\u00e8mes, pour la traduction machine: le passage par des r\u00e8gles grammaticales et syntaxiques). Des ann\u00e9es de recherche ont \u00e9t\u00e9 consacr\u00e9es \u00e0 l'exploitation et \u00e0 la transformation de ces donn\u00e9es non structur\u00e9es de mani\u00e8re \u00e0 en tirer du sens.<\/span>\r\n\r\n<span style=\"color: #000000;\">Le probl\u00e8me est que ces nouvelles repr\u00e9sentaitons des donn\u00e9es invent\u00e9s par les chercheurs se sont heurt\u00e9s \u00e0 la g\u00e9n\u00e9ralisation: \u00e0 tout texte, image ou son. Si vous avez utilis\u00e9 Google Tranlate avant 2014, ann\u00e9e \u00e0 laquelle ils ont bascul\u00e9 \u00e0 100% sur de l'apprentissage profond, vous vous souviendrez des limites \u00e9videntes \u00e0 l'\u00e9poque.<\/span>\r\n\r\n<span style=\"color: #000000;\">L'apprentissage profond propose de se placer directement au niveau des donn\u00e9es sans d\u00e9formation ou aggr\u00e9gation pr\u00e9alable. Puis, au moyen d'un nombre extr\u00eamement grand de param\u00e8tres qui s'auto-ajustent au fil de l'apprentissage, le r\u00e9seau va apprendre de lui m\u00eame les liens implicites qui existent dans les donn\u00e9es.<\/span>\r\n\r\n<span style=\"color: #000000;\">Avant de rentrer dans le d\u00e9tail de trois diff\u00e9rents algorithmes* utilis\u00e9s en apprentissage profond pour diff\u00e9rents cas d'usage, commencons par d\u00e9finir simplement le mod\u00e8le au coeur de l'apprentissage profond: le \u201cr\u00e9seau de neurones\u201d.<\/span>\r\n\r\n<span style=\"color: #000000;\">*On parle \u00e9galement d'architectures diff\u00e9rente de r\u00e9seaux.<\/span>\r\n<h3><b>1. Les R\u00e9seaux de Neurones\r\n<\/b><\/h3>\r\n<span style=\"color: #000000;\">Disons le tout de suite les r\u00e9seaux de neurones n'ont que tr\u00e8s peu \u00e0 voir avec le syst\u00e8me neuronal et le cerveau. L'analogie entre un neurone et un r\u00e9seau de neurones \u00e0 une couche est essentiellement graphique, dans la mesure o\u00f9 il y a un flux d'information d'un bout \u00e0 l'autre du r\u00e9seau.<\/span>\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_107.png\"><img class=\"aligncenter size-full wp-image-683\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_107.png\" alt=\"neural network vs neuron\" width=\"927\" height=\"411\" \/><\/a><span style=\"color: #000000;\">La premi\u00e8re couche d'un r\u00e9seau de neurones est celle d'entr\u00e9e (input). C'est par cette couche que vont rentrer les donn\u00e9es dont vous disposez. Avant de pouvoir \u201cnourrir\u201d le r\u00e9seau, il faudra pr\u00e9alablement transformer vos donn\u00e9es en nombres si elles n'en sont pas d\u00e9j\u00e0.<\/span>\r\n\r\n<span style=\"color: #000000;\">Nous allons prendre l'exemple de l'analyse de sentiment d'un texte.<\/span>\r\n\r\n<span style=\"color: #000000;\">Vous avez 10 000 commentaires sur votre site \u00e0 propos de produits vendus:<\/span>\r\n<span style=\"color: #000000;\">Avec votre \u00e9quipe vous avez labellis\u00e9 1000 d'entre eux (nous allons voir que vous pouvez \u00e9galement vous appuyer sur des r\u00e9seaux de neurones pr\u00e9-entra\u00een\u00e9s) en 3 classes (satisfait | neutre | insatisfait). Ce nombre de 3 classes, souvent utilis\u00e9 dans l'analyse de sentiment, est un exemple et vous pouvez en r\u00e9alit\u00e9 en d\u00e9finir plus.<\/span>\r\n\r\n<span style=\"color: #000000;\">- \u201cJ'ai ador\u00e9, tr\u00e8s bon go\u00fbt\u201d;<\/span>\r\n<span style=\"color: #000000;\">- \u201cJe n'ai pas beaucoup aim\u00e9 l'emballage\u201d;<\/span>\r\n<span style=\"color: #000000;\">- \u201cJ'ai trouv\u00e9 \u00e7a plut\u00f4t bon\u201d<\/span>\r\n\r\n<span style=\"color: #000000;\">La couche finale, dite de sortie (output), va vous fournir la classification \u201csatisfait \/ neutre \/ insatisfait\u201d.<\/span>\r\n\r\n<span style=\"color: #000000;\">Et toutes les couches entre la couche d'entr\u00e9e et de sortie, des couches dites \u201ccach\u00e9es\u201d, sont autant de repr\u00e9sentations diff\u00e9rentes des donn\u00e9es. Par repr\u00e9sentation cela peut \u00eatre le nombre de mots dans une phrase, le nombre de ponctuations (?!) dans une phrase, etc. Vous n'aurez pas \u00e0 pr\u00e9ciser au r\u00e9seau ces repr\u00e9sentations; si statistiquement celles-ci aident \u00e0 classifier correctement les phrases le r\u00e9seau les apprendra tout seul.<\/span>\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_100.png\"><img class=\"size-full wp-image-688\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_100.png\" alt=\"simple neural network\" width=\"743\" height=\"443\" \/><\/a>\r\n\r\n<span style=\"color: #000000;\">Pour illustrer ces couches prenons un autre exemple: celui de l'estimation du prix d'une maison.<\/span>\r\n<span style=\"color: #000000;\">Comme nous pouvons le voir on prend ici 4 variables en entr\u00e9e: la superficie de la maison, le nombre de chambres, le code postal et le degr\u00e9 de richesse du quartier. En sortie on cherche non plus \u00e0 classifier mais \u00e0 pr\u00e9dire un nombre: le prix de la maison. C'est un probl\u00e8me dit de regression.<\/span>\r\n<span style=\"color: #000000;\">Les mots en italique renvoient \u00e0 des exemples de repr\u00e9sentations que le r\u00e9seau de neurones va faire des donn\u00e9es apr\u00e8s en avoir vu un grand nombre.<\/span>\r\n\r\n<span style=\"color: #000000;\">Les param\u00e8tres du r\u00e9seau sont mis \u00e0 jour gr\u00e2ce \u00e0 un processus appel\u00e9 \"r\u00e9tropropagation\". Plus il y a de couches cach\u00e9es dans le r\u00e9seau, plus on dit qu'il est \"profond\", d'o\u00f9 le nom d'apprentissage \"profond\".<\/span>\r\n\r\n<span style=\"color: #000000;\">Voyons maintenant 3 types architectures diff\u00e9rents de r\u00e9seaux de neurone.<\/span>\r\n<h3><\/h3>\r\n<h3><span style=\"color: #000000;\"><b>2. Les r\u00e9seaux de neurones dits convolutifs (CNN)<\/b><\/span><\/h3>\r\n\u00a0\r\n\r\n<span style=\"color: #000000;\">Ces r\u00e9seaux sont utilis\u00e9s pour tout usage autour de l'image ou de la vid\u00e9o dont fait partie la reconnaissance faciale ou encore la classification d'image.<\/span>\r\n\r\n<span style=\"color: #000000;\">L'entreprise Bai Du (l'\u00e9quivalent de Google en Chine) a par exemple mis en place des portiques actionn\u00e9s par la reconnaissance visuelle qui laissent passer uniquement leurs employ\u00e9s.<\/span>\r\n\r\n<span style=\"color: #000000;\">Snapchat et de nombreuses applications mobiles ont utilis\u00e9 la perc\u00e9e de l'apprentissage et des CNN pour augmenter leurs fonctionnalit\u00e9s de \u201cfiltres\u201d.<\/span>\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_101.png\"><img class=\"aligncenter size-full wp-image-687\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_101.png\" alt=\"convolutional neural network\" width=\"928\" height=\"389\" \/><\/a><span style=\"color: #000000;\">Le nom r\u00e9seau convolutif renvoit \u00e0 un terme math\u00e9matique: le produit de convolution.<\/span>\r\n<span style=\"color: #000000;\">En termes simples, l'id\u00e9e est qu'on applique un filtre \u00e0 l'image d'entr\u00e9e, les param\u00e8tres du filtre seront appris au fur et \u00e0 mesure de l'apprentissage. Un filtre appris permettra par exemple de d\u00e9tecter les angles dans une image si les angles servent \u00e0 classifier au mieux l'image.<\/span>\r\n\r\n<span style=\"color: #000000;\">L'image est d'abord d\u00e9compos\u00e9 dans les 3 cannaux (R,G,B) pixels par pixels, on obtient donc 3 matrices de taille <strong>n x n<\/strong> (o\u00f9 <strong>n<\/strong> est le nombre de pixels).<\/span>\r\n\r\n<span style=\"color: #000000;\">Voici un exemple de convolution avec une matrice de taille<strong> 6 x 6<\/strong> :<\/span>\r\n\r\n<span style=\"color: #000000;\"><span style=\"font-weight: 400;\">:<\/span><\/span>\r\n<blockquote><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_102.png\"><img class=\"aligncenter size-full wp-image-686\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_102.png\" alt=\"neural network convolution\" width=\"843\" height=\"409\" \/><\/a><\/blockquote>\r\n<span style=\"color: #000000;\">Il est important de noter deux avantages importants inh\u00e9rents aux r\u00e9seaux convolutifs :<\/span>\r\n<ul>\r\n \t<li><span style=\"color: #000000;\">le r\u00e9seau peut apprendre par \u00e9tape \u00e0 reconnaitre les \u00e9l\u00e9ments caract\u00e9ristiques d'une image. Pour reconna\u00eetre un visage par exemple: il apprendra \u00e0 reconna\u00eetre d'abord des paupi\u00e8res, des pupilles, pour arriver \u00e0 identifier des yeux;<\/span><\/li>\r\n \t<li><span style=\"color: #000000;\">une fois un \u00e9l\u00e9ment appris \u00e0 un endroit de l'image le r\u00e9seau sera capable de le reconna\u00eetre n'importe o\u00f9 d'autre dans l'image.<\/span><\/li>\r\n<\/ul>\r\n\u00a0\r\n<h3>3. Les r\u00e9seaux de neurones dits r\u00e9currents (RNN)<\/h3>\r\n\u00a0\r\n\r\n<span style=\"color: #000000;\">Les r\u00e9seaux de neurones r\u00e9currents sont au coeur de bon nombre d'am\u00e9liorations substantiels dans des domaines aussi divers que la reconnaissance vocale, la composition automatique de musique, l'analyse de sentiments, l'analyse de s\u00e9quence ADN, la traduction automatique.<\/span>\r\n\r\n<span style=\"color: #000000;\">La diff\u00e9rence principal avec les autres r\u00e9seaux de neurones vient du fait que ces derniers tiennent compte de l'encha\u00eenement successif des donn\u00e9es, bien souvent de leur encha\u00eenement dans le temps. Par exemple dans le cas de l'analyse d'une s\u00e9rie de mesures de capteurs (s\u00e9ries temporelles) le r\u00e9seau aura encore en m\u00e9moire tout ou partie des observations pr\u00e9c\u00e9dentes.<\/span>\r\n\r\n<span style=\"color: #000000;\">Le sch\u00e9ma de ce r\u00e9seau est produit ici:<\/span>\r\n\r\n\u00a0\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_105.png\"><img class=\"aligncenter size-full wp-image-685\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_105.png\" alt=\"recurrent neural network\" width=\"827\" height=\"463\" \/><\/a>\r\n\r\n<span style=\"color: #000000;\">Au lieu de prendre en compte les donn\u00e9es d'entr\u00e9e de mani\u00e8re s\u00e9par\u00e9e (comme un CNN analyse image par image) le r\u00e9seau r\u00e9current lui prend en compte les donn\u00e9es d'entr\u00e9e pass\u00e9es.<\/span>\r\n\r\n<span style=\"color: #000000;\">Certaines architectures, dites bidirectionnelles, peuvent aussi prendre en compte les donn\u00e9es futures. Par exemple lors d'une analyse de texte o\u00f9 on cherche \u00e0 trouver des entit\u00e9s nomm\u00e9es (noms de personnes, soci\u00e9t\u00e9s, pays, etc.) n\u00e9cessite de voir les mots de toute la phrase.<\/span>\r\n\r\n<span style=\"color: #000000;\">Exemple:<\/span>\r\n<ul>\r\n \t<li><span style=\"color: #000000;\">\u201cJe vois que [Jean] Valjean t'as encore \u00e9chapp\u00e9, Javert!\u201d<\/span><\/li>\r\n \t<li><span style=\"color: #000000;\">\u201cJe vois que [Jean] R. joue \u00e9galement dans l'adaptation des Mis\u00e9rables.\u201d<\/span><\/li>\r\n<\/ul>\r\n<span style=\"color: #000000;\">Le d\u00e9but de phrase ne suffit pas \u00e0 identifier qui est \u201cJean\u201d.<\/span>\r\n\r\n\u00a0\r\n<h3><span style=\"color: #000000;\"><b>4. Les Auto encodeurs<\/b><\/span><\/h3>\r\n\u00a0\r\n\r\n<span style=\"color: #000000;\">Les auto encodeurs sont appliqu\u00e9s principalement \u00e0 la d\u00e9tection d'anomalie ( par exemple pour d\u00e9tecter la fraude en banque ou bien pour trouver des anomalies dans une ligne de production industrielle ). Ils peuvent \u00e9galement servir \u00e0 la r\u00e9duction de dimension (proche dans l'id\u00e9e d'une Analyse en Composante Principale). En effet le but des auto encodeurs est d'apprendre \u00e0 la machine en quoi consiste des observations \u201cnormales\u201d.\r\nL'architecture de notre r\u00e9seau est le suivant:<\/span>\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_106.png\"><img class=\"aligncenter size-full wp-image-684\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_106.png\" alt=\"autoencoder\" width=\"812\" height=\"474\" \/><\/a>\r\n\r\n<span style=\"color: #000000;\">Le r\u00e9seau va donc repr\u00e9senter les donn\u00e9es au moyen d'une ou plusieurs couches cach\u00e9es de sorte \u00e0 ce qu'en sortie on retrouve les m\u00eames donn\u00e9es qu'en entr\u00e9e.<\/span>\r\n\r\n<span style=\"color: #000000;\">L'objectif de retrouver en sortie les memes donn\u00e9es qu'en entr\u00e9e est caract\u00e9ristique des auto-encodeurs (analogue \u00e0 la fonction identit\u00e9 f(x)=x).<\/span>\r\n<span style=\"color: #000000;\">La phase d'encodage et de d\u00e9codage n'est elle pas propre aux auto-encodeurs. En effet, on les retrouve dans la traduction automatique dans des r\u00e9seaux de neurones r\u00e9currents.<\/span>\r\n\r\n<span style=\"color: #000000;\">Apr\u00e8s avoir entra\u00een\u00e9 le r\u00e9seau avec suffisamment de donn\u00e9es il sera possible d'identifier des observations suspectes ou anormales lorsque celles-ci d\u00e9passent un certain seuil par rapport \u00e0 la nouvelle \u201cnorme\u201d.<\/span>\r\n\r\n\u00a0\r\n\r\n<span style=\"color: #000000;\"><b>Conclusion:<\/b><\/span>\r\n\r\n<span style=\"color: #000000;\">Nous avons vu 3 grands types de r\u00e9seau de neurones :<\/span>\r\n<ol>\r\n \t<li><span style=\"color: #000000;\">les r\u00e9seaux convolutifs avec des applications dans la reconnaissance faciale et classification d'images<\/span><\/li>\r\n \t<li><span style=\"color: #000000;\">les r\u00e9seaux r\u00e9currents avec des applications dans l'analyse du texte et de la voix;<\/span><\/li>\r\n \t<li><span style=\"color: #000000;\">les auto encodeurs avec des applications \u00e0 la d\u00e9tection d'anomalie et \u00e0 la r\u00e9duction de dimension<\/span><\/li>\r\n<\/ol>\r\n<span style=\"color: #000000;\">D'autres architectures existent comme les GAN, r\u00e9seaux antagonistes g\u00e9n\u00e9ratifs, qui sont compos\u00e9s d'un mod\u00e8le g\u00e9n\u00e9rant des candidats pour une tache donn\u00e9e, par exemple synth\u00e9tiser une image, et d'un autre qui les \u00e9valuent. Ou encore l'apprentissage par renforcement, m\u00e9thode utilis\u00e9e par DeepMind pour entrainer leurs mod\u00e8les Alpha Go et Alpha Go z\u00e9ro.<\/span>\r\n\r\n<span style=\"color: #000000;\">Bien \u00e9videmment des limites existent: par exemple il est possible de tromper des r\u00e9seaux convolutifs en ajoutant un bruit particulier \u00e0 des images non d\u00e9tectable \u00e0 l'oeil humain mais qui peut etre fatal pour un mod\u00e8le qui n'a pas \u00e9t\u00e9 assez test\u00e9 en robustesse. De nouvelles architectures telles que les r\u00e9seaux de capsules ont fusionn\u00e9 pour faire face \u00e0 ce probl\u00e8me particulier.<\/span>\r\n\r\n\u00a0\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_099.png\"><img class=\"aligncenter size-full wp-image-689\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_099.png\" alt=\"\" width=\"863\" height=\"346\" \/><\/a>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">Il est certain que l'apprentissage profond a encore de beaux jours devant lui avec de nombreuses nouvelles applications pour les entreprises \u00e0 venir.<\/span>\r\n\r\n<span style=\"color: #000000;\"><span style=\"font-weight: 400;\">Ga\u00ebl Bonnardot,<\/span><span style=\"font-weight: 400;\">\r\n<\/span><em><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">Cofounder et CTO chez Datakeen<\/span><\/em><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">Chez Datakeen nous cherchons \u00e0 simplifier l'utilisation et la compr\u00e9hension des nouveaux paradigmes d'apprentissage automatique par les fonctions m\u00e9tier de toutes les industries.\r\n<\/span><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">Contact us for more information: contact@datakeen.co<\/span><\/span><!--:--><!--:en--><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/3_Algorithms_Deep_Learning.png\"><img class=\"aligncenter size-full wp-image-690\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/3_Algorithms_Deep_Learning.png\" alt=\"\" width=\"1920\" height=\"1080\" \/><\/a>\r\n\r\n\u00a0\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">\u201cDeep\" Learning has attracted much attention during these past years. And for a good reason: this subset of machine learning has stood out impressively in several research fields: facial recognition, speech synthesis, machine translation, and many others.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">These research fields have in common to be perceptual problems related to our senses and our expression. They have long represented a real challenge for researchers because it is extremely difficult to model vision or voice by means of algorithms and mathematical formulas.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">As a result, the first models that have been implemented in these fields have been constructed with a good deal of business expertise (in speech recognition: decomposition into phonemes, in machine translation: application of grammatical and syntactic rules). Years of research have been dedicated to the exploitation and processing of these non-structured data in order to derive meaning.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">The problem is that these new representations of data invented by researchers have failed to generalize at full extent to any text, sound or image. If you used Google Translate before 2014, year when they switched to a 100% deep learning model, you will remember the obvious limitations at the time.<\/span>\r\n\r\n<span style=\"color: #000000;\">Deep learning places itself directly on top of raw data without distortion or pre-aggregation. Then, thanks to a very large number of parameters that self-adjust over learning, will learn from implicit links existing in the data.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">Before going into details of three different algorithms * used in deep learning for different use cases, let\u2019s start by simply defining the model at the heart of deep learning: the \"neural network\".<\/span>\r\n\r\n<span style=\"color: #000000;\">* We also talk about different network architectures.<\/span>\r\n<h3><\/h3>\r\n<h3><b>1. Neural networks<\/b><\/h3>\r\n\u00a0\r\n\r\n\u00a0\r\n\r\n<span style=\"color: #000000;\">Let me begin by saying that neural networks have very little to do with the neural system and the brain. The analogy between a neuron and a one-neuron neural network is essentially graphic, insofar as there is a flow of information from one end to the other network.<\/span>\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_107.png\"><img class=\"aligncenter size-full wp-image-683\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_107.png\" alt=\"neural network vs neuron\" width=\"927\" height=\"411\" \/><\/a>\r\n\r\n<span style=\"color: #000000;\">The first layer of a neural network is called the input layer. It is through this layer that your data will enter the network. Prior to \"feeding\" the network with your data you will need to change it to numbers if they are not already.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">We'll take the example of sentiment analysis on textual data.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">Let\u2019s say you have 10,000 comments on your ecommerce website about products sold:<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">With your team you have labeled 1000 of them (we'll see that you can also rely on pre-trained neural networks) into 3 classes (satisfied | neutral | dissatisfied). This number of 3 classes, often taken in the sense of analysis, is an example and you can actually set more.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">- \"I loved it, very good taste\";<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">- \"I didn\u2019t like the packaging that much\";<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">- \"I thought it was pretty good\"<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">The final layer, called output layer, will provide you with the classification \"satisfied \/ neutral \/ dissatisfied\".<\/span>\r\n\r\n<span style=\"color: #000000;\">And all layers between the input and output layer, layers called \"hidden\" are all different representations of the data. A representation may be the number of words in a sentence, the number of punctuation (?!) in a sentence, etc. You will not have to specify the network these representations; if statistically they help to correctly classify the sentences the network will teach alone.<\/span>\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_100.png\"><img class=\"size-full wp-image-688\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_100.png\" alt=\"simple neural network\" width=\"743\" height=\"443\" \/><\/a>\r\n\r\n\u00a0\r\n\r\n<span style=\"color: #000000;\">To illustrate these layers take another example: that of the estimated price of a home.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">As we can see here we take four input variables: the size of the house, number of bedrooms, the postal code and the degree of richness of the area. The output is not seeking to classify but to predict a number: the price of the house. This is a problem known as regression.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">The italicized words to examples of representations that the neural network will make the data after having seen many.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">The parameters of the network are updated thanks to a process called \"backpropagation\". And the more hidden layers there are in the network the \"deeper\" it is, hence the name \"Deep\" Learning.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">Let us now see 3 different types of architectures of neural networks.<\/span>\r\n<h3><\/h3>\r\n<h3><span style=\"color: #000000;\"><b>2. Convolutional Neural Networks (CNN)<\/b><\/span><\/h3>\r\n\u00a0\r\n\r\n<span style=\"color: #000000;\">These networks are used for all use cases around image or video which include face recognition or image classification.<\/span>\r\n\r\n<span style=\"color: #000000;\">For example Bai Du (the equivalent of Google in China) has set up portals powered by face recognition to let enter only employees of the company.<\/span>\r\n\r\n<span style=\"color: #000000;\">Snapchat and many mobile applications have leveraged the breakthroughs of deep learning and CNNs to increase their face recognition capacities in order to add extra layers on your face such as funny bunny ears and a pink nose.<\/span>\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_101.png\"><img class=\"aligncenter size-full wp-image-687\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_101.png\" alt=\"convolutional neural network\" width=\"928\" height=\"389\" \/><\/a>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">The name \u201cconvolution\u201d comes from a mathematical operation: convolution between functions.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">Put simply, the convolution applies a filter to the input image, the filter parameters are learned through the learning. A learnt filter will be able of detecting features in an image, for example angles, and use them to classify at best the image.<\/span>\r\n\r\n<span style=\"color: #000000;\"><span style=\"font-weight: 400;\">The image is first decomposed into 3 channels (R, G, B) pixels per pixel, we obtain three matrices of size <\/span><b>n x n<\/b><span style=\"font-weight: 400;\"> (where <\/span><b>n<\/b><span style=\"font-weight: 400;\"> is the number of pixels).<\/span><\/span>\r\n\r\n<span style=\"color: #000000;\"><span style=\"font-weight: 400;\">Below is an example of a convolution with a <\/span><b>6 x 6<\/b><span style=\"font-weight: 400;\"> size matrix:<\/span><\/span>\r\n<blockquote><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_102.png\"><img class=\"aligncenter size-full wp-image-686\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_102.png\" alt=\"neural network convolution\" width=\"843\" height=\"409\" \/><\/a><\/blockquote>\r\n<span style=\"font-weight: 400; color: #000000;\">It is important to note two important advantages inherent to convolutional networks:<\/span>\r\n<ul>\r\n \t<li><span style=\"font-weight: 400; color: #000000;\">the network can learn by steps to recognize characteristics in an image. To recognize a face for instance: it will learn to recognize first of eyelids and pupils, and then recognize eyes;<\/span><\/li>\r\n \t<li><span style=\"color: #000000;\">once an item to a learned image place the network will be able to recognize it anywhere else in the picture.<\/span><\/li>\r\n<\/ul>\r\n\u00a0\r\n\r\n\u00a0\r\n<h3>3. Recurrent neural networks (RNN)<\/h3>\r\n\u00a0\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">Recurrent neural networks are at the heart of many substantial improvements in areas as diverse as speech recognition, automatic music composition, sentiment analysis, DNA sequence analysis, machine translation.<\/span>\r\n\r\n\u00a0\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">The main difference with other neural networks is that they take into account a sequence of data, often a sequence evolving over time. For example in the case of analyzing temporal data (time series) the network will still have in memory a part or all of the observations previous to the data being analyzed.<\/span>\r\n\r\n\u00a0\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">The pattern of this network is produced here:<\/span>\r\n\r\n\u00a0\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_105.png\"><img class=\"aligncenter size-full wp-image-685\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_105.png\" alt=\"recurrent neural network\" width=\"827\" height=\"463\" \/><\/a>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">Instead of taking into account separately input data (in the way a CNN would analyse image per image) the recurrent network takes into account data previously processed.<\/span>\r\n\r\n<span style=\"color: #000000;\">Some architectures, called bidirectional, can also take into account future data. For instance when analyzing text to identify named entities (people, companies, countries, etc.) the network would need to see the words of the whole sentence.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">Example:<\/span>\r\n<ul>\r\n \t<li><span style=\"color: #000000;\"><span style=\"font-weight: 400;\">\"<\/span><span style=\"font-weight: 400;\">I see<\/span><span style=\"font-weight: 400;\"> [Jean] Valjean still have escaped you, Javert!\"<\/span><\/span><\/li>\r\n \t<li><span style=\"color: #000000;\">\"<span style=\"font-weight: 400;\">I see <\/span><span style=\"font-weight: 400;\">[Jean] R. plays in this adaptation of \u2018Les Mis\u00e9rables\u2019\u201d.<\/span><\/span><\/li>\r\n<\/ul>\r\n<span style=\"font-weight: 400; color: #000000;\">The beginning of the sentence (underlined) is not enough to identify who is \u2018Jean\u2019.<\/span>\r\n\r\n\u00a0\r\n<h3><span style=\"color: #000000;\"><b>4. Autoencoders<\/b><\/span><\/h3>\r\n\u00a0\r\n\r\n<span style=\"color: #000000;\">Autoencoders are applied mainly to anomaly detection (for example to detect fraud in banking or to find faults in an industrial production line). They can also be used in dimensionality reduction (close to the objective of a Principal Component Analysis). Indeed the goal of autoencoders is to teach the machine what constitutes \"normal\" data.<\/span>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">The architecture of our network is the following:<\/span>\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_106.png\"><img class=\"aligncenter size-full wp-image-684\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_106.png\" alt=\"autoencoder\" width=\"812\" height=\"474\" \/><\/a>\r\n\r\n<span style=\"color: #000000;\">The network will therefore represent data through one or more hidden layers so that the output will be as close as possible to the input data.<\/span>\r\n\r\n<span style=\"color: #000000;\">The objective to find the same data back as the output of the network is characteristic of autoencoders (analogous to the identity function f (x) = x).<\/span>\r\n\r\n<span style=\"color: #000000;\">The encoding and decoding stage it is not however specific to autoencoders. Indeed, they are found in machine translation in recurrent neural networks.<\/span>\r\n\r\n<span style=\"color: #000000;\">After training the network with enough data it will be possible to identify suspicious or anomalous observations when they exceed a certain threshold compared to the new \"standard\".<\/span>\r\n\r\n\u00a0\r\n\r\n<span style=\"color: #000000;\"><b>Conclusion:<\/b><\/span>\r\n\r\n<span style=\"color: #000000;\">We saw 3 major types of neural networks:<\/span>\r\n<ul>\r\n \t<li><span style=\"color: #000000;\">Convolution networks with applications in facial recognition and image classification;<\/span><\/li>\r\n \t<li><span style=\"color: #000000;\">Recurrent networks with applications in the timeseries, text and voice analysis;<\/span><\/li>\r\n \t<li><span style=\"color: #000000;\">Autoencoders with applications to anomaly detection as well as dimensionality reduction.<\/span><\/li>\r\n<\/ul>\r\n<span style=\"color: #000000;\">Other architectures exist such as GANs, generative adversarial networks, which are composed of a model generating candidates for a given task, for example image creation, and another that evaluates the different outputs. Or Reinforcement Learning, a method used by Deepmind to train their Alpha Go and Alpha Go Zero models.<\/span>\r\n\r\n<span style=\"color: #000000;\">Obviously there are limits: for example it is possible to fool convolutional network by adding a particular sound to image undetectable to the human eye but can be fatal for a model that has not been sufficiently tested robustness. New architectures such as capsule networks have merged to face this particular problem.<\/span>\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_099.png\"><img class=\"aligncenter size-full wp-image-689\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_099.png\" alt=\"\" width=\"863\" height=\"346\" \/><\/a>\r\n\r\n<span style=\"font-weight: 400; color: #000000;\">All in all it is certain that deep learning has a bright future with many business applications to come.<\/span>\r\n\r\n<span style=\"color: #000000;\"><span style=\"font-weight: 400;\">Ga\u00ebl Bonnardot,<\/span><span style=\"font-weight: 400;\">\r\n<\/span><em><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">Cofounder and CTO at <strong>Datakeen<\/strong><\/span><\/em><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">At Datakeen we seek to simplify the use and understanding of new machine learning paradigms by the business functions of all industries.<\/span><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">Contact us for more information: contact@datakeen.co<\/span><\/span><!--:--><!--:de--><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/3_Algorithms_Deep_Learning.png\"><img class=\"aligncenter size-full wp-image-690\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/3_Algorithms_Deep_Learning.png\" alt=\"\" width=\"1920\" height=\"1080\" \/><\/a>\r\n\r\n\"Deep\" Learning hat in den letzten Jahren viel Aufmerksamkeit erregt. Und das aus gutem Grund: Diese Teilmenge des maschinellen Lernens hat sich in mehreren Forschungsbereichen eindrucksvoll hervorgetan: Gesichtserkennung, Sprachsynthese, maschinelle \u00dcbersetzung und vieles mehr.\r\n\r\nDiese Forschungsgebiete haben gemeinsam, dass es sich um Wahrnehmungsprobleme im Zusammenhang mit unseren Sinnen und unserem Ausdruck handelt. Sie stellen seit langem eine echte Herausforderung f\u00fcr die Forscher dar, da es \u00e4u\u00dferst schwierig ist, Vision oder Stimme mit Hilfe von Algorithmen und mathematischen Formeln zu modellieren.\r\n\r\nInfolgedessen wurden die ersten Modelle, die in diesen Bereichen implementiert wurden, mit viel Gesch\u00e4ftserfahrung erstellt (in der Spracherkennung: Zerlegung in Phoneme, in der maschinellen \u00dcbersetzung: Anwendung grammatikalischer und syntaktischer Regeln). Jahrelange Forschungsarbeit widmet sich der Nutzung und Verarbeitung dieser unstrukturierten Daten, um Bedeutung zu gewinnen.\r\n\r\nDas Problem ist, dass diese neuen Darstellungen von Daten, die von Forschern erfunden wurden, es vers\u00e4umt haben, Text, Ton oder Bild vollst\u00e4ndig zu verallgemeinern. Wenn Sie Google Translate vor 2014, dem Jahr, in dem sie zu einem 100%igen Deep-Learning-Modell gewechselt haben, verwendet haben, werden Sie sich an die offensichtlichen Einschr\u00e4nkungen dieser Zeit erinnern.\r\n\r\nDeep Learning stellt sich direkt auf die Rohdaten, ohne Verzerrung oder Voraggregation. Dann, dank einer sehr gro\u00dfen Anzahl von Parametern, die sich beim Lernen selbst anpassen, wird aus impliziten Verbindungen lernen, die in den Daten vorhanden sind.\r\n\r\nBevor wir auf drei verschiedene Algorithmen * eingehen, die im Deep Learning f\u00fcr verschiedene Anwendungsf\u00e4lle verwendet werden, lassen Sie uns zun\u00e4chst einfach das Modell definieren, das das Herzst\u00fcck des Deep Learning bildet: das \"neuronale Netzwerk\".\r\n\r\n* Wir sprechen auch \u00fcber verschiedene Netzwerkarchitekturen.\r\n<h3>1. NEURONALE NETZE<\/h3>\r\nLassen Sie mich zun\u00e4chst sagen, dass neuronale Netze sehr wenig mit dem neuronalen System und dem Gehirn zu tun haben. Die Analogie zwischen einem Neuron und einem neuronalen Einneuronennetzwerk ist im Wesentlichen grafisch, sofern ein Informationsfluss von einem Ende zum anderen Netzwerk stattfindet.\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_107.png\"><img class=\"aligncenter size-full wp-image-683\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_107.png\" alt=\"neural network vs neuron\" width=\"927\" height=\"411\" \/><\/a>\r\n\r\n\u00a0\r\n\r\nDie erste Schicht eines neuronalen Netzwerks wird als Eingangsschicht bezeichnet. Durch diese Schicht gelangen Ihre Daten in das Netzwerk. Bevor Sie das Netzwerk mit Ihren Daten \"f\u00fcttern\" k\u00f6nnen, m\u00fcssen Sie es in Nummern \u00e4ndern, falls dies noch nicht geschehen ist.\r\n\r\nWir nehmen das Beispiel der Stimmungsanalyse von Textdaten.\r\n\r\nNehmen wir an, Sie haben 10.000 Kommentare zu Ihrer E-Commerce-Website \u00fcber verkaufte Produkte:\r\n\r\nMit Ihrem Team haben Sie 1000 von ihnen (wir werden sehen, dass Sie sich auch auf vortrainierte neuronale Netze verlassen k\u00f6nnen) in 3 Klassen eingeteilt (zufrieden | neutral | unzufrieden). Diese Anzahl von 3 Klassen, oft im Sinne der Analyse, ist ein Beispiel und Sie k\u00f6nnen tats\u00e4chlich mehr einstellen.\r\n\r\n- \"Ich liebte es, sehr guter Geschmack\";\r\n\r\n- \"Mir gefiel die Verpackung nicht so sehr\";\r\n\r\n- \"Ich dachte, es w\u00e4re ziemlich gut.\"\r\n\r\nDie letzte Schicht, die Ausgabeschicht genannt wird, liefert Ihnen die Klassifizierung \"zufrieden \/ neutral \/ unzufrieden\".\r\n\r\nUnd alle Schichten zwischen der Ein- und Ausgabeschicht, die sogenannten \"versteckten\" Schichten, sind alle unterschiedliche Darstellungen der Daten. Eine Darstellung kann die Anzahl der W\u00f6rter in einem Satz, die Anzahl der Interpunktion (?!) in einem Satz usw. sein. Sie m\u00fcssen dem Netzwerk diese Darstellungen nicht angeben; wenn sie statistisch gesehen dazu beitragen, die S\u00e4tze, die das Netzwerk allein unterrichtet, korrekt zu klassifizieren.\r\n\r\n\u00a0\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_100.png\"><img class=\"size-full wp-image-688\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_100.png\" alt=\"simple neural network\" width=\"743\" height=\"443\" \/><\/a>\r\n\r\nUm diese Schichten zu veranschaulichen, nehmen Sie ein weiteres Beispiel: das des gesch\u00e4tzten Preises eines Hauses.\r\n\r\nWie wir hier sehen k\u00f6nnen, nehmen wir vier Eingabevariablen: die Gr\u00f6\u00dfe des Hauses, die Anzahl der Schlafzimmer, die Postleitzahl und den Grad des Reichtums der Gegend. Das Ergebnis ist nicht zu klassifizieren, sondern eine Zahl vorherzusagen: den Preis des Hauses. Dies ist ein Problem, das als Regression bekannt ist.\r\n\r\nDie kursiv gedruckten W\u00f6rter zu Beispielen von Darstellungen, dass das neuronale Netzwerk die Daten macht, nachdem es viele gesehen hat.\r\n\r\nDie Parameter des Netzwerks werden durch einen Prozess namens \"Backpropagation\" aktualisiert. Und je mehr versteckte Schichten es im Netzwerk gibt, desto \"tiefer\" ist es, daher der Name \"Deep\" Learning.\r\n\r\nLassen Sie uns nun 3 verschiedene Arten von Architekturen neuronaler Netze sehen.\r\n<h3>2. FALTUNGSNEURONALE NETZE (CNN)<\/h3>\r\nDiese Netzwerke werden f\u00fcr alle Anwendungsf\u00e4lle rund um Bild oder Video verwendet, die Gesichtserkennung oder Bildklassifizierung beinhalten.\r\n\r\nZum Beispiel hat Bai Du (das \u00c4quivalent zu Google in China) Portale mit Gesichtserkennung eingerichtet, um nur Mitarbeiter des Unternehmens betreten zu k\u00f6nnen.\r\n\r\nSnapchat und viele mobile Anwendungen haben die Durchbr\u00fcche des Deep Learning und CNNs genutzt, um ihre Gesichtserkennungskapazit\u00e4ten zu erh\u00f6hen, um zus\u00e4tzliche Schichten auf Ihrem Gesicht hinzuzuf\u00fcgen, wie lustige Hasenohren und eine rosa Nase.\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_101.png\"><img class=\"aligncenter size-full wp-image-687\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_101.png\" alt=\"convolutional neural network\" width=\"928\" height=\"389\" \/><\/a>\r\n\r\nDer Name \"Faltung\" kommt von einer mathematischen Operation: Faltung zwischen Funktionen.\r\n\r\nEinfach ausgedr\u00fcckt, die Faltung wendet einen Filter auf das Eingangsbild an, die Filterparameter werden durch das Lernen gelernt. Ein erlernter Filter ist in der Lage, Merkmale in einem Bild, z.B. Winkel, zu erkennen und damit das Bild bestenfalls zu klassifizieren.\r\n\r\nDas Bild wird zun\u00e4chst in 3 Kan\u00e4le (R, G, B) Pixel pro Pixel zerlegt, wir erhalten drei Matrizen der Gr\u00f6\u00dfe <strong>n x n<\/strong> (wobei <strong>n<\/strong> die Anzahl der Pixel ist).\r\n\r\nUnten ist ein Beispiel f\u00fcr eine Faltung mit einer<strong> 6 x 6<\/strong> gro\u00dfen Matrix:\r\n<blockquote><a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_102.png\"><img class=\"aligncenter size-full wp-image-686\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_102.png\" alt=\"neural network convolution\" width=\"843\" height=\"409\" \/><\/a><\/blockquote>\r\nEs ist wichtig, zwei wichtige Vorteile der Faltungsnetze zu beachten:\r\n<ul>\r\n \t<li>Das Netzwerk kann durch Schritte lernen, Merkmale in einem Bild zu erkennen. Um zum Beispiel ein Gesicht zu erkennen: Es wird lernen, zuerst die Augenlider und Pupillen zu erkennen und dann die Augen zu erkennen;<\/li>\r\n \t<li>sobald ein Element an einem gelernten Bildort ist, kann das Netzwerk es an einer anderen Stelle im Bild erkennen.<\/li>\r\n<\/ul>\r\n<h3>3. WIEDERKEHRENDE NEURONALE NETZE (RNN)<\/h3>\r\nWiederkehrende neuronale Netze sind der Kern vieler wesentlicher Verbesserungen in so unterschiedlichen Bereichen wie Spracherkennung, automatische Musikkomposition, Stimmungsanalyse, DNA-Sequenzanalyse, maschinelle \u00dcbersetzung.\r\n\r\nDer Hauptunterschied zu anderen neuronalen Netzen besteht darin, dass sie eine Folge von Daten ber\u00fccksichtigen, die sich oft im Laufe der Zeit entwickelt. Bei der Analyse von Zeitdaten (Zeitreihen) hat das Netzwerk beispielsweise noch einen Teil oder alle Beobachtungen vor den zu analysierenden Daten im Speicher.\r\n\r\nDas Muster dieses Netzwerks wird hier erzeugt:\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_105.png\"><img class=\"aligncenter size-full wp-image-685\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_105.png\" alt=\"recurrent neural network\" width=\"827\" height=\"463\" \/><\/a>\r\n\r\nAnstatt getrennt eingegebene Daten zu ber\u00fccksichtigen (so wie ein CNN das Bild pro Bild analysieren w\u00fcrde), ber\u00fccksichtigt das wiederkehrende Netzwerk die zuvor verarbeiteten Daten.\r\n\r\nEinige Architekturen, die als bidirektional bezeichnet werden, k\u00f6nnen auch zuk\u00fcnftige Daten ber\u00fccksichtigen. Zum Beispiel bei der Analyse von Texten, um benannte Einheiten (Personen, Unternehmen, L\u00e4nder usw.) zu identifizieren, m\u00fcsste das Netzwerk die W\u00f6rter des ganzen Satzes sehen.\r\n\r\nBeispiel:\r\n<ul>\r\n \t<li>\"Ich sehe,[Jean] Valjean ist dir immer noch entkommen, Javert!\"<\/li>\r\n \t<li>\"Ich sehe, dass[Jean] R. in dieser Adaption von'Les Mis\u00e9rables' spielt\".<\/li>\r\n<\/ul>\r\nDer Anfang des Satzes (unterstrichen) reicht nicht aus, um festzustellen, wer Jean ist.\r\n\r\n\u00a0\r\n<h3>4. AUTOENCODERS<\/h3>\r\nAutoencoder werden haupts\u00e4chlich zur Erkennung von Anomalien eingesetzt (z.B. zur Erkennung von Betrug im Bankensektor oder zur Fehlersuche in einer industriellen Produktionslinie). Sie k\u00f6nnen auch zur Reduzierung der Dimensionalit\u00e4t eingesetzt werden (nahe dem Ziel einer Hauptkomponentenanalyse). Tats\u00e4chlich ist das Ziel von Autoencodern, der Maschine beizubringen, was \"normale\" Daten sind.\r\n\r\nDie Architektur unseres Netzwerks ist die folgende:\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_106.png\"><img class=\"aligncenter size-full wp-image-684\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_106.png\" alt=\"autoencoder\" width=\"812\" height=\"474\" \/><\/a>\r\n\r\nDas Netzwerk stellt daher Daten durch eine oder mehrere versteckte Schichten dar, so dass die Ausgabe so nah wie m\u00f6glich an den Eingangsdaten liegt.\r\n\r\nDas Ziel, die gleichen Daten wie die Ausgabe des Netzwerks zur\u00fcckzufinden, ist charakteristisch f\u00fcr Autoencoder (analog zur Identit\u00e4tsfunktion f (x) = x).\r\n\r\nDie Kodierungs- und Dekodierungsstufe ist jedoch nicht spezifisch f\u00fcr Autoencoder. Tats\u00e4chlich sind sie in der maschinellen \u00dcbersetzung in wiederkehrenden neuronalen Netzen zu finden.\r\n\r\nNach dem Training des Netzwerks mit gen\u00fcgend Daten ist es m\u00f6glich, verd\u00e4chtige oder anomale Beobachtungen zu identifizieren, wenn sie einen bestimmten Schwellenwert im Vergleich zum neuen \"Standard\" \u00fcberschreiten.\r\n\r\n<strong>Fazit<\/strong>:\r\n\r\nWir sahen 3 Haupttypen von neuronalen Netzen:\r\n<ul>\r\n \t<li>Convolution vernetzt sich mit Anwendungen in der Gesichtserkennung und Bildklassifikation;<\/li>\r\n \t<li>Wiederkehrende Netzwerke mit Anwendungen in den Bereichen Zeitreihen, Text- und Sprachanalyse;<\/li>\r\n \t<li>Autoencoder mit Anwendungen zur Erkennung von Anomalien sowie zur Reduzierung der Dimensionalit\u00e4t.<\/li>\r\n<\/ul>\r\nEs gibt auch andere Architekturen wie GANs, generative gegnerische Netzwerke, die sich aus einem Modell zusammensetzen, das Kandidaten f\u00fcr eine bestimmte Aufgabe erzeugt, z.B. die Erstellung von Bildern, und einer anderen, die die verschiedenen Ergebnisse bewertet. Oder Reinforcement Learning, eine Methode, mit der Deepmind seine Alpha Go und Alpha Go Zero Modelle trainiert.\r\n\r\nNat\u00fcrlich gibt es Grenzen: Zum Beispiel ist es m\u00f6glich, das Faltungsnetzwerk zu t\u00e4uschen, indem man einen bestimmten Ton zu einem Bild hinzuf\u00fcgt, das f\u00fcr das menschliche Auge nicht wahrnehmbar ist, aber f\u00fcr ein Modell, das nicht ausreichend getestet wurde, fatal sein kann. Neue Architekturen wie Kapsel-Netzwerke haben sich zusammengeschlossen, um diesem besonderen Problem zu begegnen.\r\n\r\n<a href=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_099.png\"><img class=\"aligncenter size-full wp-image-689\" src=\"https:\/\/www.datakeen.co\/wp-content\/uploads\/2018\/02\/S\u00e9lection_099.png\" alt=\"\" width=\"863\" height=\"346\" \/><\/a>\r\n\r\nAlles in allem ist es sicher, dass tiefes Lernen eine vielversprechende Zukunft hat, da viele Gesch\u00e4ftsanwendungen kommen werden.\r\n\r\n<span style=\"color: #000000;\"><span style=\"font-weight: 400;\">Ga\u00ebl Bonnardot,<\/span><span style=\"font-weight: 400;\">\r\n<\/span><em><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">Cofounder and CTO at\u00a0<strong>Datakeen<\/strong><\/span><\/em><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">At Datakeen we seek to simplify the use and understanding of new machine learning paradigms by the business functions of all industries.<\/span><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">\r\n<\/span><span style=\"font-weight: 400;\">Contact us for more information: contact@datakeen.co<\/span><\/span><!--:-->","_et_gb_content_width":"","content-type":"","footnotes":""},"categories":[1],"tags":[195,196,197,155],"class_list":["post-3689","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-non-classifiee","tag-advanced-statitistics-en","tag-data-science-en","tag-deep-learning-en","tag-machine-learning-en"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v25.5 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>3 Deep Learning Architectures explained in Human Language - Datakeen<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"3 Deep Learning Architectures explained in Human Language - Datakeen\" \/>\n<meta property=\"og:description\" content=\"&nbsp; \u201cDeep&quot; Learning has attracted much attention during these past years. And for a good reason: this subset of machine learning has stood out impressively in several research fields: facial recognition, speech synthesis, machine translation, and many others. These research fields have in common to be perceptual problems related to our senses and our expression. [&hellip;]\" \/>\n<meta property=\"og:url\" content=\"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/\" \/>\n<meta property=\"og:site_name\" content=\"Datakeen\" \/>\n<meta property=\"article:published_time\" content=\"2018-02-25T12:53:47+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2021-08-27T10:31:22+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141504\/3_Algorithms_Deep_Learning.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1920\" \/>\n\t<meta property=\"og:image:height\" content=\"1080\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"Ga\u00ebl\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@DatakeenCO\" \/>\n<meta name=\"twitter:site\" content=\"@DatakeenCO\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Ga\u00ebl\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"8 minutes\" \/>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"3 Deep Learning Architectures explained in Human Language - Datakeen","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/","og_locale":"en_US","og_type":"article","og_title":"3 Deep Learning Architectures explained in Human Language - Datakeen","og_description":"&nbsp; \u201cDeep\" Learning has attracted much attention during these past years. And for a good reason: this subset of machine learning has stood out impressively in several research fields: facial recognition, speech synthesis, machine translation, and many others. These research fields have in common to be perceptual problems related to our senses and our expression. [&hellip;]","og_url":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/","og_site_name":"Datakeen","article_published_time":"2018-02-25T12:53:47+00:00","article_modified_time":"2021-08-27T10:31:22+00:00","og_image":[{"width":1920,"height":1080,"url":"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141504\/3_Algorithms_Deep_Learning.png","type":"image\/png"}],"author":"Ga\u00ebl","twitter_card":"summary_large_image","twitter_creator":"@DatakeenCO","twitter_site":"@DatakeenCO","twitter_misc":{"Written by":"Ga\u00ebl","Est. reading time":"8 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/#article","isPartOf":{"@id":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/"},"author":{"name":"Ga\u00ebl","@id":"https:\/\/legacy-wp.datakeen.co\/en\/#\/schema\/person\/201d4b0eea1c7a6ca576cd44f2188730"},"headline":"3 Deep Learning Architectures explained in Human Language","datePublished":"2018-02-25T12:53:47+00:00","dateModified":"2021-08-27T10:31:22+00:00","mainEntityOfPage":{"@id":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/"},"wordCount":1557,"commentCount":0,"publisher":{"@id":"https:\/\/legacy-wp.datakeen.co\/en\/#organization"},"image":{"@id":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/#primaryimage"},"thumbnailUrl":"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141504\/3_Algorithms_Deep_Learning.png","keywords":["advanced statitistics","data science","deep learning","machine learning"],"articleSection":["Non classifi\u00e9(e)"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/","url":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/","name":"3 Deep Learning Architectures explained in Human Language - Datakeen","isPartOf":{"@id":"https:\/\/legacy-wp.datakeen.co\/en\/#website"},"primaryImageOfPage":{"@id":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/#primaryimage"},"image":{"@id":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/#primaryimage"},"thumbnailUrl":"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141504\/3_Algorithms_Deep_Learning.png","datePublished":"2018-02-25T12:53:47+00:00","dateModified":"2021-08-27T10:31:22+00:00","breadcrumb":{"@id":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/#primaryimage","url":"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141504\/3_Algorithms_Deep_Learning.png","contentUrl":"https:\/\/media.datakeen.co\/wp-content\/uploads\/2018\/02\/28141504\/3_Algorithms_Deep_Learning.png","width":1920,"height":1080},{"@type":"BreadcrumbList","@id":"https:\/\/legacy-wp.datakeen.co\/en\/3-deep-learning-architectures-explained-in-human-language\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Accueil","item":"https:\/\/legacy-wp.datakeen.co\/en\/"},{"@type":"ListItem","position":2,"name":"3 Deep Learning Architectures explained in Human Language"}]},{"@type":"WebSite","@id":"https:\/\/legacy-wp.datakeen.co\/en\/#website","url":"https:\/\/legacy-wp.datakeen.co\/en\/","name":"Datakeen","description":"[:fr]Democratise l&#039;IA en Entreprise[:en]Democratize AI Solutions[:de]KI-L\u00f6sungen demokratisieren[:es]Democratizar las soluciones de IA[:ja]\u4eba\u5de5\u77e5\u80fd\u30bd\u30ea\u30e5\u30fc\u30b7\u30e7\u30f3\u306e\u6c11\u4e3b\u5316[:zh]\u6c11\u4e3b\u5316\u4eba\u5de5\u667a\u80fd\u89e3\u51b3\u65b9\u6848[:]","publisher":{"@id":"https:\/\/legacy-wp.datakeen.co\/en\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/legacy-wp.datakeen.co\/en\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/legacy-wp.datakeen.co\/en\/#organization","name":"Datakeen","url":"https:\/\/legacy-wp.datakeen.co\/en\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/legacy-wp.datakeen.co\/en\/#\/schema\/logo\/image\/","url":"https:\/\/media.datakeen.co\/wp-content\/uploads\/2017\/03\/28141638\/datakeen_large_blue_nobg.png","contentUrl":"https:\/\/media.datakeen.co\/wp-content\/uploads\/2017\/03\/28141638\/datakeen_large_blue_nobg.png","width":2000,"height":826,"caption":"Datakeen"},"image":{"@id":"https:\/\/legacy-wp.datakeen.co\/en\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/x.com\/DatakeenCO","https:\/\/www.linkedin.com\/company\/11023029"]},{"@type":"Person","@id":"https:\/\/legacy-wp.datakeen.co\/en\/#\/schema\/person\/201d4b0eea1c7a6ca576cd44f2188730","name":"Ga\u00ebl","url":"https:\/\/legacy-wp.datakeen.co\/en\/author\/admin\/"}]}},"_links":{"self":[{"href":"https:\/\/legacy-wp.datakeen.co\/en\/wp-json\/wp\/v2\/posts\/3689","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/legacy-wp.datakeen.co\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/legacy-wp.datakeen.co\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/legacy-wp.datakeen.co\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/legacy-wp.datakeen.co\/en\/wp-json\/wp\/v2\/comments?post=3689"}],"version-history":[{"count":0,"href":"https:\/\/legacy-wp.datakeen.co\/en\/wp-json\/wp\/v2\/posts\/3689\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/legacy-wp.datakeen.co\/en\/wp-json\/wp\/v2\/media\/3691"}],"wp:attachment":[{"href":"https:\/\/legacy-wp.datakeen.co\/en\/wp-json\/wp\/v2\/media?parent=3689"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/legacy-wp.datakeen.co\/en\/wp-json\/wp\/v2\/categories?post=3689"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/legacy-wp.datakeen.co\/en\/wp-json\/wp\/v2\/tags?post=3689"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}