<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Deep Learning &#8211; Deep Forgetting</title>
	<atom:link href="https://www.deepforgetting.com/category/deep-learning/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.deepforgetting.com</link>
	<description>Artificial Intelligence Conference</description>
	<lastBuildDate>Fri, 22 Oct 2021 07:45:45 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=5.8.6</generator>
	<item>
		<title>Where does our initial learning come from?</title>
		<link>https://www.deepforgetting.com/transfer-learning-in-newborns/</link>
		
		<dc:creator><![CDATA[deep_2022]]></dc:creator>
		<pubDate>Thu, 21 Oct 2021 00:21:04 +0000</pubDate>
				<category><![CDATA[Big Data]]></category>
		<category><![CDATA[Deep Learning]]></category>
		<guid isPermaLink="false">https://www.deepforgetting.com/?p=1744</guid>

					<description><![CDATA[When a baby emerges from the womb, the baby's brain already has all the learning essential for the correct functioning of the body. Isn't that interesting when our whole premise is learning based on "DATA"?]]></description>
										<content:encoded><![CDATA[
<p>When a baby emerges from the womb, the baby&#8217;s brain already has all the learning essential for the correct functioning of the body. The neural circuits for breathing, heartbeat, kicking, sleeping, blood circulation, etc., are all ready! The baby can track a moving object, orient towards Mom or Dad&#8217;s face, feed, or even has the desire to walk when you hold it up with feet touching a flat surface. Isn&#8217;t that interesting when our whole premise is learning based on &#8220;DATA&#8221;?<br></p>



<p>Let&#8217;s start from the <strong>beginning</strong>.</p>



<div class="wp-block-image"><figure class="alignleft size-large is-resized"><img loading="lazy" src="https://www.deepforgetting.com/wp-content/uploads/2021/10/Ivf2_04-1024x699.jpg" alt="" class="wp-image-1751" width="393" height="268" srcset="https://www.deepforgetting.com/wp-content/uploads/2021/10/Ivf2_04-1024x699.jpg 1024w, https://www.deepforgetting.com/wp-content/uploads/2021/10/Ivf2_04-300x205.jpg 300w, https://www.deepforgetting.com/wp-content/uploads/2021/10/Ivf2_04-768x524.jpg 768w, https://www.deepforgetting.com/wp-content/uploads/2021/10/Ivf2_04-1536x1048.jpg 1536w, https://www.deepforgetting.com/wp-content/uploads/2021/10/Ivf2_04-600x410.jpg 600w, https://www.deepforgetting.com/wp-content/uploads/2021/10/Ivf2_04.jpg 1600w" sizes="(max-width: 393px) 100vw, 393px" /></figure></div>



<p>We start from a zygote, a cell formed by the fusion of the male and female gamete (gametes are the male and female with only 23 chromosomes, unlike the regular cells with 46 chromosomes). After the fusion, the zygote has a complete set of 46 chromosomes. In case you forgot, genes are segments of DNA, and chromosomes are the cells&#8217; structures that contain the genes &#8211; hundreds to thousands of genes. So zygote contains the DNA &#8211; the complete DNA that defines us. DNA is a fascinating multidimensional code ever known to humankind. It is not your typical software function. It has the encoded instruction set for complete development.<br></p>



<p>From one cell, the zygote will soon become multicellular, as each cell will undergo mitosis. The cells divide, and the new cells have identical copies of the DNA as the original cells. The cells rearrange themselves spatially to form three layers that differentiate into different organ systems. The cells multiply and change. They produce cells that become white blood cells, red blood cells, nerve cells, eyes, liver, fingers, toes, hair, heart, skin, etc. A cell becomes a palm, an adjacent cell becomes a finger, and another becomes a nail. Not only does a cell know what to do, but it also knows what neighboring cells should do. The DNA instruction set is controlling this.<br></p>



<p>But to me, the most fascinating thing is the learning that gets developed in the brain. The cells give rise to all the neurons in the brain stem so that a newborn knows what to do – kicking, grasping, crying, sleeping, rooting, feeding, etc. The correct synaptic connections, the correct synaptic weights, the correct inhibitors, the correct circuits! Nothing is random &#8211; a well planned execution of the DNA code.</p>



<blockquote class="wp-block-quote"><p>The code in DNA builds these <strong>initial learning/inference models</strong> in the brain.</p></blockquote>



<p>Initial learning is coming from DNA …&nbsp;transfer learning. But the kicking, grasping, crying, sleeping, rooting, feeding are not the only elements of learning that a newborn is born with. There is lot more learning in the brain &#8211; the unconscious bias &#8211; the things that we will discover about the baby as the baby grows.</p>



<p><strong>Let me ask you this. What is something that you know today but you have never learned?</strong></p>



<p>When we use the term DNA, it may sound that learning is influenced only by the parental genomes, but it is a lot more interesting. Along with the genome, we also have an epigenome.</p>



<p></p>



<div class="wp-block-image"><figure class="alignleft size-full is-resized"><img loading="lazy" src="https://www.deepforgetting.com/wp-content/uploads/2021/10/epigenome.jpg" alt="" class="wp-image-1756" width="257" height="257" srcset="https://www.deepforgetting.com/wp-content/uploads/2021/10/epigenome.jpg 450w, https://www.deepforgetting.com/wp-content/uploads/2021/10/epigenome-300x300.jpg 300w, https://www.deepforgetting.com/wp-content/uploads/2021/10/epigenome-150x150.jpg 150w, https://www.deepforgetting.com/wp-content/uploads/2021/10/epigenome-100x100.jpg 100w" sizes="(max-width: 257px) 100vw, 257px" /></figure></div>



<p>The epigenome is a series of chemical tags (called epigenetics markers) that promote or repress the expression of genes without altering the genome. These epigenetic modifications are added in response to the physiological or environmental stimuli our cells receive. Diet, stress, habits, learning, exercise are a few examples that can permanently alter the epigenetic profile of an individual. Researchers use the term “imprinted genes” to refer to genes influenced by the epigenetic markers. The DNA is not modified, but these epigenetic modifications influence the instruction code contained in the DNA.</p>



<p></p>



<p><strong>Epigenome is the software layer on top of the DNA code.</strong></p>



<p></p>



<p></p>



<p>The learning in the newborn is not just coming from the DNA i.e. defined by the parental genome &#8211; the default learning of a baby is influenced by the epigenetic modifications too. These markers fully influence the zygote development.&nbsp; Our learnings, Karma, habits, environment, and lifestyle influence what our newborns will know by default. A child may have the same habits as the father. A child may know things the mother spent years learning. All these sentences make sense biologically. The genome and epigenome take part in that initial learning. It is not just the kicking, grasping, crying, sleeping, rooting, and feeding; unconscious bias, aptitude, habits, and a lot of learning is transferred from generation to generation.</p>



<blockquote class="wp-block-quote"><p>Are people responsible for the (unconscious) actions based on the learning/behaviors they are born with? That is a debate we can discuss another time.</p></blockquote>



<p><strong>Now, let me ask you this. What are the epigenetics-specific learnings that you think came from your parents?</strong></p>



<p>Regardless, this is very interesting for AI/ML engineers as brain development from the genome and epigenome offers a cue to storing model, transfer learning and incremental model change mechanisms. Brain is fascinating and to build real intelligence, we need to dig a lot deeper &#8211; not just neurons but also genome and epigenetics. <strong>Answers are in the nature, we just need to look closer.</strong></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>There are two kinds of AI: BIG Data or MIN Data</title>
		<link>https://www.deepforgetting.com/two-kinds-of-ai/</link>
		
		<dc:creator><![CDATA[deep_2022]]></dc:creator>
		<pubDate>Thu, 01 Jul 2021 02:47:00 +0000</pubDate>
				<category><![CDATA[Big Data]]></category>
		<category><![CDATA[Deep Learning]]></category>
		<guid isPermaLink="false">http://demo.ovatheme.com/imevent/?p=774</guid>

					<description><![CDATA[In next 3 years, AI will rely less on BIG DATA and more on MIN DATA. The new MIN Data approach of AI will enable a whole new set of use cases that seemed unsuited before.]]></description>
										<content:encoded><![CDATA[
<p></p>


<p>There are two kinds of people in the world — those who divide the world into groups of two and those who don’t. And, here is, now, my take on the AI. There are two kinds of AIs — those that are built using BIG DATA and others that are NOT.</p>
<div class="slate-resizable-image-embed slate-image-embed__resize-full-width"><img loading="lazy" class="aligncenter" src="https://media-exp1.licdn.com/dms/image/C5612AQFSODpKwl5EZQ/article-inline_image-shrink_1500_2232/0/1631354341203?e=1637798400&amp;v=beta&amp;t=K3hDSlooj8dcnM1wQ3aAjzu_IyLzZlAFCNQusILF_14" alt="Two types of AI - SECO MIND" width="645" height="736" data-media-urn="" data-li-src="https://media-exp1.licdn.com/dms/image/C5612AQFSODpKwl5EZQ/article-inline_image-shrink_1500_2232/0/1631354341203?e=1637798400&amp;v=beta&amp;t=K3hDSlooj8dcnM1wQ3aAjzu_IyLzZlAFCNQusILF_14" /></div>
<h2>The BIG DATA AI</h2>
<p>In simplistic terms, the BIG Data powered AI models “learn” without being explicitly programed to do so. The learning is basically a process that examines a large set of samples and creates a formula using a large number of parameters many of them seem indistinguishable to the naked eye. It is a time-consuming trial and error process where the parameters are continually adjusted until training data with the same labels consistently yield similar outputs. A classic example of this is using a data set that has 50000 images of digits (0 to 9) and the AI model learns to recognize the digit.</p>
<blockquote>
<p>The BIG DATA AI is all about &#8220;mathematically recognizing&#8221; incredibly subtle patterns within the mountains of data.</p>
</blockquote>
<p>Using this &#8220;pattern recognition technique&#8221; you can now &#8220;provide&#8221; an answer, a decision or a prediction to an input data that you have never seen before. It is a great tool but is unlike the human brain in basic learning. These algorithms require mountains of data and high CPU/GPU power to train. It has no common sense, conceptual learning, creativity, planning, human-like intuition, imagination, cross-domain thinking, self-awareness, emotions, etc. It has trouble when it comes to extreme edges of data for which it had limited samples.</p>
<blockquote>
<p>BIG DATA AI is basically an optimizer based on a large volume of data for a very specific vertical single task.</p>
</blockquote>
<h2> </h2>
<h2>The MIN DATA AI</h2>
<div class="slate-resizable-image-embed slate-image-embed__resize-right"><img loading="lazy" class="alignright" src="https://media-exp1.licdn.com/dms/image/C5612AQHCDXlaf1BHKw/article-inline_image-shrink_1000_1488/0/1630274959257?e=1637798400&amp;v=beta&amp;t=cACTtP6Vgy-Th6ivUErVfiNEJHPD2QVyL902_EUHHRA" alt="No alt text provided for this image" width="501" height="230" data-media-urn="" data-li-src="https://media-exp1.licdn.com/dms/image/C5612AQHCDXlaf1BHKw/article-inline_image-shrink_1000_1488/0/1630274959257?e=1637798400&amp;v=beta&amp;t=cACTtP6Vgy-Th6ivUErVfiNEJHPD2QVyL902_EUHHRA" /></div>
<p>In MIN DATA AI, the AI models &#8220;learn&#8221; with minimal data that is needed to learn. A formula is indeed created however with minimal data. A classic example of this is how some people say they never forget a face even if they met in passing, or, how you recognize a taste even if you have experienced it only once before.</p>
<p>MIN DATA AI is closer to how we learn and do problem-solving. I don&#8217;t need to drink 10,000 OLD FASHIONED (cocktail) to tell the drink in my hand is old fashioned (although may be in my case I already have) or to recognize that the old fashioned is from <a href="https://www.haberdashersj.com/" target="_blank" rel="nofollow noopener">HaberDasher San Jose</a> or NOT (and, yes, HaberDasher is a great place). With this AI, you will be able to make bets even though data says something else. Your learning can be super fast.</p>
<p>Here is how you can start your journey for MIN DATA AI:</p>
<ol>
<li>Stop looking for BIG DATA. Like any other addiction, it will be difficult to cope with BIG DATA addiction but you have to do it.</li>
<li>Focus more on algorithms.</li>
<li>Innovate how to create BIG DATA from MIN DATA.</li>
</ol>
<p>In next 3 years, AI will rely less on BIG DATA and more on MIN DATA. This new MIN Data approach of AI will enable a whole new set of use cases that seemed unsuited before.</p>
<blockquote>
<p>MIND is perhaps MIN D(ata)</p>
</blockquote>
<p>Future of AI making devices, processes and automation intelligent (and not just high confidence level decision making based on BIG Data) is not far. </p>]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Deep learning to avoid real time computation</title>
		<link>https://www.deepforgetting.com/deep-learning-to-avoid-real-time-computation/</link>
		
		<dc:creator><![CDATA[deep_2022]]></dc:creator>
		<pubDate>Sat, 19 Jun 2021 16:16:45 +0000</pubDate>
				<category><![CDATA[Big Data]]></category>
		<category><![CDATA[Deep Learning]]></category>
		<guid isPermaLink="false">https://www.deepforgetting.com/?p=1472</guid>

					<description><![CDATA[“The underlying physical laws necessary for the mathematical theory of a large part of physics and the whole of chemistry are thus completely known, and the difficulty is only that the exact application of these laws leads to equations much too complicated]]></description>
										<content:encoded><![CDATA[
<p>“The underlying physical laws necessary for the mathematical theory of a large part of physics and the whole of chemistry are thus completely known, and the difficulty is only that the exact application of these laws leads to equations much too complicated to be soluble. It therefore becomes desirable that approximate practical methods of applying quantum mechanics should be developed, which can lead to an explanation of the main features of complex atomic systems without too much computation” — Paul Dirac, 1929. For example:</p>



<ul><li>Navier–Stokes equations are the fundamental basis of almost all Computational Fluid Dynamics (CFD) problems. These are extremely useful to model the weather, airflow around an airplane wing, ocean currents, water flow in a pipe, and the analysis of pollution. But these cannot be used in real-time due to the time they need for computation.</li><li>Fully describing an arbitrary many-body state in quantum mechanics requires an exponential amount of information. While simulating a quantum system with 30 qubits requires just tens of gigabytes, simulating 300 qubits requires more bytes than the number of atoms in the observable universe! Even the state-of-the-art approximation methods for quantum mechanics such as Hartree-Fock Theory (HF) and Density Functional Theory (DFT) can take a long time: hours, days, or even weeks to compute.&nbsp;<a href="https://www.linkedin.com/pulse/quantum-computing-opens-new-front-cloud-ajay-malik/" target="_blank" rel="noreferrer noopener">Learn more about QuBits.</a></li><li>Even with the computing power available today, simulations or real-time analysis for acoustic transmission or absorptions in buildings can take a long time.</li><li>Computer simulation of electrons in the potential of atomic nuclei is the workhorse of modeling material properties such as phase stability, mechanical behavior, and thermal conductivity. However, these simulations are limited by their computational cost.</li></ul>



<figure class="wp-block-image"><img src="https://media-exp1.licdn.com/dms/image/C5612AQF96vX_0cUaog/article-inline_image-shrink_1500_2232/0/1572821921045?e=1637798400&amp;v=beta&amp;t=bL1tHZ7UveHlr_2sZKQieM6SYLQ_qhpcdr4Fi0VwOlM" alt="Deep Learning has the potential"/></figure>



<p>Deep learning has the potential to break through this limitation.</p>



<p>Imagine creating a deep learning model by constructing a dataset that covers the entire physically relevant set of configurations for the problem and then just using the model to completely bypass costly calculations in the future.</p>



<blockquote class="wp-block-quote"><p>You can use the predictive power of deep neural networks to cut the computation time down to a&nbsp;couple of seconds.</p></blockquote>



<p>Yes, you will do complex simulations once, but only once! Once you have built your dataset, decided the fitting methodology such as a simple FFN (feed forward neural network), RBM (a restricted-Boltzmann-machine) or any other neural network architecture, your model can serve as a template for all future work!</p>



<p>To me, this is one of the best uses of deep learning to build an explanation without too much computation!</p>



<p>People often think of AI for boosting growth by substituting humans, but actually, huge value is going to come from how humans will use AI. This is yet another perfect example how deep learning will help us advance more.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Explainable Intelligence Part 3 &#8211; The Strategy for XAI</title>
		<link>https://www.deepforgetting.com/explainable-intelligence-part-3-the-strategy-for-xai/</link>
		
		<dc:creator><![CDATA[deep_2022]]></dc:creator>
		<pubDate>Mon, 15 Jul 2019 04:35:00 +0000</pubDate>
				<category><![CDATA[Deep Learning]]></category>
		<guid isPermaLink="false">https://www.deepforgetting.com/?p=1610</guid>

					<description><![CDATA[It is not enough to say that something is true just because &#8216;I know it’s true!&#8217; – we have to have some evidence or argument that gives a justification for our belief. Explanations, justifications, and more broadly epistemology have been the focus]]></description>
										<content:encoded><![CDATA[
<p>It is not enough to say that something is true just because &#8216;I know it’s true!&#8217; – we have to have some evidence or argument that gives a justification for our belief. Explanations, justifications, and more broadly epistemology have been the focus of philosophy for thousands of years. For Plato, being puzzled and therefore wanting to be unpuzzled is the origin of philosophy.</p>



<blockquote class="wp-block-quote"><p>Theaetetus: Yes, Socrates, and I am amazed when I think of them; by the Gods I am! And I want to know what on earth they mean; and there are times when my head quite swims with the contemplation of them.</p></blockquote>



<div class="wp-block-image"><figure class="alignleft is-resized"><img loading="lazy" src="https://media-exp1.licdn.com/dms/image/C5612AQHzTPlOjpPT7Q/article-inline_image-shrink_1000_1488/0/1562446090116?e=1638403200&amp;v=beta&amp;t=MHrIUU8g4D1aFYoPBYjKtmBCu97TaiCpec5PdBQlPgw" alt="Explainable Intelligence: Plato" width="-343" height="-158"/></figure></div>



<p>And, here we are, 2500 years later, longing for clarity,&nbsp;in something we have created &#8211; artificial intelligence.</p>



<p>The &#8220;Explanation&#8221; in AI aims to enable human users to understand the reasons behind the predictions. Comprehending the reasons behind predictions is fundamental if one plans to take action based on a prediction. There are three essential needs for explainability:</p>



<ul><li><strong>Explain to Justify</strong>&nbsp;&#8211; For users to understand and trust. Why did the AI system do that? Why didn&#8217;t the AI system do something else?</li><li><strong>Explain to Control</strong>&nbsp;&#8211; To comply and sustain. In fields like medical, defense, judiciary, education etc., models must be strictly answerable for their predictions. You need to be able to explain so as to comply with accountability and regulatory requirements. When you can control, you can sustain performance and prevent things from going wrong. Bias is potentially present in any dataset and explainability help you to identify and mitigate bias. And, for Debugging/Troubleshooting.</li><li><strong>Explain to Improve &#8211;</strong>&nbsp;To Iterate and improve performance. When you understand the underlying mechanics of that technique, you will know the potential pitfalls associated with it and how to improve. Explanations help in continuous optimizations for better decision making.</li></ul>



<p>Explanations make AI &#8211; fair, robust, certifiable, ethical, privacy-preserving, and human interpretable. Explainability enables human users to effectively manage the AI systems.</p>



<p>However,&nbsp;<em>&nbsp;what makes machine learning algorithms excellent predictors, also makes them difficult to understand</em>. They look like&nbsp;&#8220;<em>black boxes&#8221;&nbsp;</em>to most users.</p>



<figure class="wp-block-image"><img src="https://media-exp1.licdn.com/dms/image/C5612AQHJt5mJVbkktw/article-inline_image-shrink_1500_2232/0/1563197952546?e=1638403200&amp;v=beta&amp;t=-NEswwBzhL-uUWQzSKdiLHInNdSa--r1tTsHVVd6X9g" alt="Accuracy vs. Explainability"/></figure>



<p>XAI aims to &#8220;produce more explainable models while maintaining a high level of learning performance (prediction accuracy); and enable human users to understand, appropriately, trust, and effectively manage the emerging generation of artificially intelligent partners&#8221; &#8211; DARPA. It aims to answer the following questions:</p>



<ol><li>Why a particular prediction was made, as opposed to others?</li><li>Will the model always do that?</li><li>When does it fail &amp; why? When to disregard the model output or how to fix the error?</li></ol>



<figure class="wp-block-image"><img src="https://media-exp1.licdn.com/dms/image/C5612AQEGL8CIYyvCfw/article-inline_image-shrink_1500_2232/0/1563197935654?e=1638403200&amp;v=beta&amp;t=D-IEIXzJ-PQa5EBmwhXLgI6G0f3iNcckVLiK0M94PFU" alt="No alt text provided for this image"/></figure>



<p>Before we go into how XAI is achieved, lets discuss how it is delivered, in what form and what is the scope of explainability.</p>



<h2>Type of Explanations</h2>



<p>The purpose of the explanations is to allow affected parties, regulators, and other non-insiders to understand, discuss, and potentially contest decisions made by black-box algorithmic models. An explanation system can provide explanations mainly in multiple forms:</p>



<ul><li><em>Why-type explanations&nbsp;</em>to describe why a result was generated for particular input. Such explanations aim to communicate what features in input data or what logic in the model resulted in a given machine output.</li><li><em>Contrastive Explanations</em>&nbsp;that highlight not only the pertinent positives but also the pertinent negatives.&nbsp;For example in medicine, a patient showing symptoms of cough, cold and fever, but no sputum or chills, will most likely be diagnosed as having flu rather than having pneumonia.</li><li><em>Explanations by Example</em>: Explaining the decisions of a model might be to report other examples the model considers to be most similar.</li><li><em>What If Explanations&nbsp;</em>is an interactive approach to help user explore what would the system do if a particular input was different. In this approach, user can simulate a few changes to explore the change in output. For example, checking if one&#8217;s mortgage request would still be declined if the requester was a male.</li><li><em>How to Explanations</em>&nbsp;is another interactive approach to help user explore how to get the system to produce a chosen output value. The goal is to provide user with the hypothetical input conditions to produce that output. For example, finding out what annual income would have helped to get the mortgage request approved.</li></ul>



<p><em>Why-type explanations</em>,&nbsp;<em>contrastive explanations</em>&nbsp;and&nbsp;<em>explanations by example</em>&nbsp;address evaluation (how a system works), while&nbsp;<em>what if</em>&nbsp;and&nbsp;<em>how to</em>&nbsp;<em>explanations</em>&nbsp;address the curiosity (build a mental model of system).</p>



<h2>How Explanations are Delivered</h2>



<p>An explanation system can deliver explanations to humans in various forms:</p>



<h3>Verbal/Textual Explanations</h3>



<p>Humans often justify decisions verbally. Verbal explanations describe the machine learning model and reasoning with words, text, or natural language. Verbal explanations are popular in applications like question answering explanations, decision lists, and explanation interfaces.</p>



<p>This form of explanation has also been implemented in recommendation systems and robotics. An example technique of building verbal explanation is by training one model to generate predictions and a separate model, such as a recurrent neural network language model, to generate an explanation.</p>



<h3>Visual Explanations</h3>



<p>Visual explanations use visual elements to describe the reasoning behind the machine learning models. Visual explanations include visualizations of learned model parameters, evaluation metrics, computational graphs, data-flow graphs etc.</p>



<p>Visualizations can be in form of scatter plots, line charts, heat-maps, node-link diagrams, hierarchical (decision trees) etc. For example, line charts for temporal metrics, heat-maps that overlay images to highlight important regions that contribute towards classification and their sensitivity, synthetic images that are representative of what the model has learned about the chosen features, visual back-propagation to visualize which parts of an image have contributed to the classification, visualizing CNN filters, prediction difference analysis to highlight features in an image to provide evidence for or against a certain class and more.</p>



<h2>Scope of Explainability</h2>



<p>Does the interpretation method explain the entire model behavior or an individual prediction? Or is the&nbsp;scope&nbsp;somewhere in between? An explanation can be either at a global level or local level.</p>



<h3><strong>Global level</strong></h3>



<p>This is all about trying to understand&nbsp;<em>&#8220;How does the model as a whole make predictions?&#8221;&nbsp;</em>It is all about being able to explain the entire reasoning leading to all the different possible outcomes.</p>



<p>Understanding of conditional interactions between the response variable(s) and the predictor features on the complete dataset provides the first measure in assessing the trust of any model. The trust is established when the list of important variables is consistent with domain expectations and can also stay stable, with slight data variations.</p>



<p>Even though a multitude of techniques (such as decision trees, rule lists, feature importance, etc.) can be used to enable global interpretability, analyzing and explaining the feature interactions (for model decisions) when there are more than three or four features is quite difficult.</p>



<p>This class of methods is helpful to explain population level decisions, such as alcohol consumption trends or ML course enrollment trend.</p>



<h3><strong>Local level</strong></h3>



<p>This is all about providing justification for a single prediction. It is about understanding&nbsp;<em>&#8220;Why did the model make specific decisions for a single instance?&#8221;&nbsp;</em>and&nbsp;<em>&#8220;Why did the model make specific decisions for a group of instances?&#8221;</em>.</p>



<p>Local explanations identify the specific variables that contributed to an individual decision. For example, when trying to explain why a machine learning algorithm declined mortgage to an individual.</p>



<p>Although global interpretability allows you to verify the hypotheses and whether the model is overfitting to noise, it is hard to diagnose specific model predictions. Local interpretability, on the other hand, tries to answer: why was this prediction made or which variables caused the prediction?</p>



<p>Several techniques such as LIME, LOCO, Anchors, Saliency maps, Local Explanation Vectors etc. can be used to enable local interpretability.</p>



<p>As you can see, XAI is not an AI that can explain itself, it is a design decision in implementation. In the next posts, I will go over the approach of designing XAI, model dependency of XAI, details of inner workings of various techniques as well as how explainability is evaluated.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Explainable Intelligence Part 2 &#8211; Illusion of the Free Will</title>
		<link>https://www.deepforgetting.com/explainable-intelligence-part-2-illusion-of-the-free-will/</link>
		
		<dc:creator><![CDATA[deep_2022]]></dc:creator>
		<pubDate>Sat, 29 Jun 2019 04:30:00 +0000</pubDate>
				<category><![CDATA[Deep Learning]]></category>
		<guid isPermaLink="false">https://www.deepforgetting.com/?p=1603</guid>

					<description><![CDATA[Explainable Artificial Intelligence&#160;(XAI) is getting a lot of attention these days, and like most people, you&#8217;re drawn to it because the very nature of neural networks &#8211; opacity induces the feeling of deprivation that arises from the perception of a gap in]]></description>
										<content:encoded><![CDATA[
<p><a href="https://www.linkedin.com/pulse/explainable-intelligence-part-1-xai-third-wave-ai-ajay-malik/" target="_blank" rel="noreferrer noopener">Explainable Artificial Intelligence</a>&nbsp;(XAI) is getting a lot of attention these days, and like most people, you&#8217;re drawn to it because the very nature of neural networks &#8211; opacity induces the feeling of deprivation that arises from the perception of a gap in knowledge and understanding.</p>



<div class="wp-block-image"><figure class="aligncenter is-resized"><img loading="lazy" src="https://media-exp1.licdn.com/dms/image/C5612AQGylfebfPHAqg/article-inline_image-shrink_1000_1488/0/1561384742661?e=1638403200&amp;v=beta&amp;t=F2dSUL1vFSb3CtUg3LihM6ip_qSDx3WVuovW_qxtino" alt="Right to explanation" width="-543" height="-238"/></figure></div>



<p>Many of us are even willing to refuse the use of artificial intelligence because we cannot explain&nbsp;how artificial intelligence gives its decision. The governments are creating laws around it. For example, GDPR Regulations prohibit any automated decision that &#8216;significantly affects&#8217; EU citizens. These new rules give citizens the&nbsp;<strong>right to review</strong>&nbsp;how digital services made specific algorithmic choices affecting people.</p>



<p>And, that does make a compelling case for investing in Explainable Artificial Intelligence but, before we demand machines to explain why they did what they did, let us look at one fact &#8211; can humans explain their choices?</p>



<p>I know I am opening the old and heated debate of the existence of free will in human behavior.&nbsp;But, it is crucial to dive into the secrets of the brain (artificial intelligence models are based after real brain after all) to understand the mechanics of making Artificial Intelligence Explainable.</p>



<p>Every day, humans face many decisions, some trivial, others more complex.&nbsp;How to dress for the day, what ice-cream flavor to eat, where to eat lunch today – the choices we make in life have a significant impact on our own lives, and the lives of others.</p>



<p>Although we think that it is the &#8220;conscious&#8221; that is making a decision, it is the &#8220;brain&#8221; that&#8217;s running the show. I am sure you have seen visual illusions. It is incredible how many of us see the same illusion. Its almost as if there is the same &#8220;program&#8221; in our brains that is constructing the image and telling us. The &#8220;reality&#8221; we experience is the result of what our brain tells us.</p>



<div class="wp-block-image"><figure class="alignleft is-resized"><img loading="lazy" src="https://media-exp1.licdn.com/dms/image/C4E12AQFenwwoDDtRPQ/article-inline_image-shrink_1000_1488/0/1561474264124?e=1638403200&amp;v=beta&amp;t=gT3gQX3ZgCddI5udfwdBfQnNi1qS1-6l6A1MQnLDn-4" alt="No alt text provided for this image" width="249" height="166"/></figure></div>



<p><em>Hermann Grid is one of the most classic examples of an&nbsp;optical illusion,&nbsp;where your mind is being tricked into seeing something that&#8217;s not there.&nbsp;I bet you see the gray blobs appearing at the intersection of black and white squares. There are none!</em></p>



<p>This is true not only of visual experiences but of all sensory perceptions. Whether we are experiencing the feeling of &#8220;hate&#8221;, the appearance of &#8220;gray blobs&#8221; or hearing things, these are the result of the &#8220;decisions&#8221; made in our brain.</p>



<p></p>



<p></p>



<p>The brain is a hugely complex, highly recurrent, and nonlinear network made up of billions of nerve cells, or neurons. Everything you do, from breathing, walking, and talking to thinking, reading, and problem-solving is the result of neurons communicating with each other.</p>



<p>For neurons to communicate, they need to transmit information both within the neuron and from one neuron to the next. This process utilizes both electrical signals as well as chemical messengers.</p>



<div class="wp-block-image"><figure class="alignright is-resized"><img loading="lazy" src="https://media-exp1.licdn.com/dms/image/C5612AQFC3IXnsNTEgw/article-inline_image-shrink_1500_2232/0/1561727174044?e=1638403200&amp;v=beta&amp;t=bK0rs3kEoTvhkg3TzUqHSx_evp4C96AOgJkAT9_QBWE" alt="neuron" width="-365" height="-229"/></figure></div>



<p>A typical neuron possesses a&nbsp;soma&nbsp;(the bulbous cell body which contains the cell nucleus),&nbsp;dendrites&nbsp;(long, feathery filaments attached to the cell body in a complex branching &#8220;dendritic tree&#8221;), a single&nbsp;axon&nbsp;(an extra-long, branched cellular filament, which may be thousands of times the length of the soma) and maintains a&nbsp;voltage gradient&nbsp;across its membrane. The tips of the branches of the axon, called axon terminals or boutons, impinge on other neurons or effectors.</p>



<div class="wp-block-image"><figure class="alignleft is-resized"><img loading="lazy" src="https://media-exp1.licdn.com/dms/image/C5612AQHJWlPl35PZJg/article-inline_image-shrink_1000_1488/0/1561727687535?e=1638403200&amp;v=beta&amp;t=9ZFify5Q4T38gHu1Iv2SMpiQi1owx_6-_tNzN3RXw5Q" alt="synapse" width="-325" height="-186"/></figure></div>



<p>Any two neurons do not touch; they have a space between them. This locus of interaction between a bouton and the cell on which it impinges is called a synapse (the actual gap, also known as the&nbsp;synaptic cleft, is of the order of 20 nanometers), and we say that the cell with the bouton synapses upon the cell with which the connection is made. Each neuron can be connected to tens of thousands of other neurons.</p>



<p>The dendrites of neurons receive information from sensory receptors or other neurons. This information is then passed down to the cell body and on to the axon. Once the information has arrived at the axon, it travels down the length of the axon in the form of&nbsp;an electrical signal known as an&nbsp;<em>action potential</em>. The action potential travels down the axon until it reaches the synapses, where it then causes the release of&nbsp;<em>neurotransmitters</em>&nbsp;of various types. The neurotransmitters cross the cell membrane into the synaptic gap between neurons. These chemicals then bind to chemical&nbsp;receptors&nbsp;in the dendrites of the receiving (<em>postsynaptic</em>) neuron. In the process, they result in a decrease or increase of the membrane potential of the postsynaptic<em>&nbsp;</em>neuron. If it causes the&nbsp;<em>membrane potential</em>&nbsp;to pass the firing threshold, then it will activate an&nbsp;<em>action potential</em>&nbsp;in the postsynaptic neuron and send it down its axon. And it goes on.</p>



<p>The number of neurotransmitter receptors at a synapse is a key element determining synaptic transmission efficacy. The changes in the number of receptors at synapses is one of the main underlying molecular events that happen as a result of numerous transient interactions between neurons. Not only synaptic strength, but the neuronal activity also modifies the morphology of the dendrites in response to the stimuli received during a learning process. Neurons grow new and retract old branches, attaching (synapsing) to other neurons in new places and removing old connections.</p>



<p>Learning is continually rewiring the brain. However, the vast majority of these cells are produced in the womb – during early pregnancy. The majority of the remaining neurons are produced during a short period of time after birth.&nbsp;Babies are not born as blank slates &#8211; we basically inherit these neural connections, synaptic weights, thresholds as a coded list of amino acids (DNA) and, many encoded in our brain in our most formative years as we are born into a world over which we have no choice.</p>



<p>Yes, there is a real world out there, and we perceive events that occur around us, however we&#8217;re not perceiving what&#8217;s out there. We&#8217;re perceiving whatever our brain tells us. Our decisions, even when we think that we are involving your conscious self, are based on the &#8220;weights&#8221;, &#8220;thresholds&#8221; and &#8220;connections&#8221; that had been burned deep into the brain&#8217;s circuitry. To quote David Eagleman, &#8220;Our decisions are not free-will choices but result of the hands of cards we&#8217;re dealt.&#8221;</p>



<figure class="wp-block-pullquote"><blockquote><p>What we have in form of DNA and brain&#8217;s circuitry is not a blueprint, but a fortune teller.</p></blockquote></figure>



<p>We are not at the center of ourselves, and we can not explain our choices. The levels of subjective wellbeing we feel, our problem solving skills, how we ideate, and so on. You might watch less TV, drink less alcohol, might have invented the next &#8216;as seen on TV&#8217; gadget, and, your wife might not have left you! Recent reviews of neuroscientific work confirm that many of Freud&#8217;s original observations, not least the pervasive influence of non-conscious processes and the organizing function of emotions for thinking, have found confirmation in laboratory studies.</p>



<p>However, I will stop right here while you ponder the question of free will.</p>



<p>AI needs to be explainable, and in the next articles, I will share my learnings of the strategy, approach, scope, and techniques for XAI.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Explainable Intelligence Part 1 &#8211; XAI, the third wave of AI</title>
		<link>https://www.deepforgetting.com/explainable-intelligence-part-1-xai-the-third-wave-of-ai/</link>
		
		<dc:creator><![CDATA[deep_2022]]></dc:creator>
		<pubDate>Sun, 23 Jun 2019 16:20:00 +0000</pubDate>
				<category><![CDATA[Deep Learning]]></category>
		<guid isPermaLink="false">https://www.deepforgetting.com/?p=1598</guid>

					<description><![CDATA[Artificial Intelligence (AI) is democratized in our everyday life. Tractica forecasts the global artificial intelligence software market revenues will grow from around 9.5 billion US dollars in 2018 to an expected 118.6 billion by 2025. Gartner predicts&#160;that by 2020, AI technologies will be virtually pervasive in almost]]></description>
										<content:encoded><![CDATA[
<p>Artificial Intelligence (AI) is democratized in our everyday life. <a rel="noreferrer noopener" href="https://www.tractica.com/" target="_blank">Tractica</a> forecasts the global artificial intelligence software market revenues will grow from around 9.5 billion US dollars in 2018 to an expected <a rel="noreferrer noopener" href="https://www.tractica.com/newsroom/press-releases/artificial-intelligence-software-market-to-reach-118-6-billion-in-annual-worldwide-revenue-by-2025/" target="_blank">118.6 billion by 2025</a>.</p>



<div class="wp-block-image"><figure class="alignleft is-resized"><img loading="lazy" src="https://media-exp1.licdn.com/dms/image/C4E12AQGDx48NfHw16A/article-inline_image-shrink_1000_1488/0/1561127359508?e=1638403200&amp;v=beta&amp;t=vSjMK4fkaZHsl6BoX7AarlfBD-NZ9Tpi4xAUBFj8nnU" alt="AI Software Revenue 2018 - 2025" width="365" height="319"/></figure></div>



<p><a href="https://www.gartner.com/en/newsroom/press-releases/2017-07-18-gartner-says-ai-technologies-will-be-in-almost-every-new-software-product-by-2020" target="_blank" rel="noreferrer noopener">Gartner predicts</a>&nbsp;that by 2020, AI technologies will be virtually pervasive in almost every new software product and service. It also predicts that the business value created by AI will reach $3.9T in 2022.&nbsp;</p>



<p>We are becoming accustomed to AI making decisions for us in our daily life, from product and movie recommendations on Netflix and Amazon to friend suggestions on Facebook and tailored advertisements on Google search result pages.</p>



<p>NetFlix uses AI for personalization of movie recommendations, auto-generation and personalization of thumbnails / artwork, location scouting for movie production (pre-production), movie editing (post-production), streaming quality and a lot more (<a href="https://www.businessinsider.com/netflix-recommendation-engine-worth-1-billion-per-year-2016-6" target="_blank" rel="noreferrer noopener">Business Insider</a>).</p>



<p>AI plays a huge role in Amazon&#8217;s recommendation engine, which&nbsp;<a href="https://www.mckinsey.com/industries/retail/our-insights/how-retailers-can-keep-up-with-consumers" target="_blank" rel="noreferrer noopener">generates 35%</a>&nbsp;of the company&#8217;s revenue as well as in Amazon&#8217;s GO hardware, which includes color and depth cameras, as well as weight sensors and algorithms.</p>



<p>23% of North American enterprises have machine learning embedded in at least one company function as of last year. 19% of enterprises in developing markets including China and 21% in Europe also have successfully integrated machine learning into functions.&nbsp;Most businesses are considering adding AI capability in their systems for certain Business Value Proposition.</p>



<figure class="wp-block-pullquote"><blockquote><p>By 2021, the term AI will no longer be considered a differentiator in marketing tech provider solutions.</p></blockquote></figure>



<p>Although AI is powerful in terms of results and predictions, AI algorithms suffer from opacity, that it is difficult to get insight into their internal mechanism of work. And, there are times when we need to know why it made a specific decision.</p>



<div class="wp-block-image"><figure class="alignleft is-resized"><img loading="lazy" src="https://media-exp1.licdn.com/dms/image/C4E12AQHnbc4m2BrdDA/article-inline_image-shrink_1000_1488/0/1561127172739?e=1638403200&amp;v=beta&amp;t=DcnFcnXEeuPLbb3hwxrh6UDPa7xmqvmVmZPzi0BSgQc" alt="Self-driving car kills a PEDESTRIAN" width="-419" height="-342"/></figure></div>



<p>In March 2018, an autonomous vehicle operated by Uber hit and killed a woman in Tempe, Ariz., as she was walking her bicycle across the street.</p>



<p>So what happened? Uber car&#8217;s computer system had spotted Ms. Herzberg six seconds before impact, but classified Ms. Herzberg, who was not in a crosswalk, first as an unrecognized object, then as another vehicle and finally as a bicycle. The self-driving software decided not to take any actions after the car’s sensors detected the pedestrian. Uber&#8217;s autonomous mode disables Volvo’s factory-installed automatic emergency braking system, according to the US National Transportation Safety Board <a rel="noreferrer noopener" href="https://www.ntsb.gov/news/press-releases/Pages/NR20180524.aspx" target="_blank">preliminary report</a> on the accident. Uber suspended testing of self-driving vehicles after the crash. In December, the vehicles returned to public roads, though at reduced speeds and in less-challenging environments.</p>



<p></p>



<div class="wp-block-image"><figure class="alignright is-resized"><img loading="lazy" src="https://media-exp1.licdn.com/dms/image/C4E12AQFFH0RvxNfjwg/article-inline_image-shrink_1000_1488/0/1561127235975?e=1638403200&amp;v=beta&amp;t=inRoRnGIscD_bbaLN37lNP8-BVPZ2p-f_FeFr4cUdNY" alt="IBM Watson Oncology Failure" width="-361" height="-281"/></figure></div>



<p></p>



<p>&#8220;This product is a piece of sh**&#8221; wrote a doctor at Florida&#8217;s Jupiter Hospital regarding IBM&#8217;s flagship AI program Watson, according to internal <a rel="noreferrer noopener" href="https://www.statnews.com/2018/07/25/ibm-watson-recommended-unsafe-incorrect-treatments/" target="_blank">documents</a> obtained by Stat. It recommended &#8216;unsafe and incorrect&#8217; cancer treatments.</p>



<p>In 2013 IBM developed Watson&#8217;s first commercial application for cancer treatment recommendation, and the company secured a number of key partnerships with hospitals and research centers over the past five years. But Watson AI Health has not impressed doctors. Some complained it gave wrong recommendations on cancer treatments that could cause severe and even fatal consequences. In February 2017, <a rel="noreferrer noopener" href="https://www.forbes.com/sites/matthewherper/2017/02/19/md-anderson-benches-ibm-watson-in-setback-for-artificial-intelligence-in-medicine/" target="_blank">Forbes reported</a> that MD Anderson had “benched” the Watson for Oncology project.</p>



<p></p>



<div class="wp-block-image"><figure class="alignleft is-resized"><img loading="lazy" src="https://media-exp1.licdn.com/dms/image/C4E12AQG2GCNRugR_jA/article-inline_image-shrink_1000_1488/0/1561127478458?e=1638403200&amp;v=beta&amp;t=iaHh_vLZ6Sh4cGzEti_-_8vhtXQ8MWrvkJEvMjaUmrY" alt="Recruiting software downgraded resumes that implied the applicant was FEMALE" width="421" height="288"/></figure></div>



<p></p>



<p>Amazon HR reportedly used an AI-enabled recruiting software between 2014 and 2017 to help review resumes and make recommendations. The software was, however, found to be more favorable to male applicants.</p>



<p>The software reportedly downgraded resumes that contain the word &#8220;women&#8221; or implied the applicant was female, for example, because they had attended a women&#8217;s college. Amazon has since abandoned the software.&nbsp;</p>



<blockquote class="wp-block-quote"><p>AI is a journey, and in no way, any of these failures are suggesting that we should do to Artificial Intelligence what Luddites tried.</p></blockquote>



<p>Although Artificial Intelligence has captured the imagination of the world since its inception in 1956, at a historic conference at Dartmouth, it has been stuck in the discovery phase with hand-crafted&nbsp;and custom engineered AI applications. Researchers worked on figuring out how to solve a particular problem, and, then do traditional coding. A key characteristic these applications shared was no learning ability and poor handling of uncertainty. It is the Deep Learning systems, starting in the early 2010&#8217;s, aided by huge amounts of training data and massive computational power that have shown us the substantive glimpses of the power and application of AI. The period in AI until 2010 essentially represents the&nbsp;<strong>&#8220;First Wave&#8221;</strong>&nbsp;and then the deep learning starting the&nbsp;<strong>&#8220;Second Wave&#8221;</strong>.</p>



<p><em>Deep learning</em>&nbsp;is an architecture modeled loosely on the human brain. It makes use of neural networks that consist of thousands or even millions of nodes (neurons) that are densely interconnected and organized into multiple layers. An individual node might be connected to several nodes in the layer beneath it, from which it receives data, and several nodes in the layer above it, to which it sends data. The essential concept is that there is a &#8220;weight&#8221; assigned to each of the incoming connection of a node, the node computes the sum of the weighted inputs based on data and weight from all connected nodes beneath it, and then uses a threshold to pass or not pass the data to all its outgoing connections to the nodes in the layer above it (akin to &#8220;firing of a neuron&#8221;).</p>



<p>For training, huge amounts of labeled data is fed into the neural network. An object recognition system, for instance, might be fed thousands of labeled images of cats, dogs, digits, and so on, and it would find visual patterns in the images that consistently correlate with particular labels. During training, the weights and thresholds are continually adjusted until training data with the same labels consistently yield similar outputs.</p>



<figure class="wp-block-pullquote"><blockquote><p>You have &#8220;learned&#8221; to mathematically recognize incredibly subtle patterns within the mountains of data</p></blockquote></figure>



<p>Using this &#8220;pattern recognition,&#8221; you can now &#8220;provide&#8221; an answer, a decision or a prediction to an input data that you have never seen before.</p>



<p>However, this AI using deep learning still has a long way to go to approach human-level learning, thinking, and problem-solving ability. It is a great tool but is unlike the human brain in basic learning. These algorithms suffer from opacity, and do not provide reasoning. They require mountains of data to train. It has no common sense, conceptual learning, creativity, planning, cross-domain thinking, self-awareness, emotions, etc. It&nbsp;is basically an<em>&nbsp;optimizer&nbsp;</em>based on data for a very specific vertical single task.</p>



<figure class="wp-block-pullquote"><blockquote><p>Deep Learning is still very weak in intelligence.</p></blockquote></figure>



<figure class="wp-block-image"><img src="https://media-exp1.licdn.com/dms/image/C5612AQGbGZ01pVS6sw/article-inline_image-shrink_1500_2232/0/1561310689901?e=1638403200&amp;v=beta&amp;t=25EuVPAb-3Z7-syH8S-R11Pj_UfAd7CMw9Tqv-7Pq24" alt="Four Waves of AI"/></figure>



<p>Until we figure out the engineering paths to the general intelligence capabilities, the&nbsp;<strong>&#8220;fourth wave&#8221;</strong>, the focus of AI is shifting to opening the blackbox.</p>



<p>We are now entering&nbsp;<strong>&#8220;third wave&#8221;</strong>&nbsp;of AI where AI systems will become capable of explaining the reasoning behind every decision made by them. The AI systems themselves will construct models that will explain how it works.</p>



<p>XAI is all about improving trust of AI-based systems. At one end it brings fairness, accountability and transparency to the front and center of AI; and on the other end it enables us to control and continuously improve our AI systems.</p>



<figure class="wp-block-image"><img src="https://media-exp1.licdn.com/dms/image/C5612AQFe-kpEI4ddcg/article-inline_image-shrink_1500_2232/0/1561309413548?e=1638403200&amp;v=beta&amp;t=tcve8jiiVXdgbEbvvdU40qsnJOWO6IVOl29V9QKeTqA" alt="Benefits of XAI"/></figure>



<p>To me, this third wave, XAI, is the sine qua non for AI to continue making steady progress without disruption. In last year and a half, I have been spending a lot of time on learning XAI.&nbsp;In this series of articles, I will share my learnings of the strategy, approach, scope, and techniques for XAI.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>

<!--
Performance optimized by W3 Total Cache. Learn more: https://www.boldgrid.com/w3-total-cache/

Page Caching using disk: enhanced 

Served from: www.deepforgetting.com @ 2023-04-12 23:47:19 by W3 Total Cache
-->