Changes

no edit summary
Line 82: Line 82:     
<div class="smallimage">[[Image:Learning Machines Seminar.png|150px|link=|Learning Machines Seminar]]</div>
 
<div class="smallimage">[[Image:Learning Machines Seminar.png|150px|link=|Learning Machines Seminar]]</div>
<h3 style="text-decoration:none;">[https://www.youtube.com/watch?v=K8LNtTUsiMI&t Learning Machines Seminar: Yoshua Bengio (Université de Montreal)]</h3>
+
<h3 style="text-decoration:none;">[https://www.youtube.com/watch?v=K8LNtTUsiMI&t Learning Machines Seminar: Extending Deep Learning to High-Level Cognition and Scientific Discovery with Amortized Bayesian Causal Modeling]</h3>
 
<p class="author">Prof. Yoshua Bengio</p>
 
<p class="author">Prof. Yoshua Bengio</p>
 
<p>(In English) How can what has been learned on previous tasks generalize quickly to new tasks or changes in distribution? The study of conscious processing in human brains (and the window into it given by natural language) suggests that we are able to decompose high-level verbalizable knowledge into reusable components (roughly corresponding to words and phrases). This has stimulated research in modular neural networks where attention mechanisms can be used to dynamically select which modules should be brought to bear in a given new context. Another source of inspiration for tackling this challenge is the body of research into causality, where changes in tasks and distributions are viewed as interventions. The crucial insight is that we need to learn to separate (somewhat like in meta-learning) what is stable across changes in distribution, environments or tasks and what may be separate to each of them or changing in non-stationary ways in time. From a causal perspective what is stable are the reusable causal mechanisms, along with the inference machinery to make probabilistic guesses about the appropriate combination of mechanisms (maybe seen as a graph) in a particular new context. What may change with time are the interventions and other random variables which are those that yield more directly to observations. If interventions are not observed (we do not have labels for fully explaining the changes in tasks in terms of the underlying modules and causal variables) we would ideally like to estimate the Bayesian posterior over the interventions, given whatever is observed. This research approach raises many interesting research questions ranging from Bayesian inference and identifiability to causal discovery, representation learning and out-of-distribution generalization and adaptation, which will be discussed in the presentation.</p>
 
<p>(In English) How can what has been learned on previous tasks generalize quickly to new tasks or changes in distribution? The study of conscious processing in human brains (and the window into it given by natural language) suggests that we are able to decompose high-level verbalizable knowledge into reusable components (roughly corresponding to words and phrases). This has stimulated research in modular neural networks where attention mechanisms can be used to dynamically select which modules should be brought to bear in a given new context. Another source of inspiration for tackling this challenge is the body of research into causality, where changes in tasks and distributions are viewed as interventions. The crucial insight is that we need to learn to separate (somewhat like in meta-learning) what is stable across changes in distribution, environments or tasks and what may be separate to each of them or changing in non-stationary ways in time. From a causal perspective what is stable are the reusable causal mechanisms, along with the inference machinery to make probabilistic guesses about the appropriate combination of mechanisms (maybe seen as a graph) in a particular new context. What may change with time are the interventions and other random variables which are those that yield more directly to observations. If interventions are not observed (we do not have labels for fully explaining the changes in tasks in terms of the underlying modules and causal variables) we would ideally like to estimate the Bayesian posterior over the interventions, given whatever is observed. This research approach raises many interesting research questions ranging from Bayesian inference and identifiability to causal discovery, representation learning and out-of-distribution generalization and adaptation, which will be discussed in the presentation.</p>
Line 145: Line 145:  
<p>(In English) Government agencies in Aotearoa New Zealand are increasingly offshoring their data, citing greater security and reduced cost as key factors.</p>  
 
<p>(In English) Government agencies in Aotearoa New Zealand are increasingly offshoring their data, citing greater security and reduced cost as key factors.</p>  
 
<p>As the government accelerates its digital transformation strategy across the public service, Māori data sovereignty requirements must be central to decision making, particularly with regard to offshoring and procurement.</p>
 
<p>As the government accelerates its digital transformation strategy across the public service, Māori data sovereignty requirements must be central to decision making, particularly with regard to offshoring and procurement.</p>
 +
<br>
 
<br>
 
<br>
 
<br>
 
<br>