HAMLET

HAMLET  (Human, Animal, and Machine Learning: Experiment and Theory)

HAMLET is an interdisciplinary proseminar series that started in 2008. The goal is to provide behavioral and computational graduate students with a common grounding in the learning sciences. Guest speakers give a talk each week, followed by discussions. It has been an outlet for fresh research results from various projects. Participants are typically from Computer Sciences, ECE, Psychology, Educational Psychology as well as other parts of the campus. Multiple federal research grants and publications at machine learning and cognitive psychology venues have resulted from the interactions of HAMLET participants.

Meetings: Fridays 3:45 p.m. – 5 p.m. Berkowitz room (338 Psychology Building)

Subscribe/Unsubscribe: HAMLET mailing list

This semester we are trying a new format for HAMLET. Speakers pursuing research at the intersection of computation and human behavior will give a short (10-15 minute) introduction/outline/description of a current research project or problem, with the aim of sparking collaborative discussion in the group about the project.

How would people from your field approach it?
Are there existing tools for solving the problem?
What is already known about the domain?
What kind of data are involved?
How does the research connect to real-world issues?

Brainstorm sessions will be interleaved with standard talks on more mature work throughout the semester.


Fall 2019 Schedule:

HAMLET this semester will have a mini-series on neural networks.

9/27 Gary Lupyan lead a group discussion on what does it mean to understand a neural network?

What does it mean to understand what a neural network is doing? How do people in different disciplines–cognition, neuroscience, machine learning–think about or answer this question? To orient the discussion, please have a look at the following recent paper by Tim Lillicrap (machine learning scientist at Deep Mind) and Konrad Kording (computational neuroscientist at Penn):

https://arxiv.org/pdf/1907.06374.pdf

We are very interested to know whether different disciplines even agree on what “interpretability” means, and whether there is useful cross-disciplinary progress to be made.

10/4 Yingyu Liang

Title: Robust Attribution Regularization
Abstract: An emerging problem in trustworthy machine learning is to train models that produce robust interpretations for their predictions. We take a step towards solving this problem through the lens of axiomatic attribution of neural networks. Our theory is grounded in the recent work, Integrated Gradients (IG), in axiomatically attributing a neural network’s output change to its input change. We propose training objectives in classic robust optimization models to achieve robust IG attributions. Our objectives give principled generalizations of previous objectives designed for robust predictions, and they naturally degenerate to classic soft-margin training for one-layer neural networks. We also generalize previous theory and prove that the objectives for different robust optimization models are closely related. Experiments demonstrate the effectiveness of our method, and also point to intriguing problems which hint at the need for better optimization techniques or better neural network architectures for robust attribution training.

This is joint work with Jiefeng Chen, Xi Wu, Vaibhav Rastogi, and Somesh Jha.

10/11 Rahul Parhi: Neural Networks Learn Splines

Where: State Room (4th floor) in Memorial Union

We develop a general framework based on splines to understand the interpolation properties of overparameterized neural networks. We prove that minimum “norm” two-layer neural networks (with appropriately chosen activation functions) that interpolate scattered data are minimal knot splines. Our results follow from understanding key relationships between notions of neural network “norms”, linear operators, and continuous-domain linear inverse problems.

10/18 Tim Rogers: Using neural networks to reverse-engineer cortical anatomy

10/25 Claudia Solis-Lemus

11/1 Mike Zinn

Where: Union South, please check the TTIU for the exact room

11/8 Peter Adamczyk

11/15 Bilge Mutlu

11/22 Yin Li

11/29 Thanksgiving – no HAMLET

12/6 Michael Gleicher


Spring 2019 Schedule:

2/8 John Curtin, Psychology

Brainstorm: Digital phenotyping with (somewhat) big data for harm reduction in alcohol and other drug use disorders

John has been applying machine learning methods to Facebook and geo-tracking data to understand and predict patterns of relapse in
people recovering from addiction. He will briefly characterize issues that motivate the research, the kinds of data his group is working with,
some of the work they have already done and some of the challenges they are currently facing. The discussion will focus on generating ideas for moving the project forward.

2/15: Data blitz: Current research in cognition, perception, and cognitive neuroscience (special event at the WID)

2/22: Paula Niedenthal, Psychology

Brainstorming connections between emotion, cognition and datascience

Paula will initiate a session concerning ways that scientists in machine learning/data science might collaborate with scientists working to understand human affective cognition. Read about some of her recent work here: Science Daily

3/1: Dimitris Papailiopoulos (ECE, WID)  LOCATION CHANGE: Union South (look at the TTIU for the room)

Dimitris Papailioupolos will spark a discussion assessing whether principles of human cognition can guide us toward more robust machine learning.

3/8 Alyssa Adams, VEDA Data Solutions

Emergence, evolution and video games

Alyssa Adams, a data scientist working at Madison startup VEDA Data Solutions. Alyssa’s work focuses on understanding what makes living systems different from non-living systems, and in particular the mechanisms that underlie emergence in social and biological systems. Toward these questions she applies computational approaches to data generated by people interacting in video games–so if you like emergence, math, human behavior and video games, you should really come. Alyssa is also a great person to talk to if you are interested in how data science is being applied in the wild, or what it is like to move from an academic to an industrial career path.

3/22  Spring Break

3/29 Karen Schloss, Psychology & WID & Laurent Lessard, WID/ECE

Using machines to learn how humans map colors to concepts

Karen Schloss and Laurent Lessard with lead a brainstorming session focusing on the use of machine learning to understand a central aspect of human cognition: what explains the conceptual inferences and associations that people draw from simple visual features, such as colors? Karen and Laurent will briefly set up the problem, describing some interesting phenomena and prior work, then will lead an interactive discussion on ways that machine learning might be developed or applied to solve key questions.

4/5 Jon Willits, Psychology at University of Illinois

A Two-Process Model of Semantic Knowledge

Jon’s work has connected large-corpus NLP-based approaches to word meaning with cognitive theories of language. He will present new work on this topic, suggesting a two-stage model of semantic processing with implications for both human language and the design of artificial knowledge systems.

Abstract: Computational models of semantic memory, and machine learning approaches to knowledge-based systems, have traditionally proposed that there is a single semantic memory – one set of factual knowledge that is
used for all activities. There are a number of problems (both theoretical and empirical) with this view. In this talk, I will propose an alternate, two-process model. Process 1 consists of a general, passive, statistical-learning system that acquires knowledge of general, context-insensitive associations and similarities in the environment.
Process 2 consists of task- or goal-specific subsystems, where stored semantic representations are changed in ways that optimize their use for specific situations. I will show how this model addresses a number of thorny theoretical and empirical issues in the study of semantic memory and building artificial knowledge systems.

4/12 Ari Rosenberg, Neuroscience

Inferring whole brain connectivity

Knowing the interconnections between brain regions is important to understand how neural functions are implemented by distributed networks, and how different networks share information. Diffusion tensor imaging (DTI) provides a technique for estimating anatomical pathways within the brain. Our preliminary attempts to create connectivity matrices for macaque cortex using DTI have proven promising. First, we find a surprisingly strong ability to predict functional connectivity data from other labs. Second, we find evidence of pathways hypothesized to exist based on electrophysiological data. A potential point of discussion is whether graph theory provides appropriate techniques for quantifying brain networks using DTI data.

4/19 Greg Zelinsky, Psychology at Stony Brook University

Predicting goal-directed attention control: A tale of two brain-inspired computational models

The ability to control the allocation of attention underlies all visuospatial goal-directed behavior; there would be no sense in having a goal if there was no way to use it to select visual inputs. He will summarize two recent efforts to computationally model this core mechanism of cognitive control. These models are similar in that both are image computable and that the design of each is inspired by mechanisms used by the attention-control system in the primate brain. But the models also differ in their type of brain inspiration and in their neurocomputational approach. One model adopts a computationally narrow approach but is closely parametrized by known neural constraints (MASC, a Model of Attention in the Superior Colliculus). The other model adopts a more powerful deep neural network approach but is less directly brain inspired (Deep-BCN, a deep network implementation of the widely accepted biased-competition theory of attention control). Together they are intended to span a continuum of brain-inspired model design, ranging from relatively local neural circuits at one end to the broader network of brain areas comprising the primate attention control system at the other.

4/26 Glenn Fung Moo, a machine learning scientist at American Family Insurance

Non-IID Model Learning via Graph Neural Networks

Most of the more commonly used machine learning methods assume that training and test data comes from independent and identically distributed (IID) random variables. However, this assumption is strongly violated when the data points available to learn/test the model are highly interdependent. Two examples of this scenario are: when the data exhibits temporal or spatial correlations and when the task is to learn relative characteristics between data points (e.g., ranking or attribute learning). In this talk we show how emerging ideas in graph neural
networks can yield a solution to various problems that broadly fall under this characteristics. More specifically, we show interesting results for the relative attribute learning problem from images. This setting, naturally benefits from exploiting the graph of dependencies among the different relative attributes of images, especially when only
partial ordering is provided at training time. Our experiments show that this simple framework is effective in achieving competitive accuracy with specialized methods for both relative attribute learning and binary attribute prediction.

5/3 Sarah Sant’Ana, Psychology

 


 

Fall 2018 Schedule:

Sep 14, Tim Rogers, Psychology

Investigating the covariance structure of widely-held false beliefs

ABSTRACT: It has become increasingly clear that our society is suffering a crisis of false belief. Across a variety of domains–including science, health, economics, politics, history, and current events–there often exists vociferous disagreement about what ought to be matters of fact. I’d like to know whether cognitive science and machine learning together can help us to understand why this is happening. In prior work we have studied the formation of false beliefs in small well-controlled lab studies with no real-world stakes. In this talk I will describe some preliminary work looking at real false beliefs occurring in the wild. In collaboration with a summer PREP student and colleagues in Psychology and in the School of Journalism, we generated a long list of incorrect but putatively widely-held beliefs spanning many different knowledge domains. Using Amazon Mechanical Turk, we asked people about the degree
to which they endorsed or rejected the belief. We also collected a variety of data about each respondent, including sociodemographic data, political leanings, media trust, IQ, and so on. The aim was to measure how susceptibility to false beliefs pattern across individuals, and to understand what properties of individuals predict susceptibility to which kinds of false beliefs. Our early results provide some provocative evidence contradicting some straight-forward hypotheses about false belief formation, but there is a lot more work to be done. We would like to connect with others interested in these problems from a machine-learning-and-social-media perspective, and in particular would like to consider ways of using social network analytics to assess the degree to which media consumption may be causing the patterns we observe.

Sep 21, LUCID faculty meeting – no HAMLET

Sep 28, no HAMLET

Oct 5, Jay Patel, Educational Psychology and Ayon Sen, Computer Science

A Novel Application of Machine Teaching to Perceptual Fluency Training, a collaboration between Ed Psych and CS

W will hear from Jay Patel and Ayon Sen, who will present new work applying neural networks and machine teaching to undergraduate chemistry learning—something for everyone!

Abstract:

In STEM domains, students are expected to acquire domain knowledge from visual representations that they may not yet be able to
interpret. Such learning requires perceptual fluency: the ability to intuitively and rapidly see which concepts visuals show and to translate among multiple visuals. Instructional problems that engage students in nonverbal, implicit learning processes enhance perceptual fluency. Such processes are highly influenced by sequence effects. Thus far, we lack a principled approach for identifying a sequence of perceptual-fluency problems that promote robust learning. Here, we describe a novel educational data mining approach that uses machine learning to generate an optimal sequence of visuals for perceptual-fluency problems. In a human experiment, we show that a machine-generated sequence outperforms both a random sequence and a sequence generated by a human domain
expert. Interestingly, the machine-generated sequence resulted in significantly lower accuracy during training, but higher posttest
accuracy. This suggests that the machine-generated sequence induced desirable difficulties. To our knowledge, our study is the first to show that an educational data mining approach can induce desirable difficulties for perceptual learning.

Oct 12, Tim Rogers, Psychology

Data science and brain imaging at UW Madison

Over the summer Tim had conversations with several groups about their work applying machine learning or other data science tools to brain imaging data. These activities are part of a broader trend across the country, with collaborations between neuroscientists, cognitive scientists, and computer scientists driving very rapid innovations in techniques for finding signal in neural data. Many of us are interested in connecting disparate efforts across campus to develop a general knowledge base, community of support, and new collaborations in this field. There is also interest in understanding how work conducted here relates to other high-profile work emerging from other labs.

Tim will present a general overview of the variety of new methods that I am aware of and some of the astonishing results they have uncovered. The overview will be targeted at both cognitive scientists/neuroscientists and members of the machine learning
community. I will also highlight some of the opportunities for new work and collaboration that I am aware of, and hope to initiate a discussion about what we can do to promote these efforts at UW.

Join the new email list to facilitate discussion, collaboration, and information-sharing on these issues.  big-lucid@lists.wisc.edu  (BIG = “Brain-imaging Interest Group”; LUCID is the broader community interested in learning, understanding, cognition, intelligence and data-science)

Oct 19, LUCID faculty meeting – no HAMLET

Oct 26, no HAMLET

Nov 2, Jerry Zhu, Computer Science

Discuusion: Can Machine Teaching Influence How Humans Do Regression?

Jerry is interested in discussing with cognitive scientists possible relationships between a hot topic in machine learning–data
poisoning, and defences against data poisoning–and approaches to human learning and teaching. Can central concepts from this area of study in machine learning help or hinder human learning?

How do people make numerical predictions (e.g. weight of a person) from observed features (height, fitness, gender, age, etc.)?  We may assume that they learn this from some sort of regression in their head, applied to training data they receive.  Can machine teaching help them learn better predictions by being careful about the training examples we give them?   Or to mess them up intentionally?  This is equivalent to ‘training data poisoning attacks’ in adversarial machine learning.   I will explain how this is done with control theory under highly simplified, likely incorrect assumptions on human learning.  This is where we need the cognitive science audience to participate in the discussion, and together we will explore the research opportunities on this topic.

Nov 9, Data Blitz

75 minutes, 9 talks, a sampling of current research at UW connecting data science and human behavior

This HAMLET will feature a cross-disciplinary data-blitz. We will hear a series of 5-minute talks on topics at the intersection of
machine learning, data science, and human behavior. This is a great opportunity to hear about some of the projects happening on campus now that you may want to get involved with.

Can neural networks help you learn chemistry? How do Trumpian false beliefs differ from other partisan false beliefs? These questions any many more will be addressed in this data blitz!

Nov 16, Gary Lupen

How translateable are languages? Quantifying semantic alignment of natural languages.

Abstract:
Do all languages convey semantic knowledge in the same way? If language  simply mirrors the structure of the world, the answer should be a qualified “yes”. If, however, languages impose structure as much as reflecting it, then even ostensibly the “same” word in different
languages may mean quite different things. We provide a first pass at a large-scale quantification of cross-linguistic semantic alignment of approximately 1000 meanings in 70+ languages. We find that the translation equivalents in some domains (e.g., Time, Quantity, and Kinship) exhibit high alignment across languages while the structure of other domains (e.g., Politics, Food, Emotions, and Animals) exhibits substantial cross-linguistic variability. Our measure of semantic alignment correlates with phylogenetic relationships between languages and with cultural distances between societies speaking the languages, suggesting a rich co-adaptation of language and culture even in domains of experience that appear most constrained by the natural world.

Nov 23, Thanksgiving – no HAMLET

Nov 30, Emily Ward, Psychology

Dec 7, Priya Kalra, Educational Psychology

 

 

 

 


Spring 2018 Schedule:

Feb 2, Blake Mason

Title: Low-Dimensional Metric Learning with Application to Perceptual Feature Selection

Abstract: I will discuss recent work investigating the theoretical foundations of metric learning, focused on four key topics: 1) how to learn a general low-dimensional (low-rank) metrics as well as sparse metrics; 2) upper and lower (minimax) bounds on the generalization error; 3) how to quantify the sample complexity of metric learning in terms of the dimension of the feature space and the dimension/rank of the underlying metric; 4) the accuracy of the learned metric relative to the underlying true generative metric. As an application of these ideas, I will discuss work with collaborators in Educational Psychology that applies metric learning for perceptual feature detection in non-verbally mediated cognitive processes.

Feb 9, Adrienne Wood

Form follows function: Emotion expressions as adaptive social tools

Emotion expressions convey people’s feelings and behavioral intentions, and influence, in turn, the feelings and behaviors of perceivers. I take a social functional approach to the study of emotion expression, examining how the physical forms of emotion expression are flexible and can be adapted to accomplish specific social tasks. In this talk, I discuss two lines of research, the first of which applies a social functional lens to smiles and laughter. I present work suggesting that smiles and laughter vary in their physical form in order to achieve three distinct tasks of social living: rewarding others, signaling openness to affiliation, and negotiating social hierarchies. My approach, which generalizes to other categories of expressive behavior, accounts for the form and context of the occurrence of the expressions, as well as the nature of their influence on social partners. My second line of research examines how cultural and historical pressures influence emotional expressiveness. Cultures arising from the intersection of many other cultures, such as in the U.S., initially lacked a clear social structure, shared norms, and a common language. Recent work from my collaborators and myself suggests such cultures increase their reliance on emotion expressions, establishing a cultural norm of expressive clarity. I conclude by presenting plans to quantify individual differences in the tendency to synchronize with and accommodate to the emotion expressive style of a social partner, and relate those differences to people’s social network positions. Given the important social functions served by emotion expression, I suggest that the ability to use it flexibly is associated with long-term social integration.

Feb 16 Nils Ringe, Professor, Political Science

Speaking in Tongues: The Politics of Language and the Language of Politics in the European Union

Politics in the European Union primarily takes place between political actors who do not share a common language, yet this key feature of EU politics has not received much attention from political scientists. This project investigates if and how multilingualism affects political processes and outcomes. Its empirical backbone is a series of in-depth interviews with almost 100 EU policy-makers, administrators, translators, and interpreters, but it also involves at least two potential components where political science, linguistic, and computational approaches overlap. The first involves the analysis of oral legislative negotiations in the European Parliament, where non-native English speakers interact using English as their shared language, and in native-English speaking parliamentary settings (in Ireland, Scotland, and/or the UK), to determine if “EU English” differs syntactically and semantically from “regular” English. The expectation is that speech in the EP is simpler, more neutral, and more utilitarian. The second component involves the identification of languages spoken in EP committee meetings using computational methods, to determine the language choices members of the EP make.

Feb 23 Andreas Obersteiner

Does 1/4 look larger than 1/3? The natural number bias in comparing symbolic and nonsymbolic fractions

When people compare the numerical values of fractions, they are often more accurate and faster when the larger fraction has the larger natural number components (e.g., 2/5 > 1/5) than when it has the smaller components (e.g., 1/3 > 1/4). However, recent studies produced conflicting evidence of this “natural number bias” when the comparison problems were more complex (e.g., 25/36 vs. 19/24). Moreover, it is unclear whether the bias also occurs when fractions are presented visually as shaded parts of rectangles rather than as numerical symbols. I will first present data from a reaction time study in which university students compared symbolic fractions. The results suggest that the occurrence and strength of the bias depends on the specific type of comparison problems and on people’s ability to activate overall fraction magnitudes. I will then present preliminary data from an eye tracking study in which university students compared rectangular fraction visualizations. Participants’ eye movements suggest that the pure presence of countable parts encouraged them to use unnecessary counting strategies, although the number of countable parts did not bias their decisions. The results have implications for mathematics education, which I will discuss in the talk.

Mar 2 Nicole Beckage, University of Kansas

Title: Multiplex network optimization to capture attention to features

Abstract: How does attention to features and current context affect people’s search in mental and physical spaces? I will examine how procedures for optimally searching through “multiplex” networks — networks with multiple layers or types of relationships — capture human search and retrieval patterns. Prior work on semantic memory, people’s memory for facts and concepts, has primarily focused on modeling similarity judgments of pairs of words as distances between points in a high-dimensional space (e.g., LSA by Laudauer et al, 1998; Word2Vec by Mikolov et al. 2013). While these decisions seem to accurately account for human similarity in some contexts, it’s very difficult to interpret high dimensional spaces, making it hard to use such representations for scientific research. Further, it is difficult to adapt these spaces to a specific context or task. Instead, I define a series of individual feature networks to construct a multiplex network, where each network in the multiplex captures a “sense” or type of similarity between items. I then optimize the “influence” of each of these feature networks within the multiplex framework, using real world search behavior on a variety of tasks. These tasks include semantic memory search in a cognitive task and information search in Wikipeida. The resulting weighting of the multiplex can capture aspects of human attention and contextual information in these diverse tasks. I explore how this method can provide interpretability to multi-relational data in psychology and other domains by developing an optimization framework that considers not only the presence or absence of relationships but also the nature of the relationships. While I focus on applications of semantic memory, I discuss mathematical proofs and simulation experiments that apply more generally to optimization problems in the multiplex network literature.

Mar 9 Psychology visit day Data Blitz! (WID Orchard Room)

The UW-Madison Psychology department will be hosting its second annual data blitz for prospective graduate students in the Orchard Room of the WID. The data blitz will feature speakers from across the department presenting their research in an easily digestible format. Each talk will be 5 minutes long with an additional 2 minutes for questions at the end of each talk. All are welcome to attend. Below you will find a list of the speakers and the titles of their talks:

    • Rista Plate “Unsupervised learning shifts emotion category boundaries”
    • Ron Pomper “Familiar object salience affects novel word learning”
    • Elise Hopman “Measuring meaning alignment between different languages”
    • Anna Bartel “How diagrams influence students’ mental models of mathematical story problems”
    • Aaron Cochrane “Chronic and phasic interactions between video game playing and addiction”
    • Pierce Edmiston “Correlations between programming languages and beliefs about programming”
    • Chris Racey “Neural processing underlying color preference judgments”
    • Sofiya Hupalo “Corticotropin-releasing factor (CRF) modulation of frontostriatal circuit function”

Mar 16 Varun Jog, Assistant Professor, ECE

Title: Mathematical models for social learning

Abstract: Individuals in a society learn about the world and form opinions not only through their own experiences, but also through interactions with other members in the society. This is an incredibly complex process, and although it is difficult to describe it completely using simple mathematical models, valuable insights may be obtained through such a study. Such social learning models have turned out to be a rich source of problems for probabilists, statisticians, information theorists, and economists. In this talk, we survey different social learning models, describe the necessary mathematical tools to analyze such models, and give examples of results that one may prove through such an approach.

Mar 23 Josh Cisler, Assistant Professor, Department of Psychiatry

Title: Real-time fMRI neurofeedback using whole-brain classifiers with an adaptive implicit emotion regulation task: analytic considerations

Most fMRI neuroimaging studies manipulate a psychological or cognitive variable (e.g., happy versus neutral faces) and observe the manipulations impact on brain function (e.g., amygdala activity is greater for happy faces). As such, the causal inferences that can be drawn from these studies is the effect of cognition on brain function, and not the effect of brain function on cognition. Real-time fMRI refers to processing of fMRI data simultaneous with data acquisition, enabling feedback of current brain states to be presented back to the participant in (near) real-time, thus enabling the participant to use the feedback signals to modify brain states. We are conducting an experiment using real-time fMRI neurofeedback where the feedback signal consists of classifier output (hyperplane distances) from a SVM trained on all grey matter voxels in the brain. The feedback signal is embedded within a commonly used implicit emotion regulation task, such that the task becomes easier or harder depending on the participant’s brain state. This type of ‘closed loop’ design allows for testing whether manipulations of brain state (via feedback) have a measurable impact on cognitive function (task performance). The purpose of this presentation will be to present the experimental design and resulting data properties for the purpose of obtaining feedback and recommendations for understanding and analyzing the complex dynamical systems relations between the feedback signal, brain state, and task performance.

Mar 30 (no meeting, spring break)

Apr 6 Student Research Presentations

Mini talk 1: Ayon Sen, Computer Sciences

For Teaching Perceptual Fluency, Machines Beat Human Experts

In STEM domains, students are expected to acquire domain knowledge from visual representations. Such learning requires perceptual fluency: the ability to intuitively and rapidly see what concepts visuals show and to translate among multiple visuals. Instructional problems that enhance perceptual fluency are highly influenced by sequence effects. Thus far, we lack a principled approach for identifying a sequence of perceptual-fluency problems that promote robust learning. Here, we describe a novel educational data mining approach that uses machine learning to generate an optimal sequence of visuals for perceptual-fluency problems. In a human experiment realted to chemistry, we show that a machine-generated sequence outperforms both a random sequence and a sequence generated by a human domain expert. To our knowledge, our study is the first to show that an educational data mining approach can yield desirable difficulties for perceptual learning.

Mini talk 2: Evan Hernandez, Ara Vartanian, Computer Sciences

Block-based programming environments are popular in computer science education, but the click-and-drag style of these environments render them inaccessible by students with motor impairments. Vocal user interfaces (VUIs) offer a popular alternative to traditional keyboard and mouse interfaces. We design a VUI for Google Blockly in the traditional Turtle/LOGOS setting and discuss the relevant design choices. We then investigate augmentations to educational programming environments. In particular, we describe a method of program synthesis for completing the partial or incorrect programs of students, and ask how educational software may leverage program synthesis to enhance student learning.

Apr 13 (no meeting)

Apr 20

Rob Nowak: All of Machine Learning

Apr 27 (STARTING AT 4PM instead of 3:45pm). Tyler Krucas, The Wisconsin Gaming Alliance.

An Industry Perspective on Data in Game Design and Development

The Wisconsin game development industry offers a surprisingly comprehensive cross section of the types of individuals and teams that develop video games. This includes everything from studios that collaborate on AAA titles such as Call of Duty and Bioshock Infinite, to studios that work largely on mobile or free-to-play games, to studios that primarily work on educational games or games for impact. In all cases, data collection and analysis is an important tool in every step of the game development process. However, the scale of the data collected and its use can vary dramatically from developer to developer. In my talk, I will provide an overview of the the game development ecosystem in Wisconsin, as well as examples of the different types data collection and use practices found in the regional industry. Critically, I’ll frame this discussion in the context of possible links with the HAMLET group – in terms of possible sources of data to address fundamental questions surrounding human learning or behavior as well as possible collaborations.

Fall 2017 Schedule:

Sept 15, Virtual and Physical: Two Frames of Mind, Bilge Mutlu (CS)

In creating interactive technologies, virtual and physical embodiments are often seen as two sides of the same coin. They utilize similar core technologies for perception, planning, and interaction and engage people in similar ways. Thus, designers consider these embodiments to be broadly interchangeable and choice of embodiment to primarily depend on the practical demands of an application. In this talk, I will make the case that virtual and physical embodiments elicit fundamentally different frames of mind in the users of the technology and follow different metaphors for interaction. These differences elicit different expectations, different forms of engagement, and eventually different interaction outcomes. I will discuss the design implications of these differences, arguing for different domains of interaction serving as appropriate context for virtual and physical embodiments.

October 13, Learning semantic representations for text: analysis of recent word embedding methods, Yingyu Liang (CS)

Recent advances in natural language processing build upon the approach of embedding words as low dimensional vectors. The fundamental observation that empirically justifies this approach is that these vectors can capture semantic relations. A probabilistic model for generating text is proposed to mathematically explain this observation and existing popular embedding algorithms. It also reveals surprising connections to classical notions such as Pointwise Mutual Information in computational linguistics, and allows to design novel, simple, and practical algorithms for applications such as embedding sentences as vectors.

October 20, Vlogging about research, Martina Rau (Ed Psych)

October 27, LUCID faculty meeting, No large group meeting

November 3, A Discussion of Open Science Practices, Martha W. Alibali (Psych)

This HAMLET session will be a discussion of open-science practices, led by Martha Alibali. We will start with brief discussion of the “replication crisis” and “questionable research practices”. We will then discuss solutions, including better research practices, data sharing and preregistration. Please read at least some of the provided papers, and come prepared to ask questions and share your experiences.

Replication crisis paper

Questionable Research Practices (QRPs) and solutions paper paper

Data sharing paper http://deevybee.blogspot.co.uk/2014/05/data-sharing-exciting-but-scary.html

Preregistration paper pic https://www.psychologicalscience.org/observer/seven-selfish-reasons-for-preregistration

November 10, Systematic misperceptions of 3D motion explained by Bayesian inference, Bas Rokers (Psych)

Abstract: Over the years, a number of surprising, but seemingly unrelated errors in 3D motion perception have been reported. Given the relevance of accurate motion perception to our everyday life, it is important to understand the cause of these perceptual errors. We considered that these perceptual errors might arise as a natural consequence of estimating motion direction given sensory noise and the geometry of 3D viewing. We characterized the retinal motion signals produced by objects moving along arbitrary trajectories through three dimensions and developed a Bayesian model of perceptual inference. The model predicted a number of known errors, including a lateral bias in the perception of motion trajectories, and a dependency of this bias on stimulus contrast and viewing distance. The model also predicted a number of previously unknown errors, including a dependency of perceptual bias on eccentricity, and a surprising tendency to misreport approaching motion as receding and vice versa. We then used standard 3D displays as well as a virtual reality (VR) headsets to test these predictions in naturalistic settings, and established that people make the predicted errors. In sum, we developed a quantitative model of 3D motion perception and provided a parsimonious account for a range of systematic perceptual errors in naturalistic environments.

November 17, Total variation regression under highly correlated designs, Becca Willett (ECE)

Abstract: I will describe a general method for solving high-dimensional linear inverse problems with highly correlated variables. This problem arises regularly in applications like neural decoding from fMRI data, where we often have two orders of magnitude more brain voxels than independent scans. Our approach leverages a graph structure that represents connections among voxels in the brain. This graph can be estimated from side sources, such as diffusion-weighted MRI, or from fMRI data itself. We will explore the underlying models, computational methods, and initial empirical results. This is joint work with Yuan Li and Garvesh Raskutti.

November 24 Thanksgiving Holiday

December 1, Micro-(Shape-And-Motion)-Scopes, Mohit Gupta (CS)

Imagine a drone looking for a safe landing site in a dense forest, or a social robot trying to determine the emotional state of a person by measuring her micro-saccade movements and skin-tremors due to pulse beats, or a surgical robot performing micro-surgery inside the body. In these applications, it is critical to resolve fine geometric details, such as tree twigs; to recover micro-motion due to biometric signals; and the precise motion of a robotic arm. Such precision is more than an order-of-magnitude beyond the capabilities of traditional vision techniques. I will talk about our recent work on designing extreme (micro) resolution 3D shape and motion sensors using unconventional, but low-cost optics, and computational techniques. These methods can measure highly subtle motions (< 10 microns), and highly detailed 3D geometry (<100 microns). These sensors can potentially detect a persons pulse or micro-saccade movements, and resolve fine geometric details such as a facial features, from a long distance.

December 8, Influence maximization in stochastic and adversarial settings, Po-Ling Loh (ECE)

We consider the problem of influence maximization in fixed networks, for both stochastic and adversarial contagion models. Such models may be used to model infection spreads in epidemiology, as well as the diffusion of information in viral marketing. In the stochastic setting, nodes are infected in waves according to linear threshold or independent cascade models. We establish upper and lower bounds for the influence of a subset of nodes in the network, where the influence is defined as the expected number of infected nodes at the conclusion of the epidemic. We quantify the gap between our upper and lower bounds in the case of the linear threshold model and illustrate the gains of our upper bounds for independent cascade models in relation to existing results. In the adversarial setting, an adversary is allowed to specify the edges through which contagion may spread, and the player chooses sets of nodes to infect in successive rounds. Our main result is to establish upper and lower bounds on the regret for possibly stochastic strategies of the adversary and player. This is joint work with Justin Khim (UPenn) and Varun Jog (UW-Madison).

For other events check out our calendar: Seminar and Events  This content and updates can be found: HAMLET


HAMLET Archives

Fall 2015 archive
Fall 2012 archive
Fall 2011 archive
Spring 2011 archive
Fall 2010 archive
Fall 2009 archive
Spring 2009 archive
Fall 2008 archive