Off-Policy Evaluation with Policy-Dependent Optimization Response
The intersection of causal inference and machine learning for decision-making is rapidly expanding, but the default decision criterion remains an average of individual causal outcomes across a population. In practice, various operational restrictions ensure that a decision-maker's utility is not realized as an average but rather as an output of a downstream decision-making problem (such as matching, assignment, network flow, minimizing predictive risk). In this work, we develop a new framework for off-policy evaluation with policy-dependent linear optimization responses: causal outcomes introduce stochasticity in objective function coefficients. Under this framework, a decision-maker's utility depends on the policy-dependent optimization, which introduces a fundamental challenge of optimization bias even for the case of policy evaluation. We construct unbiased estimators for the policy-dependent estimand by a perturbation method, and discuss asymptotic variance properties for a set of adjusted plug-in estimators.
Interpolation and Regularization for Causal Learning
Recent work shows that in complex model classes, interpolators can achieve statistical generalization and even be optimal for statistical learning. However, despite increasing interest in learning models with good causal properties, there is no understanding of whether such interpolators can also achieve causal generalization. To address this gap, we study causal learning from observational data through the lens of interpolation and its counterpart---regularization. Under a simple linear causal model, we derive precise asymptotics for the causal risk of the min-norm interpolator and ridge regressors in the high-dimensional regime. We find a large range of behavior that can be precisely characterized by a new measure of confounding strength.
Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding
Imagen builds on the power of large transformer language models in understanding text and hinges on the strength of diffusion models in high-fidelity image generation. Our key discovery is that generic large language models (e.g., T5), pretrained on text-only corpora, are surprisingly effective at encoding text for image synthesis: increasing the size of the language model in Imagen boosts both sample fidelity and image-text alignment much more than increasing the size of the image diffusion model. Imagen achieves a new state-of-the-art FID score of 7.27 on the COCO dataset, without ever training on COCO, and human raters find Imagen samples to be on par with the COCO data itself in image-text alignment. To assess text-to-image models in greater depth, we introduce DrawBench, a comprehensive and challenging benchmark for text-to-image models. With DrawBench, we compare Imagen with recent methods including VQ-GAN CLIP, Latent Diffusion Models, and DALL-E 2, and find that human raters prefer Imagen over other models in side-by-side comparisons, both in terms of sample quality and image-text alignment.
'Hey, Cortana' becomes 'Hey, Copilot' in Windows 11
Stop us if you've heard this before: You can now talk to your PC's built-in AI. But in Windows 11, Cortana has been replaced with Windows Copilot, and you can now interact with Copilot by saying "Hey, Copilot" instead. Microsoft is testing the new feature within the Windows Insider program. If your PC is unlocked, and you've configured it to accept the "Hey Copilot" wake words, you can now interact with Copilot verbally. The Copilot UI will launch as a small microphone icon.
Fair Infinitesimal Jackknife: Mitigating the Influence of Biased Training Data Points Without Refitting
In consequential decision-making applications, mitigating unwanted biases in machine learning models that yield systematic disadvantage to members of groups delineated by sensitive attributes such as race and gender is one key intervention to strive for equity. Focusing on demographic parity and equality of opportunity, in this paper we propose an algorithm that improves the fairness of a pre-trained classifier by simply dropping carefully selected training data points. We select instances based on their influence on the fairness metric of interest, computed using an infinitesimal jackknife-based approach. The dropping of training points is done in principle, but in practice does not require the model to be refit. Crucially, we find that such an intervention does not substantially reduce the predictive performance of the model but drastically improves the fairness metric.
Palm up: Playing in the Latent Manifold for Unsupervised Pretraining
Large and diverse datasets have been the cornerstones of many impressive advancements in artificial intelligence. Intelligent creatures, however, learn by interacting with the environment, which changes the input sensory signals and the state of the environment. In this work, we aim to bring the best of both worlds and propose an algorithm that exhibits an exploratory behavior whilst it utilizes large diverse datasets. Our key idea is to leverage deep generative models that are pretrained on static datasets and introduce a dynamic model in the latent space. The transition dynamics simply mixes an action and a random sampled latent.
Contrastive Learning as Goal-Conditioned Reinforcement Learning
In reinforcement learning (RL), it is easier to solve a task if given a good representation. While deep RL should automatically acquire such good representations, prior work often finds that learning representations in an end-to-end fashion is unstable and instead equip RL algorithms with additional representation learning parts (e.g., auxiliary losses, data augmentation). How can we design RL algorithms that directly acquire good representations? In this paper, instead of adding representation learning parts to an existing RL algorithm, we show (contrastive) representation learning methods are already RL algorithms in their own right. To do this, we build upon prior work and apply contrastive representation learning to action-labeled trajectories, in such a way that the (inner product of) learned representations exactly corresponds to a goal-conditioned value function.
How Black Girls Code is preparing marginalized kids for the AI revolution
Despite its global prominence, and years of investment from the tech industry's loudest voices and biggest pocketbooks, AI still has a diversity problem. Filling an increasingly worrisome gap created by the tech's creators and evangelists, diversity-based organizations have been trying to tackle that issue on their own. Black Girls Code for example -- which offers tech skill building for Black girls and other historically underrecognized groups -- has been leaning more heavily into AI as part of its tech preparedness and training curriculum, including creating the brand new position of AI Expert-in-Residence to oversee a more thoughtful approach to teaching about AI. "Most AI is built in environments that prioritize profit over people, which means bias gets baked in and the same communities left out of past tech waves are now at risk of being harmed again. It's not enough to teach people to use AI, we have to teach them to be thoughtful about the tools that they use," Black Girls Code CEO Cristina Mancini tells Mashable. What values does it reflect?
Teacher quits profession after viral rant on how AI is 'ruining' education
Hannah, a former teacher, joins'Fox & Friends' to explain why she left the classroom, saying AI tools are making it difficult to teach. A former high school English teacher went viral this week after posting a candid video on social media announcing she was quitting the teaching profession because of how technology was "ruining" education. In her video, which reached over 1 million views on TikTok, Hannah explained how AI tools have made teaching more difficult because students rely on technology to do the work for them and are unmotivated to put in effort themselves. She said that kids do not know how to read because of read-aloud tools, and have short attention spans because of the "high stimulation" of social media. "They want to use [technology] for entertainment. They don't want to use it for education," she said in a TikTok video which reached over 1 million views.
Elon Musk's Grok AI Can't Stop Talking About 'White Genocide'
A chatbot developed by Elon Musk's multibillion-dollar artificial intelligence startup xAI appeared to be suffering from a glitch Wednesday when it repeatedly brought up white genocide in South Africa in response to user queries about unrelated topics on X. Grok, which competes with other chatbots like OpenAI's ChatGPT, is directly integrated into the social media platform that Musk also owns. Numerous examples of the phenomenon could be found by searching the official Grok profile for posts containing the term "boer," a word used to refer to people from South Africa of "Dutch, German, or Huguenot descent." It is sometimes used by Black South Africans as a pejorative against white Afrikaners, or people associated with the apartheid regime. In response to topics ranging from streaming platform HBO Max's name change to Medicaid cuts proposed by US lawmakers, the chatbot often seemed to initially stay on topic before veering back to white genocide in South Africa, completely unprompted. When asked to confirm the salary of Toronto Blue Jays player Max Scherzer, for example, the generative artificial intelligence chatbot launched into an explanation of white genocide and a controversial South African anti-apartheid song.