Artificial Intelligence is the study and practice of building systems that can solve complex tasks in ways that would traditionally need human intelligence.
Yale has a number of faculty working in Artificial Intelligence. Our main areas of research are Robotics, Natural Language Processing, Machine Learning, Computer Vision, and the emerging area of AI, Ethics and Society.
Faculty working in this area:
Highlights in this area:
Yang Cai is broadly interested in the theory of computation and its interface with economics, game theory, and machine learning. His current focus includes equilibrium computation in multi-agent machine learning, wherein agents learn, choose actions and receive rewards in a shared environment. Examples of application include generative adversarial networks, adversarial examples, and multi-robot interactions. His group is exploring what equilibrium concepts are meaningful in these environments and when they can be realistically computed. Another focus of his group is on the design of information and incentive structure in markets, with an emphasis on the computational complexity, structural simplicity, and robustness of the design. He has received the Sloan Research Fellowship, the NSF CAREER Award, and the William Dawson Scholarship.
Smita Krishnaswamy – The primary focus of my research is on Machine Learning for extracting patterns and insights from scientific data in order to drive biomedical discovery. While much of AI has focused on matching known patterns for classification, there is a great need for using AI to find unknown patterns and to generate plausible scientific hypotheses. My work is at the intersection of several fields including applied math, deep learning, data geometry, topology, manifold learning, and graph signal processing, all serving to tackle key challenges in data science. The problems I address are motivated by the ubiquity of high-throughput, high-dimensional data in the biomedical sciences – a result of breakthroughs in measurement technologies like single cell sequencing, proteomics, fMRI and vast improvements in health record data collection and storage. While these large datasets, containing millions of cellular or patient observations hold great potential for understanding the generative mechanisms, the state space of the data, as well as causal interactions driving development, disease and progression, they also pose new challenges in terms of noise, missing data, measurement artifacts, and the so-called “curse of dimensionality.” My research has been addressing these issues, by developing denoised data representations that are designed for data exploration, mechanistic understanding, and hypothesis generation. My lab is at the forefront of unsupervised learning where we have uncovered a unifying principle in scientific data, particularly from biomedical systems: the manifold principle[Moon 2018], which has been previously used in mathematical constructs. Biological entities have a myriad of parts (genes, epigenetics, proteins, signals) that can be measured to result in a high ambient space, but by their nature, they intrinsically lie in low dimensional, smoothly varying spaces. For instance, single-cell RNA-sequencing (scRNA-seq) measurements often have 1000s of gene dimensions. These genes cannot be acting individually, uncoordinated with other genes, and must be informationally redundant, thus lowering intrinsic dimension to the 20-30 dimensions we see in practical datasets [van Dijk Cell 2018, Moon Nature Biotechnology 2019]. We have followed the far-reaching effects of the manifold assumption to devise both graph spectral and deep learning methods such as: MAGIC [van Dijk et al. Cell 2018] to denoise data by restoration to low frequency spectral dimensions; PHATE [Moon Nature Biotech 2019] and multiscale PHATE [Kuchroo Nature Biotechnology 2022] to create manifold, affinity-preserving low dimensional embeddings and visualizations [Moon Nature Biotechnology 2019], MELD to understand effects of experimental perturbation [Burkhardt et al. Nature Biotechnology 2021]; TrajectoryNet to learn high dimensional trajectories from static snapshot data [Tong ICML 2020], and many more. The productivity of my lab and its impact on many fields can be seen in our broad publication profile which includes: machine learning venues like ICML, NeurIPS, IDA (Intelligent Data Analysis), IJCNN, and CVPR; signal processing venues such as IEEE MLSP (Machine Learning for signal Processing), IEEE Big Data, Journal of Signal Processing; applied math venues such as SAMPTA (Sampling Theory and Applications), SIAM Data Mining, as well as biomedical journals such as Science, Cell, Nature, Nature Biotechnology, and Nature Methods. I have been recognized for these contributions with the NSF CAREER grant, Sloan Faculty Fellowship, FASEB Excellence in Science Award, two NIH (NIGMS) R01 grants, a joint NSF/NIH grant as well as grants from private foundations such as CZI, Novo Nordisk and Simons Foundation—all as PI or co-PI. In addition, I have been designated as the Dean’s Faculty Fellow of the Yale School of Medicine.
Rex Ying – My research spans 3 broad areas: deep learning for graphs, geometric representation learning, and real-world applications with relational reasoning and modeling. In the past, I created many widely used GNN algorithms such as GraphSAGE, PinSAGE and GNNExplainer. In addition, I have worked on a variety of applications of graph learning in physical simulations, social networks, knowledge graphs and biology. I developed the first billion-scale graph embedding services at Pinterest, and the graph-based anomaly detection algorithm at Amazon.
Stephen Slade – My interests include goal based decision making, explanation, philosophy of mind, interpersonal relationships, fintech, politics, and ethics. For more information, see my course on Automated Decision Systems.
Mark Gerstein – Current research foci in the lab include disease genomics (particularly neurogenomics and cancer genomics), human genome annotation, genomic privacy, network science, wearable and molecular image data analysis, text mining of the biological-science literature and macromolecular simulation.
Steven Zucker – The current research focus of the ZuckerLab includes (1) inferring manifolds of neurons in the mouse visual system, and associated theoretical problems in manifold learning; (2) applying topological methods to understand human shape inferences; and (3) applying computational modeling to understand the role of color in material and shape perception.