Skip to yearly menu bar Skip to main content


Timezone: America/Chicago

Registration Desk Mon 11 Dec 07:30 a.m.  


Affinity Workshop: Women in ML Mon 11 Dec 08:00 a.m.  

Nataša Tagasovska · Eda Okur · Hewitt Tusiime · Megha Srivastava · Jessica Sorrell · Estefany Kelly Buchanan · Shweta Khushu

Affinity Workshop: New in ML Mon 11 Dec 08:15 a.m.  

Mélisande Teng · Nishanth Anand · Reyhane Askari Hemmat · Subhrajyoti Dasgupta · Diganta Misra · Beheshteh Toloueirakhshan · Zach Xu · Isabelle Guyon · Rohan Sukumaran · Zhimeng Jiang

Affinity Workshop: Black in AI Mon 11 Dec 08:15 a.m.  

Marquita Riggins · Gelyn Watkins

Affinity Workshop: Latinx in AI Mon 11 Dec 08:15 a.m.  

Ignacio G. Lopez-Francos · CJ Barberan · Francisco Zabala · Karla Caballero · Cleber Zanchettin · Ana Maria Quintero · Vítor Lourenço · Walter M Mayor · Vinicius Caridá · Sebastian Caldas · Brayan Ortiz · Gabriela L. Vega Lopez · Abel Reyes-Angulo · Luis G. Sanchez Giraldo · Rocio Athziri Padilla Medina · Aaron Ferber · Marco Sanchez Sorondo · Laura Montoya

The 6th Annual LXAI Research Workshop, held alongside NeurIPS conference, is a one-day event that unites faculty, researchers, practitioners, and students globally to foster collaborations and exchange novel ideas in the AI field. Spotlighting the contributions of the Latinx/Hispanic community, the workshop offers a platform to discuss current research trends and showcase innovative work. The agenda includes sessions with renowned and early-career speakers, oral presentations, industry and mentoring panels, and collaborative poster sessions, culminating in networking social events. While the primary presenters are from the Latinx/Hispanic community, all are welcome to join and enrich the dialogue.


Affinity Workshop: Muslims in ML Mon 11 Dec 08:30 a.m.  

Sanae Lotfi · Hammaad Adam · Hadeel Al-Negheimish · Sarah Fakhoury · Razan Baltaji · Marzyeh Ghassemi · Shakir Mohamed · Aya Salama · S. M. Ali Eslami · Tasmie Sarker

The Muslims in ML workshop seeks to promote awareness, collaboration, and the development of mitigation strategies to ensure that machine learning and artificial intelligence advancements are implemented fairly and equitably for Muslims worldwide. By bringing together a diverse range of experts and incorporating multiple perspectives and backgrounds, our workshop aims to examine the challenges and opportunities of integrating AI/ML in the lives of Muslims and those in Muslim-majority countries. The workshop's focus extends beyond religious identification, encompassing cultural association and proximity to the Muslim identity. This broad approach acknowledges the complexity and diversity within the Muslim community and emphasizes the importance of inclusivity and understanding in addressing the potential impact of AI/ML technologies.


Affinity Workshop: North Africans in ML Mon 11 Dec 08:30 a.m.  

SOFIA BOURHIM · Oumaima Hourrane

The North Africans in ML (NAML) aims to inspire and empower North Africans, enabling them to realize their full potential in the AI field up to publishing in top AI conferences. By fostering collaboration, networking, and skill development, the workshop intends to create a supportive community that celebrates and amplifies the contributions of North Africans to the world of machine learning.


Affinity Workshop: Indigenous in AI/ML Mon 11 Dec 09:00 a.m.  

Mason Grimshaw · Andrea M. Delgado-Olson · Michael Running Wolf

Indigenous In AI’s vision is to build an international community of Native, Aboriginal, and First Nations who will collectively transform their home communities with advanced technology. By elevating the voices of Indigenous ML researchers we will inspire future impactful work and break stereotypes. Additionally, this group will strive to educate the broader NeurIPS on contemporary indigenous issues relevant to information technology and practices.


Affinity Workshop: Queer in AI Mon 11 Dec 09:30 a.m.  

Jaidev Shriram · Sharvani Jha · Ruchira Ray · Sarthak Arora

Queer in AI’s workshop + socials at NeurIPS 2023 aim to act as a gathering space for queer folks to build community + solidarity while enabling participants to learn about key issues + topics at the intersection of AI and queerness.


Tutorial: Karsten Kreis · Ruiqi Gao · Arash Vahdat

Latent Diffusion Models: Is the Generative AI Revolution Happening in Latent Space?

Diffusion models have emerged as a powerful class of generative models and demonstrated astonishing results, in particular in image synthesis. However, training high-resolution diffusion models in pixel space can be highly expensive. Overcoming these limitations, Latent Diffusion Models (LDMs) first map high-resolution data into a compressed, typically lower-dimensional latent space using an autoencoder, and then train a diffusion model in that latent space more efficiently. Thereby, LDMs enable high-quality image synthesis while avoiding excessive compute demands. Furthermore, the LDM paradigm with an autoencoder, which can be tailored to specific problems and data, and a separate diffusion model in latent space offers significant flexibility with respect to architecture and model design. This has allowed LDMs to be successfully extended to various tasks beyond image generation, such as video synthesis, 3D object and scene generation, language modeling, and more. Most prominently, the well-known text-to-image model Stable Diffusion leverages the LDM framework. LDMs have become very popular and widely used in the generative modeling literature.

In this tutorial, we aim to provide an introduction to LDMs. While the literature on diffusion models has become broad, the LDM paradigm stands out as a particularly powerful approach due to its flexibility and excellent trade-off with respect to performance and compute demands. We aim to present a tutorial on LDMs that will benefit researchers interested in efficient and flexible, yet expressive generative modeling frameworks. We will also highlight advanced techniques for accelerated sampling and controllability, and discuss various applications of LDMs beyond image synthesis. Moreover, a panel discussion will provide diverse perspectives on this dynamic field and offer an outlook for future research on LDMs.




Tutorial: Stephanie Schoch · Ruoxi Jia · Yangfeng Ji

Data Contribution Estimation for Machine Learning

Tasks enabled by data contribution estimation (DCE) aid model improvement through data improvement. While benchmark DCE evaluation tasks show application across many ML domains, DCE has limited visibility in other research domains that stand to benefit from its use cases. We propose a tutorial on data contribution for machine learning to address this. This tutorial will provide an overview of DCE for machine learning and natural language processing. Following this tutorial, attendees will have gained an understanding of 1) broadly, what questions data contribution estimation aims to answer; 2) the theory and methods that are widely in use within the DCE community that can be applied to a broad range of domains; 3) DCE from the perspectives of large language models and privacy.




Tutorial: Emily Black · Hoda Heidari · Daniel Ho

Governance & Accountability for ML: Existing Tools, Ongoing Efforts, & Future Directions

The tutorial aims to familiarize the ML community with major existing AI governance frameworks, ongoing AI policy proposals worldwide, and the concrete tools the research community has developed to adhere to standards and regulations applicable to ML systems in socially high-stakes domains. As a concrete governance challenge, we will focus on issues of bias and unfairness and overview pipeline-centric approaches to operationalize algorithmic harm prevention. As we will discuss, this approach is particularly relevant to challenges around leveraging the disparate impact doctrine for algorithmic harm prevention and recent FTC advanced notice of proposed rulemakings (ANPRMs). The concluding expert panel is an opportunity for the ML community to hear diverse perspectives on the key AI governance challenges in the near future and how the ML research community can prepare for and support efforts to address those challenges.




Tutorial: Elizabeth Bondi-Kelly · Krishnamurthy Dvijotham · Matthew Taylor

How to Work With Real Humans in Human-AI Systems

As more and more AI systems are deployed in the real world, it becomes imperative to study these systems with real humans to avoid unexpected negative consequences during deployment. Yet, this can be challenging for researchers with more experience designing algorithms and less experience running human participant experiments, or deploying systems in the real world. In this tutorial, we will discuss the state of the human-AI collaboration field, emphasizing (i) incorporating humans into AI systems, including multi-agent, machine learning, and reinforcement learning systems, (ii) investigating when to rely on human vs. AI strengths, and (iii) designing human-AI studies to evaluate algorithms with real humans.

Elizabeth Bondi-Kelly

 

Elizabeth Bondi-Kelly is an Assistant Professor of Electrical Engineering and Computer Science at the University of Michigan. She has a PhD in Computer Science at Harvard University, where she was advised by Prof. Milind Tambe, and she was formerly a Postdoctoral Fellow at MIT through the CSAIL METEOR Fellowship. Her research interests are focused on artificial intelligence for social impact, particularly spanning the fields of multi-agent systems and data science. Her work, which has been published in venues such as AAAI, AAMAS, AIES, and IJCAI, has applications in conservation and public health, and has been deployed to support conservation efforts. She has been recognized as an MIT EECS Rising Star in 2021, and has been awarded the Best Paper Runner Up at AAAI 2021, Best Application Demo Award at AAMAS 2019, Best Paper Award at SPIE DCS 2016, and an Honorable Mention for the NSF Graduate Research Fellowship Program in 2017. She also founded and currently leads Try AI, a 501(c)(3) nonprofit committed to increasing diversity, equity, inclusion, and belonging in the field of AI through community-building educational programs, largely focused on AI for social impact.



Tutorial: Zhiting Hu · Tianmin Shu

Language Models Meet World Models

Large language models (LMs) have achieved remarkable success in many language tasks.
Recent works have also shown that knowledge of the world can emerge from large LMs, enabling large LMs to assist decision-making for embodied tasks. However, the world knowledge exhibited by the current large LMs is often not robust and cannot be grounded in physical environments without additional models. This hinders large LMs’ abilities to perform complex reasoning and planning tasks reliably. For example, in creating action plans to move blocks to a target state, GPT-3 achieves a success rate of only 1%, compared to 78% for humans.

On the other hand, humans perform deliberate reasoning and planning based on the mental model of the world (i.e., world model, WMs) that enables us to simulate actions and their effects on the world’s state. WMs encoding the knowledge of the physical world can drastically improve the data efficiency and robustness of intelligent agents. However, WMs were typically studied in reinforcement learning and robotics, which are conceptually distinct from problems studied in language modeling.

This gap indicates enormous new opportunities for connecting WMs and LMs, to enhance LM capabilities of reasoning/planning in both embodied and general settings, and address the aforementioned limitations. Emerging studies on the intersection of WMs and LMs have demonstrated promising results. This tutorial aims to summarize and present a unified view of connecting WMs and LMs and highlight the various opportunities for improved machine reasoning and planning based on (or even beyond) large LMs through world modeling. We will review recent works on learning WMs and on using them to further learn and perform embodied tasks. We will show how LMs can utilize external WMs to compensate for their lack of grounded world knowledge and how LMs themselves can learn world models from embodied experiences that are beyond text data and use the internal WMs to guide complex reasoning.




Tutorial: Michal Derezinski · Michael Mahoney

Recent and Upcoming Developments in Randomized Numerical Linear Algebra for ML

Large matrices arise in many ML applications, including as representations of datasets, graphs, model weights, first and second-order derivatives, etc. Randomized Numerical Linear Algebra (RandNLA) is an area that uses randomness to develop improved algorithms for ubiquitous matrix problems. The area has reached a certain level of maturity, and current efforts of incorporating RandNLA algorithms into core numerical libraries, as well as recent advances in ML, Statistics, and Random Matrix Theory, have led to new theoretical and practical challenges. This tutorial will provide a self-contained overview of RandNLA in light of these important developments.

Michal Derezinski

 

Michal Derezinski is an Assistant Professor of Computer Science and Engineering at the University of Michigan. Previously, he was a postdoctoral fellow in the Department of Statistics at the University of California, Berkeley, and a research fellow at the Simons Institute for the Theory of Computing. He obtained his PhD in Computer Science at the University of California, Santa Cruz, with a dissertation on importance sampling methods for machine learning. His research interests include the theoretical foundations of randomized algorithms for machine learning, optimization, and data science, with a particular focus on applying randomized sampling and sketching methods to large-scale stochastic optimization. He co-authored a survey on ``Determinantal Point Processes in Randomized Numerical Linear Algebra'', published in Notices of the AMS, and he received the Best Paper Award at Neurips 2020 for his work uncovering a connection between RandNLA methods and the double descent curve.



Tutorial: James Demmel · Yang You

Contributing to an Efficient and Democratized Large Model Era

The success of the Transformer model has pushed the limits of deep learning to operate on the scale of trillions of parameters. This proliferation of large model size has outpaced the advances in hardware, resulting in an urgent need to distribute enormous models across multiple GPUs. Despite this trend, best practices for choosing an optimal strategy are still lacking due to the breadth of knowledge required across both deep learning and parallel computing.
This drives researchers to question deeply about: How to improve the training and inference efficiency of large models to reduce costs? Can we accommodate larger models with limited resources? What efforts can we make to enable more AI community members to access big models easily? In this tutorial, we investigate the efforts to solving above problems. A diverse set of parallelism is an important tool to improving the efficiency of large model training and inference. Heterogeneous memory management can enhance the model accommodation capacity of processors (e.g. GPUs).Further, deep learning systems for large AI models will significantly reduce the specialized background knowledge required from users, allowing AI users to quickly get started with larger models. We believe that with the benefits of these effective and extensive technologies for AI models, realizing an efficient and democratic big model era has become possible. We will provide participants with a systemic open-source solution and practical demonstrations for big models, in the hope of encouraging more practitioners and helping them apply mentioned technologies to their own practice.




Affinity Workshop: Global South AI Mon 11 Dec 11:00 a.m.  

Susanna Raj · Pariya Sarin · Sudha Jamthe

Global South in AI has the mission to add inclusion to Language AI. They focus on training new researchers from Global South Languages and Countries to present posters (peer reviewed selection) and bring them to NeurIPS to collaborate.


Mentorship: Education Outreach Mon 11 Dec 12:00 p.m.  

Tristan Naumann · Sanmi Koyejo · Marzyeh Ghassemi · Saadia Gabriel

This event is by invitation only

12:00 - 12:10

  • Welcoming students
  • Give out lunch

12:10 - 12:25

  • Opening remark with Tristan Naumann
  • Welcome to NeurIPS
  • Purpose/history/vision of NeurIPS

12:30 - 12:55

  • Outreach program
  • Purpose of the program
  • Last year’s success
  • How many schools/students participated
  • This year’s scope
  • NeurIPS schedule announcement
  • Affinity workshop & Tutorial
  • Welcome reception
  • Main tracks
  • Keynote speeches
  • Workshops

13:00 - 13:30

  • Fireside Chat with Sanmi Koyejo, Marzyeh Ghassemi, Saadia Gabriel
  • Theme: Becoming a successful AI researcher/engineer as a budding student

13:30 - 13:40

  • Q&A
  • Wrap-up

Mentorship: Science Communication for AI Researchers - A Quick Introduction Mon 11 Dec 12:45 p.m.  

Lucy Smith

Would you like to learn how to communicate your AI research to a general audience? In this short tutorial you will learn how to turn your research articles into blog posts, how to use social media to promote your work, and how to avoid hype when writing about your research. You will also hear from AI researchers on how science communication has helped them improve their communication skills, and made their research more visible and impactful.

One of the challenges facing the field of AI is its portrayal in the media, which leads to misconceptions among policy makers, business leaders, and the general public alike. By communicating about AI in a clear, informed, and measured manner we can help to combat the flow of misinformation and convey the reality of today’s technology.

We will guide participants on how to quickly shape the story of their AI research. We’ll focus on how to structure this research story to form a blog post. Participants will learn how to explain their research to a general audience in a clear and concise manner, and how to find suitable images to illustrate their work.

After the session we will host a two-hour drop-in session to work with you one-on-one on your sci-comm questions, ideas and stories.


Tutorial: Mihaela van der Schaar · Isabelle Guyon · Nabeel Seedat · Jennifer Wortman Vaughan · Kyunghyun Cho · Razvan Pascanu · Jim Weatherall

Data-Centric AI for reliable and responsible AI: from theory to practice

Data-Centric AI has recently been raised as an important paradigm shift in machine learning and AI — placing the previously undervalued “data work’ at the center of AI development. This tutorial aims to illuminate the fundamentals of Data-Centric AI and articulate its transformative potential. We will explore the motivation behind the data-centric approach, highlighting the power to improve model performance, engender more trustworthy, fair, and unbiased AI systems, as well as discuss benchmarking from a data-centric perspective. Our examination extends to standardized documentation frameworks, exposing how they form the backbone of this new paradigm. The tutorial will cover state-of-the-art methodologies that underscore these areas, which we will contextualize around the high-stakes setting of healthcare. A focus of this tutorial is providing participants with an interactive and hands-on experience. To this end, we provide coding/software tools and resources, thereby enabling practical engagement. The panel discussion, with experts spanning diverse industries, will provide a dynamic platform for discourse, enabling a nuanced understanding of the implications and limitations of Data-Centric AI across different contexts. Ultimately, our goal is that participants gain a practical foundation in data-centric AI, such that they can use or contribute to Data-Centric AI research.

Isabelle Guyon

 

Isabelle Guyon recently joined Google Brain as a research scientist. She is also professor of artificial intelligence at Université Paris-Saclay (Orsay). Her areas of expertise include computer vision, bioinformatics, and power systems. She is best known for being a co-inventor of Support Vector Machines. Her recent interests are in automated machine learning, meta-learning, and data-centric AI.  She has been a strong promoter of challenges and benchmarks, and is president of ChaLearn, a non-profit dedicated to organizing machine learning challenges. She is community lead of Codalab competitions, a challenge platform used both in academia and industry. She co-organized the “Challenges in Machine Learning Workshop” @ NeurIPS between 2014 and 2019, launched the "NeurIPS challenge track" in 2017 while she was general chair, and pushed the creation of the "NeurIPS datasets and benchmark track" in 2021, as a NeurIPS board member.



Tutorial: Aditya Gopalan · Aadirupa Saha · Yoshua Bengio · Craig Boutilier · Elad Hazan · Robert Nowak · Tobias Schnabel

Do You Prefer Learning with Preferences?

AI desires to imitate human intelligence for designing efficient decision-making systems, but are we really training them the way humans learn every day or take decisions? Studies have shown humans are inherently more comfortable making decisions on a relative scale or choosing alternatives from a set, which often helps us converge to an optimal decision faster. In recent times, as we are employing more and more AI tools for executing everyday tasks, it’s becoming necessary to align machine behavior with human-like decisions. Another critical challenge in training user-friendly systems lies in the requirement of a huge amount of human feedback, which is often costly and hard to obtain. The solution lies in learning to train our machines through human preferences! Our tutorial aims to address the critical need for educating researchers on different types of preference models by exploring real-world problems and showcasing how training systems through preference feedback can provide cutting-edge solutions. We will equip attendees with a comprehensive understanding of diverse preference models and inference techniques. Another goal of the tutorial is to encourage collaboration among various communities that have significant connections to preference-based learning, including bandits, multiagent games, econometrics, social choice theory, RL, optimization, robotics, and more. We will consider our tutorial a success if it inspires researchers to embark on novel insights in the general area of preference-based learning, bringing attention from different communities to foster dissemination, cross-fertilization, and discussion at scale. Let’s learn to train our machines like humans: Machine Learning meets Human Learning through preference feedback!

Tutorial website: https://sites.google.com/view/pref-learning-tutorial-neurips/home




Tutorial: Zhangir Azerbayev · Emily First · Albert Q. Jiang · Kaiyu Yang · Anima Anandkumar · Noah Goodman · Alex Sanchez-Stern · Dawn Song · Sean Welleck

Machine Learning for Theorem Proving

Machine learning, especially large language models (LLMs), has shown promise in proving formal theorems using proof assistants such as Coq, Isabelle, and Lean. Theorem proving is an important challenge for machine learning: Formal proofs are computer programs whose correctness can be verified. Therefore, theorem proving is a form of code generation with rigorous evaluation and no room for the model to hallucinate, opening up a new avenue for addressing LLMs’ flaws in factuality. Despite its potential, learning-based theorem proving has significant entry barriers, primarily due to the steep learning curve for proof assistants. This tutorial aims to bridge this gap and make theorem proving accessible to researchers with a general machine learning background. To that end, our presentation will contextualize theorem proving from a machine learning perspective and demonstrate how to develop LLMs for theorem proving, using newly available open-source tools that provide interfaces to proof assistants without requiring in-depth knowledge of their internals. Furthermore, we will cover advanced topics and open problems in learning-based theorem proving, including its synergies with natural language processing and software verification. Throughout the presentation, we will highlight several conceptual themes recurring in theorem proving that are also critical for machine learning, such as mathematical reasoning, code generation, and hallucination prevention. The panel will complement the presentation through a broader discussion of related topics such as trustworthy machine learning, LLMs for code, reasoning, and program synthesis.

Kaiyu Yang

 

Kaiyu Yang is a postdoctoral researcher at Caltech in the Computing + Mathematical Sciences (CMS) Department, working with Prof. Anima Anandkumar. His research aims to make machine learning capable of symbolic reasoning. It includes (1) applying machine learning to symbolic reasoning tasks, such as automated theorem proving; and (2) introducing symbolic components into machine learning models to make them more interpretable, verifiable, and data-efficient. In addition, he has also worked on constructing and analyzing machine learning datasets, especially focusing on fairness, privacy, and mitigating dataset bias. His research is recognized with a Siebel Scholar award. Before joining Caltech, he received his Ph.D. from the Department of Computer Science at Princeton University, advised by Prof. Jia Deng.



Tutorial: Nihar Shah · Hugo Larochelle · Andrew McCallum · Alice Oh · - Mausam · Charvi Rastogi

What can we do about NeurIPS Reviewer #2?Challenges, Solutions, Experiments and Open Problemsin Peer Review

Peer review is fundamental to scientific research, impacting scientific progress, grant funding allocation, researcher well-being, career paths, and the public's view of science. This tutorial provides a scientific lens on the systemic issues in peer review. It aims to stimulate discussions and inform policy-making based on scientific evidence (rather than individual opinions or anecdotes), addressing a topic that directly affects us all. To this end, the tutorial will delve into various inherent challenges, drawing on experiments on the peer-review process in diverse scientific disciplines. It will also discuss viable solutions and important open problems. The tutorial material will be available at https://cs.cmu.edu/~nihars/tutorials/NeurIPS2023. Finally, the presenter is excited about two things—peer review and minions—and both of these will be reflected generously in the tutorial.




Tutorial: Jiashuo Liu · Tianhui Cai · Peng Cui · Hongseok Namkoong

Modeling and Exploiting Data Heterogeneity under Distribution Shifts

Data heterogeneity is a key determinant of the performance of ML systems. Standard algorithms that optimize for average-case performance do not consider the presence of diversity within data. As a result, variations in data sources, data generation mechanisms, and sub-populations lead to unreliable decision-making, poor generalization, unfairness, and false scientific discoveries. Carefully modeling data heterogeneity is a necessary step in building reliable data-driven systems. Its rigorous study is a nascent field of research spanning several disciplines, including statistics, causal inference, machine learning, economics, and operations research. In this tutorial, we develop a unified view of the disparate intellectual threads developed by different communities. We aim to foster interdisciplinary research by providing a unified view based on a shared language. Drawing upon several separate literatures, we establish a taxonomy of heterogeneity and present quantitative measures and learning algorithms that consider heterogeneous data. To spur empirical progress, we conclude by discussing validation protocols and benchmarking practices.




Tutorial: Spencer Frei · Vidya Muthukumar · Fanny Yang · Arash Amini · Kamalika Chaudhuri · Daniel Hsu · Nati Srebro · Chiyuan Zhang

Reconsidering Overfitting in the Age of Overparameterized Models

Large, overparameterized models such as neural networks are now the workhorses of modern machine learning. These models are often trained to near-zero error on noisy datasets and simultaneously generalize well to unseen data, in contrast to the textbook intuition regarding the perils of overfitting. At the same time, near-perfect data-fitting can have severe issues in the context of robustness, privacy, and fairness. Classical theoretical frameworks provide little guidance for navigating these questions due to overparameterization. It is thus crucial to develop new intuition regarding overfitting and generalization that are reflective of these empirical observations. In this tutorial, we discuss recent work in the learning theory literature that provides theoretical insights into these phenomena.

See <https://www.cs.columbia.edu/~djhsu/>



Tutorial: Andrew Ng · Isa Fulford

Application Development using Large Language Models

The rise of large language models (LLMs) offers a new approach for quickly building AI applications. While LLMs such as ChatGPT, Bard, and Bing chat are widely understood as consumer tools, the best practices for developers to effectively use these models through API calls remain poorly understood. This tutorial will share with the NeurIPS audience best practices for building AI applications using LLMs.
This course will include, but also go significantly beyond, “prompt engineering.” We will share best practices for integrating LLMs into more complex software systems, evaluating and continually improving their performance, and enhancing their safety. We will discuss best practices for using LLMs in common operations such as summarizing, making inferences, transforming text, and expanding text, as well as in-context learning, fine-tuning, and the utilization of both open-source and proprietary cloud-hosted LLMs.
LLMs are transforming the development process of AI applications. For example, a sentiment classifier that used to take weeks to build, via a process of collecting and labeling training examples, tuning a supervised model, and then finally deploying the model to make inferences, can now be built in hours by prompting an LLM API.
Through this tutorial, we hope to connect research and practice, and also inspire researchers to pursue new directions relevant to how LLMs are being used today.




Affinity Poster Session: Joint Poster Session Mon 11 Dec 03:30 p.m.  


Remarks Mon 11 Dec 05:00 p.m.  


Invited Talk: Björn Ommer

NextGenAI: The Delusion of Scaling and the Future of Generative AI

Björn Ommer

 

Björn Ommer is a full professor at University of Munich where he is heading the Computer Vision & Learning Group. Before he was a full professor in the department of mathematics and computer science at Heidelberg University and a co-director of its Interdisciplinary Center for Scientific Computing. He received his diploma in computer science from University of Bonn, his PhD from ETH Zurich, and he was a postdoc at UC Berkeley. Björn serves as an associate editor for IEEE T-PAMI. His research interests include semantic scene understanding and retrieval, generative AI and visual synthesis, self-supervised metric and representation learning, and explainable AI. Moreover, he is applying this basic research in interdisciplinary projects within neuroscience and the digital humanities. His group has published a series of generative approaches, including "VQGAN" and "Stable Diffusion", which are now democratizing the creation of visual content and have already opened up an abundance of new directions in research, industry, the media, and beyond.



Welcome Reception Mon 11 Dec 06:15 p.m.  


Creative AI Performances 1 Mon 11 Dec 06:30 p.m.  

Jean Oh · Isabelle Guyon
Presentation
David R Rokeby
Abstract

Voice Scroll is a real-time voice to panorama generator. It can be used either in performance or as an interactive installation where the audience generates a continuously unfolding panorama by speaking.

Presentation
Abstract

“Emergent Rhythm” is an audio-visual DJ performance using real-time AI audio generation. Artist/DJ Tokui manipulates multiple models on stage to spontaneously generate rhythms and melodies. He then combines and mixes the generated audio loops to create musical developments. We employ AI audio synthesis models in real-time and faces unprecedented challenges: Everything heard during this performance is purely AI-generated sound.

As the title suggests, we focus on the musical and visual "rhythms" and recurring patterns that emerge in the interaction between multiple AI models and the artist. The accompanying visuals feature not only the periodicity over time but also the common patterns across multiple scales ranging from the extreme large-scale of the universe to the extreme small-scale of cell and atomic structures.

Aligning with the visual theme, we extracted loops from natural and man-made environmental sounds and used them as training data for audio generation. We also employ real-time timbre transfer that converts incoming audio into various singing voices, such as Buddhist chants. This highlights the diversity and commonality within the human cultural heritage.

We adapted the GAN (Generative Adversarial Networks) architecture for audio synthesis. StyleGAN models trained on spectrograms of various sound materials generate spectrograms, and vocoder GAN models (MelGAN) …

Video Presentation
Haru Ji
Abstract

How much work must the universe do, and how many dreams does it have to nurture, in order to grow a single tree? Then, how much of the universe does a forest harbor?

Entanglement, inspired by the motif of the forest, is a large-scale (16x16x4m) immersive artwork that invites spectators into a multi-sensory environment where visible and invisible worlds are interconnected and symbiotic. The artwork consists of three elements: the growth of trees through procedural modeling, generative AI that dreams images of trees and forests, and the operation of dynamic systems that connect tree roots with the mechanisms of fungi and bacteria–or of neural networks within a brain. Through the entanglement of microcosmic and simultaneous connections, it offers a sensory opportunity for contemplation and inspiration regarding ways of connecting with the world beyond ourselves, and a vision of an AI future that is fully present in its environment, as a diverse, living system in ecosystemic balance with the world. To borrow a phrase from Ursula Le Guin, our word for world is forest.

The artwork was produced using extensive custom software authored by the artists as well as SideFX Houdini and Stable Diffusion/ControlNet. Here we are using generative AI non-conventionally …

Video Presentation
Abstract

Fusion: Landscape and Beyond is an interdisciplinary art project that explores the relationship between memory, imagination, and Artificial Intelligence (AI) embodied in the century-long practices and discourse of Shan-Shui-Hua – Chinese landscape painting. It draws inspiration from the concept of Cultural Memory, where memories are selectively retrieved and updated based on present circumstances. The project considers text-to-image AI algorithms as analogous to Cultural Memory, as they generate diverse and imaginative images using pre-existing knowledge. In response to this analogy, the project introduces the concept of "AI memory" and situates it in the culturally significant Chinese landscape painting — a synthetic embodiment of creativity derived from the artist's memory.

Diversity plays both as a driving force and major inspiration for this project, which delves deeply into addressing the bias and the necessity for cultural diversity within the realm of machine-learning generative models for creative art. Recognizing that machines inherently exhibit bias stemming from their design and predominant use, it becomes essential to acknowledge and rectify such prejudices, particularly from a cultural standpoint. The initial phase of this project involves the fine-tuning of the Stable Diffusion model. The necessity for fine-tuning stems from the imperative to infuse a deeper cultural resonance within …

Video Presentation
Adam Cole
Abstract

Kiss/Crash is a multi-screen work exploring the subject of AI-imagery and representation as well as the autobiographical themes of loneliness, desire, and intimacy in the digital age. The installation consists of three individual works in a shared space, Kiss/Crash, Me Kissing Me, and Crash Me, Gently, all of which play with augmenting, inverting, and negating the iconic image of the kiss using AI image translation. Repurposing a classic Hollywood aesthetic through a queer lens, the piece reflects on the nature of images and places AI models within a history of image-production technologies meant to arouse and homogenize our desires. In the process, it reveals the logic of AI imagery and hints at how our relationship to reality will continue to be stretched and shaped by artificial representations at an accelerating pace. This piece celebrates diversity by bringing a unique queer perspective to generative AI, questioning how homogenous representations of love might haunt our AI-mediated future and how LGBT artists can playfully resist and invert that dominant narrative.

Video Presentation
Yonatan Bitton · Nitzan Bitton Guetta · Jack Hessel · · Yuval Elovici
Abstract

The WHOOPS! art gallery presents 500 AI-generated images that challenge common sense perceptions. Resulting from a collaboration between AI researchers and human designers, the collection underscores disparities in visual commonsense reasoning between machines and humans. While humans readily identify the anomalies, contemporary AI models struggle, highlighting gaps in AI understanding. This study offers insights into the evolving interplay between human cognition, art, and artificial intelligence.