Towards a Responsible Thinking in the New Era of Gen AI: Walking the Walk
Mona Diab, Carnegie Mellon University, United States
Scalable Vector Analytics: A Story of Twists and Turns
Themis Palpanas, University Paris Cite, France
Past, Present, Future, and Far Future of AI
Jürgen Schmidhuber, KAUST AI Initiative, Kingdom of Saudi Arabia; Swiss AI Lab IDSIA, Switzerland and NNAISENSE, Switzerland
Towards a Responsible Thinking in the New Era of Gen AI: Walking the Walk
Mona Diab
Carnegie Mellon University
United States
Brief Bio
Mona Diab is recently selected as ACL Fellow, and is Full Professor and the Director of the Language Technologies Institute within the school of Computer Science at Carnegie Mellon University (CMU), in Pittsburgh, USA. She directs the R3LIT (read relit) Lab. Prior to joining CMU, she was a Lead Responsible AI Research Scientist with Meta, and full Professor of Computer Science at the George Washington University, where she directed the CARE4Lang NLP Lab. Before joining Meta, she led the Lex Conversational AI project within Amazon AWS AI. Her current focus is on Responsible AI and how to operationalize it for NLP technologies. Her interests span building robust technologies for low resource scenarios with a special interest in Arabic technologies, (mis) information propagation, computational socio-pragmatics, computational psycholinguistics, NLG evaluation metrics, Language modeling and resource creation. Mona has served the community in several capacities: Elected President of SIGLEX and SIGSemitic, and the elected President of ACL SIGDAT, the board supporting EMNLP conferences. She helped establish two research trends in NLP, namely computational approaches to Code Switching and Semantic Textual Similarity. She is also a founding member of the SEM conference, one of the top tier conferences in NLP. Mona has published more than 250 peer reviewed articles.
Abstract
In a world of racing to get the best systems on leaderboards, winning best shared tasks, building the largest LLM, are we losing our soul as a scientific enterprise? Do we need to re-orient and re-pivot NLP? If so, what is needed to make this happen? Can we chart together a program where we ensure that science is the pivotal ingredient in CL/NLP? Could Responsible NLP be an avenue that could lead us back towards that goal? In this talk, in the spirit of my Responsible Thinking mission, I will explore some "practical" ideas around framing a Responsible NLP vision hoping to achieve a higher scientific standard for our field, addressing issues from the "how" we conduct our research and venturing into the "what" we work on and produce using tenets from responsible mindset perspective.
Scalable Vector Analytics: A Story of Twists and Turns
Themis Palpanas
University Paris Cite
France
Brief Bio
Themis Palpanas is an elected Senior Member of the French University
Institute (IUF), a distinction that recognizes excellence across all
academic disciplines, and Distinguished Professor of computer science
at the University Paris Cite (France), where he is director of the Data
Intelligence Institute of Paris (diiP), and director of the data
management group, diNo. He received the BS degree from the National
Technical University of Athens, Greece, and the MSc and PhD degrees from
the University of Toronto, Canada. He has previously held positions at
the University of California at Riverside, University of Trento, and at
IBM T.J. Watson Research Center, and visited Microsoft Research, and
the IBM Almaden Research Center. His interests include problems related
to data science (big data analytics and machine learning applications).
He is the author of 14 patents. He is the recipient of 3 Best Paper
awards, and the IBM Shared University Research (SUR) Award. His service
includes the VLDB Endowment Board of Trustees (2018-2023),
Editor-in-Chief for PVLDB Journal (2024-2025) and BDR Journal (2016-
2021), PC Chair for IEEE BigData 2023 and ICDE 2023 Industry and
Applications Track, General Chair for VLDB 2013, Associate Editor for
the TKDE Journal (2014-2020), and Research PC Vice Chair for ICDE 2020.
Abstract
Similarity search in high-dimensional data spaces was a relevant and
challenging data management problem in the early 1970s, when the first
solutions to this problem were proposed. Today, fifty years later, we
can safely say that the exact same problem is more relevant (from Time
Series Management Systems to Vector Databases) and challenging than
ever. Very large amounts of high-dimensional data are now omnipresent
(ranging from traditional multidimensional data to time series and deep
embeddings), and the performance requirements (i.e., response-time and
accuracy) of a variety of applications that need to process and analyze
these data have become very stringent and demanding. In these past fifty
years, high-dimensional similarity search has been studied in its many
flavors. Similarity search algorithms for exact and approximate, one-off
and progressive query answering. Approximate algorithms with and without
(deterministic or probabilistic) quality guarantees. Solutions for
on-disk and in-memory data, static and streaming data. Approaches based
on multidimensional space-partitioning and metric trees, random
projections and locality-sensitive hashing (LSH), product quantization
(PQ) and inverted files, k-nearest neighbor graphs and optimized linear
scans. Surprisingly, the work on data-series (or time-series) similarity
search has recently been shown to achieve the state-of-the-art
performance for several variations of the problem, on both time-series
and general high-dimensional vector data. In this talk, we will touch
upon the different aspects of this interesting story, present some of
the state-of-the-art solutions, and discuss open research directions.
Past, Present, Future, and Far Future of AI
Jürgen Schmidhuber
KAUST AI Initiative, Kingdom of Saudi Arabia; Swiss AI Lab IDSIA, Switzerland and NNAISENSE
Switzerland
Brief Bio
The New York Times headlined: "When A.I. Matures, It May Call Jürgen Schmidhuber 'Dad'." Since age 15, his main goal has been to build a self-improving A.I. smarter than himself, then retire. His lab's deep learning artificial neural networks have revolutionised machine learning and A.I. By 2017, they were on over 3 billion smartphones, and used billions of times per day, for Facebook’s automatic translation, Google’s speech recognition, Google Translate, Apple’s Siri & QuickType, Amazon’s Alexa, etc. He pioneered the principles of artificial curiosity & generative adversarial networks (1990, now widely used), neural network distillation (1991, now widely used), self-supervised pre-training for deep learning (1991, the "P" in "ChatGPT" stands for "pre-trained"), unnormalised linear Transformers (1991, the "T" in "ChatGPT" stands for "Transformer"), and meta-learning machines that learn to learn (since 1987). His lab also produced LSTM, the most cited AI of the 20th century, and the LSTM-inspired Highway Net, the first very deep feedforward net with hundreds of layers (ResNet, the most cited AI of the 21st century, is an open-gated Highway Net). Elon Musk tweeted: "Schmidhuber invented everything." He is recipient of numerous awards, Director of the AI Initiative at KAUST in KSA, Scientific Director of the Swiss AI Lab IDSIA, Adj. Prof. of A.I. at Univ. Lugano, and Co-Founder & Chief Scientist of the company NNAISENSE. He is a frequent keynote speaker at major events, and advising various governments on A.I. strategies.
Abstract
I’ll discuss modern Machine Learning, its historic context, and its expected impact on the future of the universe.