This is an international workshop on weakly supervised learning. The main goal of this workshop is to discuss challenging research topics in weakly supervised learning areas such as semi-supervised learning, positive-unlabeled learning, label noise learning, partial label learning, and self-supervised learning, as well as the foster collaborations among universities and institutes.
Program
Venue & Date
Venue: OpenSpace of RIKEN Center for Advanced Intelligence Project (RIKEN AIP), Tokyo, Japan. Please click here for more access details of RIKEN AIP.
Date: November 23 09:30--17:40, and November 24 09:30--16:00 (Asia/Tokyo time (UTC+9)).
URL for Online Attending: https://c5dc59ed978213830355fc8978.doorkeeper.jp/events/164877
Schedule
The workshop will use Asia/Tokyo time (UTC+9) for scheduling, and it will be combined with invited talks and contributed talks.
Time | Event |
---|---|
November 23 | |
09:30 -- 09:35 | Opening Ceremony |
Host: Masashi Sugiyama (RIKEN / The University of Tokyo) | |
09:35 -- 10:35 | Session 1 (25min talk + 5min Q&A for each speaker) |
Title: Enhancing Language Models through Improved Pre-Training and Fine-Tuning Speaker: James Tin Yau Kwok (Hong Kong University of Science and Technology) | |
Title: The "Gene" of Machine Learning: Make Machines Learn Like Humans Speaker: Xin Geng (Southeast University) | |
10:35 -- 10:55 | Bio Breaks and Social |
10:55 -- 11:55 | Session 2 (25min talk + 5min Q&A for each speaker) |
Title: Towards Safe Abductive Learning with Weak Label and Weak Rules Speaker: Yu-Feng Li (Nanjing University) | |
Title: Visual Recovery towards Understanding Speaker: Chang Xu (The University of Sydney) | |
12:00 -- 13:00 | Lunch Time |
13:00 -- 14:00 | Session 3 (25min talk + 5min Q&A for each speaker) |
Title: Late stopping for weakly supervised learning Speaker: Tongliang Liu (The University of Sydney) | |
Title: Open-World Learning: Challenges and Solutions Speaker: Chen Gong (Nanjing University of Science and Technology) | |
14:00 -- 14:20 | Bio Breaks and Social |
14:20 -- 15:50 | Session 4 (25min talk + 5min Q&A for each speaker) |
Title: Exploring Trustworthy Machine Learning under Imperfect Data Speaker: Bo Han (Hong Kong Baptist University) | |
Title: Label-Agnostic Unlearning in Deep Models Speaker: Miao Xu (The University of Queensland) | |
Title: An Information-Theoretic Analysis of Learning under General Data Corruption Speaker: Nan Lu (The University of Tübingen) | |
15:50 -- 16:10 | Bio Breaks and Social |
16:10 -- 17:10 | Session 5 (25min talk + 5min Q&A for each speaker) |
Title: Robust Contrastive Learning and Its Applications Speaker: Shuo Chen (RIKEN) | |
Title: Is the Performance of My Deep Network Too Good to Be True? A Direct Approach to Estimating the Bayes Error in Binary Classification Speaker: Takashi Ishida (RIKEN / The University of Tokyo) | |
17:10 -- 18:00 | Open Discussion |
18:00 -- 20:30 | Closed Discussion / Dinner |
November 24 | |
09:30 -- 10:30 | Session 1 (25min talk + 5min Q&A for each speaker) |
Title: Emerging Drug Interaction Prediction from Biomedical Network Speaker: Quanming Yao (Tsinghua University) | |
Title: Domain Generalization via Content Factors Isolation: A Two-level Latent Variable Modelling Approach Speaker: Mingming Gong (The University of Melbourne) | |
10:30 -- 10:50 | Bio Breaks and Social |
10:50 -- 11:50 | Session 2 (25min talk + 5min Q&A for each speaker) |
Title: Unmasking and Improving Data Credibility: A Study with Datasets for Training Harmless Language Models Speaker: Yang Liu (University of California, Santa Cruz) | |
Title: Towards Robust Foundation Models: Adversarial Contrastive Learning Speaker: Jingfeng Zhang (The University of Auckland) | |
11:50 -- 13:00 | Lunch Time |
13:00 -- 14:00 | Session 3 (25min talk + 5min Q&A for each speaker) |
Title: Towards Continuous Adaptation in Non-stationary Environments Speaker: Zhen-Yu Zhang (RIKEN) | |
Title: Towards Robust Deep Learning under Distribution Shift: an Importance Weighting Approach Speaker: Tongtong Fang (The University of Tokyo) | |
14:00 -- 14:20 | Bio Breaks and Social |
14:20 -- 15:20 | Session 4 (25min talk + 5min Q&A for each speaker) |
Title: Weakly Supervised Disentanglement Speaker: Yivan Zhang (The University of Tokyo) | |
Title: On the Relation between Complementary-Label Learning and Negative-Unlabeled Learning Speaker: Wei Wang (The University of Tokyo) | |
15:20 -- 16:00 | Closing Ceremony |
Host: Masashi Sugiyama (RIKEN / The University of Tokyo) |
Topics
Overview
Machine learning should not be accessible only to those who can pay. Specifically, modern machine learning is migrating to the era of complex models (e.g., deep neural networks), which require a plethora of well-annotated data. Giant companies have enough money to collect well-annotated data. However, for startups or non-profit organizations, such data is barely acquirable due to the cost of labeling data or the intrinsic scarcity in the given domain. These practical issues motivate us to research and pay attention to weakly supervised learning (WSL), since WSL does not require such a huge amount of annotated data. We define WSL as the collection of machine learning problem settings and algorithms that share the same goals as supervised learning but can only access to less supervised information than supervised learning. In this workshop, we discuss both theoretical and applied aspects of WSL.
This workshop is a series of our previous workshops at ACML 2019, SDM 2020, ACML 2020, IJCAI 2021, and ACML 2021. Our particular technical emphasis at this workshop is incomplete supervision, inexact supervision, inaccurate supervision, cross-domain supervision, imperfect demonstration, and weak adversarial supervision. Meanwhile, this workshop will also focus on WSL for Science and Social Good, such as WSL for COVID-19, WSL for healthcare, WSL for climate change, WSL for remote sensing, and new public WSL datasets regarding the scientific scenarios.
Topics of Interest
WSL workshop includes (but not limited to) the following topics:
- Algorithms and theories of incomplete supervision, e.g., semi-supervised learning, active learning, and positive-unlabeled learning;
- Algorithms and theories of inexact supervision, e.g., multi-instance learning, complementary learning, and open-set learning;
- Algorithms and theories of inaccurate supervision, e.g., crowdsourced learning and label-noise learning;
- Algorithms and theories of cross-domain supervision, e.g., zero-/one-/few-shot learning, transferable learning, and multi-task learning;
- Algorithms and theories of imperfect demonstration, e.g., inverse reinforcement learning and imitation learning with non-expert demonstrations;
- Algorithms and theories of adversarial weakly-supervised learning, e.g., adversarial semi-supervised learning and adversarial label-noisy learning;
- Algorithms and theories of self-supervision, e.g., contrastive learning and autoencoder learning;
- Broad applications of weakly supervised learning in the field of computer science, such as weakly supervised object detection (computer vision), weakly supervised sequence modeling (natural language processing), and weakly supervised cross-media retrieval (information retrieval).
- WSL for science and social good, such as WSL for COVID-19, WSL for healthcare, WSL for climate change, and WSL for remote sensing, meanwhile, new public datasets regarding the above WSL research directions (new focus).
Further Descriptions
The focus of this workshop is six types of weak supervision: incomplete supervision, inexact supervision, inaccurate supervision, cross-domain supervision, imperfect demonstration, and weak adversarial supervision, which are briefly introduced below.
- Incomplete supervision considers a subset of training data given with ground-truth labels while the other data remain unlabeled, such as semi-supervised learning and positive-unlabeled learning.
- Inexact supervision considers the situation where some supervision information is given but not as exacted as desired, i.e., only coarse-grained labels are available. For example, if we are considering to classify every pixel of an image, rather than the image itself, then ImageNet becomes a benchmark with inexact supervision. Besides, multi-instance learning belongs to inexact supervision, where we do not exactly know which instance in the bag corresponds to the given ground-truth label.
- Inaccurate supervision considers the situation where the supervision information is not always the ground-truth, such as label-noise learning.
- Cross-domain supervision considers the situation where the supervision information is scarce or even non-existent in the current domain but can be possibly derived from other domains. Examples of cross-domain supervision appear in zero-/one-/few-shot learning, where external knowledge from other domains is usually used to overcome the problem of too few or even no supervision in the original domain.
- Imperfect demonstration considers the situation for inverse reinforcement learning and imitation learning, where the agent learns with imperfect or non-expert demonstrations. For example, AlphaGo learns a policy from a sequence of states and actions (expert demonstration). Even if an expert player wins a game, it is not guaranteed that every action in the sequence is optimal.
- Weak adversarial supervision considers the situation where weak supervision meets adversarial robustness. Since machine learning models are increasingly deployed in real-world applications, their security attracts more and more attention from both academia and industry. Therefore, many robust learning algorithms aim to prevent various evasion attacks, e.g., adversarial attacks, privacy attacks, model stealing attacks, and so on. However, almost all those robust algorithms (against evasion attacks) implicitly assume the strong supervision signals (no noisy labels in the training data), which hardly meets the requirements in practice. Therefore, when we develop evasion-robust algorithms, it is very practical/urgent to consider the supervision signals are imperfect.
- Self-supervision considers the unsupervised situation, and it pre-trains a generic feature representation by autonomously building the pseudo supervision (e.g., the similarity contrast and sample reconstruction) from the raw data, the learned representation can be applied in various downstream tasks such as classification, retrieval, and clustering.
Meanwhile, this workshop will continue discussing broad applications of weakly supervised learning in the field of computer science, such as weakly supervised object detection (computer vision), weakly supervised sequence modeling (natural language processing), and weakly supervised cross-media retrieval (information retrieval).
Organizers
Program Chairs
Shuo Chen, RIKEN, Japan.
Yang Liu, University of California, Santa Cruz, USA.
Feng Liu, The University of Melbourne, Australia.
Bo Han, Hong Kong Baptist University, Hong Kong SAR, China.
Tongliang Liu, The University of Sydney, Australia.
Gang Niu, RIKEN, Japan.
Masashi Sugiyama, RIKEN / The University of Tokyo, Japan.
Advisory Board (alphabetical order by last name)
Chen Gong, Nanjing University of Science and Technology, China.
Mingming Gong, The University of Melbourne, Australia.
Yu-Feng Li, Nanjing University, China.
Miao Xu, The University of Queensland, Australia.
Quanming Yao, Tsinghua University / 4Paradigm Inc., China.
Jingfeng Zhang, The University of Auckland, New Zealand.
Min-Ling Zhang, Southeast University, China.
Previous Workshops
ACML2022 WSL Workshop, Online.
ACML2021 WSL Workshop, Online.
IJCAI2021 WSRL Workshop, Online.
ACML2020 WSRL Workshop, Online.
SDM2020 WSUL Workshop, Ohio, United States.
ACML2019 WSL Workshop, Nagoya, Japan.