Détail du document
Identifiant

oai:arXiv.org:2402.15505

Sujet
Computer Science - Machine Learnin... Computer Science - Artificial Inte... Computer Science - Computer Vision...
Auteur
Liu, Yuejiang Alahi, Alexandre
Catégorie

Computer Science

Année

2024

Date de référencement

28/02/2024

Mots clés
weak-to-strong computer
Métrique

Résumé

Steering the behavior of a strong model pre-trained on internet-scale data can be difficult due to the scarcity of competent supervisors.

Recent studies reveal that, despite supervisory noises, a strong student model may surpass its weak teacher when fine-tuned on specific objectives.

Yet, the effectiveness of such weak-to-strong generalization remains limited, especially in the presence of large capability gaps.

In this paper, we propose to address this challenge by harnessing a diverse set of specialized teachers, instead of a single generalist one, that collectively supervises the strong student.

Our approach resembles the classical hierarchical mixture of experts, with two components tailored for co-supervision: (i) we progressively alternate student training and teacher assignment, leveraging the growth of the strong student to identify plausible supervisions; (ii) we conservatively enforce teacher-student and local-global consistency, leveraging their dependencies to reject potential annotation noises.

We validate the proposed method through visual recognition tasks on the OpenAI weak-to-strong benchmark and additional multi-domain datasets.

Our code is available at \url{https://github.com/yuejiangliu/csl}.

;Comment: Preprint

Liu, Yuejiang,Alahi, Alexandre, 2024, Co-Supervised Learning: Improving Weak-to-Strong Generalization with Hierarchical Mixture of Experts

Document

Ouvrir

Partager

Source

Articles recommandés par ES/IODE IA

Systematic druggable genome-wide Mendelian randomization identifies therapeutic targets for lung cancer
agphd1 subtypes replication hykk squamous cell gene carcinoma causal targets mendelian randomization cancer analysis