Back to Search Start Over

Doppelg\'anger's Watch: A Split Objective Approach to Large Language Models

Authors :
Ghasemlou, Shervin
Katiyar, Ashish
Saraf, Aparajita
Moon, Seungwhan
Pujari, Mangesh
Donmez, Pinar
Damavandi, Babak
Kumar, Anuj
Publication Year :
2024

Abstract

In this paper, we investigate the problem of "generation supervision" in large language models, and present a novel bicameral architecture to separate supervision signals from their core capability, helpfulness. Doppelg\"anger, a new module parallel to the underlying language model, supervises the generation of each token, and learns to concurrently predict the supervision score(s) of the sequences up to and including each token. In this work, we present the theoretical findings, and leave the report on experimental results to a forthcoming publication.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2409.06107
Document Type :
Working Paper