top of page
Search

Look Ma, No RLHF!

  • Writer: VC Ramesh
    VC Ramesh
  • May 24, 2023
  • 1 min read

Superficial Alignment Hypothesis: A model’s knowledge and capabilities are learnt almost entirely during pretraining, while alignment teaches it which subdistribution of formats should be used when interacting with users. If this hypothesis is correct, and alignment is largely about learning style, then a corollary of the Superficial Alignment Hypothesis is that one could sufficiently tune a pretrained language model with a rather small set of examples.


 
 
 

Recent Posts

See All
Ask ChatGPT: Seq2Seq vs Causal LLM

I am not sure what to make of this conversation with ChatGPT. Is this what they mean by a "stochastic parrot"? :-) Choosing LLM...

 
 
 
Chatbots in the fast food industry

https://finance.yahoo.com/news/ai-making-name-itself-fast-162303086.html https://finance.yahoo.com/news/ordering-fast-food-robot-may-1100...

 
 
 
Rebirth of chatbots

A key significance of ChatGPT is that it has ignited the rebirth of chatbots. Chatbots, which first emerged around 2016, have been widely...

 
 
 

Comments


©2023 by vcrsoft.

bottom of page