Viewing a single comment thread. View all comments

WigglyHypersurface t1_j4f1r8b wrote

Did you forget to change the tokenizer?

8

GasZealousideal8691 OP t1_j4g8djf wrote

No, both use the GPT2 tokenizer. GPT-Neo uses GPT2Tokenizer.from_pretrained(‘EleutherAI/gpt-neo-1.3B)”, and GPT2 uses GPT2Tokenizer.from_pretrained(‘gpt2-xl’).

1

WigglyHypersurface t1_j4gpm5i wrote

What kind of head is on the models for the task?

1

GasZealousideal8691 OP t1_j4gpu8j wrote

GPT Neo is GPTNeoForCausalLM, and GPT2 is GPT2LMHeadModel. Like I said, I am not 100% familiar with these, but the huggingface docs listed both as “GPT-neo/GPT2 with an LM head”, so I figured they were analogous.

1