If you want in line completions, you need a model that is trained on “fill in the middle” tasks.
On their Huggingface page they even say that this is not supported and needs fine tuning:
We do not recommend using base language models for conversations. Instead, you can apply post-training, e.g., SFT, RLHF, continued pretraining, etc., or fill in the middle tasks on this model.
A model that can do it is:
starcoder2
codegemma
codellama
Another option is to just use the qwen model, but instead of only adding a few lines let it rewrite the entire function each time.
If you want in line completions, you need a model that is trained on “fill in the middle” tasks. On their Huggingface page they even say that this is not supported and needs fine tuning:
A model that can do it is:
Another option is to just use the qwen model, but instead of only adding a few lines let it rewrite the entire function each time.