Context window 90% of adapter model full after single user prompt

I have been able to train an adapter on Google's Colaboratory.

I am able to start a LanguageModelSession and load it with my adapter.

The problem is that after one simple prompt, the context window is 90% full.

If I start the session without the adapter, the same simple prompt consumes only 1% of the context window.

Has anyone encountered this? I asked Claude AI and it seems to think that my training script needs adjusting. Grok on the other hand is (wrongly, I tried) convinced that I just need to tweak some parameters of LanguageModelSession or SystemLanguageModel.

Thanks for any tips.

Hey Michael! To better understand your issue, what parameters did you use when training your adapter (max seq length, pack sequence, batch size, etc)? Do you mind copying your AdapterTrainingConfiguration?

Are you using tools in your training data system message? Do you mind sharing a training sample entry?

Hi! Just want to check if this is still an issue after tool calling succeeded from your other thread post?

Context window 90% of adapter model full after single user prompt
 
 
Q