couple of things i am wondering:
1.it is generally 4G larger in terms of disk usage for a 72b-sized model,even without considering pissa init folder size. nou sure why
2.i can just funtune this model directly on a pre-processed 4 bit model and saved chekpoint will also be a 4 bit model ,yes?
3.last thing ,do u have pre-processed Qwen2.5 series models ? only saw Qwen2 on huggingface,not sure how much GPU i need to process a large 72b sized model
thanks ,for u attention on this matter