Can not generate proper Bangla
Even it is not instruction fine-tuned, it should have the ability to generate Bangla sentence correctly. It can not even generate correct Bangla sentences. But it can generate English sentences very well, even paragraph/story.
I'm interested in knowing which dataset is used to fine-tune this model.
Here is the test result of this model:
https://www.kaggle.com/code/muhammadmazed/check-banglallm-bangla-llama-13b-base-v0-1
Hey, thanks for the comment. This is a base model (pretrained). Base models are not good at specific task. You need to look into the instruct finetuned model. One of the best is possibly https://huggingface.co/BanglaLLM/BanglaLLama-3.1-8b-bangla-alpaca-orca-instruct-v0.0.1
It's not surprising that these models can generate well in English (because more than 90% of the language it's trained on is with English). The reason it's not good in Bangla is because it has not seen enough Bangla examples. Feel free to contribute by taking these models and tuning it. :)
So what exactly you did to this model? It's the same as the original llama base model. Your Bangla base model should generate proper sentences in Bangla. Base models cannot follow instructions, answer to questions, we all know that. But Base models must be able to generate complete sentences. In English, your model can do what it should but in Bangla it just the same as the original llama model. Seems like it is not even fine-tuned in Bangla.
AGAIN, IT'S BANGLA GENERATION CAPABILITY IS THE SAME AS THE ORIGINAL MODEL FROM META.