How to run inference after SFT tuning ? #2459
-
Hi, this is how I load the model: I am using Qwen1.5 for the finetune. thanks |
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 3 replies
-
use CLI demo to infer trained models |
Beta Was this translation helpful? Give feedback.
-
I tried increasing the n_epochs, from 3 to eventually 800, finally getting some meaningful response. However, I worry that this will have model overfitting. The loss is pretty low now : appreciate your work and support very much, @hiyouga ! |
Beta Was this translation helpful? Give feedback.
-
Hello, Is there any way to pass the entire dataset to the inference code of LLama-Factory and get the response for that. Any python code would help? The command llamafactory-cli chat examples/inference/llama3_lora_sft.yaml prompts the user to enter an input. I am looking for some code where I pass my test data to the finetuned model as entries of a json file and model gives me a response that I could save for my analysis. My test dataset is has 11K entries in .json file. |
Beta Was this translation helpful? Give feedback.
use CLI demo to infer trained models