-
Notifications
You must be signed in to change notification settings - Fork 43
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
After doing all the process to generate the project. There are no Checkpoints. #23
Comments
This might be due to OneDrive interfering with file locks and stuff (I see it in the path). If I were you, I might try moving everything to some folder that isn't tied to OneDrive. |
@venturaEffect, can you make sure the dataset file is getting copied to the dataset folder? Also did you increased the max_steps to match your needs on the dataset? |
@vriveras Oh, I have include it manually in the dataset. Does it make it automatically? No, haven't increased the max_steps. How do I know which are the required? |
Well it is on my Desktop. Should I move it maybe to User folder? |
@venturaEffect could it be model dependent? Using
more output
|
As for the error on line 40 in
Windows side: |
The max steps are low of the default dataset which means you are not triggering the eval steps where the checkpoint is created. Once you are done the final adapter should still be available under model Thanks for pointing out the bug with the inference on line 40, we have now fixed the templates. |
The project is not saved inside WSL instance but on To continue finetuning copy the project from |
After the fine-tuning is done just run one of the inferencing files console_chat.py or gradio_chat.py they will automatically load the model with the adapter. You can take a look here https://github.com/microsoft/windows-ai-studio/blob/main/walkthrough-simple-dataset.md#inferencing-with-the-fine-tuned-model |
Ok, thanks, will try. In any case this seems to be a bug. |
I appreciate your response. Just can't understand why this new workaround, and opening the bar with a default error PATH. Non sense. |
This has nothing to do with the checkpoint. Projects are always created in Windows unless you placed them in WSL when creating it. WSL mounts your drives automatically and we use that as the location. C:\users\ becomes /mnt/c/users automatically when launching. You can manually open the project after they are created by just opening the folder on a WSL remote. Go to the folder in WSL and type 'code .' from the folder and it will open it. You can follow the instructions after that to do inference. |
Following your steps. The Path is good (if it is how you say.... ) But no, Also have done all the other things you have said: changed your code error (from true to True), max_steps also (even if not necessary), tried to run it even if there are no checkpoints because there is the adapter and ti will work (??) Nothing, errors, wrong paths, icons showing forbidden when hovering in the path input (?). Tried like over 20 times. One time it stayed like 24 hours (on Windows 11 NVIDIA 4090). The promise where great and should be legit coming from Windows (??). But start to miss Langchain. |
First of all congrats, this looks promising.
But, after trying 5 times to finetune with the Mistral 7b model, I see there are no Checkpoints. I've set batch_size to 4 fl16, gradient accumulation to 8,... My JSON file is set on key:value pairs like "instruction" and "output", have changed that where it is "phrase" and "tone". Everything went through just with one Warning but nothing serious. Waited to it finnishes the finetunning. I see models folder with qlora and checkpoints. In QLora I have
run_history_gpu-cpu.txt
:There is also another subfolder called again "qlora". Inside are on .json file "gpu-cpu_model.json" and another folder called "gpu-cpu_model". Inside this folder is another folder called adapter with a .json file "adapter_model.json" and "adapter_model.bin"
But in Checkpoints folder there is nothing. As said, tried several times. Different .json datasets. No clue why it doesn't work.
I'm running it on Windows 11 with a NVIDIA RTX 4090. And it shows it is running it on my GPU.
So, what is going on???
When I run
python gradio_chat.py
I get:(mistral-7b-env) zasear@zaesarius:/mnt/c/Users/zaesa/OneDrive/Escritorio/AI/Lawyer Mistral Agent/inference$ python gradio_chat.py Number of GPUs available: 1 Running on device: cuda CPU threads: 16 Loading checkpoint shards: 100%|████████████████████████████████████████████| 2/2 [02:44<00:00, 82.49s/it] Traceback (most recent call last): File "/mnt/c/Users/zaesa/OneDrive/Escritorio/AI/Lawyer Mistral Agent/inference/gradio_chat.py", line 40, in <module> usingAdapter = true NameError: name 'true' is not defined
Appreciate any help!
The text was updated successfully, but these errors were encountered: