Nir Rashti
unread,Jun 10, 2024, 9:10:29 AM6/10/24Sign in to reply to author
Sign in to forward
You do not have permission to delete messages in this group
Either email addresses are anonymous for this group or you need the view member email addresses permission to view the original message
to users, Odai Odeh, Omri Sabach
Hey,
I was using the basic Workshop VM (created by using Red Hat Demo Platform | InstructLab RHEL VM) in order to get started with InstructLab.
First of all just as a starting point I serve the basic model and chat with it (all works well btw).
Second, I tried 2 things, first was to add Q&A file that covers what is instructLab (same as was demonstrated in the instruction lab section), I ran the ilab diff command of course and saw that the new file is there as expected.
Than, ran: 'ilab generate' and right after 'ilab train --device=cuda', it was clear via the train logs file (generate by the ilab train command) that the model actually being improved, seeing in the logs the previous answer before the train and the new answer which indeed based on the generate files as expected.
The problem was that when the ilab train command finished and a new model was established, I ran the ilab serve command with the newly create model, than I was chatting with this model (I make sure few times that the chat session in indeed with the new model) but when asking the new model what is InstructLab I start to get very weird answers (not what I accept to get, as appear in the logs files generate by ilab train e.g).
Same thing happened when I try to teach ilab with some Oscar questions (as was shown in the github repo example), when I trained the model with the Oscar Q&A, the logs of the train command looks good. But after serving the newly created model and asking him basic questions like When did the 2024 Oscar took place or something similar it start to hallucinate completely (the answers appear to be worse, compare to the answer I got for the same question with the basic model, not the trained one).
Can you pinpoint what can be wrong? Should I expect those basic example to work well with instructlab newly created models (the trained models)?
Thanks you very much, Nir.