-
Notifications
You must be signed in to change notification settings - Fork 13
ci: Add workflow to test python script examples #45
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
33da7a4
to
2f72a0d
Compare
Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
2f72a0d
to
e0e0145
Compare
Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think it is failing with the 8B model. Maybe that can be fixed somehow with caching.
Otherwise, the 2B should make CI run much better, but then I think our "thinking" test will fail until we get some parser fixes for 2B.
8000 | - name: Install Ollama and start server | |
shell: bash | ||
run: | | ||
curl -fsSL https://ollama.com/install.sh | sudo -E sh |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Also, I was looking at examples of this and was always wondering if we need to try to kill the ollama server at the end (capture the pid or pkill). But I wasn't finding good examples so perhaps CI shuts everything down okay? I mean it should shut down all subprocesses and not hang CI resources, right?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hmmm .. not sure. I'll investigate this further.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is it ok to do this post PR merge?
- model_chat_with_thinking.py - watsonx_litellm.py Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
1 Python version is good enough for now. Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
This more efficient instead of separate runners using matrix in workflow. Co-authored-by: Mark Sturdevant <mark.sturdevant@ibm.com> Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
6ad67a3
to
a396a0a
Compare
Its ok now. It might have been at start when adding intermediate fixes. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks!
No description provided.