Skip to content

Helm chart for Ollama #774

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 17 commits into from
Feb 14, 2025
Merged

Conversation

jonminkin97
Copy link
Contributor

Description

Add Helm chart to deploy Ollama as a model server.

Issues

n/a

Type of change

List the type of change like below. Please delete options that are not relevant.

  • New feature (non-breaking change which adds new functionality)

Dependencies

List the newly introduced 3rd party dependency if exists.

Tests

  • Manual deployment of Helm chart, verifying installation and inference
  • Helm test pod to automatically test the Helm release using helm test [RELEASE]

Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Copy link
Collaborator

@lianhao lianhao left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks very much for contributing this. Besides the embedded comment, there is something missing in the helm template(e.g. HPA, service monitor). I would suggest you check https://github.com/opea-project/GenAIInfra/tree/main/helm-charts/common/lvm-serve to add the missing gap. Thanks!

Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
@jonminkin97
Copy link
Contributor Author

@lianhao Thank you for the feedback. I've made the requested changes, and they are now pending your approval. Let me know if there are any other enhancements you would like to see.

Copy link
Collaborator

@lianhao lianhao left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

2 more things beside the embedding comments:

  • Add a cpu-values.yaml file which is used to trigger the CI, just like what lvm-serve does. These kind of xxx-values contains additional configuration for CI to pass(could be empty if there no additional requirements at all).

  • Could you name this chart ollama instead of ollama-service? One of our 1.3 task is to make sure the helm chart name is consistent with the corresponding names in GenAIComps.

Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Signed-off-by: Jonathan Minkin <minkinj@amazon.com>
Copy link
Collaborator

@lianhao lianhao left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for your contribution @jonminkin97

@lianhao lianhao requested a review from poussa February 12, 2025 01:38
@lianhao lianhao merged commit 7d66afb into opea-project:main Feb 14, 2025
10 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants