Skip to content

Add modelcar#126

Open
fyuan1316 wants to merge 2 commits intomasterfrom
add-modelcar
Open

Add modelcar#126
fyuan1316 wants to merge 2 commits intomasterfrom
add-modelcar

Conversation

@fyuan1316
Copy link
Contributor

@fyuan1316 fyuan1316 commented Mar 6, 2026

Summary by CodeRabbit

  • Documentation
    • Added an end-to-end guide for using KServe Modelcar with OCI-based model storage: overview and benefits, prerequisites, two packaging approaches with examples, steps to build and push OCI images, deploying an InferenceService with recommended runtime and storage settings, verification commands, best practices, and troubleshooting guidance.

@coderabbitai
Copy link

coderabbitai bot commented Mar 6, 2026

No actionable comments were generated in the recent review. 🎉

ℹ️ Recent review info
⚙️ Run configuration

Configuration used: Organization UI

Review profile: CHILL

Plan: Pro

Run ID: 75901080-3f72-4222-960f-2861a24793b0

📥 Commits

Reviewing files that changed from the base of the PR and between c2a3e34 and ec76a99.

📒 Files selected for processing (1)
  • docs/en/model_inference/inference_service/how_to/using_modelcar.mdx
🚧 Files skipped from review as they are similar to previous changes (1)
  • docs/en/model_inference/inference_service/how_to/using_modelcar.mdx

Walkthrough

Adds a new MDX documentation file that documents how to use KServe Modelcar (OCI container-based model storage) with the Alauda AI platform, including prerequisites, two packaging approaches, building and pushing OCI images, an InferenceService YAML example, verification, best practices, and troubleshooting.

Changes

Cohort / File(s) Summary
Documentation - Modelcar Guide
docs/en/model_inference/inference_service/how_to/using_modelcar.mdx
Added a new end-to-end how-to for using KServe Modelcar on Alauda: overview, prerequisites, enabling in AmlCluster, packaging options (Busybox & UBI Micro), Dockerfile/Containerfile samples, build/push commands, InferenceService YAML with annotations/resources/storageUri, apply/verify steps, best practices, and troubleshooting.

Estimated code review effort

🎯 2 (Simple) | ⏱️ ~12 minutes

Poem

🐇 I hopped through lines of docs today,
Packaged models in OCI array,
Containers tidy, yaml bright,
Inference ready, day and night,
Hooray—Modelcar finds its way!

🚥 Pre-merge checks | ✅ 2 | ❌ 1

❌ Failed checks (1 inconclusive)

Check name Status Explanation Resolution
Title check ❓ Inconclusive The title 'Add modelcar' is vague and lacks specificity about what is being added. While it relates to the changeset, it doesn't convey meaningful information about the documentation being added or its purpose. Consider a more descriptive title such as 'Add documentation for KServe Modelcar usage in Alauda AI platform' or 'Add guide for using KServe Modelcar with InferenceService'.
✅ Passed checks (2 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing Touches
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Post copyable unit tests in a comment
  • Commit unit tests in branch add-modelcar

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

Copy link

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 3

🧹 Nitpick comments (1)
docs/en/model_inference/inference_service/how_to/using_modelcar.mdx (1)

35-44: Clarify the YAML configuration path.

The instruction says to add configuration "under kserve.values", but the YAML snippet starts from spec:. This may confuse users about exactly what to add and where. Consider either:

  1. Showing only the portion to add under kserve.values, or
  2. Clarifying that this is the complete path from the root of the AmlCluster values
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/en/model_inference/inference_service/how_to/using_modelcar.mdx` around
lines 35 - 44, The YAML path is ambiguous; update the text to clarify where to
place the snippet by either (a) replacing the current snippet with only the
fragment to add under kserve.values (showing: kserve: storage: enableModelcar:
true) and adjusting the surrounding sentence to say "add the following under
kserve.values", or (b) explicitly state that the provided snippet is the full
AmlCluster values path from root and keep the existing
spec/components/kserve/values/kserve/... tree; reference kserve.values, spec,
components, kserve, storage, and enableModelcar so readers know whether to paste
the short fragment under kserve.values or the full spec tree at the AmlCluster
root.
🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In `@docs/en/model_inference/inference_service/how_to/using_modelcar.mdx`:
- Line 226: The sentence in the doc content currently reads "Using KServe
Modelcar (OCI container-based model storage) provides a efficient way to deploy
models..." — change "a efficient" to "an efficient" so it reads "provides an
efficient way to deploy models"; update the phrase in the same paragraph where
"Using KServe Modelcar (OCI container-based model storage)..." appears to fix
the grammar.
- Around line 54-67: The Dockerfile's COPY source path (COPY
Qwen2.5-0.5B-Instruct/ /models/) is inconsistent with the build instructions
that expect a generic models/ directory; update the Dockerfile to use the
generic source directory used by the build steps (e.g., change the COPY line to
copy from models/ into /models/) and adjust the accompanying comment to reflect
that it copies the local models/ folder contents so users following the build
instructions won't encounter a missing-source error.
- Line 167: Update the callout link target for "Extend Inference Runtimes":
locate the sentence containing `aml.cpaas.io/runtime-type: vllm` and change the
link target currently pointing to `./external_access_inference_service.mdx` to
`./custom_inference_runtime.mdx` so the "Extend Inference Runtimes" link
correctly points to the runtime extension documentation.

---

Nitpick comments:
In `@docs/en/model_inference/inference_service/how_to/using_modelcar.mdx`:
- Around line 35-44: The YAML path is ambiguous; update the text to clarify
where to place the snippet by either (a) replacing the current snippet with only
the fragment to add under kserve.values (showing: kserve: storage:
enableModelcar: true) and adjusting the surrounding sentence to say "add the
following under kserve.values", or (b) explicitly state that the provided
snippet is the full AmlCluster values path from root and keep the existing
spec/components/kserve/values/kserve/... tree; reference kserve.values, spec,
components, kserve, storage, and enableModelcar so readers know whether to paste
the short fragment under kserve.values or the full spec tree at the AmlCluster
root.

ℹ️ Review info
⚙️ Run configuration

Configuration used: Organization UI

Review profile: CHILL

Plan: Pro

Run ID: 513cbe3c-8655-448e-a7c2-0d2d9bc98d47

📥 Commits

Reviewing files that changed from the base of the PR and between b451348 and c2a3e34.

📒 Files selected for processing (1)
  • docs/en/model_inference/inference_service/how_to/using_modelcar.mdx


## Conclusion

Using KServe Modelcar (OCI container-based model storage) provides a efficient way to deploy models in Alauda AI platform. By following the steps outlined in this guide, you can package your models as OCI images and deploy them with faster startup times and improved resource utilization. No newline at end of file
Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟡 Minor

Minor grammar fix needed.

"a efficient" should be "an efficient".

📝 Suggested fix
-Using KServe Modelcar (OCI container-based model storage) provides a efficient way to deploy models in Alauda AI platform. By following the steps outlined in this guide, you can package your models as OCI images and deploy them with faster startup times and improved resource utilization.
+Using KServe Modelcar (OCI container-based model storage) provides an efficient way to deploy models in Alauda AI platform. By following the steps outlined in this guide, you can package your models as OCI images and deploy them with faster startup times and improved resource utilization.
📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
Using KServe Modelcar (OCI container-based model storage) provides a efficient way to deploy models in Alauda AI platform. By following the steps outlined in this guide, you can package your models as OCI images and deploy them with faster startup times and improved resource utilization.
Using KServe Modelcar (OCI container-based model storage) provides an efficient way to deploy models in Alauda AI platform. By following the steps outlined in this guide, you can package your models as OCI images and deploy them with faster startup times and improved resource utilization.
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/en/model_inference/inference_service/how_to/using_modelcar.mdx` at line
226, The sentence in the doc content currently reads "Using KServe Modelcar (OCI
container-based model storage) provides a efficient way to deploy models..." —
change "a efficient" to "an efficient" so it reads "provides an efficient way to
deploy models"; update the phrase in the same paragraph where "Using KServe
Modelcar (OCI container-based model storage)..." appears to fix the grammar.

@cloudflare-workers-and-pages
Copy link

cloudflare-workers-and-pages bot commented Mar 6, 2026

Deploying alauda-ai with  Cloudflare Pages  Cloudflare Pages

Latest commit: ec76a99
Status: ✅  Deploy successful!
Preview URL: https://f6dd1e45.alauda-ai.pages.dev
Branch Preview URL: https://add-modelcar.alauda-ai.pages.dev

View logs

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant