From sagemaker-ai
Generates Jupyter notebook deploying LoRA fine-tuned Nova/OSS models from SageMaker Serverless Model Customization to SageMaker endpoints or Bedrock.
npx claudepluginhub awslabs/agent-plugins --plugin sagemaker-aiThis skill uses the workspace's default tool permissions.
Identifies the correct deployment pathway based on model characteristics and generates deployment code.
Generates design tokens/docs from CSS/Tailwind/styled-components codebases, audits visual consistency across 10 dimensions, detects AI slop in UI.
Records polished WebM UI demo videos of web apps using Playwright with cursor overlay, natural pacing, and three-phase scripting. Activates for demo, walkthrough, screen recording, or tutorial requests.
Delivers idiomatic Kotlin patterns for null safety, immutability, sealed classes, coroutines, Flows, extensions, DSL builders, and Gradle DSL. Use when writing, reviewing, refactoring, or designing Kotlin code.
Identifies the correct deployment pathway based on model characteristics and generates deployment code.
This skill supports deploying Nova and OSS models that were fine-tuned through SageMaker Serverless Model Customization only.
Not supported:
You need the training job name or ARN. Check the conversation history first — the user may have already mentioned it, or it may be available from earlier steps in the workflow (e.g., fine-tuning). If not, ask the user.
Once you have the training job name or ARN, use the AWS MCP tool to look it up:
describe-training-job and extract:
ModelArtifacts.S3ModelArtifacts or OutputDataConfig.S3OutputPath)RoleArn)list-tags on the training job ARN and extract:
sagemaker-studio:jumpstart-model-id tagUnsupported models: This skill only supports OSS and Nova models that were LoRA fine-tuned through SageMaker Serverless Model Customization. If the model doesn't match, tell the user this skill can't help and suggest the finetuning skill.
Use the following table:
| Model Type | Eligible Targets |
|---|---|
| OSS | SageMaker, Bedrock |
| Nova | SageMaker, Bedrock |
If only one target is eligible, confirm it with the user. Use details from Step 5.
If multiple targets are eligible, help the user decide. Use details from Step 5.
If no targets are eligible, tell the user and explain why.
Present the eligible options to the user. Present these details to help them decide between SageMaker and Bedrock, if both are available options:
SageMaker Endpoint:
Bedrock:
Do NOT make a recommendation. Let the user choose.
Do NOT mention technical details like merged/unmerged weights, reference files, or APIs, unless the user asks.
⏸ Wait for user to select a deployment option.
Before proceeding to deployment, display the model's license or service terms to the user.
references/model-licenses.md and look up the model by its model ID (determined in Step 1).⏸ Wait for the user to confirm before proceeding.
Read the reference file for the selected pathway and follow its instructions.
| Model Type | Deployment Target | Reference |
|---|---|---|
| OSS | SageMaker | references/deploy-oss-sagemaker.md |
| OSS | Bedrock | references/deploy-oss-bedrock.md |
| Nova | SageMaker | references/deploy-nova-sagemaker.md |
| Nova | Bedrock | references/deploy-nova-bedrock.md |
After deployment completes, provide the user with a summary. Cover these topics, using details from the pathway reference doc you followed in Step 5:
If deployment fails unexpectedly, the model may have been full fine-tuned (FFT) rather than LoRA. To check, download the training job's hydra config from its S3 output path at .hydra/config.yaml:
peft_config populated (r, alpha, dropout, etc.) → LoRA (supported)peft_config: null → FFT (not supported by this skill)