LogoTensorFusion 文档
LogoTensorFusion 文档
首页

快速开始

TensorFusion概览在Kubernetes安装在虚拟机/服务器安装(K3S)Helm本地安装在宿主机/虚拟机安装TensorFusion架构

应用操作

创建AI应用配置自动扩缩容迁移现有应用最佳实践

自定义AI基础设施

生产级部署指南QoS级别与计价云厂商集成(BYOC)管理许可证

维护与优化

组件更新配置告警GPU热迁移预加载模型优化GPU效率

故障排除

问题处理手册链路追踪/性能分析查询指标和日志

参考

工作负载配置
TensorFusionWorkloadTensorFusionConnectionWorkloadProfile

对比

与NVIDIA vGPU比较与MIG/MPS对比与趋动科技对比与 Run.AI 对比与HAMi的对比
终端用户参考Kubernetes资源定义

TensorFusionWorkload

TensorFusionWorkload is the Schema for the tensorfusionworkloads API.

TensorFusionWorkload is the Schema for the tensorfusionworkloads API.

Resource Information

FieldValue
API Versiontensor-fusion.ai/v1
KindTensorFusionWorkload
ScopeNamespaced

Spec

WorkloadProfileSpec defines the desired state of WorkloadProfile.

PropertyTypeDescription
autoScalingConfigobjectAutoScalingConfig configured here will override Pool's schedulingConfig This field can not be fully supported in annotation, if user want to enable auto-scaling in annotation, user can set tensor-fusion.ai/auto-resources
gpuCountinteger<int32>The number of GPUs to be used by the workload, default to 1
gpuModelstringGPUModel specifies the required GPU model (e.g., "A100", "H100")
isLocalGPUbooleanSchedule the workload to the same GPU server that runs vGPU worker for best performance, default to false
nodeAffinityobjectNodeAffinity specifies the node affinity requirements for the workload
poolNamestring
qosstringQos defines the quality of service level for the client. Allowed values: low, medium, high, critical
replicasinteger<int32>If replicas not set, it will be dynamic based on pending Pod If isLocalGPU set to true, replicas must be dynamic, and this field will be ignored
resourcesobject
sidecarWorkerbooleanWhen set to sidecar worker mode, its always Local GPU mode, and hard-isolated with shared memory default to false, indicates the workload's embedded worker is same process, soft-isolated
workerPodTemplateobjectWorkerPodTemplate is the template for the worker pod, only take effect in remote vGPU mode

Status

TensorFusionWorkloadStatus defines the observed state of TensorFusionWorkload.

PropertyTypeDescription
activeCronScalingRuleobjectThe currently active cron scaling rule
appliedRecommendedReplicasinteger<int32>The number of replicas currently applied based on the latest recommendation
conditionsarrayRepresents the latest available observations of the workload's current state.
phasestring(default: Pending) Allowed values: Pending, Running, Failed, Unknown
podTemplateHashstringHash of the pod template used to create worker pods
readyWorkersinteger<int32>readyWorkers is the number of vGPU workers ready
recommendationobjectThe most recently GPU resources recommended by the autoscaler
workerCount *integer<int32>workerCount is the number of vGPU workers

工作负载配置

本文档解释了如何使用Pod Annotation和WorkloadProfile自定义资源为AI应用分配vGPU资源。

TensorFusionConnection

TensorFusionConnection is the Schema for the tensorfusionconnections API.

目录

Resource Information
Spec
Status