@@ -144,7 +144,7 @@ const (
144
144
// to exist at request time, the error is processed by the Inference Gateway
145
145
// and emitted on the appropriate InferenceModel object.
146
146
type TargetModel struct {
147
- // Name is the name of the adapter as expected by the ModelServer.
147
+ // Name is the name of the adapter or base model, as expected by the ModelServer.
148
148
//
149
149
// +kubebuilder:validation:MaxLength=253
150
150
// +kubebuilder:validation:Required
@@ -174,10 +174,31 @@ type TargetModel struct {
174
174
175
175
// InferenceModelStatus defines the observed state of InferenceModel
176
176
type InferenceModelStatus struct {
177
- // Conditions track the state of the InferencePool .
177
+ // Conditions track the state of the InferenceModel .
178
178
Conditions []metav1.Condition `json:"conditions,omitempty"`
179
179
}
180
180
181
+ // InferenceModelConditionType is a type of condition for the InferenceModel.
182
+ type InferenceModelConditionType string
183
+
184
+ // InferenceModelConditionReason is the reason for a given InferenceModelConditionType.
185
+ type InferenceModelConditionReason string
186
+
187
+ const (
188
+ // This condition indicates whether the model is ready for traffic or not, and why.
189
+ ModelConditionReady InferenceModelConditionType = "Ready"
190
+
191
+ // Desired state. Model is ready for serving with no conflicts or issues.
192
+ ModelReasonReady InferenceModelConditionReason = "Ready"
193
+
194
+ // This reason is used when a given ModelName already exists within the pool.
195
+ // Details about naming conflict resolution are on the ModelName field itself.
196
+ ModelReasonNameInUse InferenceModelConditionReason = "ModelNameInUse"
197
+
198
+ // This reason is the initial state, and indicates that the controller has not yet reconciled the InferenceModel.
199
+ PoolReasonPending InferenceModelConditionReason = "Pending"
200
+ )
201
+
181
202
func init () {
182
203
SchemeBuilder .Register (& InferenceModel {}, & InferenceModelList {})
183
204
}
0 commit comments