83 lines
3.6 KiB
Go
83 lines
3.6 KiB
Go
package openai
|
|
|
|
import (
|
|
"errors"
|
|
"strings"
|
|
)
|
|
|
|
var (
|
|
// Deprecated: use ErrReasoningModelMaxTokensDeprecated instead.
|
|
ErrO1MaxTokensDeprecated = errors.New("this model is not supported MaxTokens, please use MaxCompletionTokens") //nolint:lll
|
|
ErrCompletionUnsupportedModel = errors.New("this model is not supported with this method, please use CreateChatCompletion client method instead") //nolint:lll
|
|
ErrCompletionStreamNotSupported = errors.New("streaming is not supported with this method, please use CreateCompletionStream") //nolint:lll
|
|
ErrCompletionRequestPromptTypeNotSupported = errors.New("the type of CompletionRequest.Prompt only supports string and []string") //nolint:lll
|
|
)
|
|
|
|
var (
|
|
ErrO1BetaLimitationsMessageTypes = errors.New("this model has beta-limitations, user and assistant messages only, system messages are not supported") //nolint:lll
|
|
ErrO1BetaLimitationsTools = errors.New("this model has beta-limitations, tools, function calling, and response format parameters are not supported") //nolint:lll
|
|
// Deprecated: use ErrReasoningModelLimitations* instead.
|
|
ErrO1BetaLimitationsLogprobs = errors.New("this model has beta-limitations, logprobs not supported") //nolint:lll
|
|
ErrO1BetaLimitationsOther = errors.New("this model has beta-limitations, temperature, top_p and n are fixed at 1, while presence_penalty and frequency_penalty are fixed at 0") //nolint:lll
|
|
)
|
|
|
|
var (
|
|
//nolint:lll
|
|
ErrReasoningModelMaxTokensDeprecated = errors.New("this model is not supported MaxTokens, please use MaxCompletionTokens")
|
|
ErrReasoningModelLimitationsLogprobs = errors.New("this model has beta-limitations, logprobs not supported") //nolint:lll
|
|
ErrReasoningModelLimitationsOther = errors.New("this model has beta-limitations, temperature, top_p and n are fixed at 1, while presence_penalty and frequency_penalty are fixed at 0") //nolint:lll
|
|
)
|
|
|
|
// ReasoningValidator handles validation for reasoning model requests.
|
|
type ReasoningValidator struct{}
|
|
|
|
// NewReasoningValidator creates a new validator for reasoning models.
|
|
func NewReasoningValidator() *ReasoningValidator {
|
|
return &ReasoningValidator{}
|
|
}
|
|
|
|
// Validate performs all validation checks for reasoning models.
|
|
func (v *ReasoningValidator) Validate(request ChatCompletionRequest) error {
|
|
o1Series := strings.HasPrefix(request.Model, "o1")
|
|
o3Series := strings.HasPrefix(request.Model, "o3")
|
|
o4Series := strings.HasPrefix(request.Model, "o4")
|
|
gpt5Series := strings.HasPrefix(request.Model, "gpt-5")
|
|
|
|
if !o1Series && !o3Series && !o4Series && !gpt5Series {
|
|
return nil
|
|
}
|
|
|
|
if err := v.validateReasoningModelParams(request); err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// validateReasoningModelParams checks reasoning model parameters.
|
|
func (v *ReasoningValidator) validateReasoningModelParams(request ChatCompletionRequest) error {
|
|
if request.MaxTokens > 0 {
|
|
return ErrReasoningModelMaxTokensDeprecated
|
|
}
|
|
if request.LogProbs {
|
|
return ErrReasoningModelLimitationsLogprobs
|
|
}
|
|
if request.Temperature > 0 && request.Temperature != 1 {
|
|
return ErrReasoningModelLimitationsOther
|
|
}
|
|
if request.TopP > 0 && request.TopP != 1 {
|
|
return ErrReasoningModelLimitationsOther
|
|
}
|
|
if request.N > 0 && request.N != 1 {
|
|
return ErrReasoningModelLimitationsOther
|
|
}
|
|
if request.PresencePenalty > 0 {
|
|
return ErrReasoningModelLimitationsOther
|
|
}
|
|
if request.FrequencyPenalty > 0 {
|
|
return ErrReasoningModelLimitationsOther
|
|
}
|
|
|
|
return nil
|
|
}
|