Skip to content

QAEvaluator does not pass through is_reasoning_model #43842

@hirparap1

Description

@hirparap1

I was trying to use the QAEvaluator w/ a reasoning model w/ the is_reasoning_model: true parameter and ran into the following error:

openai.BadRequestError: Error code: 400 - {'error': {'message': "Unsupported parameter: 'max_tokens' is not supported with this model. Use 'max_completion_tokens' instead.", 'type': 'invalid_request_error', 'param': 'max_tokens', 'code': 'unsupported_parameter'}}

This doesn't happen for individual evaluators that the QAEvaluator makes use of. It seems to be due to the fact that the QAEvaluator does not pass through the is_reasoning_model for the evaluators under the hood > https://github.com/Azure/azure-sdk-for-python/blob/main/sdk/evaluation/azure-ai-evaluation/azure/ai/evaluation/_evaluators/_qa/_qa.py#L104-L111

Metadata

Metadata

Assignees

Labels

ClientThis issue points to a problem in the data-plane of the library.EvaluationIssues related to the client library for Azure AI EvaluationService AttentionWorkflow: This issue is responsible by Azure service team.bugThis issue requires a change to an existing behavior in the product in order to be resolved.customer-reportedIssues that are reported by GitHub users external to the Azure organization.needs-team-attentionWorkflow: This issue needs attention from Azure service team or SDK team

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions