Replies: 1 comment
-
|
Turns out the issue was a missing/wrong configuration for the used AzureOpenAI model. It just silently failed and therefore no turns were generated. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
To test a multi-agent system (with openai-agents-python), I wanted to try DeepEval for qualitative evaluation/testing.
Unfortunately, the ConversationSimulator doesn't seem to generate or propagate the generated turns for the test case as expected. Sometimes, like 20% of the time, it works though - perhaps a timing issue.
I tried to stay as close as possible to the documentation of DeepEval (Conv Simulator, CI/CD, E2E), but it seems partially outdated as many argument names changed meanwhile.
I'm using DeepEval 3.8.8
Error:
TypeError: 'turns' must not be emptyCode:
Function definition for model_callback:
async def model_callback(input: str, thread_id: str) -> Turn:Adding
turnsto the function definition didn't change anything either.Calling the test with
deepeval test run tests/agent-tests/test_knowledge.pyand yes, I also tried the non-pytest way withevaluate()- same result. Also disabling async didn't help.Am I doing something wrong or is there a bug? Thanks!
Beta Was this translation helpful? Give feedback.
All reactions