-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Consolidate perf testing tools #23
Comments
My llmperf performance testing was conducted in Istio service mesh, and I have made some adaptions to make the inference gateway to run in service mesh. However, when I was trying to reproduce my performance test base on the community version, some issues blocked my test.
So I have to resolve those issues fist or the performance testing with llmperf can not proceed. cc @danehans |
Hi @Kuromesi 👋 |
The model is not in the request body. Can you share the details of the manual request being sent? |
My request like: curl ${GATEWAY_IP}:${GATEWAY_PORT}/metrics -d '{"test":"test"}' Yes I did not set the the request model, I just want to illustrate that the error BTW, does this work as expected? Should we require the request to have request body and request model? Can we randomly return a pod if they are not set? |
Sorry my bad, I falsely set some configurations which cause llmperf not working as expected, now the llmperf works fine. But I think we should provide a way to random return a pod if target model is not specified? |
We made some enhancement to based on the istio community version, which supports to add a header like |
If the @ahg-g @kfswain thoughts on how this issue should be handled, e.g. update |
xref: kubernetes-sigs/gateway-api-inference-extension#332
Heya Perf folks! We have a need for specific perf testing in GIE. And this issue is just to centralize discussion in the inf-perf repo, so that we can all be on the same page and working towards the same goals. Thanks!!
cc: @liu-cong @Kuromesi
The text was updated successfully, but these errors were encountered: