R
RunPod6mo ago
antoniog

Issue with Request Count Scale Type

Request Count is set to 15 and there are more than 15 requests but an additional worker is not being added. It's an A1111 worker (https://github.com/ashleykleynhans/runpod-worker-a1111) with runpod>=0.10.0. See screenshots attached.
No description
No description
6 Replies
nerdylive
nerdylive6mo ago
Your max workers check it, if the running workers is already 3 then it wont be added anymore max workers limits the current running workers
flash-singh
flash-singh6mo ago
you set request count to 15, in order for it to spawn 4th worker, you would need > 45 requests in queue & in progress math is (19+3)/15, that logic only needs 2 workers, you already have 3 running i would set request count to 5, 4th worker will spawn when > 15
ashleyk
ashleyk6mo ago
Thanks, this isn't very obvious, would be good to add it to the docs.
flash-singh
flash-singh6mo ago
its shared there in description
ashleyk
ashleyk6mo ago
Oh my bad, missed that, sorry
antoniog
antoniog6mo ago
got it, thanks
Want results from more Discord servers?
Add your server
More Posts
billing not adding upThe time listed in billing doesn't add up Hello guys. I see a bunch of charges listed as 1 minute bDo I need to keep Pod open after using it to setup serverless APIs for stable diffusion?Hi I'm following this tutorial on building serverless endpoints for running txt2img with ControlNet SSH key not workingHello, im trying to get SSH working. My pod is pre-configured. I added my key to the pod variables. how do you access the endpoint of a deployed llm on runpod webui and access it through Python?how do you access the endpoint of a deployed llm on runpod webui and access it through Python?Is runpod UI accurate when saying all workers are throttled?To be honest, I cannot tell if the image I see is correct? I have two endpoints both with max 3 workserverless: any way to figure out what gpu type a job ran on?trying to get data on speeds across gpu types for our jobs, and i'm wondering if the api exposes thiIs it possible to build an API for an automatic1111 extension to be used through Runpod serverless?I want to use the faceswaplab extension for automatic1111 as a serverless endpoint on Runpod. I manhosting mistral model in productionhi, I wish to host mistral model in runpod for production. what will happen to the app during scheduJobs suddenly queuing up: only 1 worker active, 9 jobs queued** Endpoint: vieo12phdoc8kh** Hi, are there any known issues at the moment with 4090s? Our processiIssues with building the new `worker-vllm` Docker ImageI've been using the previous version of `worker-vllm` with the `awq` model in production, and it rec