Delay Time
Hello,
I'm wondering if those Delay Time are normal?
If not, what should I do?
I'm wondering if those Delay Time are normal?
If not, what should I do?


FROM nvidia/cuda:12.4.1-cudnn-runtime-ubuntu20.04
RUN apt-get update && apt-get install -y \
python3-pip \
wget \
&& rm -rf /var/lib/apt/lists/*
WORKDIR /app
RUN pip install runpod==1.6.2 torch torchvision torchaudio sentence-transformers==2.7.0
# caching the model
RUN python3 -c "from sentence_transformers import SentenceTransformer; SentenceTransformer('clip-ViT-B-32', device='cpu')"
COPY builder/requirements.txt .
RUN pip install --no-cache-dir -r requirements.txt
COPY src/ .
CMD ["python3", "handler.py"]import numpy as np
# Can I load the model here??
def handler(job):
return True
runpod.serverless.start({"handler": handler})# Base image
FROM runpod/base:0.4.2-cuda11.8.0
ENV HF_HUB_ENABLE_HF_TRANSFER=0
# Install Python dependencies (Worker Template)
COPY builder/requirements.txt /requirements.txt
RUN python3.11 -m pip install --upgrade pip && \
python3.11 -m pip install --upgrade -r /requirements.txt --no-cache-dir && \
rm /requirements.txt
# Cache Models
COPY builder/cache_models.py /cache_models.py
RUN python3.11 /cache_models.py && \
rm /cache_models.py
# Add src files (Worker Template)
ADD src .
CMD python3.11 -u /rp_handler.py# Cache Models
COPY builder/cache_models.py /cache_models.py
RUN python3.11 /cache_models.py && \
rm /cache_models.py