Skip to main content
Filter by
Sorted by
Tagged with
0 votes
0 answers
64 views

I'm trying to deploy a custom PyTorch model to a SageMaker Multi-Model Endpoint (MME). My model is saved as a state_dict using torch.save(), so it requires a custom inference.py script to load the ...
Miguel Oliveira's user avatar
0 votes
0 answers
109 views

I'm attempting to serve a pretrained Stable Diffusion model with LoRA weights applied using TorchServe. However, the LoRA weights don't seem to load properly, and I'm not sure why. Could anyone help ...
박연수's user avatar
0 votes
1 answer
225 views

I have an image classifier model that I plan to deploy via torch serve. My question is, what is the ideal way to load as well write images from / to s3 buckets instead of from local filesystem for ...
mtoto's user avatar
  • 24.3k
2 votes
1 answer
266 views

I am using a custom inference script for a Huggingface embedding model in an AWS SageMaker TorchServe container. My script accepts JSON input in the following format: { "inputs": ["...
Hassan Rahamathullah's user avatar
1 vote
0 answers
320 views

I am running torchserve container with pipeline and 2 models. if I am sending python async requests to the pipeline, for more than 14 requests torchserve got stuck for long time and than fail. But if ...
Benny Koren's user avatar
-1 votes
1 answer
199 views

My goal is to deploy a model I trained with Fastai into Torchserve. I was following this tutorial but got stuck on the part where he created the model class for pytorch. He mentions that to run our ...
Antonio Tapia's user avatar
0 votes
1 answer
131 views

I am creating a custom pytorch serve handler for my image enhancement GAN model. The server successfully loads the model but gives no output when I make a request. It neither shows an error in logs. ...
Muhammad Danial Khan's user avatar
0 votes
1 answer
285 views

I think my torchserve loaded config.properties correctly because the number of worker is 2 as I set. But the batch_size is 1 instead of 20. Anyone has an idea what might go wrong ? Thanks ! I have ...
Lu9999's user avatar
  • 11
0 votes
1 answer
120 views

I'm running TorchServe in WSL2. There are three issues with the metrics: Even if metrics_config parameter in ts.config points to non-existing file everything works without any problems. It looks like ...
feeeper's user avatar
  • 3,047
2 votes
1 answer
702 views

I'm currently developing a Rails application that interacts with a TorchServe instance for machine learning inference. The TorchServe server is hosted on-premises and equipped with 4 GPUs. We're ...
Aj Gu's user avatar
  • 1,439
0 votes
1 answer
313 views

I'm running a Yolov8 object detector with TorchServe. In my custom_handler, I'm trying to grab the detection output JSON and also get the image of the annotated bounding boxes. When I run the code ...
Eddie E Massey III's user avatar
0 votes
0 answers
423 views

We deployed NER Model with n1-standard-8 machine without GPU with below config properties. when we kept batch size as 1, it is taking more time to process the simultaneous requests. when we try to ...
RajeshM's user avatar
  • 79
0 votes
1 answer
627 views

I have this docker image: # syntax = docker/dockerfile:1.2 FROM continuumio/miniconda3 # install os dependencies RUN mkdir -p /usr/share/man/man1 RUN apt-get update && \ DEBIAN_FRONTEND=...
AynonT's user avatar
  • 323
1 vote
1 answer
951 views

I'm attempting to serve a pretrained huggingface model with torchserve and i've managed to save the model as a torchscript file (.pt). However, I do not know what the handler would look like for such ...
maxwellspi's user avatar
1 vote
1 answer
69 views

I've read TorchServe's default handlers' sources and found that the BaseHandler is inherited from abc.ABC and doesn't have any abstract method. The VisionHandler is the same. What could be the reason ...
feeeper's user avatar
  • 3,047
1 vote
0 answers
913 views

Every once in a while a TorchServe worker dies with the following message io.netty.handler.codec.CorruptedFrameException: Message size exceed limit: 16. When I rerun the request in question, it ...
sereneSentry's user avatar
1 vote
0 answers
314 views

I have a torchserve service running on kubernetes and I am already able to track metrics with it on port 8082. My problem is that from the kubernetes pod I can see it logs hardware metrics like: [INFO ...
Prosciutt0's user avatar
1 vote
1 answer
906 views

Currently, I am working with a PyTorch model locally using the following code: from transformers import pipeline classify_model = pipeline("zero-shot-classification", model='models/...
Роман Сергеевич's user avatar
0 votes
1 answer
855 views

I am trying to create a custom handler in torchserve and want to also use torchserve's batch capability for parallelism for optimum use of resources. I am not able to find out how to write custom ...
Paras Bansal's user avatar
2 votes
2 answers
418 views

Can TorchServe run separate Python environments for each model in TorchServe? I have four models in production right now, and they all use transformers==3.1.0 package. A new model about to be put in ...
sereneSentry's user avatar
1 vote
1 answer
744 views

I am trying to send a singular request to a Torchserve server and retrieve a stream of responses. The processing of the request takes some time and I would like to receive intermeddiate updates over ...
P_Andre's user avatar
  • 781
1 vote
1 answer
633 views

I have deployed a trained PyTorch model to a Google Vertex AI Prediction endpoint. The endpoint is working fine, giving me predictions, but when I examine its logs in Logs Explorer, I see: INFO 2023-...
urig's user avatar
  • 16.9k
0 votes
1 answer
956 views

I ran the following commands torch-model-archiver --model-name "bert" --version 1.0 --serialized-file ./bert_model/pytorch_model.bin --extra-files "./bert_model/config.json,./bert_model/...
Varun Sappa's user avatar
0 votes
1 answer
1k views

I'v recently worked with TensorFlow Extended(TFX) platform. Since my development background is on Pytorch stack, I'm looking for well-compatible alternatives of TFX for Pytorch. While searching for ...
Mohsen Mahmoodzadeh's user avatar
1 vote
0 answers
335 views

I have been tasked with updating kserve from 0.7 to 0.9. Our company mar files run fine on 0.7 but when I update to kserve 0.9 the pods are brought up without issue. However, when I when a request is ...
Waqas Shah's user avatar
1 vote
1 answer
1k views

I have deployed Bert Hugging Face models via TorchServe on the AWS EC2 GPU instance. There are enough resources provisioned, usage of everything is consistently below 50%. TorchServe performs ...
sereneSentry's user avatar
8 votes
3 answers
7k views

NVIDIA Triton vs TorchServe for SageMaker inference? When to recommend each? Both are modern, production grade inference servers. TorchServe is the DLC default inference server for PyTorch models. ...
juvchan's user avatar
  • 6,263
1 vote
1 answer
760 views

I'm trying to access an API with aiohttp but something is causing this code to block each iteration. def main(): async with aiohttp.ClientSession() as session: for i, (image, target) in ...
Terv's user avatar
  • 11
2 votes
1 answer
4k views

I am trying to create a custom handler on Torchserve. The custom handler has been modified as follows # custom handler file # model_handler.py """ ModelHandler defines a custom model ...
Takayama-Shin's user avatar
0 votes
1 answer
404 views

Please bear with me as I am new to docker and have never used torchserve before, so any feedback will help. I am trying to create a .mar file in an existing docker container from a model.pt file, ...
Cashmoosef's user avatar
2 votes
2 answers
2k views

I have written a custom handler for an DL model using torch-serve and am trying to understand how to add manual log messages to the handler. I know that I can simply print any messages and it will ...
cotrane's user avatar
  • 199
-1 votes
1 answer
1k views

I deployed a pytorch model using sagemaker and can successfully query it on a CPU. Deploying it on a GPU leads to a InternalServerError client-side though. Checking the CloudWatch Logs shows that the ...
cotrane's user avatar
  • 199
1 vote
0 answers
760 views

I am trying to build a container running torchserve with the pretrained fast-rcnn model for object detection in a all-in-one Dockerfile, based on this example: https://github.com/pytorch/serve/tree/...
VEHC's user avatar
  • 11
1 vote
1 answer
1k views

I have a model that is served using TorchServe. I'm communicating with the TorchServe server using gRPC. The final postprocess method of the custom handler defined returns a list which is converted ...
Mohit Motwani's user avatar
0 votes
1 answer
524 views

I try to create a torchserve on google colab but it took forever to load and it seem that i can't even connect to the serve. Is this possible to create a torchserve on colab? Here is what it show when ...
Dzung's user avatar
  • 1
0 votes
1 answer
113 views

I'm deploying torchserve on heroku free dyno. Deploy works fine but the app isn't running properly. LOG1: 2022-03-17T02:11:10.352655+00:00 heroku[web.1]: Starting process with command `torchserve --...
Dionisis G's user avatar
0 votes
1 answer
2k views

I have a running torchserve service. According to the docs, I can register a new model at port 8081 with the ManagementAPI. When running curl -X OPTIONS http://localhost:8081, the output also states ...
c0mr4t's user avatar
  • 361
0 votes
1 answer
1k views

I have a torchscript model (.pt) that I can successfully load and serve with torch serve on my local machine. On the other side, when trying to deploy it in the oficial torchserve docker it will ...
eljiwo's user avatar
  • 856
0 votes
1 answer
1k views

I'm trying to deploy a TorchServe instance on Google Vertex AI platform but as per their documentation (https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#...
Timon. Z's user avatar
  • 681