Skip to content
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 5 additions & 0 deletions Makefile
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
start-gpu:
docker compose -f docker-compose.yml -f docker-compose-gpu.yml up

start-cpu:
docker compose -f docker-compose.yml -f docker-compose-cpu.yml up
8 changes: 8 additions & 0 deletions docker-compose-cpu.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,8 @@
version: '3.8'

services:
serving:
deploy:
resources:
reservations:
devices: []
11 changes: 11 additions & 0 deletions docker-compose-gpu.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
version: '3.8'

services:
serving:
deploy:
resources:
reservations:
devices:
- driver: nvidia
device_ids: ["${GPU_DEVICE}"]
capabilities: [ gpu ]
9 changes: 2 additions & 7 deletions docker-compose.yml
Original file line number Diff line number Diff line change
Expand Up @@ -9,13 +9,8 @@ services:
userns_mode: "host"
volumes:
- ./models:/models:z
deploy:
resources:
reservations:
devices:
- driver: nvidia
device_ids: ["${GPU_DEVICE}"]
capabilities: [ gpu ]
ports:
- "8501:8501"
shm_size: 8GB

develop:
Expand Down
25 changes: 25 additions & 0 deletions models/3dmolms/3dmolms_test/1/model.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
import json
import triton_python_backend_utils as pb_utils


class TritonPythonModel:
def __init__(self):
super().__init__()
self.output_dtype = []

def initialize(self, args):
model_config = json.loads(args["model_config"])
output0_config = pb_utils.get_output_config_by_name(model_config, "charge_norm")
self.output_dtype = pb_utils.triton_string_to_numpy(output0_config["data_type"])

def execute(self, requests):
responses = []
for request in requests:
raw = pb_utils.get_input_tensor_by_name(request, "charge_raw")
norm = raw.as_numpy() * 0.1
ce_tensor = pb_utils.Tensor("charge_norm", norm.astype(self.output_dtype))
responses.append(pb_utils.InferenceResponse(output_tensors=[ce_tensor]))
return responses

def finalize(self):
pass
15 changes: 15 additions & 0 deletions models/3dmolms/3dmolms_test/config.pbtxt
Original file line number Diff line number Diff line change
@@ -0,0 +1,15 @@
max_batch_size: 1000
input[
{
name: 'charge_raw',
data_type: TYPE_INT32,
dims: [1]
}
]
output [
{
name: 'charge_norm',
data_type: TYPE_FP32,
dims: [1]
}
]