im using docker image triton-server-20.02
i have exposed port 800, 8001 , 8002
if i use http port 8000 , though curl i get response
curl 3.144.31.75:8000/v2/models/IndianVehicle_model/config
{"name":"IndianVehicle_model","platform":"tensorrt_plan","backend":"","version_policy":{"latest":{"num_versions":1}},"max_batch_size":16,"input":[{"name":"Input","data_type":"TYPE_FP32","format":"FORMAT_NCHW","dims":[3,384,1248],"is_shape_tensor":false,"allow_ragged_batch":false}],"output":[{"name":"BatchedNMS","data_type":"TYPE_INT32","dims":[1],"label_filename":"","is_shape_tensor":false},{"name":"BatchedNMS_1","data_type":"TYPE_FP32","dims":[200,4],"label_filename":"","is_shape_tensor":false},{"name":"BatchedNMS_2","data_type":"TYPE_FP32","dims":[200],"label_filename":"","is_shape_tensor":false},{"name":"BatchedNMS_3","data_type":"TYPE_FP32","dims":[200],"label_filename":"","is_shape_tensor":false}],"batch_input":[],"batch_output":[],"optimization":{"priority":"PRIORITY_DEFAULT","input_pinned_memory":{"enable":true},"output_pinned_memory":{"enable":true},"gather_kernel_buffer_threshold":0,"eager_batching":false},"dynamic_batching":{"preferred_batch_size":[16],"max_queue_delay_microseconds":0,"preserve_ordering":false,"priority_levels":0,"default_priority_level":0,"priority_queue_policy":{}},"instance_group":[{"name":"IndianVehicle_model_0","kind":"KIND_GPU","count":1,"gpus":[0],"secondary_devices":[],"profile":[],"passive":false,"host_policy":""}],"default_model_filename":"trt.engine","cc_model_filenames":{},"metric_tags":{},"parameters":{},"model_warmup":[]}root@6b55a2214e5a:/opt/nvidia/deepstream/deepstream-6.0/sources/client/src/python/examples
but if i use through grpc 8001
curl 3.144.31.75:8001/v2/models/IndianVehicle_model/config
curl: (1) Received HTTP/0.9 when not allowed