Thread 3 "tritonserver" received signal SIGSEGV, Segmentation fault. [Switching to Thread 0x7f0817fff000 (LWP 577)] __strlen_avx2 () at ../sysdeps/x86_64/multiarch/strlen-avx2.S:65 65 ../sysdeps/x86_64/multiarch/strlen-avx2.S: No such file or directory. (gdb) bt #0 __strlen_avx2 () at ../sysdeps/x86_64/multiarch/strlen-avx2.S:65 #1 0x00007f089c70e1a5 in nvidia::inferenceserver::Metrics::UUIDForCudaDevice(int, std::__cxx11::basic_string, std::allocator >*) () from /opt/tritonserver/lib/libtritonserver.so #2 0x00007f089c707f7e in nvidia::inferenceserver::MetricModelReporter::GetMetricLabels(std::map, std::allocator >, std::__cxx11::basic_string, std::allocator >, std::less, std::allocator > >, std::allocator, std::allocator > const, std::__cxx11::basic_string, std::allocator > > > >*, std::__cxx11::basic_string, std::allocator > const&, long, int, google::protobuf::Map, std::allocator >, std::__cxx11::basic_string, std::allocator > > const&) () from /opt/tritonserver/lib/libtritonserver.so #3 0x00007f089c708aab in nvidia::inferenceserver::MetricModelReporter::Create(std::__cxx11::basic_string, std::allocator > const&, long, int, google::protobuf::Map, std::allocator >, std::__cxx11::basic_string, std::allocator > > const&, std::shared_ptr*) () from /opt/tritonserver/lib/libtritonserver.so #4 0x00007f089c8b26d7 in nvidia::inferenceserver::TritonModelInstance::TritonModelInstance(nvidia::inferenceserver::TritonModel*, std::__cxx11::basic_string, std::allocator > const&, unsigned long, TRITONSERVER_instancegroupkind_enum, int) () from /opt/tritonserver/lib/libtritonserver.so #5 0x00007f089c8b27f8 in nvidia::inferenceserver::TritonModelInstance::CreateInstance(nvidia::inferenceserver::TritonModel*, std::__cxx11::basic_string, std::allocator > const&, unsigned long, TRITONSERVER_instancegroupkind_enum, int, std::vector >, std::allocator > > >*) () from /opt/tritonserver/lib/libtritonserver.so #6 0x00007f089c8b2ffc in nvidia::inferenceserver::TritonModelInstance::CreateInstances(nvidia::inferenceserver::TritonModel*, inference::ModelConfig const&, std::vector >, std::allocator > > >*) () from /opt/tritonserver/lib/libtritonserver.so #7 0x00007f089c8b110c in nvidia::inferenceserver::TritonModel::Create(nvidia::inferenceserver::InferenceServer*, std::__cxx11::basic_string, std::allocator > const&, std::unordered_map, std::allocator >, std::vector, std::allocator >, std::__cxx11::basic_string, std::allocator > >, std::allocator, std::allocator >, std::__cxx11::basic_string, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator > const, std::vector, std::allocator >, std::__cxx11::basic_string, std::allocator > >, std::allocator, std::allocator >, std::__cxx11::basic_string, std::allocator > > > > > > > const&, std::__cxx11::basic_string, std::allocator > const&, long, inference::ModelConfig const&, std::unique_ptr >*) () from /opt/tritonserver/lib/libtritonserver.so #8 0x00007f089c7349c0 in nvidia::inferenceserver::ModelRepositoryManager::BackendLifeCycle::CreateInferenceBackend(std::__cxx11::basic_string, std::allocator > const&, long, nvidia::inferenceserver::ModelRepositoryManager::BackendLifeCycle::BackendInfo*) () from /opt/tritonserver/lib/libtritonserver.so #9 0x00007f089c742641 in std::thread::_State_impl, std::allocator > const&, long, nvidia::inferenceserver::ModelRepositoryManager::BackendLifeCycle::BackendInfo*), nvidia::inferenceserver::ModelRepositoryManager::BackendLifeCycle*, std::__cxx11::basic_string, std::allocator >, long, nvidia::inferenceserver::ModelRepositoryManager::BackendLifeCycle::BackendInfo*> > >::_M_run() () from /opt/tritonserver/lib/libtritonserver.so #10 0x00007f089c0e8de4 in ?? () from /lib/x86_64-linux-gnu/libstdc++.so.6 #11 0x00007f089c582609 in start_thread (arg=) at pthread_create.c:477 #12 0x00007f089bdd6293 in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:95 (gdb)