Hi,
I would like to ask about VLM Inference Service:
- Where is the source code this service? It would be great if we could do some customization in the code if possible.
- Is there any plan to open-source it?
- Is there any similar pipeline that could be run on dGPU?
Thank you!