How to get an output from a custom trained objected detection model

Hey guys, so I have my object detection programming that checks if an object is a ping pong ball using detect net. I then need to get this data, i.e if the detect net program detects a ping pong ball with high enough accuracy it should received some sort of output which then I can use to send my Arduino to control the movement of the servos. I have seen the object detection program running, but I am not sure how I am suppose to get a tangible output from it that I can send off as input to the arduino.

Any advice on how I would achieve this would be appreciated. Thank you.

Hi,

You can find how to get the confidence value and detected class below:
(detections[n].ClassID and detections[n].Confidence)

Thanks.

So once I modify the detect net script to obtain these values in real time, do you know any methods of using this data to have serial communication with the arduino? Would I be able to add a function in the detectnet script or would I need a separate script for this?

Note that the Python bindings for detectNet already have access to these values - these are the members of the detectNet.Detection object:

Detection = <class 'jetson.inference.detectNet.Detection'>
Object Detection Result
 
----------------------------------------------------------------------
Data descriptors defined here:
 
Area
    Area of bounding box
 
Bottom
    Bottom bounding box coordinate
 
Center
    Center (x,y) coordinate of bounding box
 
ClassID
    Class index of the detected object
 
TrackID
    Unique tracking ID (or -1 if untracked)
 
Confidence
    Confidence value of the detected object
 
Height
    Height of bounding box
 
Left
    Left bounding box coordinate
 
Right
    Right bounding box coordinate
 
Top
    Top bounding box coordinate
 
Width
     Width of bounding box

So you can do things like this for example:

detections = net.Detect(img)

for detection in detections:
    print(f"class {detection.ClassID} found at ({detection.Left}, {detection.Top}, {detection.Right}, {detection.Bottom}")

Yes, you would need to add your own code to the detectnet.py main loop for doing this (i.e. using pyserial or some other Python module for interfacing with the serial port). If the serial port functions are blocking and impact performance, you might want to run them in a separate Python thread inside your customized script.

Do you know where the main detectnet.py script is located? I see a detectnet.py in the jetson-inference/python/examples directory.

#!/usr/bin/env python3
#
# Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved.
#
# Permission is hereby granted, free of charge, to any person obtaining a
# copy of this software and associated documentation files (the "Software"),
# to deal in the Software without restriction, including without limitation
# the rights to use, copy, modify, merge, publish, distribute, sublicense,
# and/or sell copies of the Software, and to permit persons to whom the
# Software is furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in
# all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
# THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
# DEALINGS IN THE SOFTWARE.
#

import sys
import argparse

from jetson_inference import detectNet
from jetson_utils import videoSource, videoOutput, Log

# parse the command line
parser = argparse.ArgumentParser(description="Locate objects in a live camera stream using an object detection DNN.", 
                                 formatter_class=argparse.RawTextHelpFormatter, 
                                 epilog=detectNet.Usage() + videoSource.Usage() + videoOutput.Usage() + Log.Usage())

parser.add_argument("input", type=str, default="", nargs='?', help="URI of the input stream")
parser.add_argument("output", type=str, default="", nargs='?', help="URI of the output stream")
parser.add_argument("--network", type=str, default="ssd-mobilenet-v2", help="pre-trained model to load (see below for options)")
parser.add_argument("--overlay", type=str, default="box,labels,conf", help="detection overlay flags (e.g. --overlay=box,labels,conf)\nvalid combinations are:  'box', 'labels', 'conf', 'none'")
parser.add_argument("--threshold", type=float, default=0.5, help="minimum detection threshold to use") 

is_headless = ["--headless"] if sys.argv[0].find('console.py') != -1 else [""]

try:
	args = parser.parse_known_args()[0]
except:
	print("")
	parser.print_help()
	sys.exit(0)

# create video sources and outputs
input = videoSource(args.input, argv=sys.argv)
output = videoOutput(args.output, argv=sys.argv+is_headless)
	
# load the object detection network
net = detectNet(args.network, sys.argv, args.threshold)

# note: to hard-code the paths to load a model, the following API can be used:
#
# net = detectNet(model="model/ssd-mobilenet.onnx", labels="model/labels.txt", 
#                 input_blob="input_0", output_cvg="scores", output_bbox="boxes", 
#                 threshold=args.threshold)

# process frames until EOS or the user exits
while True:
    # capture the next image
    img = input.Capture()

    if img is None: # timeout
        continue  
        
    # detect objects in the image (with overlay)
    detections = net.Detect(img, overlay=args.overlay)

    # print the detections
    print("detected {:d} objects in image".format(len(detections)))

    for detection in detections:
        print(detection)

    # render the image
    output.Render(img)

    # update the title bar
    output.SetStatus("{:s} | Network {:.0f} FPS".format(args.network, net.GetNetworkFPS()))

    # print out performance info
    net.PrintProfilerTimes()

    # exit on input/output EOS
    if not input.IsStreaming() or not output.IsStreaming():
        break

I assume I put my custom code to send out serial data within this while loop, and the values Im looking for should be in the detections array. Just making sure.

Thank you.

Yes, the main detectnet.py is located in jetson-inference/python/examples
When you run cmake ../ from the build directory, it gets copied to jetson-inference/build/aarch64/bin
And then when you run sudo make install, it gets copied to /usr/local/bin

What I’d actually recommend is creating your own script with your own filename to avoid any confusion or overwriting.

So I edited the detectnet.py script, so far I am just trying to get the script so that it writes to a output text file when it sees a ping pong ball. But it doesn’t seem to be creating the text file in the current directory which is the jetson-inference/python/examples folder.

These were the additions I added to the code:

Before the main while True: loop, I added import os, and output_file = open(“output.txt”, “w”)

And then within the while loop I added inside the following in the for loop:

for detection in detections:
         print(f"class {detection.ClassID}")
         output_file.write(str(detection.ClassID) + "\n")

and lastly after the while loop ends I just added output_file.close().

So the output file is not being created, and I was also wondering where do all the print statements print to? As I also see several print statements in the unmodified detectnet.py script.

The way I run the detectnet program is first I cd to jetson-inference, and then I run docker/run.sh to enter the container. Then I cd to jetson-inference/python/training/detection/ssd. Then I run the command

detectnet --model=models/pingpongballs/ssd-mobilenet.onnx --labels=models/pingpongballs/labels.txt --input-blob=input_0 --output-cvg=scores --output-bbox=boxes csi://0

I wanted to ask if this is running the detectnet.py script as I know there is also the detectnet.cpp script. Along with that where do all the print statements output to as when I run that command above it does not show the prints, it just shows the Timing Report models/pingpongballs/ssd-mobilenet.onnx charts.

Thank you in advance.

The apps inside the container only exist inside the container, unless you mount a folder with them into the container. So those are separate files. Also detectnet refers to the C++ executable, the python version is detectnet.py. And if you want to run the python file from your current directory, do python3 detectnet.py (otherwise if you just execute detectnet.py it will run the one from /usr/local/bin)

Since you are using the container, what I recommend doing is creating a folder on your device (outside of the container) that stores your scripts. Then mount that folder into the container when you start it.

mkdir ~/my_scripts
cp jetson-inference/python/examples/detectnet.py ~/my_scripts/my_detectnet.py
# make modifications to my_detectnet.py

# start the container, mount your folder into it
cd jetson-inference
docker/run.sh -v ~/my_scripts:/my_scripts

# now running inside the container
cd /my_scripts
python3 my_detection.py

Also, if you want you can hardcode the paths to your custom model in your script like this (then you don’t need to specify them on the command-line every time): https://github.com/dusty-nv/jetson-inference/blob/2ec121e545a8bdaee11241cefee8003ad28cb9b3/python/examples/detectnet.py#L57

This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.