Object detection with Faster RCNN Deep Learning in C#
The sample walks through how to run a pretrained Faster R-CNN object detection ONNX model using the ONNX Runtime C# API.
The source code for this sample is available here.
Contents
Prerequisites
To run this sample, you’ll need the following things:
- Install .NET Core 3.1 or higher for you OS (Mac, Windows or Linux).
- Download the Faster R-CNN ONNX model to your local system.
- Download this demo image to test the model. You can also use any image you like.
Get started
Now we have everything set up, we can start adding code to run the model on the image. We’ll do this in the main method of the program for simplicity.
Read paths
Firstly, let’s read the path to the model, path to the image we want to test, and path to the output image:
string modelFilePath = args[0];
string imageFilePath = args[1];
string outImageFilePath = args[2];
Read image
Next, we will read the image in using the cross-platform image library ImageSharp:
using Image<Rgb24> image = Image.Load<Rgb24>(imageFilePath, out IImageFormat format);
Note, we’re specifically reading the Rgb24
type so we can efficiently preprocess the image in a later step.
Resize image
Next, we will resize the image to the appropriate size that the model is expecting; it is recommended to resize the image such that both height and width are within the range of [800, 1333].
float ratio = 800f / Math.Min(image.Width, image.Height);
using Stream imageStream = new MemoryStream();
image.Mutate(x => x.Resize((int)(ratio * image.Width), (int)(ratio * image.Height)));
image.Save(imageStream, format);
Preprocess image
Next, we will preprocess the image according to the requirements of the model:
var paddedHeight = (int)(Math.Ceiling(image.Height / 32f) * 32f);
var paddedWidth = (int)(Math.Ceiling(image.Width / 32f) * 32f);
var mean = new[] { 102.9801f, 115.9465f, 122.7717f };
// Preprocessing image
// We use DenseTensor for multi-dimensional access
DenseTensor<float> input = new(new[] { 3, paddedHeight, paddedWidth });
image.ProcessPixelRows(accessor =>
{
for (int y = paddedHeight - accessor.Height; y < accessor.Height; y++)
{
Span<Rgb24> pixelSpan = accessor.GetRowSpan(y);
for (int x = paddedWidth - accessor.Width; x < accessor.Width; x++)
{
input[0, y, x] = pixelSpan[x].B - mean[0];
input[1, y, x] = pixelSpan[x].G - mean[1];
input[2, y, x] = pixelSpan[x].R - mean[2];
}
}
});
Here, we’re creating a Tensor of the required size (channels, paddedHeight, paddedWidth)
, accessing the pixel values, preprocessing them and finally assigning them to the tensor at the appropriate indicies.
Setup inputs
// Pin DenseTensor memory and use it directly in the OrtValue tensor // It will be unpinned on ortValue disposal
using var inputOrtValue = OrtValue.CreateTensorValueFromMemory(OrtMemoryInfo.DefaultInstance,
input.Buffer, new long[] { 3, paddedHeight, paddedWidth });
Next, we will create the inputs to the model:
var inputs = new Dictionary<string, OrtValue>
{
{ "image", inputOrtValue }
};
To check the input node names for an ONNX model, you can use Netron to visualise the model and see input/output names. In this case, this model has image
as the input node name.
Run inference
Next, we will create an inference session and run the input through it:
using var session = new InferenceSession(modelFilePath);
using var runOptions = new RunOptions();
using IDisposableReadOnlyCollection<OrtValue> results = session.Run(runOptions, inputs, session.OutputNames);
Postprocess output
Next, we will need to postprocess the output to get boxes and associated label and confidence scores for each box:
var boxesSpan = results[0].GetTensorDataAsSpan<float>();
var labelsSpan = results[1].GetTensorDataAsSpan<long>();
var confidencesSpan = results[2].GetTensorDataAsSpan<float>();
const float minConfidence = 0.7f;
var predictions = new List<Prediction>();
for (int i = 0; i < boxesSpan.Length - 4; i += 4)
{
var index = i / 4;
if (confidencesSpan[index] >= minConfidence)
{
predictions.Add(new Prediction
{
Box = new Box(boxesSpan[i], boxesSpan[i + 1], boxesSpan[i + 2], boxesSpan[i + 3]),
Label = LabelMap.Labels[labelsSpan[index]],
Confidence = confidencesSpan[index]
});
}
}
Note, we’re only taking boxes that have a confidence above 0.7 to remove false positives.
View prediction
Next, we’ll draw the boxes and associated labels and confidence scores on the image to see how the model went:
using var outputImage = File.OpenWrite(outImageFilePath);
Font font = SystemFonts.CreateFont("Arial", 16);
foreach (var p in predictions)
{
image.Mutate(x =>
{
x.DrawLines(Color.Red, 2f, new PointF[] {
new PointF(p.Box.Xmin, p.Box.Ymin),
new PointF(p.Box.Xmax, p.Box.Ymin),
new PointF(p.Box.Xmax, p.Box.Ymin),
new PointF(p.Box.Xmax, p.Box.Ymax),
new PointF(p.Box.Xmax, p.Box.Ymax),
new PointF(p.Box.Xmin, p.Box.Ymax),
new PointF(p.Box.Xmin, p.Box.Ymax),
new PointF(p.Box.Xmin, p.Box.Ymin)
});
x.DrawText($"{p.Label}, {p.Confidence:0.00}", font, Color.White, new PointF(p.Box.Xmin, p.Box.Ymin));
});
}
image.Save(outputImage, format);
For each box prediction, we’re using ImageSharp to draw red lines to create the boxes, and drawing the label and confidence text.
Running the program
Now the program is created, we can run it will the following command:
dotnet run [path-to-model] [path-to-image] [path-to-output-image]
e.g. running:
dotnet run ~/Downloads/FasterRCNN-10.onnx ~/Downloads/demo.jpg ~/Downloads/out.jpg
detects the following objects in the image: