yolov8使用
导出onnx yolo export model=D:/zou/wjzz/workDoc/screwsDetech/runs/detect/train/weights/best.pt format=onnx opset=12 dynamic=True simplify=True
内容 | 解释 |
---|---|
蓝色 Conv | 卷积层,提取特征 |
黑色 Add / Mul | 做位置偏移、网格映射 |
绿色 Gather / Shape | 从 shape 里提取宽、高等维度信息 |
红色 Concat / Reshape | 拼接张量、变形输出 |
紫色 Softmax / Transpose | 用于分类概率或格式变换 |
因为这部分在 ONNX 里:
-
很难理解 / 不好调试
-
不利于优化和加速
-
有时还不能使用 GPU 推理
所以很多人导出的时候会加上 export(onnx, simplify=True, dynamic=False, nms=False)
,再用自己写的代码(像你 Python 写的那样)做后处理。
✅ 什么是张量(Tensor)?
张量(Tensor)就是一个多维数组。 你可以把它看作是:
维度 | 举例 | 类比(在 NumPy 中) |
---|---|---|
标量(0维) | 3.14 | np.array(3.14) |
向量(1维) | [1, 2, 3] | np.array([1, 2, 3]) |
矩阵(2维) | [[1,2],[3,4]] | np.array([[1,2],[3,4]]) |
三维张量(3维) | 彩色图像(H, W, C) | img[height][width][channel] |
四维张量(4维) | 多张图片:[batch, channel, height, width] | img[batch][channel][H][W] |
所以你可以记住一句话: 张量 = “有维度的数组”,是用来装图片、特征图、模型输入输出的容器。
✅ NCHW 和 NHWC 是什么?
它们是 图像数据的维度排列方式,在模型输入输出中经常见到。
格式 含义 解释 举例(YOLO) NCHW Batch, Channels, Height, Width PyTorch 默认格式 [1, 3, 640, 640]
NHWC Batch, Height, Width, Channels TensorFlow 默认格式 [1, 640, 640, 3]
比如,一张 RGB 图片输入模型时:
NCHW:先通道、后高宽:
[1, 3, 640, 640]
NHWC:先高宽、后通道:
[1, 640, 640, 3]
假设你有 100 张图片要输入模型,但你的显卡内存有限,不能一次处理 100 张,那就可以每次处理 16 张,这个 16 就是 Batch Size(N)。
NCHW: (16, 3, 640, 640)
N = 16:一次处理 16 张图片
C = 3:每张图有 3 个通道(RGB)
H = 640:高度 640 像素
W = 640:宽度 640 像素
已经可以跑通的程序、没有优化
using OpenCvSharp; using Microsoft.ML.OnnxRuntime; using Microsoft.ML.OnnxRuntime.Tensors; using System; using System.Linq;
class Program { static void Main(string[] args) { // 1. 加载模型 var session = new InferenceSession(@"D:\zou\wjzz\workDoc\screwsDetech\runs\detect\train\weights\best.onnx");
// 2. 读取图片string imgPath = @"D:\zou\wjzz\workDoc\screwsDetech\test_images\screw_keshang_bad_130.png";Mat img = Cv2.ImRead(imgPath);Mat originalImg = img.Clone();int h0 = img.Rows;int w0 = img.Cols; // 3. 图片预处理:resize到640x640,BGR转RGB,归一化Mat imgResized = new Mat();Cv2.Resize(img, imgResized, new Size(640, 640));Cv2.CvtColor(imgResized, imgResized, ColorConversionCodes.BGR2RGB);imgResized.ConvertTo(imgResized, MatType.CV_32FC3, 1.0 / 255); // 4. HWC转CHW,并变成1x3x640x640的Tensorvar inputData = new float[1 * 3 * 640 * 640];for (int y = 0; y < 640; y++){for (int x = 0; x < 640; x++){Vec3f pixel = imgResized.At<Vec3f>(y, x);inputData[0 * 3 * 640 * 640 + 0 * 640 * 640 + y * 640 + x] = pixel.Item0; // R从0开始inputData[0 * 3 * 640 * 640 + 1 * 640 * 640 + y * 640 + x] = pixel.Item1; // G从640*640inputData[0 * 3 * 640 * 640 + 2 * 640 * 640 + y * 640 + x] = pixel.Item2; // B从640*640*2}}var inputTensor = new DenseTensor<float>(inputData, new[] { 1, 3, 640, 640 }); // 5. 推理var inputs = new List<NamedOnnxValue>{NamedOnnxValue.CreateFromTensor(session.InputMetadata.Keys.First(), inputTensor)};using var results = session.Run(inputs);var output = results.First().AsTensor<float>(); // 6. 解析输出var dims = output.Dimensions; // 应该是 (1, 5, N),比如(1,5,8400)int numPreds = dims[2];float[,] preds = new float[numPreds, 5];for (int i = 0; i < 5; i++){for (int j = 0; j < numPreds; j++){preds[j, i] = output[0, i, j];}} // 7. 按置信度筛选float confThreshold = 0.25f;var boxes = new List<Rect2d>();var scores = new List<float>();for (int i = 0; i < numPreds; i++){float objness = preds[i, 4];if (objness > confThreshold){float cx = preds[i, 0];float cy = preds[i, 1];float w = preds[i, 2];float h = preds[i, 3];float x1 = cx - w / 2;float y1 = cy - h / 2; float x2 = cx + w / 2;float y2 = cy + h / 2; // 还原到原图尺度x1 *= (float)w0 / 640;x2 *= (float)w0 / 640;y1 *= (float)h0 / 640;y2 *= (float)h0 / 640; boxes.Add(new Rect2d(x1, y1, x2 - x1, y2 - y1));scores.Add(objness);}} // 8. 绘制框for (int i = 0; i < boxes.Count; i++){var rect = boxes[i];Cv2.Rectangle(originalImg, new Point(rect.X, rect.Y), new Point(rect.X + rect.Width, rect.Y + rect.Height), Scalar.Green, 2);Cv2.PutText(originalImg, $"{scores[i]:0.00}", new Point((int)rect.X, (int)rect.Y - 10), HersheyFonts.HersheySimplex, 0.5, Scalar.Green, 1);} // 9. 显示Cv2.ImShow("Result", originalImg);Cv2.WaitKey();Cv2.DestroyAllWindows(); }
}
处理逻辑优化使用NMS(非极大值抑制)
using OpenCvSharp; using Microsoft.ML.OnnxRuntime; using Microsoft.ML.OnnxRuntime.Tensors; using System; using System.Collections.Generic; using System.Linq;
class Program { static readonly int inputWidth = 640; static readonly int inputHeight = 640; static readonly float confThreshold = 0.25f; static readonly float iouThreshold = 0.45f;
static readonly InferenceSession session = new InferenceSession(@"D:\zou\wjzz\workDoc\screwsDetech\runs\detect\train\weights\best.onnx"); static void Main(string[] args) {string imgPath = @"D:\zou\wjzz\workDoc\screwsDetech\test_images\screw_keshang_bad_130.png";Mat img = Cv2.ImRead(imgPath);Mat originalImg = img.Clone(); int h0 = img.Rows;int w0 = img.Cols; var inputTensor = Preprocess(img);var output = Inference(inputTensor);var (boxes, scores) = Postprocess(output, w0, h0);var keepIdx = NMS(boxes, scores, iouThreshold); foreach (int i in keepIdx){var box = boxes[i];Cv2.Rectangle(originalImg, new Point(box.X, box.Y), new Point(box.X + box.Width, box.Y + box.Height), Scalar.Green, 2);Cv2.PutText(originalImg, $"{scores[i]:0.00}", new Point((int)box.X, (int)box.Y - 10), HersheyFonts.HersheySimplex, 0.5, Scalar.Green, 1);} Cv2.ImShow("Result", originalImg);Cv2.WaitKey();Cv2.DestroyAllWindows(); } static DenseTensor<float> Preprocess(Mat img) {Mat resized = new Mat();Cv2.Resize(img, resized, new Size(inputWidth, inputHeight));Cv2.CvtColor(resized, resized, ColorConversionCodes.BGR2RGB);resized.ConvertTo(resized, MatType.CV_32FC3, 1.0 / 255); Mat[] channels = Cv2.Split(resized);float[] inputData = new float[3 * inputHeight * inputWidth];for (int c = 0; c < 3; c++){var indexer = channels[c].GetGenericIndexer<float>();for (int y = 0; y < inputHeight; y++)for (int x = 0; x < inputWidth; x++)inputData[c * inputHeight * inputWidth + y * inputWidth + x] = indexer[y, x];} return new DenseTensor<float>(inputData, new[] { 1, 3, inputHeight, inputWidth }); } static Tensor<float> Inference(DenseTensor<float> inputTensor) {var inputs = new List<NamedOnnxValue>{NamedOnnxValue.CreateFromTensor(session.InputMetadata.Keys.First(), inputTensor)};using var results = session.Run(inputs);return results.First().AsTensor<float>(); } static (List<Rect2d>, List<float>) Postprocess(Tensor<float> output, int w0, int h0) {var dims = output.Dimensions;int numPreds = dims[2];var boxes = new List<Rect2d>();var scores = new List<float>(); for (int i = 0; i < numPreds; i++){float obj = output[0, 4, i];if (obj < confThreshold) continue; float cx = output[0, 0, i];float cy = output[0, 1, i];float w = output[0, 2, i];float h = output[0, 3, i]; float x1 = (cx - w / 2) * w0 / inputWidth;float y1 = (cy - h / 2) * h0 / inputHeight;float x2 = (cx + w / 2) * w0 / inputWidth;float y2 = (cy + h / 2) * h0 / inputHeight; boxes.Add(new Rect2d(x1, y1, x2 - x1, y2 - y1));scores.Add(obj);} return (boxes, scores); } static List<int> NMS(List<Rect2d> boxes, List<float> scores, float iouThreshold) {var indices = scores.Select((score, idx) => new { score, idx }).OrderByDescending(s => s.score).Select(s => s.idx).ToList(); var keep = new List<int>();var removed = new bool[boxes.Count]; for (int i = 0; i < indices.Count; i++){int idx = indices[i];if (removed[idx]) continue;keep.Add(idx); for (int j = i + 1; j < indices.Count; j++){int idx2 = indices[j];if (removed[idx2]) continue;if (ComputeIOU(boxes[idx], boxes[idx2]) > iouThreshold)removed[idx2] = true;}}return keep; } static double ComputeIOU(Rect2d box1, Rect2d box2) {double xx1 = Math.Max(box1.X, box2.X);double yy1 = Math.Max(box1.Y, box2.Y);double xx2 = Math.Min(box1.X + box1.Width, box2.X + box2.Width);double yy2 = Math.Min(box1.Y + box1.Height, box2.Y + box2.Height); double interArea = Math.Max(0, xx2 - xx1) * Math.Max(0, yy2 - yy1);double unionArea = box1.Width * box1.Height + box2.Width * box2.Height - interArea; return interArea / unionArea; }
}