1 Star 0 Fork 17

非一般的赏 / MobileNetv2-YOLOV3

forked from dog-qiuqiu / MobileNetv2-YOLOV3 
加入 Gitee
与超过 1200万 开发者一起发现、参与优秀开源项目,私有仓库也完全免费 :)
免费加入
克隆/下载
贡献代码
同步代码
取消
提示: 由于 Git 不支持空文件夾,创建文件夹后会生成空的 .keep 文件
Loading...
README

路过的大佬给颗星星吧

待办

  • 开放MNN模型转换脚本
  • 开放andooid示例项目

MobileNetv2-YOLOv3-SPP Darknet

A darknet implementation of MobileNetv2-YOLOv3-SPP detection network

Network COCO mAP(0.5) Resolution FLOPS Weight size
MobileNetV2-YOLOv3-SPP 42.6 416 6.1BFlops 17.6MB
YOLOv4-Tiny 40.2 416 6.9BFlops 23.1MB

*emmmm...这个懒得训练,mAP就凑合这样吧

Darknet Group convolution is not well supported on some GPUs such as NVIDIA PASCAL!!! The MobileNetV2-YOLOv3-SPP inference time is 100ms at GTX1080ti, but RTX2080 inference time is 5ms!!!

MobileNetV2-YOLOv3-Lite&Nano Darknet

Mobile inference frameworks benchmark (4*ARM_CPU)

Network VOC mAP(0.5) COCO mAP(0.5) Resolution Inference time (NCNN/Kirin 990) Inference time (MNN arm82/Kirin 990) FLOPS Weight size
MobileNetV2-YOLOv3-Lite 72.61 36.57 320 33 ms 18 ms 1.8BFlops 8.0MB
MobileNetV2-YOLOv3-Nano 65.27 30.13 320 13 ms 5 ms 0.5BFlops 3.0MB
MobileNetV2-YOLOv3-Fastest & & 320 & ms & ms 0.13BFlops 1.0MB
YOLOv3-Tiny-Prn & 33.1 416 44 ms & ms 3.5BFlops 18.8MB
YOLO-Nano 69.1 & 416 & ms & ms 4.57BFlops 4.0MB
  • Darknet Train Configuration: CUDA-version: 10010 (10020), cuDNN: 7.6.4,OpenCV version: 4 GPU:RTX2080ti
  • Support mobile inference frameworks such as NCNN&MNN

MobileNetV2-YOLO-Fastest

Network Resolution Inference time (DarkNet/i7-6700) Inference time (NCNN/Kirin 990) Inference time (MNN arm82/Kirin 990) FLOPS Weight size
MobileNetV2-YOLOv3-Fastest 320 28 ms & ms & ms 0.1BFlops 500KB
  • 快就完了,要啥mAP:sunglasses:

Reference&Framework instructions&How to Train

  • https://github.com/AlexeyAB/darknet
  • You must use a pre-trained model to train your own data set. You can make a pre-trained model based on the weights of COCO training in this project to initialize the network parameters
  • 交流qq群:1062122604

About model selection

  • MobileNetV2-YOLOv3-SPP: Nvidia Jeston, Intel Movidius, TensorRT,NPU,OPENVINO...High-performance embedded side
  • MobileNetV2-YOLOv3-Lite: High Performance ARM-CPU,Qualcomm Adreno GPU, ARM82...High-performance mobile
  • MobileNetV2-YOLOv3-NANO: ARM-CPU...Computing resources are limited
  • MobileNetV2-YOLOv3-Fastest: ....... Can you do personal face detection???It’s better than nothing

MobileNetV2-YOLOv3-Lite-COCO Test results

image

DarkNet2Caffe tutorial

Environmental requirements

  • Python2.7
  • python-opencv
  • Caffe(add upsample layer https://github.com/dog-qiuqiu/caffe-upsample)
      cd darknet2caffe/
      python darknet2caffe.py MobileNetV2-YOLOv3-Nano-voc.cfg MobileNetV2-YOLOv3-Nano-voc.weights MobileNetV2-YOLOv3-Nano-voc.prototxt MobileNetV2-YOLOv3-Nano-voc.caffemodel
      cp MobileNetV2-YOLOv3-Nano-voc.prototxt sample
      cp MobileNetV2-YOLOv3-Nano-voc.caffemodel sample
      cd sample
      python detector.py

MNN conversion tutorial

  • Convert darknet model to caffemodel through darknet2caffe
  • Manually replace the upsample layer in prototxt with the interp layer
  • Take the modification of MobileNetV2-YOLOv3-Nano-voc.prototxt as an example
#layer {
#    bottom: "layer71-route"
#    top: "layer72-upsample"
#    name: "layer72-upsample"
#    type: "Upsample"
#    upsample_param {
#        scale: 2
#    }
#}
layer {
    bottom: "layer71-route"
    top: "layer72-upsample"
    name: "layer72-upsample"
    type: "Interp"
    interp_param {
        height:20 
	      width:20
    }
}

NCNN conversion tutorial

NCNN Android Sample

  • 白嫖中....

Thanks

空文件

简介

MobileNetV2-YOLOv3-Nano的Darknet实现:移动终端设计的目标检测网络,计算量0.5BFlops!华为P40:MNN_ARM82单次推理时间6ms 模型大小:3MB! 展开 收起
取消

发行版

暂无发行版

贡献者

全部

近期动态

加载更多
不能加载更多了
C
1
https://gitee.com/tanweishang/MobileNetv2-YOLOV3.git
git@gitee.com:tanweishang/MobileNetv2-YOLOV3.git
tanweishang
MobileNetv2-YOLOV3
MobileNetv2-YOLOV3
master

搜索帮助