91欧美超碰AV自拍|国产成年人性爱视频免费看|亚洲 日韩 欧美一厂二区入|人人看人人爽人人操aV|丝袜美腿视频一区二区在线看|人人操人人爽人人爱|婷婷五月天超碰|97色色欧美亚州A√|另类A√无码精品一级av|欧美特级日韩特级

0
  • 聊天消息
  • 系統(tǒng)消息
  • 評(píng)論與回復(fù)
登錄后你可以
  • 下載海量資料
  • 學(xué)習(xí)在線課程
  • 觀看技術(shù)視頻
  • 寫文章/發(fā)帖/加入社區(qū)
會(huì)員中心
創(chuàng)作中心

完善資料讓更多小伙伴認(rèn)識(shí)你,還能領(lǐng)取20積分哦,立即完善>

3天內(nèi)不再提示

計(jì)算機(jī)視覺CV領(lǐng)域圖像分類方向文獻(xiàn)和代碼的超全總結(jié)和列表!

新機(jī)器視覺 ? 來源:新機(jī)器視覺 ? 作者:新機(jī)器視覺 ? 2020-11-03 10:08 ? 次閱讀
加入交流群
微信小助手二維碼

掃碼添加小助手

加入工程師交流群

今天給大家介紹自 2014 年以來,計(jì)算機(jī)視覺 CV 領(lǐng)域圖像分類方向文獻(xiàn)和代碼的超全總結(jié)和列表!總共涉及 36 種 ConvNet 模型。該 GitHub 項(xiàng)目作者是 weiaicunzai,項(xiàng)目地址是:

https://github.com/weiaicunzai/awesome-image-classification

背景

我相信圖像識(shí)別是深入到其它機(jī)器視覺領(lǐng)域一個(gè)很好的起點(diǎn),特別是對(duì)于剛剛?cè)腴T深度學(xué)習(xí)的人來說。當(dāng)我初學(xué) CV 時(shí),犯了很多錯(cuò)。我當(dāng)時(shí)非常希望有人能告訴我應(yīng)該從哪一篇論文開始讀起。到目前為止,似乎還沒有一個(gè)像 deep-learning-object-detection 這樣的 GitHub 項(xiàng)目。因此,我決定建立一個(gè) GitHub 項(xiàng)目,列出深入學(xué)習(xí)中關(guān)于圖像分類的論文和代碼,以幫助其他人。

對(duì)于學(xué)習(xí)路線,我的個(gè)人建議是,對(duì)于那些剛?cè)腴T深度學(xué)習(xí)的人,可以試著從 vgg 開始,然后是 googlenet、resnet,之后可以自由地繼續(xù)閱讀列出的其它論文或切換到其它領(lǐng)域。

性能表

基于簡化的目的,我只從論文中列舉出在 ImageNet 上準(zhǔn)確率最高的 top1 和 top5。注意,這并不一定意味著準(zhǔn)確率越高,一個(gè)網(wǎng)絡(luò)就比另一個(gè)網(wǎng)絡(luò)更好。因?yàn)橛行┚W(wǎng)絡(luò)專注于降低模型復(fù)雜性而不是提高準(zhǔn)確性,或者有些論文只給出 ImageNet 上的 single crop results,而另一些則給出模型融合或 multicrop results。

關(guān)于性能表的標(biāo)注:

ConvNet:卷積神經(jīng)網(wǎng)絡(luò)的名稱

ImageNet top1 acc:論文中基于 ImageNet 數(shù)據(jù)集最好的 top1 準(zhǔn)確率

ImageNet top5 acc:論文中基于 ImageNet 數(shù)據(jù)集最好的 top5 準(zhǔn)確率

Published In:論文發(fā)表在哪個(gè)會(huì)議或期刊

論文&代碼

1. VGG

Very Deep Convolutional Networks for Large-Scale Image Recognition.

Karen Simonyan, Andrew Zisserman

pdf: https://arxiv.org/abs/1409.1556

code: torchvision :

https://github.com/pytorch/vision/blob/master/torchvision/models/vgg.py

code: keras-applications :

https://github.com/keras-team/keras-applications/blob/master/keras_applications/vgg16.py

code: keras-applications :

https://github.com/keras-team/keras-applications/blob/master/keras_applications/vgg19.py

2. GoogleNet

Going Deeper with Convolutions

Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, Andrew Rabinovich

pdf:https://arxiv.org/abs/1409.4842

code: unofficial-tensorflow :

https://github.com/conan7882/GoogLeNet-Inception

code: unofficial-caffe :

https://github.com/lim0606/caffe-googlenet-bn

3.PReLU-nets

Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun

pdf:https://arxiv.org/abs/1502.01852

code: unofficial-chainer :

https://github.com/nutszebra/prelu_net

4.ResNet

Deep Residual Learning for Image Recognition

Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun

pdf:https://arxiv.org/abs/1512.03385

code: facebook-torch :

https://github.com/facebook/fb.resnet.torch

code: torchvision :

https://github.com/pytorch/vision/blob/master/torchvision/models/resnet.py

code: keras-applications :

https://github.com/keras-team/keras-applications/blob/master/keras_applications/resnet.py

code: unofficial-keras :

https://github.com/raghakot/keras-resnet

code: unofficial-tensorflow :

https://github.com/ry/tensorflow-resnet

5.PreActResNet

Identity Mappings in Deep Residual Networks

Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun

pdf:https://arxiv.org/abs/1603.05027

code: facebook-torch :

https://github.com/facebook/fb.resnet.torch/blob/master/models/preresnet.lua

code: official :

https://github.com/KaimingHe/resnet-1k-layers

code: unoffical-pytorch :

https://github.com/kuangliu/pytorch-cifar/blob/master/models/preact_resnet.py

code: unoffical-mxnet :

https://github.com/tornadomeet/ResNet

6.Inceptionv3

Rethinking the Inception Architecture for Computer Vision

Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jonathon Shlens, Zbigniew Wojna

pdf:https://arxiv.org/abs/1512.00567

code: torchvision :

https://github.com/pytorch/vision/blob/master/torchvision/models/inception.py

code: keras-applications :

https://github.com/keras-team/keras-applications/blob/master/keras_applications/inception_v3.py

7.Inceptionv4 && Inception-ResNetv2

Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning

Christian Szegedy, Sergey Ioffe, Vincent Vanhoucke, Alex Alemi

pdf:https://arxiv.org/abs/1602.07261

code: unofficial-keras :

https://github.com/kentsommer/keras-inceptionV4

code: unofficial-keras :

https://github.com/titu1994/Inception-v4

code: unofficial-keras :

https://github.com/yuyang-huang/keras-inception-resnet-v2

8. RIR

Resnet in Resnet: Generalizing Residual Architectures

Sasha Targ, Diogo Almeida, Kevin Lyman

pdf:https://arxiv.org/abs/1603.08029

code: unofficial-tensorflow :

https://github.com/SunnerLi/RiR-Tensorflow

code: unofficial-chainer :

https://github.com/nutszebra/resnet_in_resnet

9.Stochastic Depth ResNet

Deep Networks with Stochastic Depth

Gao Huang, Yu Sun, Zhuang Liu, Daniel Sedra, Kilian Weinberger

pdf:https://arxiv.org/abs/1603.09382

code: unofficial-torch :

https://github.com/yueatsprograms/Stochastic_Depth

code: unofficial-chainer :

https://github.com/yasunorikudo/chainer-ResDrop

code: unofficial-keras :

https://github.com/dblN/stochastic_depth_keras

10.WRN

Wide Residual Networks

Sergey Zagoruyko, Nikos Komodakis

pdf:https://arxiv.org/abs/1605.07146

code: official :

https://github.com/szagoruyko/wide-residual-networks

code: unofficial-pytorch :

https://github.com/xternalz/WideResNet-pytorch

code: unofficial-keras :

https://github.com/asmith26/wide_resnets_keras

code: unofficial-pytorch :

https://github.com/meliketoy/wide-resnet.pytorch

11.squeezenet

SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size?

Forrest N. Iandola, Song Han, Matthew W. Moskewicz, Khalid Ashraf, William J. Dally, Kurt Keutzer

pdf:https://arxiv.org/abs/1602.07360

code: torchvision :

https://github.com/pytorch/vision/blob/master/torchvision/models/squeezenet.py

code: unofficial-caffe :

https://github.com/DeepScale/SqueezeNet

code: unofficial-keras :

https://github.com/rcmalli/keras-squeezenet

code: unofficial-caffe :

https://github.com/songhan/SqueezeNet-Residual

12.GeNet

Genetic CNN

Lingxi Xie, Alan Yuille

pdf:https://arxiv.org/abs/1703.01513

code: unofficial-tensorflow :

https://github.com/aqibsaeed/Genetic-CNN

12.MetaQNN

Designing Neural Network Architectures using Reinforcement Learning

Bowen Baker, Otkrist Gupta, Nikhil Naik, Ramesh Raskar

pdf:https://arxiv.org/abs/1703.01513

code: official :https://github.com/bowenbaker/metaqnn

13.PyramidNet

Deep Pyramidal Residual Networks

Dongyoon Han, Jiwhan Kim, Junmo Kim

pdf:https://arxiv.org/abs/1610.02915

code: official :

https://github.com/jhkim89/PyramidNet

code: unofficial-pytorch :

https://github.com/dyhan0920/PyramidNet-PyTorch

14.DenseNet

Densely Connected Convolutional Networks

Gao Huang, Zhuang Liu, Laurens van der Maaten, Kilian Q. Weinberger

pdf:https://arxiv.org/abs/1608.06993

code: official :

https://github.com/liuzhuang13/DenseNet

code: unofficial-keras :

https://github.com/titu1994/DenseNet

code: unofficial-caffe :

https://github.com/shicai/DenseNet-Caffe

code: unofficial-tensorflow :

https://github.com/YixuanLi/densenet-tensorflow

code: unofficial-pytorch :

https://github.com/YixuanLi/densenet-tensorflow

code: unofficial-pytorch :

https://github.com/bamos/densenet.pytorch

code: unofficial-keras :

https://github.com/flyyufelix/DenseNet-Keras

15.FractalNet

FractalNet: Ultra-Deep Neural Networks without Residuals

Gustav Larsson, Michael Maire, Gregory Shakhnarovich

pdf:https://arxiv.org/abs/1605.07648

code: unofficial-caffe :

https://github.com/gustavla/fractalnet

code: unofficial-keras :

https://github.com/snf/keras-fractalnet

code: unofficial-tensorflow :

https://github.com/tensorpro/FractalNet

16.ResNext

Aggregated Residual Transformations for Deep Neural Networks

Saining Xie, Ross Girshick, Piotr Dollár, Zhuowen Tu, Kaiming He

pdf:https://arxiv.org/abs/1611.05431

code: official :

https://github.com/facebookresearch/ResNeXt

code: keras-applications :

https://github.com/keras-team/keras-applications/blob/master/keras_applications/resnext.py

code: unofficial-pytorch :

https://github.com/prlz77/ResNeXt.pytorch

code: unofficial-keras :

https://github.com/titu1994/Keras-ResNeXt

code: unofficial-tensorflow :

https://github.com/taki0112/ResNeXt-Tensorflow

code: unofficial-tensorflow :

https://github.com/wenxinxu/ResNeXt-in-tensorflow

17.IGCV1

Interleaved Group Convolutions for Deep Neural Networks

Ting Zhang, Guo-Jun Qi, Bin Xiao, Jingdong Wang

pdf:https://arxiv.org/abs/1707.02725

code official :

https://github.com/hellozting/InterleavedGroupConvolutions

18.Residual Attention Network

Residual Attention Network for Image Classification

Fei Wang, Mengqing Jiang, Chen Qian, Shuo Yang, Cheng Li, Honggang Zhang, Xiaogang Wang, Xiaoou Tang

pdf:https://arxiv.org/abs/1704.06904

code: official :

https://github.com/fwang91/residual-attention-network

code: unofficial-pytorch :

https://github.com/tengshaofeng/ResidualAttentionNetwork-pytorch

code: unofficial-gluon :

https://github.com/PistonY/ResidualAttentionNetwork

code: unofficial-keras :

https://github.com/koichiro11/residual-attention-network

19.Xception

Xception: Deep Learning with Depthwise Separable Convolutions

Fran?ois Chollet

pdf:https://arxiv.org/abs/1610.02357

code: unofficial-pytorch :

https://github.com/jfzhang95/pytorch-deeplab-xception/blob/master/modeling/backbone/xception.py

code: unofficial-tensorflow :

https://github.com/kwotsin/TensorFlow-Xception

code: unofficial-caffe :

https://github.com/yihui-he/Xception-caffe

code: unofficial-pytorch :

https://github.com/tstandley/Xception-PyTorch

code: keras-applications :

https://github.com/keras-team/keras-applications/blob/master/keras_applications/xception.py

20.MobileNet

MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications

Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, Hartwig Adam

pdf:https://arxiv.org/abs/1704.04861

code: unofficial-tensorflow :

https://github.com/Zehaos/MobileNet

code: unofficial-caffe :

https://github.com/shicai/MobileNet-Caffe

code: unofficial-pytorch :

https://github.com/marvis/pytorch-mobilenet

code: keras-applications :

https://github.com/keras-team/keras-applications/blob/master/keras_applications/mobilenet.py

21.PolyNet

PolyNet: A Pursuit of Structural Diversity in Very Deep Networks

Xingcheng Zhang, Zhizhong Li, Chen Change Loy, Dahua Lin

pdf:https://arxiv.org/abs/1611.05725

code: official :

https://github.com/open-mmlab/polynet

22.DPN

Dual Path Networks

Yunpeng Chen, Jianan Li, Huaxin Xiao, Xiaojie Jin, Shuicheng Yan, Jiashi Feng

pdf:https://arxiv.org/abs/1707.01629

code: official :

https://github.com/cypw/DPNs

code: unoffical-keras :

https://github.com/titu1994/Keras-DualPathNetworks

code: unofficial-pytorch :

https://github.com/oyam/pytorch-DPNs

code: unofficial-pytorch :

https://github.com/rwightman/pytorch-dpn-pretrained

23.Block-QNN

Practical Block-wise Neural Network Architecture Generation

Zhao Zhong, Junjie Yan, Wei Wu, Jing Shao, Cheng-Lin Liu

pdf:https://arxiv.org/abs/1708.05552

24.CRU-Net

Sharing Residual Units Through Collective Tensor Factorization in Deep Neural Networks

Chen Yunpeng, Jin Xiaojie, Kang Bingyi, Feng Jiashi, Yan Shuicheng

pdf:https://arxiv.org/abs/1703.02180

code official :

https://github.com/cypw/CRU-Net

code unofficial-mxnet :

https://github.com/bruinxiong/Modified-CRUNet-and-Residual-Attention-Network.mxnet

25.ShuffleNet

ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices

Xiangyu Zhang, Xinyu Zhou, Mengxiao Lin, Jian Sun

pdf:https://arxiv.org/abs/1707.01083

code: unofficial-tensorflow :

https://github.com/MG2033/ShuffleNet

code: unofficial-pytorch :

https://github.com/jaxony/ShuffleNet

code: unofficial-caffe :

https://github.com/farmingyard/ShuffleNet

code: unofficial-keras :

https://github.com/scheckmedia/keras-shufflenet

26.CondenseNet

CondenseNet: An Efficient DenseNet using Learned Group Convolutions

Gao Huang, Shichen Liu, Laurens van der Maaten, Kilian Q. Weinberger

pdf:https://arxiv.org/abs/1711.09224

code: official :

https://github.com/ShichenLiu/CondenseNet

code: unofficial-tensorflow :

https://github.com/markdtw/condensenet-tensorflow

27.NasNet

Learning Transferable Architectures for Scalable Image Recognition

Barret Zoph, Vijay Vasudevan, Jonathon Shlens, Quoc V. Le

pdf:https://arxiv.org/abs/1707.07012

code: unofficial-keras :

https://github.com/titu1994/Keras-NASNet

code: keras-applications :

https://github.com/keras-team/keras-applications/blob/master/keras_applications/nasnet.py

code: unofficial-pytorch :

https://github.com/wandering007/nasnet-pytorch

code: unofficial-tensorflow :

https://github.com/yeephycho/nasnet-tensorflow

28.MobileNetV2

MobileNetV2: Inverted Residuals and Linear Bottlenecks

Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, Liang-Chieh Chen

pdf:https://arxiv.org/abs/1801.04381

code: unofficial-keras :

https://github.com/xiaochus/MobileNetV2

code: unofficial-pytorch :

https://github.com/Randl/MobileNetV2-pytorch

code: unofficial-tensorflow :

https://github.com/neuleaf/MobileNetV2

29.IGCV2

IGCV2: Interleaved Structured Sparse Convolutional Neural Networks

Guotian Xie, Jingdong Wang, Ting Zhang, Jianhuang Lai, Richang Hong, Guo-Jun Qi

pdf:https://arxiv.org/abs/1804.06202

30.hier

Hierarchical Representations for Efficient Architecture Search

Hanxiao Liu, Karen Simonyan, Oriol Vinyals, Chrisantha Fernando, Koray Kavukcuoglu

pdf:https://arxiv.org/abs/1711.00436

31.PNasNet

Progressive Neural Architecture Search

Chenxi Liu, Barret Zoph, Maxim Neumann, Jonathon Shlens, Wei Hua, Li-Jia Li, Li Fei-Fei, Alan Yuille, Jonathan Huang, Kevin Murphy

pdf:https://arxiv.org/abs/1712.00559

code: tensorflow-slim :

https://github.com/tensorflow/models/blob/master/research/slim/nets/nasnet/pnasnet.py

code: unofficial-pytorch :

https://github.com/chenxi116/PNASNet.pytorch

code: unofficial-tensorflow :

https://github.com/chenxi116/PNASNet.TF

32.AmoebaNet

Regularized Evolution for Image Classifier Architecture Search

Esteban Real, Alok Aggarwal, Yanping Huang, Quoc V Le

pdf:https://arxiv.org/abs/1802.01548

code: tensorflow-tpu :

https://github.com/tensorflow/tpu/tree/master/models/official/amoeba_net

33.SENet

Squeeze-and-Excitation Networks

Jie Hu, Li Shen, Samuel Albanie, Gang Sun, Enhua Wu

pdf:https://arxiv.org/abs/1709.01507

code: official :

https://github.com/hujie-frank/SENet

code: unofficial-pytorch :

https://github.com/moskomule/senet.pytorch

code: unofficial-tensorflow :

https://github.com/taki0112/SENet-Tensorflow

code: unofficial-caffe :

https://github.com/shicai/SENet-Caffe

code: unofficial-mxnet :

https://github.com/bruinxiong/SENet.mxnet

34.ShuffleNetV2

ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design

Ningning Ma, Xiangyu Zhang, Hai-Tao Zheng, Jian Sun

pdf:https://arxiv.org/abs/1807.11164

code: unofficial-pytorch :

https://github.com/Randl/ShuffleNetV2-pytorch

code: unofficial-keras :

https://github.com/opconty/keras-shufflenetV2

code: unofficial-pytorch :

https://github.com/Bugdragon/ShuffleNet_v2_PyTorch

code: unofficial-caff2:

https://github.com/wolegechu/ShuffleNetV2.Caffe2

35.IGCV3

IGCV3: Interleaved Low-Rank Group Convolutions for Efficient Deep Neural Networks

Ke Sun, Mingjie Li, Dong Liu, Jingdong Wang

pdf:https://arxiv.org/abs/1806.00178

code: official :

https://github.com/homles11/IGCV3

code: unofficial-pytorch :

https://github.com/xxradon/IGCV3-pytorch

code: unofficial-tensorflow :

https://github.com/ZHANG-SHI-CHANG/IGCV3

36.MNasNet

MnasNet: Platform-Aware Neural Architecture Search for Mobile

Mingxing Tan, Bo Chen, Ruoming Pang, Vijay Vasudevan, Quoc V. Le

pdf:https://arxiv.org/abs/1807.11626

code: unofficial-pytorch :

https://github.com/AnjieZheng/MnasNet-PyTorch

code: unofficial-caffe :

https://github.com/LiJianfei06/MnasNet-caffe

code: unofficial-MxNet :

https://github.com/chinakook/Mnasnet.MXNet

code: unofficial-keras :

https://github.com/Shathe/MNasNet-Keras-Tensorflow

責(zé)任編輯:lq

聲明:本文內(nèi)容及配圖由入駐作者撰寫或者入駐合作網(wǎng)站授權(quán)轉(zhuǎn)載。文章觀點(diǎn)僅代表作者本人,不代表電子發(fā)燒友網(wǎng)立場(chǎng)。文章及其配圖僅供工程師學(xué)習(xí)之用,如有內(nèi)容侵權(quán)或者其他違規(guī)問題,請(qǐng)聯(lián)系本站處理。 舉報(bào)投訴
  • CV
    CV
    +關(guān)注

    關(guān)注

    0

    文章

    54

    瀏覽量

    17615
  • 圖像分類
    +關(guān)注

    關(guān)注

    0

    文章

    96

    瀏覽量

    12489
  • 計(jì)算機(jī)視覺
    +關(guān)注

    關(guān)注

    9

    文章

    1715

    瀏覽量

    47631

原文標(biāo)題:?CV 圖像分類常見的 36 個(gè)模型匯總!附完整論文和代碼

文章出處:【微信號(hào):vision263com,微信公眾號(hào):新機(jī)器視覺】歡迎添加關(guān)注!文章轉(zhuǎn)載請(qǐng)注明出處。

收藏 人收藏
加入交流群
微信小助手二維碼

掃碼添加小助手

加入工程師交流群

    評(píng)論

    相關(guān)推薦
    熱點(diǎn)推薦

    圖像采集卡分類、接口與性能指標(biāo)詳解

    圖像采集卡是機(jī)器視覺、工業(yè)檢測(cè)、安防監(jiān)控、醫(yī)療影像等領(lǐng)域的核心硬件,核心作用是將相機(jī)等圖像源輸出的模擬/數(shù)字信號(hào)轉(zhuǎn)化為計(jì)算機(jī)可識(shí)別、處理的數(shù)
    的頭像 發(fā)表于 01-29 16:55 ?468次閱讀
    <b class='flag-5'>圖像</b>采集卡<b class='flag-5'>分類</b>、接口與性能指標(biāo)詳解

    【上海晶珩睿莓 1 單板計(jì)算機(jī)】人臉識(shí)別

    計(jì)算機(jī)視覺庫,廣泛應(yīng)用于圖像處理、視頻分析和機(jī)器學(xué)習(xí)等領(lǐng)域。 為了避免影響系統(tǒng) Python,采用虛擬環(huán)境的方案。 創(chuàng)建并激活虛擬環(huán)境 mkdir ~/
    發(fā)表于 01-04 20:22

    上海計(jì)算機(jī)視覺企業(yè)行學(xué)術(shù)沙龍走進(jìn)西井科技

    12月5日,由中國圖象圖形學(xué)學(xué)會(huì)青年工作委員會(huì)(下簡稱“青工委”)、上海市計(jì)算機(jī)學(xué)會(huì)計(jì)算機(jī)視覺專委會(huì)(下簡稱“專委會(huì)”)聯(lián)合主辦,上海西井科技股份有限公司、江蘇路街道商會(huì)承辦的“上海計(jì)算機(jī)
    的頭像 發(fā)表于 12-16 15:39 ?585次閱讀

    Amphenol Rugged SOM:緊湊型嵌入式計(jì)算機(jī)模塊的卓越之選

    Amphenol Rugged SOM:緊湊型嵌入式計(jì)算機(jī)模塊的卓越之選 在嵌入式計(jì)算機(jī)模塊領(lǐng)域,Amphenol Rugged SOM 以其獨(dú)特的設(shè)計(jì)和強(qiáng)大的功能脫穎而出。今天,我
    的頭像 發(fā)表于 12-10 14:15 ?375次閱讀

    圖像采集卡:機(jī)器視覺時(shí)代的圖像數(shù)據(jù)核心樞紐

    的采集、轉(zhuǎn)換、傳輸與初步處理,搭建起圖像傳感器(相機(jī)/攝像頭)與計(jì)算機(jī)之間的高速數(shù)據(jù)鏈路。其核心價(jià)值在于解決“原始圖像信號(hào)如何高效、精準(zhǔn)、穩(wěn)定地轉(zhuǎn)化為計(jì)算機(jī)可處理數(shù)據(jù)
    的頭像 發(fā)表于 11-12 15:15 ?635次閱讀
    <b class='flag-5'>圖像</b>采集卡:機(jī)器<b class='flag-5'>視覺</b>時(shí)代的<b class='flag-5'>圖像</b>數(shù)據(jù)核心樞紐

    STM32計(jì)算機(jī)視覺開發(fā)套件:B-CAMS-IMX攝像頭模塊技術(shù)解析

    CMOS RGB圖像傳感器、ISM330DLC慣性運(yùn)動(dòng)單元和VL53L5CX ToF傳感器。B-CAMS-IMX可用于任何具有MIPI CSI-2? 接口(帶22引腳FFC連接器)的STM32開發(fā)板,輕松在STM32微控制器和微處理器上實(shí)現(xiàn)全功能計(jì)算機(jī)
    的頭像 發(fā)表于 10-20 09:46 ?1266次閱讀
    STM32<b class='flag-5'>計(jì)算機(jī)</b><b class='flag-5'>視覺</b>開發(fā)套件:B-CAMS-IMX攝像頭模塊技術(shù)解析

    什么是機(jī)器視覺及其功能?

    機(jī)器視覺本質(zhì)上是一個(gè)系統(tǒng)(例如一臺(tái)計(jì)算機(jī))查看的能力。系統(tǒng)通過該能力分析圖像,然后作出決策或進(jìn)行分類
    的頭像 發(fā)表于 09-10 17:23 ?754次閱讀
    什么是機(jī)器<b class='flag-5'>視覺</b>及其功能?

    【作品合集】賽昉科技VisionFive 2單板計(jì)算機(jī)開發(fā)板測(cè)評(píng)

    【VisionFive 2單板計(jì)算機(jī)試用體驗(yàn)】人臉識(shí)別1——交叉編譯環(huán)境搭建 作者:voidpbq【VisionFive 2單板計(jì)算機(jī)試用體驗(yàn)】1. 環(huán)境搭建與代碼整編【VisionFive 2單板
    發(fā)表于 09-04 09:08

    易控智駕榮獲計(jì)算機(jī)視覺頂會(huì)CVPR 2025認(rèn)可

    近日,2025年國際計(jì)算機(jī)視覺與模式識(shí)別頂級(jí)會(huì)議(IEEE/CVF Conference on Computer Vision and Pattern Recognition,CVPR 2025)在美國田納西州納什維爾召開。
    的頭像 發(fā)表于 07-29 16:54 ?1220次閱讀

    基于LockAI視覺識(shí)別模塊:手寫數(shù)字識(shí)別

    手寫數(shù)字識(shí)別是一種經(jīng)典的模式識(shí)別和圖像處理問題,旨在通過計(jì)算機(jī)自動(dòng)識(shí)別用戶手寫的數(shù)字。 本文將教會(huì)你如何使用基于RV1106的 LockAI視覺識(shí)別模塊進(jìn)行手寫數(shù)字識(shí)別。 源代碼:ht
    發(fā)表于 06-30 16:45

    工業(yè)相機(jī)圖像采集卡:機(jī)器視覺的核心樞紐

    工業(yè)相機(jī)圖像采集卡是用于連接工業(yè)相機(jī)與計(jì)算機(jī)的關(guān)鍵硬件設(shè)備,主要負(fù)責(zé)將相機(jī)輸出的圖像信號(hào)轉(zhuǎn)換為計(jì)算機(jī)可處理的數(shù)字信號(hào),并實(shí)現(xiàn)高速、穩(wěn)定的數(shù)據(jù)傳輸。它在工業(yè)自動(dòng)化、機(jī)器
    的頭像 發(fā)表于 05-21 12:13 ?858次閱讀
    工業(yè)相機(jī)<b class='flag-5'>圖像</b>采集卡:機(jī)器<b class='flag-5'>視覺</b>的核心樞紐

    基于LockAI視覺識(shí)別模塊:C++使用圖像的統(tǒng)計(jì)信息

    ./Test-Image-information-statistics 5.2 運(yùn)行效果 在運(yùn)行上述代碼時(shí),會(huì)輸出以下結(jié)果: 6. 總結(jié) 通過上述內(nèi)容,我們介紹了如何使用 OpenCV 提取圖像的 ROI、轉(zhuǎn)換顏色空間、
    發(fā)表于 05-08 10:31

    基于LockAI視覺識(shí)別模塊:C++圖像的基本運(yùn)算

    圖像處理中,理解圖像的基本操作是掌握計(jì)算機(jī)視覺技術(shù)的關(guān)鍵。本文章將介紹 基于LockAI視覺識(shí)別模塊下OpenCV 中
    發(fā)表于 05-06 16:56

    基于LockAI視覺識(shí)別模塊:C++圖像采集例程

    Vision Library)是一個(gè)開源的計(jì)算機(jī)視覺庫,提供豐富的圖像處理和視頻捕獲功能。通過其VideoCapture類,開發(fā)者可以輕松調(diào)用攝像頭設(shè)備并獲取視頻流。 1.2 VideoCapture
    發(fā)表于 04-30 10:52

    英飛凌邊緣AI平臺(tái)通過Ultralytics YOLO模型增加對(duì)計(jì)算機(jī)視覺的支持

    對(duì)計(jì)算機(jī)視覺的支持,擴(kuò)大了當(dāng)前對(duì)音頻、雷達(dá)和其他時(shí)間序列信號(hào)數(shù)據(jù)的支持范圍。在增加這項(xiàng)支持后,該平臺(tái)將能夠用于開發(fā)低功耗、低內(nèi)存的邊緣AI視覺模型。這將給諸多應(yīng)用領(lǐng)域的機(jī)器學(xué)習(xí)開發(fā)人員
    的頭像 發(fā)表于 03-11 15:11 ?814次閱讀
    英飛凌邊緣AI平臺(tái)通過Ultralytics YOLO模型增加對(duì)<b class='flag-5'>計(jì)算機(jī)</b><b class='flag-5'>視覺</b>的支持