Giter Club home page Giter Club logo

awesome-backbones's Introduction

Awesome backbones for image classification

BILIBILI

GitHub forks GitHub stars

写在前面

  • 若训练效果不佳,首先需要调整学习率和Batch size,这俩超参很大程度上影响收敛。其次,从关闭图像增强手段(尤其小数据集)开始,有的图像增强方法会污染数据,如

  如何去除增强?如efficientnetv2-b0配置文件中train_pipeline可更改为如下

train_pipeline = [
    dict(type='LoadImageFromFile'),
    dict(
        type='RandomResizedCrop',
        size=192,
        efficientnet_style=True,
        interpolation='bicubic'),
    dict(type='Normalize', **img_norm_cfg),
    dict(type='ImageToTensor', keys=['img']),
    dict(type='ToTensor', keys=['gt_label']),
    dict(type='Collect', keys=['img', 'gt_label'])
]

  若你的数据集提前已经将shape更改为网络要求的尺寸,那么Resize操作也可以去除。

更新日志

2023.12.02

  • 新增Issue中多人提及的输出Train AccVal loss

    • metrics_outputs.csv保存每周期train_loss, train_acc, train_precision, train_recall, train_f1-score, val_loss, val_acc, val_precision, val_recall, val_f1-score方便各位绘图
    • 终端由原先仅输出Val相关metrics升级为Train与Val都输出

2023.08.05

  • 新增TinyViT(预训练权重不匹配)、DeiT3EdgeNeXtRevVisionTransformer

2023.03.07

  • 新增MobileViTDaViTRepLKNetBEiTEVAMixMIMEfficientNetV2

2022.11.20

  • 新增是否将测试集用作验证集选项,若不使用,从训练集按ratio划分验证集数量,随机从训练集某fold挑选作为验证集(类似k-fold但不是,可自己稍改达到k-fold目的),详见Training tutorial

2022.11.06

  • 新增HorNet, EfficientFormer, SwinTransformer V2, MViT模型

测试环境

  • Pytorch 1.7.1+
  • Python 3.6+

资料

数据集 视频教程 人工智能技术探讨群
花卉数据集 提取码:0zat 点我跳转 1群:78174903
3群:584723646

快速开始

python tools/single_test.py datas/cat-dog.png models/mobilenet/mobilenet_v3_small.py --classes-map datas/imageNet1kAnnotation.txt

教程

模型

预训练权重

名称 权重 名称 权重 名称 权重
LeNet5 None AlexNet None VGG VGG-11
VGG-13
VGG-16
VGG-19
VGG-11-BN
VGG-13-BN
VGG-16-BN
VGG-19-BN
ResNet ResNet-18
ResNet-34
ResNet-50
ResNet-101
ResNet-152
ResNetV1C ResNetV1C-50
ResNetV1C-101
ResNetV1C-152
ResNetV1D ResNetV1D-50
ResNetV1D-101
ResNetV1D-152
ResNeXt ResNeXt-50
ResNeXt-101
ResNeXt-152
SEResNet SEResNet-50
SEResNet-101
SEResNeXt None
RegNet RegNetX-400MF
RegNetX-800MF
RegNetX-1.6GF
RegNetX-3.2GF
RegNetX-4.0GF
RegNetX-6.4GF
RegNetX-8.0GF
RegNetX-12GF
MobileNetV2 MobileNetV2 MobileNetV3 MobileNetV3-Small
MobileNetV3-Large
ShuffleNetV1 ShuffleNetV1 ShuffleNetV2 ShuffleNetV2 EfficientNet EfficientNet-B0
EfficientNet-B1
EfficientNet-B2
EfficientNet-B3
EfficientNet-B4
EfficientNet-B5
EfficientNet-B6
EfficientNet-B7
EfficientNet-B8
RepVGG RepVGG-A0
RepVGG-A1
RepVGG-A2
RepVGG-B0
RepVGG-B1
RepVGG-A1
RepVGG-B1g2
RepVGG-B1g4
RepVGG-B2
RepVGG-B2g4
RepVGG-B2g4
RepVGG-B3
RepVGG-B3g4
RepVGG-D2se
Res2Net Res2Net-50-14w-8s
Res2Net-50-26w-8s
Res2Net-101-26w-4s
ConvNeXt ConvNeXt-Tiny
ConvNeXt-Small
ConvNeXt-Base
ConvNeXt-Large
ConvNeXt-XLarge
HRNet HRNet-W18
HRNet-W30
HRNet-W32
HRNet-W40
HRNet-W44
HRNet-W48
HRNet-W64
ConvMixer ConvMixer-768/32
ConvMixer-1024/20
ConvMixer-1536/20
CSPNet CSPDarkNet50
CSPResNet50
CSPResNeXt50
Swin Transformer tiny-224
small-224
base-224
large-224
base-384
large-384
Vision Transformer vit_base_p16_224
vit_base_p32_224
vit_large_p16_224
vit_base_p16_384
vit_base_p32_384
vit_large_p16_384
Transformer in Transformer TNT-small
MLP Mixer base_p16
large_p16
Deit DeiT-tiny
DeiT-tiny distilled
DeiT-small
DeiT-small distilled
DeiT-base
DeiT-base distilled
DeiT-base 384px
DeiT-base distilled 384px
Conformer Conformer-tiny-p16
Conformer-small-p32
Conformer-small-p16
Conformer-base-p16
T2T-ViT T2T-ViT_t-14
T2T-ViT_t-19
T2T-ViT_t-24
Twins PCPVT-small
PCPVT-base
PCPVT-large
SVT-small
SVT-base
SVT-large
PoolFormer PoolFormer-S12
PoolFormer-S24
PoolFormer-S36
PoolFormer-M36
PoolFormer-M48
DenseNet DenseNet121
DenseNet161
DenseNet169
DenseNet201
Visual Attention Network(VAN) VAN-Tiny
VAN-Small
VAN-Base
VAN-Large
Wide-ResNet WRN-50
WRN-101
HorNet HorNet-Tiny
HorNet-Tiny-GF
HorNet-Small
HorNet-Small-GF
HorNet-Base
HorNet-Base-GF
HorNet-Large
HorNet-Large-GF
HorNet-Large-GF384
EfficientFormer efficientformer-l1
efficientformer-l3
efficientformer-l7
Swin Transformer v2 tiny-256 window 8
tiny-256 window 16
small-256 window 8
small-256 window 16
base-256 window 8
base-256 window 16
large-256 window 16
large-384 window 24
MViTv2 MViTv2-Tiny
MViTv2-Small
MViTv2-Base
MViTv2-Large
MobileVit MobileViT-XXSmall
MobileViT-XSmall
MobileViT-Small
DaViT DaViT-T
DaViT-S
DaViT-B
RepLKNet RepLKNet-31B-224
RepLKNet-31B-384
RepLKNet-31L-384
RepLKNet-XL
BEiT BEiT-base EVA EVA-G-p14-224
EVA-G-p14-336
EVA-G-p14-560
EVA-G-p16-224
EVA-L-p14-224
EVA-L-p14-196
EVA-L-p14-336
MixMIM mixmim-base EfficientNetV2 EfficientNetV2-b0
EfficientNetV2-b1
EfficientNetV2-b2
EfficientNetV2-b3
EfficientNetV2-s
EfficientNetV2-m
EfficientNetV2-l
EfficientNetV2-xl
DeiT3 deit3_small_p16
deit3_small_p16_384
deit3_base_p16
deit3_base_p16_384
deit3_medium_p16
deit3_large_p16
deit3_large_p16_384
deit3_huge_p16
EdgeNeXt edgenext-base
edgenext-small
edgenext-X-small
edgenext-XX-small
RevVisionTransformer revvit-small
revvit-base

我维护的其他项目

参考

@repo{2020mmclassification,
    title={OpenMMLab's Image Classification Toolbox and Benchmark},
    author={MMClassification Contributors},
    howpublished = {\url{https://github.com/open-mmlab/mmclassification}},
    year={2020}
}

awesome-backbones's People

Contributors

fafa-dl avatar psy-lzh avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

awesome-backbones's Issues

训练得到的pth文件,想在C++环境下调用模型,libtorch要求先转化模型,在使用torch.jit.trace(model, example)时报错

报错如下:
(torch) E:\python\Awesome-Backbones>python tools/pth2pt.py models\repvgg\repvgg_A0.py
Loading Train_Epoch042-Loss0.012.pth

Traceback (most recent call last):
File "tools/pth2pt.py", line 125, in
main()
File "tools/pth2pt.py", line 121, in main
traced_script_module = torch.jit.trace(model, example)
File "D:\anaconda3\envs\torch\lib\site-packages\torch\jit_trace.py", line 742, in trace
_module_class,
File "D:\anaconda3\envs\torch\lib\site-packages\torch\jit_trace.py", line 940, in trace_module
_force_outplace,
File "D:\anaconda3\envs\torch\lib\site-packages\torch\nn\modules\module.py", line 725, in _call_impl
result = self._slow_forward(*input, **kwargs)
File "D:\anaconda3\envs\torch\lib\site-packages\torch\nn\modules\module.py", line 709, in _slow_forward
result = self.forward(*input, **kwargs)
File "E:\python\Awesome-Backbones\models\build.py", line 125, in forward
return self.forward_train(x,**kwargs)
TypeError: forward_train() missing 1 required positional argument: 'targets'

请问博主有做这方面的工作不,期待您的回复

在单机多显卡环境下,中断训练后resume报错,

Traceback (most recent call last):
File "tools/train1.py", line 178, in
main()
File "tools/train1.py", line 172, in main
train(model,runner, lr_update_func, device, epoch, data_cfg.get('train').get('epoches'), meta)
File "/hy-tmp/Awe/utils/train_utils.py", line 215, in train
losses = model(images,targets=targets,return_loss=True)
File "/usr/local/lib/python3.8/dist-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/hy-tmp/Awe/models/build.py", line 125, in forward
return self.forward_train(x,**kwargs)
File "/hy-tmp/Awe/models/build.py", line 130, in forward_train
x = self.extract_feat(x)
File "/hy-tmp/Awe/models/build.py", line 113, in extract_feat
x = self.backbone(img)
File "/usr/local/lib/python3.8/dist-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/hy-tmp/Awe/configs/backbones/swin_transformer.py", line 409, in forward
x, hw_shape = self.patch_embed(x)
File "/usr/local/lib/python3.8/dist-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/hy-tmp/Awe/configs/common/transformer.py", line 221, in forward
x = self.projection(x)
File "/usr/local/lib/python3.8/dist-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/usr/local/lib/python3.8/dist-packages/torch/nn/modules/conv.py", line 423, in forward
return self._conv_forward(input, self.weight)
File "/usr/local/lib/python3.8/dist-packages/torch/nn/modules/conv.py", line 419, in _conv_forward
return F.conv2d(input, weight, self.bias, self.stride,
RuntimeError: Input type (torch.cuda.FloatTensor) and weight type (torch.FloatTensor) should be the same


图片
加model.to(device),后报另一个错误
图片

mobilevit模型评估的结果不一致问题

在用mobilevit网络进行评估时,在配置文件中已经把路径给修改成训练过程中保存的val集acc最高的那个模型权重文件了,窗口也显示加载的是这个模型,但是评估的结果和保存的结果不一致,并且多次执行python tools/evaluation.py models/mobilevit/mobilevit_s.py时每次结果都不一样,按理说模型确定,test集确定,每次的预测结果应该都一样的,可以看到从模型名称看到结果应该是84%才对,找不到是哪个步骤出了问题

QQ图片20230329125953

模型评估

请问可以得到模型的推理速度,如对多张模型的推理耗时ms和fps吗?希望可以添加

训练利用率问题

image
我导,这个训练利用率有问题吗,上一回跑swin transformer v2的时候没有出现偶尔是0的现象

训练时的acc-loss图像

请问这个实时图像的创建在哪一个文件中 (想知道横纵坐标,颜色等详细信息)

预训练权重使用

您好,我想问一下存储库中提供的预训练权重是在imagenet上预训练得到的吗?

swintransformerv2预训练权重似乎加载不进去

+Model info---------+----------------------+---------------+-----------------+
| Backbone | Neck | Head | Loss |
+-------------------+----------------------+---------------+-----------------+
| SwinTransformerV2 | GlobalAveragePooling | LinearClsHead | LabelSmoothLoss |
+-------------------+----------------------+---------------+-----------------+
Initialize the weights.
Loading swinv2_base_patch4_window8_256.pth
The model and loaded state dict do not match exactly

unexpected key in source state_dict: model

missing keys in source state_dict: backbone.patch_embed.projection.weight, backbone.patch_embed.projection.bias, backbone.patch_embed.norm.weight, backbone.patch_embed.norm.bias, backbone.stages.0.blocks.0.attn.w_msa.logit_scale, backbone.stages.0.blocks.0.attn.w_msa.q_bias, backbone.stages.0.blocks.0.attn.w_msa.v_bias, backbone.stages.0.blocks.0.attn.w_msa.relative_coords_table, backbone.stages.0.blocks.0.attn.w_msa.relative_position_index, backbone.stages.0.blocks.0.attn.w_msa.cpb_mlp.0.weight, backbone.stages.0.blocks.0.attn.w_msa.cpb_mlp.0.bias, backbone.stages.0.blocks.0.attn.w_msa.cpb_mlp.2.weight, backbone.stages.0.blocks.0.attn.w_msa.qkv.weight, backbone.stages.0.blocks.0.attn.w_msa.proj.weight, backbone.stages.0.blocks.0.attn.w_msa.proj.bias, backbone.stages.0.blocks.0.norm1.weight, backbone.stages.0.blocks.0.norm1.bias, backbone.stages.0.blocks.0.ffn.layers.0.0.weight, backbone.stages.0.blocks.0.ffn.layers.0.0.bias, backbone.stages.0.blocks.0.ffn.layers.1.weight, backbone.stages.0.blocks.0.ffn.layers.1.bias, backbone.stages.0.blocks.0.norm2.weight, backbone.stages.0.blocks.0.norm2.bias, backbone.stages.0.blocks.1.attn.w_msa.logit_scale, backbone.stages.0.blocks.1.attn.w_msa.q_bias, backbone.stages.0.blocks.1.attn.w_msa.v_bias, backbone.stages.0.blocks.1.attn.w_msa.relative_coords_table, backbone.stages.0.blocks.1.attn.w_msa.relative_position_index, backbone.stages.0.blocks.1.attn.w_msa.cpb_mlp.0.weight, backbone.stages.0.blocks.1.attn.w_msa.cpb_mlp.0.bias, backbone.stages.0.blocks.1.attn.w_msa.cpb_mlp.2.weight, backbone.stages.0.blocks.1.attn.w_msa.qkv.weight, backbone.stages.0.blocks.1.attn.w_msa.proj.weight, backbone.stages.0.blocks.1.attn.w_msa.proj.bias, backbone.stages.0.blocks.1.norm1.weight, backbone.stages.0.blocks.1.norm1.bias, backbone.stages.0.blocks.1.ffn.layers.0.0.weight, backbone.stages.0.blocks.1.ffn.layers.0.0.bias, backbone.stages.0.blocks.1.ffn.layers.1.weight, backbone.stages.0.blocks.1.ffn.layers.1.bias, backbone.stages.0.blocks.1.norm2.weight, backbone.stages.0.blocks.1.norm2.bias, backbone.stages.1.downsample.reduction.weight, backbone.stages.1.downsample.norm.weight, backbone.stages.1.downsample.norm.bias, backbone.stages.1.blocks.0.attn.w_msa.logit_scale, backbone.stages.1.blocks.0.attn.w_msa.q_bias, backbone.stages.1.blocks.0.attn.w_msa.v_bias, backbone.stages.1.blocks.0.attn.w_msa.relative_coords_table, backbone.stages.1.blocks.0.attn.w_msa.relative_position_index, backbone.stages.1.blocks.0.attn.w_msa.cpb_mlp.0.weight, backbone.stages.1.blocks.0.attn.w_msa.cpb_mlp.0.bias, backbone.stages.1.blocks.0.attn.w_msa.cpb_mlp.2.weight, backbone.stages.1.blocks.0.attn.w_msa.qkv.weight, backbone.stages.1.blocks.0.attn.w_msa.proj.weight, backbone.stages.1.blocks.0.attn.w_msa.proj.bias, backbone.stages.1.blocks.0.norm1.weight, backbone.stages.1.blocks.0.norm1.bias, backbone.stages.1.blocks.0.ffn.layers.0.0.weight, backbone.stages.1.blocks.0.ffn.layers.0.0.bias, backbone.stages.1.blocks.0.ffn.layers.1.weight, backbone.stages.1.blocks.0.ffn.layers.1.bias, backbone.stages.1.blocks.0.norm2.weight, backbone.stages.1.blocks.0.norm2.bias, backbone.stages.1.blocks.1.attn.w_msa.logit_scale, backbone.stages.1.blocks.1.attn.w_msa.q_bias, backbone.stages.1.blocks.1.attn.w_msa.v_bias, backbone.stages.1.blocks.1.attn.w_msa.relative_coords_table, backbone.stages.1.blocks.1.attn.w_msa.relative_position_index, backbone.stages.1.blocks.1.attn.w_msa.cpb_mlp.0.weight, backbone.stages.1.blocks.1.attn.w_msa.cpb_mlp.0.bias, backbone.stages.1.blocks.1.attn.w_msa.cpb_mlp.2.weight, backbone.stages.1.blocks.1.attn.w_msa.qkv.weight, backbone.stages.1.blocks.1.attn.w_msa.proj.weight, backbone.stages.1.blocks.1.attn.w_msa.proj.bias, backbone.stages.1.blocks.1.norm1.weight, backbone.stages.1.blocks.1.norm1.bias, backbone.stages.1.blocks.1.ffn.layers.0.0.weight, backbone.stages.1.blocks.1.ffn.layers.0.0.bias, backbone.stages.1.blocks.1.ffn.layers.1.weight, backbone.stages.1.blocks.1.ffn.layers.1.bias, backbone.stages.1.blocks.1.norm2.weight, backbone.stages.1.blocks.1.norm2.bias, backbone.stages.2.downsample.reduction.weight, backbone.stages.2.downsample.norm.weight, backbone.stages.2.downsample.norm.bias, backbone.stages.2.blocks.0.attn.w_msa.logit_scale, backbone.stages.2.blocks.0.attn.w_msa.q_bias, backbone.stages.2.blocks.0.attn.w_msa.v_bias, backbone.stages.2.blocks.0.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.0.attn.w_msa.relative_position_index, backbone.stages.2.blocks.0.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.0.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.0.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.0.attn.w_msa.qkv.weight, backbone.stages.2.blocks.0.attn.w_msa.proj.weight, backbone.stages.2.blocks.0.attn.w_msa.proj.bias, backbone.stages.2.blocks.0.norm1.weight, backbone.stages.2.blocks.0.norm1.bias, backbone.stages.2.blocks.0.ffn.layers.0.0.weight, backbone.stages.2.blocks.0.ffn.layers.0.0.bias, backbone.stages.2.blocks.0.ffn.layers.1.weight, backbone.stages.2.blocks.0.ffn.layers.1.bias, backbone.stages.2.blocks.0.norm2.weight, backbone.stages.2.blocks.0.norm2.bias, backbone.stages.2.blocks.1.attn.w_msa.logit_scale, backbone.stages.2.blocks.1.attn.w_msa.q_bias, backbone.stages.2.blocks.1.attn.w_msa.v_bias, backbone.stages.2.blocks.1.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.1.attn.w_msa.relative_position_index, backbone.stages.2.blocks.1.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.1.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.1.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.1.attn.w_msa.qkv.weight, backbone.stages.2.blocks.1.attn.w_msa.proj.weight, backbone.stages.2.blocks.1.attn.w_msa.proj.bias, backbone.stages.2.blocks.1.norm1.weight, backbone.stages.2.blocks.1.norm1.bias, backbone.stages.2.blocks.1.ffn.layers.0.0.weight, backbone.stages.2.blocks.1.ffn.layers.0.0.bias, backbone.stages.2.blocks.1.ffn.layers.1.weight, backbone.stages.2.blocks.1.ffn.layers.1.bias, backbone.stages.2.blocks.1.norm2.weight, backbone.stages.2.blocks.1.norm2.bias, backbone.stages.2.blocks.2.attn.w_msa.logit_scale, backbone.stages.2.blocks.2.attn.w_msa.q_bias, backbone.stages.2.blocks.2.attn.w_msa.v_bias, backbone.stages.2.blocks.2.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.2.attn.w_msa.relative_position_index, backbone.stages.2.blocks.2.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.2.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.2.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.2.attn.w_msa.qkv.weight, backbone.stages.2.blocks.2.attn.w_msa.proj.weight, backbone.stages.2.blocks.2.attn.w_msa.proj.bias, backbone.stages.2.blocks.2.norm1.weight, backbone.stages.2.blocks.2.norm1.bias, backbone.stages.2.blocks.2.ffn.layers.0.0.weight, backbone.stages.2.blocks.2.ffn.layers.0.0.bias, backbone.stages.2.blocks.2.ffn.layers.1.weight, backbone.stages.2.blocks.2.ffn.layers.1.bias, backbone.stages.2.blocks.2.norm2.weight, backbone.stages.2.blocks.2.norm2.bias, backbone.stages.2.blocks.3.attn.w_msa.logit_scale, backbone.stages.2.blocks.3.attn.w_msa.q_bias, backbone.stages.2.blocks.3.attn.w_msa.v_bias, backbone.stages.2.blocks.3.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.3.attn.w_msa.relative_position_index, backbone.stages.2.blocks.3.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.3.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.3.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.3.attn.w_msa.qkv.weight, backbone.stages.2.blocks.3.attn.w_msa.proj.weight, backbone.stages.2.blocks.3.attn.w_msa.proj.bias, backbone.stages.2.blocks.3.norm1.weight, backbone.stages.2.blocks.3.norm1.bias, backbone.stages.2.blocks.3.ffn.layers.0.0.weight, backbone.stages.2.blocks.3.ffn.layers.0.0.bias, backbone.stages.2.blocks.3.ffn.layers.1.weight, backbone.stages.2.blocks.3.ffn.layers.1.bias, backbone.stages.2.blocks.3.norm2.weight, backbone.stages.2.blocks.3.norm2.bias, backbone.stages.2.blocks.4.attn.w_msa.logit_scale, backbone.stages.2.blocks.4.attn.w_msa.q_bias, backbone.stages.2.blocks.4.attn.w_msa.v_bias, backbone.stages.2.blocks.4.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.4.attn.w_msa.relative_position_index, backbone.stages.2.blocks.4.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.4.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.4.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.4.attn.w_msa.qkv.weight, backbone.stages.2.blocks.4.attn.w_msa.proj.weight, backbone.stages.2.blocks.4.attn.w_msa.proj.bias, backbone.stages.2.blocks.4.norm1.weight, backbone.stages.2.blocks.4.norm1.bias, backbone.stages.2.blocks.4.ffn.layers.0.0.weight, backbone.stages.2.blocks.4.ffn.layers.0.0.bias, backbone.stages.2.blocks.4.ffn.layers.1.weight, backbone.stages.2.blocks.4.ffn.layers.1.bias, backbone.stages.2.blocks.4.norm2.weight, backbone.stages.2.blocks.4.norm2.bias, backbone.stages.2.blocks.5.attn.w_msa.logit_scale, backbone.stages.2.blocks.5.attn.w_msa.q_bias, backbone.stages.2.blocks.5.attn.w_msa.v_bias, backbone.stages.2.blocks.5.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.5.attn.w_msa.relative_position_index, backbone.stages.2.blocks.5.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.5.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.5.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.5.attn.w_msa.qkv.weight, backbone.stages.2.blocks.5.attn.w_msa.proj.weight, backbone.stages.2.blocks.5.attn.w_msa.proj.bias, backbone.stages.2.blocks.5.norm1.weight, backbone.stages.2.blocks.5.norm1.bias, backbone.stages.2.blocks.5.ffn.layers.0.0.weight, backbone.stages.2.blocks.5.ffn.layers.0.0.bias, backbone.stages.2.blocks.5.ffn.layers.1.weight, backbone.stages.2.blocks.5.ffn.layers.1.bias, backbone.stages.2.blocks.5.norm2.weight, backbone.stages.2.blocks.5.norm2.bias, backbone.stages.2.blocks.6.attn.w_msa.logit_scale, backbone.stages.2.blocks.6.attn.w_msa.q_bias, backbone.stages.2.blocks.6.attn.w_msa.v_bias, backbone.stages.2.blocks.6.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.6.attn.w_msa.relative_position_index, backbone.stages.2.blocks.6.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.6.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.6.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.6.attn.w_msa.qkv.weight, backbone.stages.2.blocks.6.attn.w_msa.proj.weight, backbone.stages.2.blocks.6.attn.w_msa.proj.bias, backbone.stages.2.blocks.6.norm1.weight, backbone.stages.2.blocks.6.norm1.bias, backbone.stages.2.blocks.6.ffn.layers.0.0.weight, backbone.stages.2.blocks.6.ffn.layers.0.0.bias, backbone.stages.2.blocks.6.ffn.layers.1.weight, backbone.stages.2.blocks.6.ffn.layers.1.bias, backbone.stages.2.blocks.6.norm2.weight, backbone.stages.2.blocks.6.norm2.bias, backbone.stages.2.blocks.7.attn.w_msa.logit_scale, backbone.stages.2.blocks.7.attn.w_msa.q_bias, backbone.stages.2.blocks.7.attn.w_msa.v_bias, backbone.stages.2.blocks.7.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.7.attn.w_msa.relative_position_index, backbone.stages.2.blocks.7.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.7.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.7.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.7.attn.w_msa.qkv.weight, backbone.stages.2.blocks.7.attn.w_msa.proj.weight, backbone.stages.2.blocks.7.attn.w_msa.proj.bias, backbone.stages.2.blocks.7.norm1.weight, backbone.stages.2.blocks.7.norm1.bias, backbone.stages.2.blocks.7.ffn.layers.0.0.weight, backbone.stages.2.blocks.7.ffn.layers.0.0.bias, backbone.stages.2.blocks.7.ffn.layers.1.weight, backbone.stages.2.blocks.7.ffn.layers.1.bias, backbone.stages.2.blocks.7.norm2.weight, backbone.stages.2.blocks.7.norm2.bias, backbone.stages.2.blocks.8.attn.w_msa.logit_scale, backbone.stages.2.blocks.8.attn.w_msa.q_bias, backbone.stages.2.blocks.8.attn.w_msa.v_bias, backbone.stages.2.blocks.8.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.8.attn.w_msa.relative_position_index, backbone.stages.2.blocks.8.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.8.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.8.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.8.attn.w_msa.qkv.weight, backbone.stages.2.blocks.8.attn.w_msa.proj.weight, backbone.stages.2.blocks.8.attn.w_msa.proj.bias, backbone.stages.2.blocks.8.norm1.weight, backbone.stages.2.blocks.8.norm1.bias, backbone.stages.2.blocks.8.ffn.layers.0.0.weight, backbone.stages.2.blocks.8.ffn.layers.0.0.bias, backbone.stages.2.blocks.8.ffn.layers.1.weight, backbone.stages.2.blocks.8.ffn.layers.1.bias, backbone.stages.2.blocks.8.norm2.weight, backbone.stages.2.blocks.8.norm2.bias, backbone.stages.2.blocks.9.attn.w_msa.logit_scale, backbone.stages.2.blocks.9.attn.w_msa.q_bias, backbone.stages.2.blocks.9.attn.w_msa.v_bias, backbone.stages.2.blocks.9.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.9.attn.w_msa.relative_position_index, backbone.stages.2.blocks.9.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.9.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.9.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.9.attn.w_msa.qkv.weight, backbone.stages.2.blocks.9.attn.w_msa.proj.weight, backbone.stages.2.blocks.9.attn.w_msa.proj.bias, backbone.stages.2.blocks.9.norm1.weight, backbone.stages.2.blocks.9.norm1.bias, backbone.stages.2.blocks.9.ffn.layers.0.0.weight, backbone.stages.2.blocks.9.ffn.layers.0.0.bias, backbone.stages.2.blocks.9.ffn.layers.1.weight, backbone.stages.2.blocks.9.ffn.layers.1.bias, backbone.stages.2.blocks.9.norm2.weight, backbone.stages.2.blocks.9.norm2.bias, backbone.stages.2.blocks.10.attn.w_msa.logit_scale, backbone.stages.2.blocks.10.attn.w_msa.q_bias, backbone.stages.2.blocks.10.attn.w_msa.v_bias, backbone.stages.2.blocks.10.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.10.attn.w_msa.relative_position_index, backbone.stages.2.blocks.10.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.10.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.10.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.10.attn.w_msa.qkv.weight, backbone.stages.2.blocks.10.attn.w_msa.proj.weight, backbone.stages.2.blocks.10.attn.w_msa.proj.bias, backbone.stages.2.blocks.10.norm1.weight, backbone.stages.2.blocks.10.norm1.bias, backbone.stages.2.blocks.10.ffn.layers.0.0.weight, backbone.stages.2.blocks.10.ffn.layers.0.0.bias, backbone.stages.2.blocks.10.ffn.layers.1.weight, backbone.stages.2.blocks.10.ffn.layers.1.bias, backbone.stages.2.blocks.10.norm2.weight, backbone.stages.2.blocks.10.norm2.bias, backbone.stages.2.blocks.11.attn.w_msa.logit_scale, backbone.stages.2.blocks.11.attn.w_msa.q_bias, backbone.stages.2.blocks.11.attn.w_msa.v_bias, backbone.stages.2.blocks.11.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.11.attn.w_msa.relative_position_index, backbone.stages.2.blocks.11.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.11.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.11.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.11.attn.w_msa.qkv.weight, backbone.stages.2.blocks.11.attn.w_msa.proj.weight, backbone.stages.2.blocks.11.attn.w_msa.proj.bias, backbone.stages.2.blocks.11.norm1.weight, backbone.stages.2.blocks.11.norm1.bias, backbone.stages.2.blocks.11.ffn.layers.0.0.weight, backbone.stages.2.blocks.11.ffn.layers.0.0.bias, backbone.stages.2.blocks.11.ffn.layers.1.weight, backbone.stages.2.blocks.11.ffn.layers.1.bias, backbone.stages.2.blocks.11.norm2.weight, backbone.stages.2.blocks.11.norm2.bias, backbone.stages.2.blocks.12.attn.w_msa.logit_scale, backbone.stages.2.blocks.12.attn.w_msa.q_bias, backbone.stages.2.blocks.12.attn.w_msa.v_bias, backbone.stages.2.blocks.12.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.12.attn.w_msa.relative_position_index, backbone.stages.2.blocks.12.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.12.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.12.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.12.attn.w_msa.qkv.weight, backbone.stages.2.blocks.12.attn.w_msa.proj.weight, backbone.stages.2.blocks.12.attn.w_msa.proj.bias, backbone.stages.2.blocks.12.norm1.weight, backbone.stages.2.blocks.12.norm1.bias, backbone.stages.2.blocks.12.ffn.layers.0.0.weight, backbone.stages.2.blocks.12.ffn.layers.0.0.bias, backbone.stages.2.blocks.12.ffn.layers.1.weight, backbone.stages.2.blocks.12.ffn.layers.1.bias, backbone.stages.2.blocks.12.norm2.weight, backbone.stages.2.blocks.12.norm2.bias, backbone.stages.2.blocks.13.attn.w_msa.logit_scale, backbone.stages.2.blocks.13.attn.w_msa.q_bias, backbone.stages.2.blocks.13.attn.w_msa.v_bias, backbone.stages.2.blocks.13.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.13.attn.w_msa.relative_position_index, backbone.stages.2.blocks.13.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.13.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.13.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.13.attn.w_msa.qkv.weight, backbone.stages.2.blocks.13.attn.w_msa.proj.weight, backbone.stages.2.blocks.13.attn.w_msa.proj.bias, backbone.stages.2.blocks.13.norm1.weight, backbone.stages.2.blocks.13.norm1.bias, backbone.stages.2.blocks.13.ffn.layers.0.0.weight, backbone.stages.2.blocks.13.ffn.layers.0.0.bias, backbone.stages.2.blocks.13.ffn.layers.1.weight, backbone.stages.2.blocks.13.ffn.layers.1.bias, backbone.stages.2.blocks.13.norm2.weight, backbone.stages.2.blocks.13.norm2.bias, backbone.stages.2.blocks.14.attn.w_msa.logit_scale, backbone.stages.2.blocks.14.attn.w_msa.q_bias, backbone.stages.2.blocks.14.attn.w_msa.v_bias, backbone.stages.2.blocks.14.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.14.attn.w_msa.relative_position_index, backbone.stages.2.blocks.14.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.14.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.14.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.14.attn.w_msa.qkv.weight, backbone.stages.2.blocks.14.attn.w_msa.proj.weight, backbone.stages.2.blocks.14.attn.w_msa.proj.bias, backbone.stages.2.blocks.14.norm1.weight, backbone.stages.2.blocks.14.norm1.bias, backbone.stages.2.blocks.14.ffn.layers.0.0.weight, backbone.stages.2.blocks.14.ffn.layers.0.0.bias, backbone.stages.2.blocks.14.ffn.layers.1.weight, backbone.stages.2.blocks.14.ffn.layers.1.bias, backbone.stages.2.blocks.14.norm2.weight, backbone.stages.2.blocks.14.norm2.bias, backbone.stages.2.blocks.15.attn.w_msa.logit_scale, backbone.stages.2.blocks.15.attn.w_msa.q_bias, backbone.stages.2.blocks.15.attn.w_msa.v_bias, backbone.stages.2.blocks.15.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.15.attn.w_msa.relative_position_index, backbone.stages.2.blocks.15.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.15.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.15.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.15.attn.w_msa.qkv.weight, backbone.stages.2.blocks.15.attn.w_msa.proj.weight, backbone.stages.2.blocks.15.attn.w_msa.proj.bias, backbone.stages.2.blocks.15.norm1.weight, backbone.stages.2.blocks.15.norm1.bias, backbone.stages.2.blocks.15.ffn.layers.0.0.weight, backbone.stages.2.blocks.15.ffn.layers.0.0.bias, backbone.stages.2.blocks.15.ffn.layers.1.weight, backbone.stages.2.blocks.15.ffn.layers.1.bias, backbone.stages.2.blocks.15.norm2.weight, backbone.stages.2.blocks.15.norm2.bias, backbone.stages.2.blocks.16.attn.w_msa.logit_scale, backbone.stages.2.blocks.16.attn.w_msa.q_bias, backbone.stages.2.blocks.16.attn.w_msa.v_bias, backbone.stages.2.blocks.16.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.16.attn.w_msa.relative_position_index, backbone.stages.2.blocks.16.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.16.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.16.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.16.attn.w_msa.qkv.weight, backbone.stages.2.blocks.16.attn.w_msa.proj.weight, backbone.stages.2.blocks.16.attn.w_msa.proj.bias, backbone.stages.2.blocks.16.norm1.weight, backbone.stages.2.blocks.16.norm1.bias, backbone.stages.2.blocks.16.ffn.layers.0.0.weight, backbone.stages.2.blocks.16.ffn.layers.0.0.bias, backbone.stages.2.blocks.16.ffn.layers.1.weight, backbone.stages.2.blocks.16.ffn.layers.1.bias, backbone.stages.2.blocks.16.norm2.weight, backbone.stages.2.blocks.16.norm2.bias, backbone.stages.2.blocks.17.attn.w_msa.logit_scale, backbone.stages.2.blocks.17.attn.w_msa.q_bias, backbone.stages.2.blocks.17.attn.w_msa.v_bias, backbone.stages.2.blocks.17.attn.w_msa.relative_coords_table, backbone.stages.2.blocks.17.attn.w_msa.relative_position_index, backbone.stages.2.blocks.17.attn.w_msa.cpb_mlp.0.weight, backbone.stages.2.blocks.17.attn.w_msa.cpb_mlp.0.bias, backbone.stages.2.blocks.17.attn.w_msa.cpb_mlp.2.weight, backbone.stages.2.blocks.17.attn.w_msa.qkv.weight, backbone.stages.2.blocks.17.attn.w_msa.proj.weight, backbone.stages.2.blocks.17.attn.w_msa.proj.bias, backbone.stages.2.blocks.17.norm1.weight, backbone.stages.2.blocks.17.norm1.bias, backbone.stages.2.blocks.17.ffn.layers.0.0.weight, backbone.stages.2.blocks.17.ffn.layers.0.0.bias, backbone.stages.2.blocks.17.ffn.layers.1.weight, backbone.stages.2.blocks.17.ffn.layers.1.bias, backbone.stages.2.blocks.17.norm2.weight, backbone.stages.2.blocks.17.norm2.bias, backbone.stages.3.downsample.reduction.weight, backbone.stages.3.downsample.norm.weight, backbone.stages.3.downsample.norm.bias, backbone.stages.3.blocks.0.attn.w_msa.logit_scale, backbone.stages.3.blocks.0.attn.w_msa.q_bias, backbone.stages.3.blocks.0.attn.w_msa.v_bias, backbone.stages.3.blocks.0.attn.w_msa.relative_coords_table, backbone.stages.3.blocks.0.attn.w_msa.relative_position_index, backbone.stages.3.blocks.0.attn.w_msa.cpb_mlp.0.weight, backbone.stages.3.blocks.0.attn.w_msa.cpb_mlp.0.bias, backbone.stages.3.blocks.0.attn.w_msa.cpb_mlp.2.weight, backbone.stages.3.blocks.0.attn.w_msa.qkv.weight, backbone.stages.3.blocks.0.attn.w_msa.proj.weight, backbone.stages.3.blocks.0.attn.w_msa.proj.bias, backbone.stages.3.blocks.0.norm1.weight, backbone.stages.3.blocks.0.norm1.bias, backbone.stages.3.blocks.0.ffn.layers.0.0.weight, backbone.stages.3.blocks.0.ffn.layers.0.0.bias, backbone.stages.3.blocks.0.ffn.layers.1.weight, backbone.stages.3.blocks.0.ffn.layers.1.bias, backbone.stages.3.blocks.0.norm2.weight, backbone.stages.3.blocks.0.norm2.bias, backbone.stages.3.blocks.1.attn.w_msa.logit_scale, backbone.stages.3.blocks.1.attn.w_msa.q_bias, backbone.stages.3.blocks.1.attn.w_msa.v_bias, backbone.stages.3.blocks.1.attn.w_msa.relative_coords_table, backbone.stages.3.blocks.1.attn.w_msa.relative_position_index, backbone.stages.3.blocks.1.attn.w_msa.cpb_mlp.0.weight, backbone.stages.3.blocks.1.attn.w_msa.cpb_mlp.0.bias, backbone.stages.3.blocks.1.attn.w_msa.cpb_mlp.2.weight, backbone.stages.3.blocks.1.attn.w_msa.qkv.weight, backbone.stages.3.blocks.1.attn.w_msa.proj.weight, backbone.stages.3.blocks.1.attn.w_msa.proj.bias, backbone.stages.3.blocks.1.norm1.weight, backbone.stages.3.blocks.1.norm1.bias, backbone.stages.3.blocks.1.ffn.layers.0.0.weight, backbone.stages.3.blocks.1.ffn.layers.0.0.bias, backbone.stages.3.blocks.1.ffn.layers.1.weight, backbone.stages.3.blocks.1.ffn.layers.1.bias, backbone.stages.3.blocks.1.norm2.weight, backbone.stages.3.blocks.1.norm2.bias, backbone.norm3.weight, backbone.norm3.bias, head.fc.weight, head.fc.bias
/home/user/miniconda3/envs/wmh/lib/python3.8/site-packages/torch/functional.py:445: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at ../aten/src/ATen/native/TensorShape.cpp:2157.)
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]

功能完善建议

1)可以考虑把每个epoch训练的loss,精度,验证的精度和loss等相关信息存放在一个CSV文件中,这样可以方便之后画图
2)可以考虑把每次训练的控制台输出同时添加到某个日志文件中
只是友善的建议,博主可以视情况采纳

Transformer相关网络Loss=nan的问题

最近在训练自己的数据集时,发现几个Transformer网络有时会出现loss=nan的情况,并且一旦出现这个情况,训练结果就不会发生任何的update.

想知道loss=nan会在怎样的情况下发生?谢谢

请问模型图片输入大小如何修改

作者您好,最近在使用您的这个项目在测试实验,请问如何在哪里可以修改输入图片大小,我的图片是自制的比如39、189。模型的默认输入是224224。直接调用也没问题,但是好像是resize图片把图片变成了224224,我想使用原本图片的大小作为输入

反馈一个潜在小Bug,在使用您的图像分类程序时,如果annoations填写顺序反向的话,在测试时可能会输出错误,如:high为2在第一行,low为0但在第三行,就会在单张图片测试时出现label和class对不上的bug

为了修改这个小bug,我修改了utils中的inference中部分语句以及单张图片测试的部分代码,目前已解决该bug

修改处为:
inference代码中新增label_names
def inference_model(model, image, val_pipeline, classes_names,label_names):

以及result['pred_class'] = classes_names[label_names.index(result['pred_label'])]

单张图片测试代码中修改了
classes_names, label_names = get_info(args.classes_map)

result = inference_model(model, args.img, val_pipeline, classes_names,label_names)

经测试修改后的代码已不会出现annoation顺序影响测试结果显示的小bug

多标签分类

up, 关于多标签分类 ,我将如何处理自己的数据集

not registry

(pytorch-gpu) C:\ai-project\Awesome-Backbones-main>python tools/evaluation.py models/shufflenet/shufflenet_v2.py
tools/evaluation.py:1: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
import imp
Loading Val_Epoch001-Acc100.000.pth

Traceback (most recent call last):
File "tools/evaluation.py", line 169, in
main()
File "tools/evaluation.py", line 146, in main
test_dataset = Mydataset(test_datas, val_pipeline)
File "C:\ai-project\Awesome-Backbones-main\utils\dataloader.py", line 16, in init
self.pipeline = Compose(self.cfg)
File "C:\ai-project\Awesome-Backbones-main\core\datasets\compose.py", line 23, in init
transform = build_from_cfg(transform, PIPELINES)
File "C:\ai-project\Awesome-Backbones-main\core\datasets\build.py", line 61, in build_from_cfg
f'{obj_type} is not in the {registry.name} registry')
KeyError: 'RandomHorizontalFlip is not in the pipeline registry'

我想把Top5修改成Top2的

可以在你模型里修改吗,具体是怎么修改呢?在head上把topk=(1, 5)变为topk=(1,2)吗?这样也不对

EfficientNetV2 is not working

Hello!

First of all, I would like to thank you for great job! But it seems that EfficientNetV2 don't work. I tried this:
python tools/train.py models/efficientnetv2/efficientnetv2_xl_512.py
And received an error:
"AttributeError: 'EnhancedConvModule' object has no attribute 'in_channels'"
Could you fix it, please?

如何将head中的分类问题转化为回归问题

感谢您提供的代码库,我目前的任务是关于含量多少的分类问题,如何通过修改head输出,使其能够成为解决回归问题的head,比如说我现在建立了三个梯度1,10,100分类模型,那我怎样预测5或者50的梯度,需要怎样在您的基础上修改

单张图像检测报错

Loading Train_Epoch077-Loss0.423.pth <All keys matched successfully> Traceback (most recent call last): File "tools/single_test.py", line 42, in <module> main() File "tools/single_test.py", line 36, in main result = inference_model(model, args.img, val_pipeline, classes_names) File "/home/Awesome-Backbones/utils/inference.py", line 48, in inference_model if val_pipeline[0]['type'] != 'LoadImageFromFile': TypeError: 'NoneType' object is not subscriptable

运行 python tools/evaluation.py,能正常显示,测试 python tools/single_test.py 就报错了... 打印 val_pipeline,是None,想问有什么原因导致发生的呢?谢谢

数据集多类别问题

你好 想请教一下这种已经分好类的数据集我想合并类别需要怎么修改 比如说把car bus truck都合并成car
QQ截图20230322173943

功能建议:输出feature.npy

可以增加输出backbone提取的feature的功能,并保存npy文件到指定文件夹。这样可以方便做分类降维可视化。还有就是可以提供打印网络结构的功能,最后再次谢谢大佬提供的框架!

使用感受➕一些小建议

这个库非常棒,我在前几天使用你的库完成了一个小任务,我在resnet的基础上又增加了一个head,在这部分做了些修改(包括一些中间件),这块我看原本是只支持更换head,但是没有做对增加的支持,不知道是不是我没有仔细看,这块是否需要完善一下。还有single test部分,我看只对单一的图片做预测,这块是否需要实现对一个batch的支持,并且是否需要将val_pipeline这部分拆分出来,这样可以提前对数据进行处理。以上是我的使用感受,欢迎探讨,最终感谢你的框架。

Initialize the weights. Loading mobilenet_v3_small-8427ecf0.pth The model and loaded state dict do not match exactly size mismatch for head.layers.1.fc.weight: copying a param with shape torch.Size([1000, 1024]) from checkpoint, the shape in current model is torch.Size([12, 1024]). size mismatch for head.layers.1.fc.bias: copying a param with shape torch.Size([1000]) from checkpoint, the shape in current model is torch.Size([12]).

您好,我制作了自己的数据集,有12个分类,但是为什么会报这个错误,迁移学习那里已经设置了true了

怎么进行多卡训练

目前的源码是只支持单卡训练,但是我想加快训练生成模型,如何进行多卡训练呢

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.