Giter Club home page Giter Club logo

yolov's People

Contributors

angstcj avatar yuhengsss avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar

yolov's Issues

训练过程报错

您好,请问一下为什么每个epoch训练都会报以下这个错误然后训练就被中断,而且每次都是训练到7020iter的时候。

2023-03-04 06:39:27.342 | INFO | yolox.core.vid_trainer:after_iter:279 - epoch: 1/7, iter: 7020/9366, mem: 8055Mb,`iter_time: 4.363s, data_time: 3.605s, total_loss: 1.1, iou_loss: 0.7, l1_loss: 0.0, conf_loss: 0.2, cls_loss: 0.1, lr: 2.247e-03, size: 480, ETA: 2 days, 21:48:48
2023-03-04 06:39:38.261 | INFO | yolox.core.vid_trainer:after_train:198 - Training of experiment is done and the best AP is 0.00
2023-03-04 06:39:38.262 | ERROR | yolox.core.launch:launch:98 - An error has been caught in function 'launch', process 'MainProcess' (267170), thread 'MainThread' (140410779206464):
Traceback (most recent call last):

File "tools/vid_train.py", line 151, in
args=(exp, args),
│ └ Namespace(batch_size=128, cache=False, ckpt='/media/user/A0F260D9F260B566/qsy/YOLOV/weights/yoloxs_vid.pth', devices=1, dist_...
└ ╒═══════════════════╤════════════════════════════════════════════════════════════════════════════════════════════════════════...

File "./yolox/core/launch.py", line 98, in launch
main_func(*args)
│ └ (╒═══════════════════╤═══════════════════════════════════════════════════════════════════════════════════════════════════════...
└ <function main at 0x7fb305add4d0>

File "tools/vid_train.py", line 128, in main
trainer.train()
│ └ <function Trainer.train at 0x7fb305ae15f0>
└ <yolox.core.vid_trainer.Trainer object at 0x7fb3ed55a990>

File "./yolox/core/vid_trainer.py", line 85, in train
self.train_in_epoch()
│ └ <function Trainer.train_in_epoch at 0x7fb305ae1b90>
└ <yolox.core.vid_trainer.Trainer object at 0x7fb3ed55a990>

File "./yolox/core/vid_trainer.py", line 94, in train_in_epoch
self.train_in_iter()
│ └ <function Trainer.train_in_iter at 0x7fb305ae1dd0>
└ <yolox.core.vid_trainer.Trainer object at 0x7fb3ed55a990>

File "./yolox/core/vid_trainer.py", line 100, in train_in_iter
self.train_one_iter()
│ └ <function Trainer.train_one_iter at 0x7fb305ae4d40>
└ <yolox.core.vid_trainer.Trainer object at 0x7fb3ed55a990>

File "./yolox/core/vid_trainer.py", line 107, in train_one_iter
inps = inps.to(self.data_type)
│ │ └ torch.float16
│ └ <yolox.core.vid_trainer.Trainer object at 0x7fb3ed55a990>
└ None

AttributeError: 'NoneType' object has no attribute 'to'

Train Error "IndexError: The shape of the mask [60] at index 0 does not match the shape of the indexed tensor [30, 30] at index 0"

Dear author, thank you very much for sharing your excellent research. It is very innovative and gets outstanding results. I'm trying to training yolov_s model with exp file yolov_s_online.py. But I encountered a problem "IndexError: The shape of the mask [60] at index 0 does not match the shape of the indexed tensor [30, 30] at index 0". Can you tell me how to fix it. Looking forward to your reply. Thank you very much.

Unable to load pretrained checkpoints for yolov_x and yolov_l

Dear Author,

Thanks for sharing your great work. I'm trying to train yolov_l and yolov_x, but it looks like it's unable to load the checkpoints for yoloxl_vid.pth and yoloxx_vid.pth provided in google drive.

Here's my training command
python tools/vid_train.py -f exps/yolov/yolov_l.py -c pretrained_weights/yoloxl_vid.pth --fp16
python tools/vid_train.py -f exps/yolov/yolov_x.py -c pretrained_weights/yoloxx_vid.pth --fp16

And I'm getting this error

image

My command below for training yolov_s is working fine. Is it possible that the large and extra large checkpoints got corrupted?
python tools/vid_train.py -f exps/yolov/yolov_s.py -c pretrained_weights/yoloxs_vid.pth --fp16

Thanks for the help!

what will happen if batch_size != gframe (lframe=0) ?

Dear auther, I found that if batch_size != gframe (e.g., batchsize=32, gframe=16 and lframe=0) which means one batch contains two groups (16 frames each x2), all the predictions of two groups of 32 frames will be used to conduct the MSA related processing. But the two groups of frames possibly come from two different videos, won't it have some problems ? If setting batch_size = gframe = a small value (e.g. 16), the problem does not exist, but the GPU Memory utilization will be very low. If setting batch_size = gframe = a big value (e.g. 64), that means making a decision during inference needs to consider 64 frames. So I don't know if I missed something ? How should I set batch_size and gframe\lframe on my 2x24GB gpus ?

why use only part of the data?

Hello, thx for your great work. I notice that you use only part of the data in dataloader.
image

does this mean “use 10% of VID”?short-time video may not appear in the dataset.

Custom data

Hi, it is unclear how I can format my custom dataset to your expected format, Can possibly provide information on how I can convert my dataset, and what are you expected format? For E.g. it is unclear to me what the context is of the video sequences.

Will the Argoverse dataset work?

thx for the great work, and I found the Argoverse dataset has been implemented in vid.py. So I want to know does this dataset work? Or should I make some minor revisions on it? Many thx for your reply ~

Online video inference

Dear authors, really impressive approach and great results! Thank you for publishing your work. First of all, is it possible to run online/real-time inference on a video (stream), in a way that the feature aggregation is done on the previous frames and inference is run on the last frame? Does the script ./tools/yolov_demo_online.py target this purpose? Thanks in advance!

How to start video training?

I've seen the docs but didn't find guidance on how to start video training. Should I start with vid_train.py?
And if I have a trained detector, what should I do then to continue the video training part ?

损失函数和预训练模型的选择问题

你好,我注意到 loss 设置为3reg_loss+2ref_loss+obj_loss+cls_loss,而YoloX的loss是5*reg_loss+obj_loss+cls_loss。这个参数3和2是经过实验得来的么?(我并没有看到论文中有提到)

另外YoloX在ImageNet Det数据上训练时使用到了COCO的预训练权重么?

Error(s) in loading state_dict for YOLOX:

Dear Author,

Thanks for sharing your work. I'm trying to run vid_demo.py and use yoloxx_vid.pth and yoloxs_vid.pth, but it looks like it's unable to load the checkpoints for yoloxs_vid.pth and yoloxx_vid.pth provided in google drive. And I'm getting this error:
issue01

can you tell me what's wrong?

Thanks for the help!

AttributeError: 'MosaicDetection_VID' object has no attribute 'dataset_pth'

!! Iteration after 3 epoch. !! Running for yolov_x stopped. !!

But, following the bottom structure could smoothly running for yolov_s.
pytorch 1.20, cuda 11.6.

The 'AttributeError' seems to related the path of dataset ? I am not sure.
But, running for yolov_s is OK.
Does anyone here could give a help ? Thank you so much.

TypeError: YOLOX.forward() got an unexpected keyword argument 'nms_thresh'

when trying to do the command in the readme python tools/vid_demo.py -f ./exps/yolov/yoloxs_vid.py -c ./yoloxs_vid.pth --path ./189_4_1_20230217042816000.tmp.mp4 --conf 0.25 --nms 0.5 --tsize 600 I get the following error:

File "/Users/joy/jtest/YOLOV/tools/vid_demo.py", line 303, in <module>
    main(exp, args)
  File "/Users/joy/jtest/YOLOV/tools/vid_demo.py", line 296, in main
    imageflow_demo(predictor, vis_folder, current_time, args)
  File "/Users/joy/jtest/YOLOV/tools/vid_demo.py", line 224, in imageflow_demo
    outputs.extend(predictor.inference(ele))
  File "/Users/joy/jtest/YOLOV/tools/vid_demo.py", line 137, in inference
    outputs,outputs_ori = self.model(img, nms_thresh=self.nmsthre)
  File "/Users/joy/jtest/YOLOV/.venv/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
    return forward_call(*args, **kwargs)
TypeError: YOLOX.forward() got an unexpected keyword argument 'nms_thresh'

Unable to train

When I try to run tools/train.py, I met a issue:

/root/hhh/YOLO-master
unknown host, exit

How can I figure out it?

the use of the json file?

In "README", it is mentioned that the two files "YOLOV/annotations/vid_train_coco.json" and "YOLOv/yolox/data/dataset/train_seq.npy" are required, but I only see train_seq.npy is used in the project, json What and Where is the use of the json file? I‘m looking for reply,thanks

About the model's random select frame input

Hi! I notice that the input of this model is randomly selected from the original video. May I know how you make use of the non-keyframe information? And is the result of this model the full result of the video(including key and non-key frames) or just keyframe results? Thanks!

Unable load .pth file

super(_open_zipfile_reader, self).__init__(torch._C.PyTorchFileReader(name_or_buffer))

RuntimeError: [enforce fail at inline_container.cc:145] . PytorchStreamReader failed reading zip archive: failed finding central directory

Is the best result from the first fine-turing epoch ?

Dear author, I've tried to train your yolov on my dataset and an obvious improvement was observed almost at the first epoch of fine-turing, but in later epoches, the performance never got better. If a big lr is used, the performace will get worse quickly from the first epoch. Did you meet similar situation during the fine-turing training on pretrained model ?

How to make sure the output features_cls meet the required size [b*50, 5, 128]?

Hi, I have the problem that if the post function finds nothing. Maybe the input is a background image

pred_result, pred_idx = self.postpro_woclass(decode_res, num_classes=30, nms_thre=self.nms_thresh,
topK=self.Afternum) # postprocess(decode_res,num_classes=30)

How to make sure the output feature meets the required shape?

features_cls, features_reg, cls_scores, fg_scores = self.find_feature_score(cls_feat_flatten, pred_idx,
reg_feat_flatten, imgs,
pred_result) # [b*50,5,128]

Pre-training with DET and VID dataset

Hey, I had a question about what training methodology was used to pre-train the yolox baseline models. I added the images from DET dataset with similar classes as the VID dataset and trained yolox-s model but I was not able to replicate your results. So could you elaborate on how you pretrained the yolox model?

questions about mAP and fps

您好,我想请教您以下两个问题:
1.在训练模型时输出的测试集mAP和单独运行vid_eval.py时得到的mAP不一致是怎么回事呢,而且发现两种操作生成的refined_pred.json文件的大小并不相同,我初步猜测是不是因为一些阈值的设置不同,导致产生的预测框数量不同,最终的测试结果不一样呢,应该以哪个为准呢
2.如果想要测试模型的FPS有相应的文件么,还是将Average forward time与Average inference time相加得到总共需要的时间
期待您的解答

BN freeze

Dear author, I noticed the following code:
# def fix_bn(m):
# classname = m.class.name
# if classname.find('BatchNorm') != -1:
# m.eval()
self.model.apply(init_yolo)
# self.model.apply(fix_bn)
In your experiments, is there big difference between self.model.apply(init_yolo) and self.model.apply(fix_bn) ?

AP50% and Times

Hi! which file did you calculate the AP and Times? I want to reproduce them and calculate other metrics.

Thanks a lot!

OVIS数据集如何评估?

您好,非常感谢您的工作!我有一些问题,我没找到评估OVIS数据集的方法,vid_eval.py 评估的加载的VID_dataset数据集的验证集。OVIS数据集如何评估?我是在自定义的数据集上实验是COCO格式标注,没办法评估模型的好坏?谢谢!

per frame annotations will be used in YOLOV?

great job! i have a question about the loss, all the annotations (reference frames and key frame) will be used for YOLOV? if that, i wanna only use information of references for key frame detection, Is it feasible?thx

无法复现结果,急需帮助

请问下可以告诉我们完整的训练命令吗,如何可以训练出87.5的结果呢,我torch版本和cuda已经保持了一致,但是无法复现结果,如果不是多节点训练的命令应该怎么实现呢?

Where is yolov7 bases

I saw above that yolov7 was added and didn't find the code for that section, can you tell me which section it is in please? Also, any thoughts on adding a detector for yolov6, v6 is a frameless detector.

Error in online video demo

Getting the following error while running the online demo script:

Traceback (most recent call last):
  File "tools/yolov_demo_online.py", line 320, in <module>
    main(exp, args)
  File "tools/yolov_demo_online.py", line 313, in main
    imageflow_demo(predictor, vis_folder, current_time, args)
  File "tools/yolov_demo_online.py", line 226, in imageflow_demo
    N = int(res_dict['cls_scores'].shape[0] / len(tmp_imgs))
TypeError: list indices must be integers or slices, not str

I have not changed anything in the script. Although I did check that pred_result, and res_dict are equal which are obtained from the inference step.

pred_result, res_dict = predictor.inference(imgs, other_result)

Can someone please guide me

ILSVRC2015 VID Data

Since the dataset is very large, can you provide a part of the video files in the datasets for testing the demo?

how FSM module works on ppyoloe?

hello,I have a question, I find that ppyoloe don't design obj prediction on the regression branch of the detection head, how the FSM module works? Only use class_conf to select?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.