Comments (6)
- The process of choosing the detected consumer clothing item is to choose a detected clothing item among all the detected clothing items in an image to represent one ground truth consumer clothing item for retrieval.
- For each detected consumer clothing item, its feature will be fed into matching network to retrieve commerical clothing gallery item. But only the retrieved result of the chosen detected consumer clothing item in step 1 will be evaluated.
from deepfashion2.
@geyuying Thanks for the response!
I got a couple more questions:
- Specifically what is the evaluation of step 1 based on? Is it based on the quote below? And so if the result of the proposed method in the below quote is positive, does that mean that the evaluation of step 1 will be positive?
If IoU between retrieved item from shop images and one of the ground truth corresponding gallery item is over the thresh(we set thresh as 0.5), the retrieved result is positive.
- I still don't understand fully how the gallery item is retrieved? The matching network only outputs a similarity score, it doesn't retrieve anything. So do you compare the similarity score of ALL gallery items with the chosen detected consumer clothing item, and choose the gallery item with the highest similarity score?
- Also where are the gallery items in the training dataset? Do you have to construct them? And do the gallery set contain ALL the commerical clothing items?
- Since you used the top-k accuracy metric for clothing retrieval, does that mean that you have to choose k gallery items with the highest similarity score for a given chosen detected consumer clothing item?
from deepfashion2.
I am also re-implementing the Match-Net. From my perspective:
- Specifically what is the evaluation of step 1 based on? Is it based on the quote below? And so if the result of the proposed method in the below quote is positive, does that mean that the evaluation of step 1 will be positive?
I think you can refer to the retrieval test script.
- I still don't understand fully how the gallery item is retrieved? The matching network only outputs a similarity score, it doesn't retrieve anything. So do you compare the similarity score of ALL gallery items with the chosen detected consumer clothing item, and choose the gallery item with the highest similarity score?
I think yes. You need to compare the similarity score of ALL (detected) gallery items with the chosen detected consumer clothing item.
- Also where are the gallery items in the training dataset? Do you have to construct them? And do the gallery set contain ALL the commerical clothing items?
I think gallery items are the images with source = shop. You can construct the image pairs (and instance pairs) between user and shop images (instances). I think the gallery set contains all the commercial clothing items.
- Since you used the top-k accuracy metric for clothing retrieval, does that mean that you have to choose k gallery items with the highest similarity score for a given chosen detected consumer clothing item?
Yes, I think so.
Still, I hope that the author can help to confirm the correctness of my answers.
from deepfashion2.
Cool, thanks for the answers!
I got a question if you can answer it:
During evaluation time, when performing the commercial clothing retrieval, the author states in issue #14 the following:
If IoU between retrieved item from shop images and one of the ground truth corresponding gallery item is over the thresh(we set thresh as 0.5), the retrieved result is positive.(If none detected item is assigned the given query item label, this query item is counted as missed. )
My question is do you know how to get this ground truth corresponding gallery item? If I understand correctly, the only thing the model is doing at the clothing retrieval task is:
- Retrieve a query item (detected consumer clothing item)
- Use the retrieved query item to get a gallery item (commercial clothing item with the highest similarity score).
- And then to evaluate the result, you want to compare this retrieved gallery item with the ground truth corresponding gallery item.
But the ground truth corresponding gallery item is not given? You have to somehow find the ground truth corresponding gallery item to compare it with the retrieved gallery item?
from deepfashion2.
Well, in the retrieval test script, the evaluation pipeline is like this:
-
Calculate the IoU between all pred_query(s) and gt_query(s). For each pred_query, the gt_query with the highest IoU is assigned as corresponding gt_query.
-
For each gt_query, find all the pred_query(s) which has the same corresponding gt_query. Then choose the one from the found pred_query(s) which has the same class and the highest score. Thus, we can build a (gt_query, pred_query) pair for all gt_query(s).
-
For each (gt_query, pred_query) pair above, we can know the pred_gallery(s) from the results given the pred_query. Besides, we can also obtain the style and pair_id given the gt_query. Using the style and pair_id, we can find all gt_gallery(s) that have the same style and pair_id.
-
For top-k evaluation, we choose top-k pred_gallery(s). Next, if any one of the top-k pred_gallery(s) has IoU >= 0.5 with any one of the gt_gallery(s) in the previous step, we will consider that for current gt_query, we can find a gt_gallery that matches. It will contribute to a positive match. On the contrary, if all top-k pred_gallery(s) have IoU < 0.5 with all the gt_gallery(s) in the previous step, we will say that for current gt_query, we CANNOT find a gt_gallery that matches. It will contribute to a negative match.
-
top-k accuracy = #positive matches / (#positive matches + #negative matches) = #positive matches / #gt_query(s)
from deepfashion2.
Okay, I'll look more closely into that, but thank you very much for the elaborated answer! Very much appreciated!
from deepfashion2.
Related Issues (20)
- Need Models to build Advance solution on top of this
- Interested in using this dataset for a larger project HOT 1
- Which tool did you use to annotate the image? HOT 3
- How to train own model and annotations for this? HOT 2
- Clarification on keypoints handling in Match R-CNN
- What is the standard for outer?
- How to convert anno files to mask images HOT 1
- Consumer to Shop Retrieval Results
- how to get bounding box annotations using deep fashion dataset? HOT 2
- detect clothing color? HOT 2
- Deepfashion train or validation images were used in Deepfashion 2 test dataset?
- Invalid password for unzipping the dataset HOT 1
- Signed field in Google form
- Error while convert deepfashion2 to coco format dataset
- install dependencies
- How to convert anno files to mask images HOT 1
- Problems with training
- Demography of Dataset Subjects
- What about shoes?
- I don't know the password to unzip the file.
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from deepfashion2.