-
Notifications
You must be signed in to change notification settings - Fork 73
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Congratulations for greatly reproducing #77
Comments
Is search config used in obtain_search_args? if prev_layer is None:
prev_layer = net AutoML is as follows: if prev_prev_fmultiplier == -1 and j == 0:
op = None
So I email the author and the reply is as follows:
Yuan Fang, how do you see? |
Thanks, We will fix it just like the author reports! |
@HeathHose |
@mrluin for example, 8H1 is the first node in level 8 so that it don't have s0 and s1. So the author preprocess stem0 and stem1 in stride 2,4 as s0 and s1 respectively |
@mrluin hi, I've bounced back and forth about how is best to do this. The authors doing it this way doesn't automatically make it the best approach. I think there may be a better way. However, It wont be implemented soon, and if you think you can improve on our results please try to do so. Should you get better results I would love for you to make a pull request and become a contributer to the project. |
I'm getting on it and see how the result will be. I'll report the result here either it gets better or not. |
architecture search results: [1 1 2 2 2 3 3 2 1 2 3 3] This is the one I've added pre_pre_input for those edge tensors. Despite the mIoU slightly decreases, the stem has only two layers (to keep consistent with the paper). In addition, all downsampling are implemented with stride 2 and 4. (edited)As we can see that the search had not converged yet because the cell structures derived are all sep conv, not even one delated conv, which shouldn't be. But I think padding pre_pre_input for those edge tensors is necessary. Searching with a larger epochs number is a straightforward option. An alternative is to boost the robustness of the search, I recently found some interesting research on it: https://openreview.net/pdf?id=H1gDNyrKDS. |
@HankKung Hi, Thanks for your hard work!
Does the way you add pre_pre_input for those edge tensors like the following one? And how do you perform downsampling with stride4, also use the FactorizedReduce?
But after I read the official code of autodeeplab (derived model), I found that the stem has three conv_layers rather than two. And how do you implement the derived model, the same as the official code? (if pre_pre_input is None, consider it as a copy of the pre_input) Looking forward to your reply! |
About the stem, they used a two-layer one during the search and used a three-layer one during the weights training (retrain) as you mentioned. I haven't evaluated the performance of the derived model, but it's supposed the pre_pre_inputs always exist because the pre_pre_input is simply the last two node's output (e.g., stem2's output as level 1's pre_pre_input), no need for the same spatial resolution one. Glad to help! If you have any idea or questions, we are happy to discuss. |
Oh, that's very helpful!
Now, I think I got your idea of this repetition. Thank you very much! |
@HankKung Thank you for providing the clarification about the network. I am wondering what does each number mean in the output of cell structure? or how we can print(genotype) instead of decode?
|
hey,I have the same problem with you ,Have you found the answer? |
Thanks for your greate AutoML! Could you pls release the architect found in search?
The text was updated successfully, but these errors were encountered: