-
Notifications
You must be signed in to change notification settings - Fork 7
[Re] Network Deconvolution #89
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Thanks for your submission. We'll assign an editor soon. |
By the way is this submision part of the ICLR reproducibility challenge? If yes, are there any open review somewhere? |
Thank You! |
No, our work was not submitted to ICLR reproducibility challenge. |
Very sorry for such a long delay, hopefully things will get better for 2025. I'll edit your review and assign reviewers soon hopefully. In the meantime, can you have a look at other submissions and propose yourself to review? |
@birdortyedi @MiWeiss Coud lyou review this submission? |
Thank you for the update! We understand how busy things can get. |
Sounds like an interesting paper and a good match for me. Unfortunately, though, I won't be able to review a paper in the next months. |
@ReScience/reviewers We're looking for two reviewers (Machine learnong / Python ICLR), any takers? |
@rochanaro Don't hesitate to post here to ask for update |
@rougier I would like to review this one. |
@alchemi5t Many thanks! You can start the review now and let's target mid-March (or sooner if you can) |
@rougier I am interested in reviewing this one if a second review is needed. |
@jsta That would be great, many thanks. Do you think mid-March would work for you? |
Yes |
@rougier I'm not getting responses to PR's and Issues in the code repo. Can we delay the due date pending a response from @rochanaro or team? |
@jsta Apologies for the delayed response. We encountered an issue with notification delivery for |
@alchemi5t @jsta Any progress on your reviews? |
@rougier I'll have my review by the end of the week. |
@rougier Yes, I have some general comments prepared. I will post them once I have finished a couple tests to verify general behavior. I do not plan to run all the tests due to computational constraints. |
Dear Authors, Here's my review of your work. Review of "[Re] Network Deconvolution"This reproduction study provides a thorough evaluation of the network deconvolution technique introduced by Ye et al. (2020). After examining both the original paper and this reproduction, I find that the authors have validated the primary claim about network deconvolution for most parts. While there are many cases where network deconvolution improves model performance compared to batch normalization, the reproduction results show few exceptions that contradict the original paper's universal claim. For example, in the ResNet-18 architecture with CIFAR-100 at 100 epochs, batch normalization (97.42%) actually outperformed network deconvolution (94.31%), which contradicts both the original paper's results and its central claim. Strengths
Key ObservationsThe few exceptions where BN seems to have outperformed has not been noted and there is no analysis around it. Furthermore, this statement is not true given the results: Another notable finding is that the reproduced accuracy values were often significantly higher than those reported in the original paper. For example:
The authors attribute this systematic improvement to:
While these explanations are plausible, the magnitude of improvement (sometimes exceeding 20%) suggests there might be additional factors at play that weren't fully investigated. Given that this improvement has increased most accuracy (both ND and BN) to 99.xx where the comparison comes down to 10e-2 (vs the 1-2 pts in the original paper), I expected deeper analysis and stronger evidence for these claims. CodeAs far as I understand, running training runs seems to be the only way to test the codebase. Due to lack of hardware I am unable to do so, but to anyone who can and would like to quickly verify the claims, releasing the trained weights at 100 epochs and a minimal script to infer the results would be greatly helpful. OpinionThis reproduction study provides valuable insights but reveals important discrepancies with the original paper's claims. While network deconvolution often outperforms batch normalization, the reproduction found notable exceptions. The reproduction yielded substantially higher accuracy values for both techniques compared to those reported in the original paper. These significant discrepancies make it difficult to draw direct comparisons with the original results, and the proposed explanations for these differences (library improvements, optimization algorithms) remain speculative without rigorous empirical validation. Rather than enhancing confidence in network deconvolution as a universal improvement over batch normalization, this reproduction suggests a more nuanced view: network deconvolution appears to be a viable alternative that performs better in many but not all scenarios. The authors' detailed reporting of computational costs and performance characteristics across architectures provides essential practical context for researchers considering which normalization technique to employ. |
Ok, I have completed my review. Please see the comments below. In addition to reading the paper and git repository, I verified that I could run 2 arbitrary architectures (vgg16 and pnasnetA) for 100 epochs for both the BN and ND cases, and got similar relative accuracy results. major comments
minor comments
highly optional and/or minor curiosity questions
|
@alchemi5t @jsta Many thanks for your detailed reviews. |
Thank you @alchemi5t and @jsta for the reviews. @rougier It would be challenging to redo the experiment using the exact same dependencies as the original paper. During our study, we attempted to contact the original authors to obtain the precise library versions, as the original study repository does not mention them, but we did not receive a response. Therefore, we adopted the following approach: we used the latest available versions of each library as of 2020 and only opted for more recent versions when compatibility issues arose. We will soon get back with answers, revisions, and updates addressing the above concerns in our upcoming response. |
Full response to ReScience C reviews (submission #89)Dear editor (@rougier) and the reviewers (@alchemi5t, @jsta) , We appreciate the reviewers’ careful evaluation of our reproduction study. Below is our detailed response addressing each of the reviewers’ points: Performance Discrepancies and Exceptional Cases
We appreciate the observation that batch normalization (BN) outperformed network deconvolution in certain cases (e.g., ResNet-18 on CIFAR-100 at 100 epochs), and we acknowledge these exceptions. Our study aimed to carefully reproduce the experiments rather than optimize hyperparameters or modify architectural choices. In the literature, such variations are not uncommon when replicating older experiments (see, e.g., Examining the Effect of Implementation Factors on Deep Learning Reproducibility, A Study on Reproducibility and Replicability of Table Structure Recognition Methods, Towards training reproducible deep learning models). We do not claim that network deconvolution universally outperforms BN; rather, our findings suggest that while ND often provides benefits, its effectiveness may be context dependent (explained in manuscript section 6 paragraph 4). Systematic Accuracy Improvements
We acknowledge the reviewer’s observation regarding systematically higher accuracy values (e.g., VGG-16 with CIFAR-100) and agree on the importance of understanding these differences. We believe that several factors contribute to these improvements. Notably, advancements in numerical stability, as documented in recent studies (e.g., Recent advances and applications of deep learning methods in materials science, A Comprehensive Review of Deep Learning: Architectures, Recent Advances, and Applications ), and enhancements in optimization algorithms may play a significant role. Additionally, prior research supports the idea that updated frameworks can improve model performance. Mienye et al. (2024), and Qin et al. 2024 and Vaishnav et al. (2022) demonstrated that deep learning models consistently achieve higher accuracy when trained with updated frameworks due to improvements in optimization techniques and regularization strategies. Similarly, Coakley et al. (2023) found that updated hardware and software environments alone can introduce accuracy differences of over 6%, highlighting the impact of implementation-level changes on reproducibility. While the observed accuracy gains (e.g., 75% → 99% for VGG-16 on CIFAR-100) are substantial, we emphasize that both network deconvolution (ND) and batch normalization (BN) baselines improved comparably, suggesting that these gains are driven by systemic factors rather than being specific to any particular method. Although we have provided plausible explanations based on updates in software libraries, we acknowledge that these factors may not fully account for the magnitude of the observed differences. Further investigation into these aspects is warranted, but we reiterate that the primary objective was to assess reproducibility rather than to deconstruct every underlying cause of performance variation. We provided detailed explanations in Section 6 Paragraphs 2 and 3 in the revised manuscript. Sharing trained weights for verification
We acknowledge the reviewer's (@alchemi5t) request and have taken steps to facilitate easy verification of our results. The trained model weights at 100 epochs have been made publicly available at https://osf.io/hp3ab/files/osfstorage , corresponding to the results presented in Tables 1 and 2. Additionally, we provide two minimal inference scripts ( The steps have been explained in the newly added section, For example, the following command can be used to verify results for the VGG-16 architecture on CIFAR-10 with batch normalization:
Clarification on Software Versions (Reviewer 2’s Comment @jsta)
In response to Reviewer 2’s remark regarding our statement in Section 1, we have revised the language. The original phrase “with the most recent versions of software libraries” has been updated to “with the most recent versions of software libraries as of 2020” to prevent any misunderstanding (Section 1 Page 3). During our study, we attempted to contact the original authors to obtain the precise library versions used in their experiments, as the repository did not include this information. Unfortunately, we did not receive a response. Therefore, we adopted the following approach: we used the latest available versions of each library as of 2020 and opted for more recent versions only when compatibility issues arose. The details are explained in Section 4.5 of the revised manuscript. Summary and a list of changes made based on the reviewers’ and editor’s comments*All the updates in the manuscript are highlighted in color blue. Reviewer 1 (@alchemi5t):
Reviewer 2 (@jsta):
Editor (@rougier):
Revised Manuscript URL: https://github.com/lamps-lab/rep-network-deconvolution/blob/master/article.pdf |
@rochanaro Thanks for your detailed answer. @jsta I guess your thumb means you're satisfied with the answer and the revised manuscript. @alchemi5t are you ok with the answer and revised manuscript? |
@rougier yes, I am satisfied with the response. |
1 similar comment
@rougier yes, I am satisfied with the response. |
Thank you @alchemi5t and @jsta, we appreciate the replies. |
Great, then we can accept your replication! Congratulations. For the actual publicaton, I would need a link, to a github repo with the sources of your articles. And you'll need to save you code repo on software heritage such as to get a SWID. |
Thank you, @rougier. The requested items are listed below. Please let us know if anything additional is needed. GitHub link for the article source files: https://github.com/lamps-lab/rep-network-deconvolution/tree/master/documents/manuscript swh:1:dir:680efae9c7c628f93ee03731e73c34cf00b0e245; |
Sorry I'm late in handling your article. Can you fill in the metadata.yaml with as much information as you can ? By the way, you may need to take the new template that is available here: https://github.com/ReScience/template (with a swh entry). For editor/reviewers, you'll find orcid numbers on https://rescience.github.io/board/ |
Dear Editor (@rougier), We have updated the metadata.yaml file and the template with SWH entry. Please let us know if there is anything additional we can do. Manuscript source files: https://github.com/lamps-lab/rep-network-deconvolution/tree/master/documents/manuscript Thank you! |
Thanks! Sandbox version is online at https://sandbox.zenodo.org/record/203035. Can you check everything is ok (including links pointing to the right place) and tell me such that I can publish the final version? |
Dear Editor (@rougier) We checked the sandbox version, and everything seems to be in order. Thank you. |
Great, thanks. Your paper is online at https://zenodo.org/record/15172014/ and will soon appear on ReScience website. |
Dear Editor (@rougier), Kindly note that there has been a mistake in the previous comment. The link https://zenodo.org/record/15172014/ refers to a different article and not our reproducibility study. |
Sorry, I meant https://zenodo.org/records/15321683 |
Dear @rougier, @alchemi5t, @jsta Thank you very much for your time, guidance, and support throughout the review process. We truly appreciate your efforts in handling our manuscript. |
It's online: https://rescience.github.io/read/#volume-10-2025 |
Original article: C. Ye, M. Evanusa, H. He, A. Mitrokhin, T. Goldstein, J. A. Yorke, C. Fermüller, and Y. Aloimonos. “Network
Deconvolution.” In: ICLR (2020).
PDF URL: https://github.com/lamps-lab/rep-network-deconvolution/blob/master/article.pdf
Metadata URL: https://github.com/lamps-lab/rep-network-deconvolution/blob/master/metadata.yaml
Code URL: https://github.com/lamps-lab/rep-network-deconvolution
Scientific domain: Machine Learning
Programming language: Python
Suggested editor:
The text was updated successfully, but these errors were encountered: