Skip to content

Commit ab8e046

Browse files
committed
Migrate paper to official TMLR template for submission
- Switch from article class to official TMLR template (tmlr.sty, tmlr.bst) - Add double-blind anonymization (author block hidden automatically) - Convert inline thebibliography to references.bib with natbib - Replace GitHub URLs with anonymous mirror placeholders - Add LLM disclosure footnote per TMLR 2025 policy - Remove TODO/FIXME macros from macros.tex - Resolve xcolor option clash via PassOptionsToPackage - Update Makefile with bibtex step and nonstopmode - Paper compiles cleanly: 31 pages, anonymous header
1 parent 8dbb984 commit ab8e046

13 files changed

Lines changed: 1971 additions & 227 deletions

paper/Makefile

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,12 @@
11
MAIN = main
2-
LATEX = pdflatex
2+
LATEX = pdflatex -interaction=nonstopmode
3+
BIBTEX = bibtex
34

45
all: $(MAIN).pdf
56

6-
$(MAIN).pdf: $(MAIN).tex preamble.tex macros.tex bibliography.tex
7+
$(MAIN).pdf: $(MAIN).tex preamble.tex macros.tex references.bib
8+
$(LATEX) $(MAIN)
9+
$(BIBTEX) $(MAIN)
710
$(LATEX) $(MAIN)
811
$(LATEX) $(MAIN)
912

paper/bibliography.tex

Lines changed: 2 additions & 198 deletions
Original file line numberDiff line numberDiff line change
@@ -1,199 +1,3 @@
11
% Shared bibliography — all venues use the same references
2-
3-
\bibliographystyle{plain}
4-
% \bibliography{references} % Uncomment when references.bib exists
5-
6-
% Temporary inline references for skeleton
7-
\begin{thebibliography}{99}
8-
9-
\bibitem{wang2023bitnet}
10-
H.~Wang et al.
11-
\newblock BitNet: Scaling 1-bit Transformers for Large Language Models.
12-
\newblock \emph{arXiv:2310.11453}, 2023.
13-
14-
\bibitem{ma2024bitnetb158}
15-
S.~Ma et al.
16-
\newblock The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits.
17-
\newblock \emph{arXiv:2402.17764}, 2024.
18-
19-
\bibitem{courbariaux2015binaryconnect}
20-
M.~Courbariaux et al.
21-
\newblock BinaryConnect: Training Deep Neural Networks with Binary Weights during Propagations.
22-
\newblock \emph{NeurIPS}, 2015.
23-
24-
\bibitem{rastegari2016xnornet}
25-
M.~Rastegari et al.
26-
\newblock XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks.
27-
\newblock \emph{ECCV}, 2016.
28-
29-
\bibitem{li2016ternary}
30-
F.~Li et al.
31-
\newblock Ternary Weight Networks.
32-
\newblock \emph{arXiv:1605.04711}, 2016.
33-
34-
\bibitem{krizhevsky2009cifar}
35-
A.~Krizhevsky.
36-
\newblock Learning Multiple Layers of Features from Tiny Images.
37-
\newblock Technical report, 2009.
38-
39-
\bibitem{devries2017cutout}
40-
T.~DeVries and G.~Taylor.
41-
\newblock Improved Regularization of Convolutional Neural Networks with Cutout.
42-
\newblock \emph{arXiv:1708.04552}, 2017.
43-
44-
\bibitem{cubuk2019autoaugment}
45-
E.~Cubuk et al.
46-
\newblock AutoAugment: Learning Augmentation Strategies from Data.
47-
\newblock \emph{CVPR}, 2019.
48-
49-
\bibitem{cubuk2020randaugment}
50-
E.~Cubuk et al.
51-
\newblock RandAugment: Practical Automated Data Augmentation with a Reduced Search Space.
52-
\newblock \emph{NeurIPS}, 2020.
53-
54-
\bibitem{he2016resnet}
55-
K.~He et al.
56-
\newblock Deep Residual Learning for Image Recognition.
57-
\newblock \emph{CVPR}, 2016.
58-
59-
\bibitem{hinton2015distilling}
60-
G.~Hinton, O.~Vinyals, and J.~Dean.
61-
\newblock Distilling the Knowledge in a Neural Network.
62-
\newblock \emph{NeurIPS Deep Learning Workshop}, 2015.
63-
64-
\bibitem{zhu2017ttq}
65-
C.~Zhu, S.~Han, H.~Mao, and W.~Dally.
66-
\newblock Trained Ternary Quantization.
67-
\newblock \emph{ICLR}, 2017.
68-
69-
\bibitem{kim2019qkd}
70-
J.~Kim, Y.~Bhalgat, J.~Lee, C.~Patel, and N.~Kwak.
71-
\newblock QKD: Quantization-aware Knowledge Distillation.
72-
\newblock \emph{arXiv:1911.12491}, 2019.
73-
74-
\bibitem{wang2019haq}
75-
K.~Wang, Z.~Liu, Y.~Lin, J.~Lin, and S.~Han.
76-
\newblock HAQ: Hardware-Aware Automated Quantization with Mixed Precision.
77-
\newblock \emph{CVPR}, 2019.
78-
79-
\bibitem{nielsen2024bitnetreloaded}
80-
J.~Nielsen and P.~Schneider-Kamp.
81-
\newblock BitNet b1.58 Reloaded: State-of-the-art Performance Also on Smaller Networks.
82-
\newblock \emph{arXiv:2407.09527}, 2024.
83-
84-
\bibitem{kim2025bdnet}
85-
D.~Kim, J.-S.~Lee, N.-r.~Kim, S.~Lee, and J.-H.~Lee.
86-
\newblock BD-Net: Has Depth-Wise Convolution Ever Been Applied in Binary Neural Networks?
87-
\newblock \emph{AAAI}, 2026.
88-
89-
\bibitem{zhou2016dorefa}
90-
S.~Zhou, Y.~Wu, Z.~Ni, X.~Zhou, H.~Wen, and Y.~Zou.
91-
\newblock DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients.
92-
\newblock \emph{arXiv:1606.06160}, 2016.
93-
94-
\bibitem{dong2019hawq}
95-
Z.~Dong, Z.~Yao, A.~Gholami, M.~Mahoney, and K.~Keutzer.
96-
\newblock HAWQ: Hessian AWare Quantization of Neural Networks with Mixed-Precision.
97-
\newblock \emph{ICCV}, 2019.
98-
99-
\bibitem{elthakeb2020dcq}
100-
A.~T.~Elthakeb, P.~Pilligundla, F.~Mireshghallah, A.~Alaghi, and H.~Esmaeilzadeh.
101-
\newblock Divide and Conquer: Leveraging Intermediate Feature Representations for Quantized Training of Neural Networks.
102-
\newblock \emph{arXiv:1906.06033}, 2020.
103-
104-
\bibitem{le2015tinyimagenet}
105-
Y.~Le and X.~Yang.
106-
\newblock Tiny ImageNet Visual Recognition Challenge.
107-
\newblock CS 231N, Stanford University, 2015.
108-
109-
\bibitem{liu2020reactnet}
110-
Z.~Liu, Z.~Shen, M.~Savvides, and K.-T.~Cheng.
111-
\newblock ReActNet: Towards Precise Binary Neural Network with Generalized Activation Functions.
112-
\newblock \emph{ECCV}, 2020.
113-
114-
\bibitem{guo2022bnext}
115-
N.~Guo, J.~Bethge, C.~Meinel, and H.~Yang.
116-
\newblock BNext: Any Precision Binary Neural Network.
117-
\newblock \emph{arXiv:2211.12933}, 2022.
118-
119-
\bibitem{dong2020hawqv2}
120-
Z.~Dong, Z.~Yao, D.~Arfeen, A.~Gholami, M.~W.~Mahoney, and K.~Keutzer.
121-
\newblock HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks.
122-
\newblock \emph{NeurIPS}, 2020.
123-
124-
\bibitem{qin2020irnet}
125-
H.~Qin, R.~Gong, X.~Liu, M.~Shen, Z.~Wei, F.~Yu, and J.~Song.
126-
\newblock Forward and Backward Information Retention for Accurate Binary Neural Networks.
127-
\newblock \emph{CVPR}, 2020.
128-
129-
\bibitem{zhao2024sqakd}
130-
K.~Zhao and M.~Zhao.
131-
\newblock SQAKD: Self-supervised Quantization-Aware Knowledge Distillation.
132-
\newblock \emph{AISTATS}, 2024.
133-
134-
\bibitem{bengio2013estimating}
135-
Y.~Bengio, N.~Léonard, and A.~Courville.
136-
\newblock Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation.
137-
\newblock \emph{arXiv:1308.3432}, 2013.
138-
139-
\bibitem{kim2021relaxloss}
140-
J.~Kim, S.~Bhattacharjee, S.~Park, S.~Jung, and Y.~M.~Kim.
141-
\newblock RelaxLoss: Penalty-Free Quantization through Normalized Loss.
142-
\newblock \emph{arXiv:2105.00944}, 2021.
143-
144-
\bibitem{zhang2018mixup}
145-
H.~Zhang, M.~Cisse, Y.~N.~Dauphin, and D.~Lopez-Paz.
146-
\newblock mixup: Beyond Empirical Risk Minimization.
147-
\newblock \emph{ICLR}, 2018.
148-
149-
\bibitem{gundersen2018reproducibility}
150-
O.~E.~Gundersen and S.~Kjensmo.
151-
\newblock State of the Art: Reproducibility in Artificial Intelligence.
152-
\newblock \emph{AAAI}, 2018.
153-
154-
\bibitem{hutson2018ai}
155-
M.~Hutson.
156-
\newblock Artificial Intelligence Faces Reproducibility Crisis.
157-
\newblock \emph{Science}, 359(6377):725--726, 2018.
158-
159-
\bibitem{dodge2019mlchecklist}
160-
J.~Dodge, S.~Gururangan, D.~Card, R.~Schwartz, and N.~A.~Smith.
161-
\newblock Show Your Work: Improved Reporting of Experimental Results.
162-
\newblock \emph{EMNLP}, 2019.
163-
164-
\bibitem{tflite}
165-
TensorFlow Lite Team.
166-
\newblock TensorFlow Lite: Deploy machine learning models on mobile and edge devices.
167-
\newblock \url{https://www.tensorflow.org/lite}, 2023.
168-
169-
\bibitem{pytorch_mobile}
170-
PyTorch Team.
171-
\newblock PyTorch Mobile: End-to-end workflow from training to deployment.
172-
\newblock \url{https://pytorch.org/mobile/}, 2023.
173-
174-
\bibitem{bitblas}
175-
L.~Wang, L.~Lei, L.~Ye, Y.~Zhao, W.~Chen, D.~Lin, X.~Zheng, and C.~Yu.
176-
\newblock BitBLAS: A High-Performance Library for Quantized Deep Learning.
177-
\newblock \emph{arXiv:2410.16144}, 2024.
178-
179-
\bibitem{pytorch_cifar_kuangliu}
180-
K.~Liu.
181-
\newblock Train CIFAR10 with PyTorch.
182-
\newblock \url{https://github.com/kuangliu/pytorch-cifar}, 2017.
183-
184-
\bibitem{pytorch_cifar100_weiaicunzai}
185-
W.~Zhang.
186-
\newblock PyTorch CIFAR-100 Benchmark.
187-
\newblock \url{https://github.com/weiaicunzai/pytorch-cifar100}, 2019.
188-
189-
\bibitem{tiny_imagenet_benchmark}
190-
Y.~Le and X.~Yang.
191-
\newblock Tiny ImageNet Visual Recognition Challenge.
192-
\newblock CS 231N, Stanford University, 2015.
193-
194-
\bibitem{cover2006information}
195-
T.~M.~Cover and J.~A.~Thomas.
196-
\newblock Elements of Information Theory, 2nd Edition.
197-
\newblock Wiley-Interscience, 2006.
198-
199-
\end{thebibliography}
2+
\bibliographystyle{tmlr}
3+
\bibliography{references}

paper/macros.tex

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -24,10 +24,6 @@
2424
\newcommand{\real}{\mathbb{R}}
2525
\newcommand{\expect}{\mathbb{E}}
2626

27-
% Highlight for TODOs (remove before submission)
28-
\newcommand{\todo}[1]{\textcolor{red}{[TODO: #1]}}
29-
\newcommand{\fixme}[1]{\textcolor{orange}{[FIXME: #1]}}
30-
3127
% Recovery tier colors for Table 8
3228
\definecolor{recovhigh}{HTML}{C8E6C9} % green - recovery >= 80%
3329
\definecolor{recovmed}{HTML}{FFF9C4} % yellow - recovery 50-79%

paper/main.pdf

-96.5 KB
Binary file not shown.

paper/main.tex

Lines changed: 13 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,8 @@
1-
% TMLR Submission: Understanding and Closing the 1.58-bit Quantization Gap in CNNs
2-
% New version with modular sections
3-
41
\documentclass[10pt]{article}
5-
6-
% Page geometry
7-
\usepackage[margin=1in, top=0.9in, bottom=0.9in]{geometry}
2+
\PassOptionsToPackage{table}{xcolor}
3+
\usepackage{tmlr}
4+
% If accepted, instead use the following line for the camera-ready submission:
5+
%\usepackage[accepted]{tmlr}
86

97
% Import packages and macros
108
\input{preamble}
@@ -13,19 +11,20 @@
1311
% Graphics path for figures
1412
\graphicspath{{figures/}}
1513

14+
\def\month{MM}
15+
\def\year{YYYY}
16+
\def\openreview{\url{https://openreview.net/forum?id=XXXX}}
17+
1618
% =============================================================================
1719
% METADATA
1820
% =============================================================================
1921

20-
\title{Understanding and Closing the 1.58-bit Quantization Gap in CNNs:\\An Empirical Study}
21-
22-
\author{
23-
Dario Cazzani\\
24-
Cisco Systems Inc.\\
25-
\texttt{dcazzani@cisco.com}
26-
}
22+
\title{Understanding and Closing the 1.58-bit Quantization Gap in CNNs: An Empirical Study}
2723

28-
\date{}
24+
% Authors must not appear in the submitted version.
25+
% Non-anonymous submissions will be rejected without review.
26+
\author{\name Dario Cazzani \email dcazzani@cisco.com \\
27+
\addr Cisco Systems Inc.}
2928

3029
% =============================================================================
3130
% DOCUMENT

paper/preamble.tex

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,14 +1,12 @@
1-
% Shared preamble — common packages for all venues
2-
% Excludes \documentclass and geometry (venue-specific)
1+
% Shared preamble — TMLR-compatible packages
2+
% Note: tmlr.sty already provides fontenc[T1], natbib, and fancyhdr
33

44
% Core packages
55
\usepackage[utf8]{inputenc}
6-
\usepackage[T1]{fontenc}
76
\usepackage{amsmath,amssymb,amsfonts}
87
\usepackage{graphicx}
98
\usepackage{booktabs}
109
\usepackage{multirow}
11-
\usepackage[table]{xcolor}
1210
\usepackage{hyperref}
1311
\usepackage{cleveref}
1412

0 commit comments

Comments
 (0)