|
1 | 1 | % Shared bibliography — all venues use the same references |
2 | | - |
3 | | -\bibliographystyle{plain} |
4 | | -% \bibliography{references} % Uncomment when references.bib exists |
5 | | - |
6 | | -% Temporary inline references for skeleton |
7 | | -\begin{thebibliography}{99} |
8 | | - |
9 | | -\bibitem{wang2023bitnet} |
10 | | -H.~Wang et al. |
11 | | -\newblock BitNet: Scaling 1-bit Transformers for Large Language Models. |
12 | | -\newblock \emph{arXiv:2310.11453}, 2023. |
13 | | - |
14 | | -\bibitem{ma2024bitnetb158} |
15 | | -S.~Ma et al. |
16 | | -\newblock The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits. |
17 | | -\newblock \emph{arXiv:2402.17764}, 2024. |
18 | | - |
19 | | -\bibitem{courbariaux2015binaryconnect} |
20 | | -M.~Courbariaux et al. |
21 | | -\newblock BinaryConnect: Training Deep Neural Networks with Binary Weights during Propagations. |
22 | | -\newblock \emph{NeurIPS}, 2015. |
23 | | - |
24 | | -\bibitem{rastegari2016xnornet} |
25 | | -M.~Rastegari et al. |
26 | | -\newblock XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks. |
27 | | -\newblock \emph{ECCV}, 2016. |
28 | | - |
29 | | -\bibitem{li2016ternary} |
30 | | -F.~Li et al. |
31 | | -\newblock Ternary Weight Networks. |
32 | | -\newblock \emph{arXiv:1605.04711}, 2016. |
33 | | - |
34 | | -\bibitem{krizhevsky2009cifar} |
35 | | -A.~Krizhevsky. |
36 | | -\newblock Learning Multiple Layers of Features from Tiny Images. |
37 | | -\newblock Technical report, 2009. |
38 | | - |
39 | | -\bibitem{devries2017cutout} |
40 | | -T.~DeVries and G.~Taylor. |
41 | | -\newblock Improved Regularization of Convolutional Neural Networks with Cutout. |
42 | | -\newblock \emph{arXiv:1708.04552}, 2017. |
43 | | - |
44 | | -\bibitem{cubuk2019autoaugment} |
45 | | -E.~Cubuk et al. |
46 | | -\newblock AutoAugment: Learning Augmentation Strategies from Data. |
47 | | -\newblock \emph{CVPR}, 2019. |
48 | | - |
49 | | -\bibitem{cubuk2020randaugment} |
50 | | -E.~Cubuk et al. |
51 | | -\newblock RandAugment: Practical Automated Data Augmentation with a Reduced Search Space. |
52 | | -\newblock \emph{NeurIPS}, 2020. |
53 | | - |
54 | | -\bibitem{he2016resnet} |
55 | | -K.~He et al. |
56 | | -\newblock Deep Residual Learning for Image Recognition. |
57 | | -\newblock \emph{CVPR}, 2016. |
58 | | - |
59 | | -\bibitem{hinton2015distilling} |
60 | | -G.~Hinton, O.~Vinyals, and J.~Dean. |
61 | | -\newblock Distilling the Knowledge in a Neural Network. |
62 | | -\newblock \emph{NeurIPS Deep Learning Workshop}, 2015. |
63 | | - |
64 | | -\bibitem{zhu2017ttq} |
65 | | -C.~Zhu, S.~Han, H.~Mao, and W.~Dally. |
66 | | -\newblock Trained Ternary Quantization. |
67 | | -\newblock \emph{ICLR}, 2017. |
68 | | - |
69 | | -\bibitem{kim2019qkd} |
70 | | -J.~Kim, Y.~Bhalgat, J.~Lee, C.~Patel, and N.~Kwak. |
71 | | -\newblock QKD: Quantization-aware Knowledge Distillation. |
72 | | -\newblock \emph{arXiv:1911.12491}, 2019. |
73 | | - |
74 | | -\bibitem{wang2019haq} |
75 | | -K.~Wang, Z.~Liu, Y.~Lin, J.~Lin, and S.~Han. |
76 | | -\newblock HAQ: Hardware-Aware Automated Quantization with Mixed Precision. |
77 | | -\newblock \emph{CVPR}, 2019. |
78 | | - |
79 | | -\bibitem{nielsen2024bitnetreloaded} |
80 | | -J.~Nielsen and P.~Schneider-Kamp. |
81 | | -\newblock BitNet b1.58 Reloaded: State-of-the-art Performance Also on Smaller Networks. |
82 | | -\newblock \emph{arXiv:2407.09527}, 2024. |
83 | | - |
84 | | -\bibitem{kim2025bdnet} |
85 | | -D.~Kim, J.-S.~Lee, N.-r.~Kim, S.~Lee, and J.-H.~Lee. |
86 | | -\newblock BD-Net: Has Depth-Wise Convolution Ever Been Applied in Binary Neural Networks? |
87 | | -\newblock \emph{AAAI}, 2026. |
88 | | - |
89 | | -\bibitem{zhou2016dorefa} |
90 | | -S.~Zhou, Y.~Wu, Z.~Ni, X.~Zhou, H.~Wen, and Y.~Zou. |
91 | | -\newblock DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients. |
92 | | -\newblock \emph{arXiv:1606.06160}, 2016. |
93 | | - |
94 | | -\bibitem{dong2019hawq} |
95 | | -Z.~Dong, Z.~Yao, A.~Gholami, M.~Mahoney, and K.~Keutzer. |
96 | | -\newblock HAWQ: Hessian AWare Quantization of Neural Networks with Mixed-Precision. |
97 | | -\newblock \emph{ICCV}, 2019. |
98 | | - |
99 | | -\bibitem{elthakeb2020dcq} |
100 | | -A.~T.~Elthakeb, P.~Pilligundla, F.~Mireshghallah, A.~Alaghi, and H.~Esmaeilzadeh. |
101 | | -\newblock Divide and Conquer: Leveraging Intermediate Feature Representations for Quantized Training of Neural Networks. |
102 | | -\newblock \emph{arXiv:1906.06033}, 2020. |
103 | | - |
104 | | -\bibitem{le2015tinyimagenet} |
105 | | -Y.~Le and X.~Yang. |
106 | | -\newblock Tiny ImageNet Visual Recognition Challenge. |
107 | | -\newblock CS 231N, Stanford University, 2015. |
108 | | - |
109 | | -\bibitem{liu2020reactnet} |
110 | | -Z.~Liu, Z.~Shen, M.~Savvides, and K.-T.~Cheng. |
111 | | -\newblock ReActNet: Towards Precise Binary Neural Network with Generalized Activation Functions. |
112 | | -\newblock \emph{ECCV}, 2020. |
113 | | - |
114 | | -\bibitem{guo2022bnext} |
115 | | -N.~Guo, J.~Bethge, C.~Meinel, and H.~Yang. |
116 | | -\newblock BNext: Any Precision Binary Neural Network. |
117 | | -\newblock \emph{arXiv:2211.12933}, 2022. |
118 | | - |
119 | | -\bibitem{dong2020hawqv2} |
120 | | -Z.~Dong, Z.~Yao, D.~Arfeen, A.~Gholami, M.~W.~Mahoney, and K.~Keutzer. |
121 | | -\newblock HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks. |
122 | | -\newblock \emph{NeurIPS}, 2020. |
123 | | - |
124 | | -\bibitem{qin2020irnet} |
125 | | -H.~Qin, R.~Gong, X.~Liu, M.~Shen, Z.~Wei, F.~Yu, and J.~Song. |
126 | | -\newblock Forward and Backward Information Retention for Accurate Binary Neural Networks. |
127 | | -\newblock \emph{CVPR}, 2020. |
128 | | - |
129 | | -\bibitem{zhao2024sqakd} |
130 | | -K.~Zhao and M.~Zhao. |
131 | | -\newblock SQAKD: Self-supervised Quantization-Aware Knowledge Distillation. |
132 | | -\newblock \emph{AISTATS}, 2024. |
133 | | - |
134 | | -\bibitem{bengio2013estimating} |
135 | | -Y.~Bengio, N.~Léonard, and A.~Courville. |
136 | | -\newblock Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation. |
137 | | -\newblock \emph{arXiv:1308.3432}, 2013. |
138 | | - |
139 | | -\bibitem{kim2021relaxloss} |
140 | | -J.~Kim, S.~Bhattacharjee, S.~Park, S.~Jung, and Y.~M.~Kim. |
141 | | -\newblock RelaxLoss: Penalty-Free Quantization through Normalized Loss. |
142 | | -\newblock \emph{arXiv:2105.00944}, 2021. |
143 | | - |
144 | | -\bibitem{zhang2018mixup} |
145 | | -H.~Zhang, M.~Cisse, Y.~N.~Dauphin, and D.~Lopez-Paz. |
146 | | -\newblock mixup: Beyond Empirical Risk Minimization. |
147 | | -\newblock \emph{ICLR}, 2018. |
148 | | - |
149 | | -\bibitem{gundersen2018reproducibility} |
150 | | -O.~E.~Gundersen and S.~Kjensmo. |
151 | | -\newblock State of the Art: Reproducibility in Artificial Intelligence. |
152 | | -\newblock \emph{AAAI}, 2018. |
153 | | - |
154 | | -\bibitem{hutson2018ai} |
155 | | -M.~Hutson. |
156 | | -\newblock Artificial Intelligence Faces Reproducibility Crisis. |
157 | | -\newblock \emph{Science}, 359(6377):725--726, 2018. |
158 | | - |
159 | | -\bibitem{dodge2019mlchecklist} |
160 | | -J.~Dodge, S.~Gururangan, D.~Card, R.~Schwartz, and N.~A.~Smith. |
161 | | -\newblock Show Your Work: Improved Reporting of Experimental Results. |
162 | | -\newblock \emph{EMNLP}, 2019. |
163 | | - |
164 | | -\bibitem{tflite} |
165 | | -TensorFlow Lite Team. |
166 | | -\newblock TensorFlow Lite: Deploy machine learning models on mobile and edge devices. |
167 | | -\newblock \url{https://www.tensorflow.org/lite}, 2023. |
168 | | - |
169 | | -\bibitem{pytorch_mobile} |
170 | | -PyTorch Team. |
171 | | -\newblock PyTorch Mobile: End-to-end workflow from training to deployment. |
172 | | -\newblock \url{https://pytorch.org/mobile/}, 2023. |
173 | | - |
174 | | -\bibitem{bitblas} |
175 | | -L.~Wang, L.~Lei, L.~Ye, Y.~Zhao, W.~Chen, D.~Lin, X.~Zheng, and C.~Yu. |
176 | | -\newblock BitBLAS: A High-Performance Library for Quantized Deep Learning. |
177 | | -\newblock \emph{arXiv:2410.16144}, 2024. |
178 | | - |
179 | | -\bibitem{pytorch_cifar_kuangliu} |
180 | | -K.~Liu. |
181 | | -\newblock Train CIFAR10 with PyTorch. |
182 | | -\newblock \url{https://github.com/kuangliu/pytorch-cifar}, 2017. |
183 | | - |
184 | | -\bibitem{pytorch_cifar100_weiaicunzai} |
185 | | -W.~Zhang. |
186 | | -\newblock PyTorch CIFAR-100 Benchmark. |
187 | | -\newblock \url{https://github.com/weiaicunzai/pytorch-cifar100}, 2019. |
188 | | - |
189 | | -\bibitem{tiny_imagenet_benchmark} |
190 | | -Y.~Le and X.~Yang. |
191 | | -\newblock Tiny ImageNet Visual Recognition Challenge. |
192 | | -\newblock CS 231N, Stanford University, 2015. |
193 | | - |
194 | | -\bibitem{cover2006information} |
195 | | -T.~M.~Cover and J.~A.~Thomas. |
196 | | -\newblock Elements of Information Theory, 2nd Edition. |
197 | | -\newblock Wiley-Interscience, 2006. |
198 | | - |
199 | | -\end{thebibliography} |
| 2 | +\bibliographystyle{tmlr} |
| 3 | +\bibliography{references} |
0 commit comments