YNICL Journal 2025 Journal Article
A robust deep learning framework for cerebral microbleeds recognition in GRE and SWI MRI
- Tahereh Hassanzadeh
- Sonal Sachdev
- Wei Wen
- Perminder S. Sachdev
- Arcot Sowmya
Author name cluster
Possible papers associated with this exact author name in Arrow. This page groups case-insensitive exact name matches and is not a full identity disambiguation profile.
YNICL Journal 2025 Journal Article
ICLR Conference 2025 Conference Paper
Evaluating large language models (LLMs) in diverse and challenging scenarios is essential to align them with human preferences. To mitigate the prohibitive costs associated with human evaluations, utilizing a powerful LLM as a judge has emerged as a favored approach. Nevertheless, this methodology encounters several challenges, including substantial expenses, concerns regarding privacy and security, and reproducibility. In this paper, we propose a straightforward, replicable, and accurate automated evaluation method by leveraging a lightweight LLM as the judge, named RocketEval. Initially, we identify that the performance disparity between lightweight and powerful LLMs in evaluation tasks primarily stems from their ability to conduct comprehensive analyses, which is not easily enhanced through techniques such as chain-of-thought reasoning. By reframing the evaluation task as a multi-faceted Q\&A using an instance-specific checklist, we demonstrate that the limited judgment accuracy of lightweight LLMs is largely attributes to high uncertainty and positional bias. To address these challenges, we introduce an automated evaluation process grounded in checklist grading, which is designed to accommodate a variety of scenarios and questions. This process encompasses the creation of checklists, the grading of these checklists by lightweight LLMs, and the reweighting of checklist items to align with the supervised annotations. Our experiments carried out on the automated evaluation benchmarks, MT-Bench and WildBench datasets, reveal that RocketEval, when using $\textit{Gemma-2-2B}$ as the judge, achieves a high correlation (0.965) with human preferences, which is comparable to $\textit{GPT-4o}$. Moreover, RocketEval provides a cost reduction exceeding 50-fold for large-scale evaluation and comparison scenarios. Our code is available at https://github.com/Joinn99/RocketEval-ICLR.
YNIMG Journal 2025 Journal Article
YNIMG Journal 2024 Journal Article
YNIMG Journal 2022 Journal Article
YNIMG Journal 2021 Journal Article
YNIMG Journal 2021 Journal Article
YNICL Journal 2020 Journal Article
YNIMG Journal 2020 Journal Article
IJCAI Conference 2020 Conference Paper
To enable DNNs on edge devices like mobile phones, low-rank approximation has been widely adopted because of its solid theoretical rationale and efficient implementations. Several previous works attempted to directly approximate a pre-trained model by low-rank decomposition; however, small approximation errors in parameters can ripple over a large prediction loss. As a result, performance usually drops significantly and a sophisticated effort on fine-tuning is required to recover accuracy. Apparently, it is not optimal to separate low-rank approximation from training. Unlike previous works, this paper integrates low rank approximation and regularization into the training process. We propose Trained Rank Pruning (TRP), which alternates between low rank approximation and training. TRP maintains the capacity of the original network while imposing low-rank constraints during training. A nuclear regularization optimized by stochastic sub-gradient descent is utilized to further promote low rank in TRP. The TRP trained network inherently has a low-rank structure, and is approximated with negligible performance loss, thus eliminating the fine-tuning process after low rank decomposition. The proposed method is comprehensively evaluated on CIFAR-10 and ImageNet, outperforming previous compression methods using low rank approximation.
YNICL Journal 2019 Journal Article
YNICL Journal 2018 Journal Article
YNIMG Journal 2018 Journal Article
YNIMG Journal 2018 Journal Article
NeurIPS Conference 2017 Conference Paper
High network communication cost for synchronizing gradients and parameters is the well-known bottleneck of distributed training. In this work, we propose TernGrad that uses ternary gradients to accelerate distributed deep learning in data parallelism. Our approach requires only three numerical levels {-1, 0, 1}, which can aggressively reduce the communication time. We mathematically prove the convergence of TernGrad under the assumption of a bound on gradients. Guided by the bound, we propose layer-wise ternarizing and gradient clipping to improve its convergence. Our experiments show that applying TernGrad on AlexNet does not incur any accuracy loss and can even improve accuracy. The accuracy loss of GoogLeNet induced by TernGrad is less than 2% on average. Finally, a performance model is proposed to study the scalability of TernGrad. Experiments show significant speed gains for various deep neural networks. Our source code is available.
NeurIPS Conference 2016 Conference Paper
High demand for computation resources severely hinders deployment of large-scale Deep Neural Networks (DNN) in resource constrained devices. In this work, we propose a Structured Sparsity Learning (SSL) method to regularize the structures (i. e. , filters, channels, filter shapes, and layer depth) of DNNs. SSL can: (1) learn a compact structure from a bigger DNN to reduce computation cost; (2) obtain a hardware-friendly structured sparsity of DNN to efficiently accelerate the DNN’s evaluation. Experimental results show that SSL achieves on average 5. 1X and 3. 1X speedups of convolutional layer computation of AlexNet against CPU and GPU, respectively, with off-the-shelf libraries. These speedups are about twice speedups of non-structured sparsity; (3) regularize the DNN structure to improve classification accuracy. The results show that for CIFAR-10, regularization on layer depth reduces a 20-layer Deep Residual Network (ResNet) to 18 layers while improves the accuracy from 91. 25% to 92. 60%, which is still higher than that of original ResNet with 32 layers. For AlexNet, SSL reduces the error by ~1%.
YNIMG Journal 2015 Journal Article
YNIMG Journal 2014 Journal Article
YNICL Journal 2014 Journal Article
YNIMG Journal 2013 Journal Article
YNIMG Journal 2013 Journal Article
YNIMG Journal 2012 Journal Article
YNIMG Journal 2012 Journal Article
YNIMG Journal 2012 Journal Article
YNIMG Journal 2012 Journal Article
YNIMG Journal 2011 Journal Article
YNIMG Journal 2011 Journal Article
YNIMG Journal 2011 Journal Article
YNIMG Journal 2010 Journal Article
YNIMG Journal 2010 Journal Article
YNIMG Journal 2007 Journal Article
YNIMG Journal 2006 Journal Article
YNIMG Journal 2004 Journal Article
YNIMG Journal 2004 Journal Article