Obserwuj
Sheng-Chun Kao
Tytuł
Cytowane przez
Cytowane przez
Rok
Gamma: Automating the hw mapping of dnn models on accelerators via genetic algorithm
SC Kao, T Krishna
Proceedings of the 39th International Conference on Computer-Aided Design, 1-9, 2020
1042020
Confuciux: Autonomous hardware resource assignment for dnn accelerators using reinforcement learning
SC Kao, G Jeong, T Krishna
2020 53rd Annual IEEE/ACM International Symposium on Microarchitecture …, 2020
892020
Magma: An optimization framework for mapping multiple dnns on multiple accelerator cores
SC Kao, T Krishna
2022 IEEE International Symposium on High-Performance Computer Architecture …, 2022
322022
FLAT: An Optimized Dataflow for Mitigating Attention Performance Bottlenecks
SC Kao, S Subramanian, G Agrawal, T Krishna
arXiv preprint arXiv:2107.06419, 2021
21*2021
Digamma: Domain-aware genetic algorithm for hw-mapping co-optimization for dnn accelerators
SC Kao, M Pellauer, A Parashar, T Krishna
2022 Design, Automation & Test in Europe Conference & Exhibition (DATE), 232-237, 2022
162022
Extending sparse tensor accelerators to support multiple compression formats
E Qin, G Jeong, W Won, SC Kao, H Kwon, S Srinivasan, D Das, GE Moon, ...
2021 IEEE International Parallel and Distributed Processing Symposium (IPDPS …, 2021
152021
Reinforcement learning based interconnection routing for adaptive traffic optimization
SC Kao, CHH Yang, PY Chen, X Ma, T Krishna
Proceedings of the 13th IEEE/ACM international symposium on networks-on-chip …, 2019
152019
Dynamically updatable ternary segmented aging bloom filter for openflow-compliant low-power packet processing
SC Kao, DY Lee, TS Chen, AY Wu
IEEE/ACM Transactions on Networking 26 (2), 1004-1017, 2018
142018
Demystifying map space exploration for NPUs
SC Kao, A Parashar, PA Tsai, T Krishna
2022 IEEE International Symposium on Workload Characterization (IISWC), 269-281, 2022
82022
DNNFuser: Generative pre-trained transformer as a generalized mapper for layer fusion in dnn accelerators
SC Kao, X Huang, T Krishna
arXiv preprint arXiv:2201.11218, 2022
52022
Domain-specific genetic algorithm for multi-tenant dnnaccelerator scheduling
SC Kao, T Krishna
arXiv preprint arXiv:2104.13997, 2021
52021
E3: A hw/sw co-design neuroevolution platform for autonomous learning in edge device
SC Kao, T Krishna
2021 IEEE International Symposium on Performance Analysis of Systems and …, 2021
52021
Training recipe for n: M structured sparsity with decaying pruning mask
SC Kao, A Yazdanbakhsh, S Subramanian, S Agrawal, U Evci, T Krishna
arXiv preprint arXiv:2209.07617, 2022
42022
JaxPruner: A concise library for sparsity research
JH Lee, W Park, NE Mitchell, J Pilault, JSO Ceron, HB Kim, N Lee, ...
Conference on Parsimony and Learning, 515-528, 2024
32024
Training Recipe for N: M Structured Sparsity with Decaying Pruning Mask
A Yazdanbakhsh, SC Kao, S Agrawal, S Subramanian, T Krishna, U Evci
arXiv preprint arXiv:2209.07617, 2022
12022
ATTACC the Quadratic Bottleneck of Attention Layers.
SC Kao, S Subramanian, G Agrawal, T Krishna
ArXiv, abs/2107.06419, 2021
12021
NonGEMM Bench: Understanding the Performance Horizon of the Latest ML Workloads with NonGEMM Workloads
R Karami, H Kota, SC Kao, H Kwon
arXiv preprint arXiv:2404.11788, 2024
2024
Progressive Gradient Flow for Robust N: M Sparsity Training in Transformers
AR Bambhaniya, A Yazdanbakhsh, S Subramanian, SC Kao, S Agrawal, ...
arXiv preprint arXiv:2402.04744, 2024
2024
A Formalism of DNN Accelerator Flexibility
SC Kao, H Kwon, M Pellauer, A Parashar, T Krishna
Proceedings of the ACM on Measurement and Analysis of Computing Systems 6 (2 …, 2022
2022
Generative Design of Hardware-aware DNNs
SC Kao, A Ramamurthy, T Krishna
arXiv preprint arXiv:2006.03968, 2020
2020
Nie można teraz wykonać tej operacji. Spróbuj ponownie później.
Prace 1–20