Posts by Collection

publications

Rethinking Optimal Verification Granularity for Compute-Efficient Test-Time Scaling

Hao (Mark) Chen, Guanxi Lu, Yasuyuki Okoshi, Zhiwen Mo, Masato Motomura, Hongxiang Fan

Efficient and Flexible FP-INTx Accelerator for Weight-only Quantized LLM Inference

Zhican Wang, Hongxiang Fan, Guanxi Lu, Chen Zhang, Haroon Waris, Hao (Mark) Chen, Guanghui He

(Invited Paper) Enhancing Trustworthiness with Mixed Precision: Benchmarks, Opportunities, and Challenges

Guanxi Lu, Hao (Mark) Chen, Zhiqiang Que, Wayne Luk, Hongxiang Fan

FastTTS: Accelerating Test-Time Scaling for Edge LLM Reasoning

Hao (Mark) Chen, Zhiwen Mo, Guanxi Lu, Shuang Liang, Lingxiao Ma, Wayne Luk, Hongxiang Fan

AdaBlock-dLLM: Semantic-Aware Diffusion LLM Inference via Adaptive Block Size

Guanxi Lu, Hao (Mark) Chen, Yuto Karashima, Zhican Wang, Daichi Fujiki, Hongxiang Fan

talks

teaching

Teaching experience 1

Undergraduate course, University 1, Department, 2014.01

This is a description of a teaching experience. You can use markdown like any other post.

Teaching experience 2

Workshop, University 1, Department, 2015.01

This is a description of a teaching experience. You can use markdown like any other post.