Boyuan Chen is a PhD candidate studying faster inference for large language models. He graduated from Pomona College in USA with computer science and math double major. He is interested in the fields of large language models and faster inference.
Current Research
My research is centered on optimizing the inference time of large language models (LLMs), particularly in the context of code generation, to reduce operational costs and improve user experience through reduced latency. Feel free to reach out to me for collaborations and ideas (I have a lot of computing resources).