Ph.D. student@UIUC-DMG || prev. undergrad@SJTU-IEEE
Highlights
- Pro
Stars
4
results
for sponsorable starred repositories
Clear filter
A high-throughput and memory-efficient inference and serving engine for LLMs
A trend starts from "Chain of Thought Prompting Elicits Reasoning in Large Language Models".
Paper collections of retrieval-based (augmented) language model.