Publications

You can also find my articles on my Google Scholar profile.

MerryQuery: A Trustworthy LLM-Powered Tool Providing Personalized Support for Educators and Students

Benyamin Tabarsi, Aditya Basarkar, Xukun Liu, Dongkuan Xu, Tiffany Barnes, 2024

The potential of Large Language Models (LLMs) in education is not trivial, but concerns about academic misconduct, misinformation, and overreliance limit their adoption. To address these issues, we introduce MerryQuery, an AI-powered educational assistant using Retrieval-Augmented Generation (RAG), to provide contextually relevant, course-specific responses. MerryQuery features guided dialogues and source citation to ensure trust and improve student learning. Additionally, it enables instructors to monitor student interactions, customize response granularity, and input multimodal materials without compromising data fidelity. By meeting both student and instructor needs, MerryQuery offers a responsible way to integrate LLMs into educational settings.

MerryQuery: A Trustworthy LLM-Powered Tool Providing Personalized Support for Educators and Students

Adaptive Draft-Verification for Efficient Large Language Model Decoding

Xukun Liu, Bowen Lie, Ruqi Zhang, Dongkuan Xu, 2024

We introduce an LLM decoding acceleration method that requires no fine-tuning. Our approach involves an adaptive draft-verification process that evolves over time to improve efficiency. We utilize a tri-gram matrixbased LLM representation to dynamically approximate the output distribution of the LLM, allowing the model to adjust to changing token probabilities during the decoding process.

Adaptive Draft-Verification for Efficient Large Language Model Decoding

Download here

LawLLM: Law Large Language Model for the US Legal System

Dong Shu, Haoran Zhao, Xukun Liu, David Demeter, Mengnan Du, Yongfeng Zhang, 2024

We introduce LawLLM, a multi-task model designed for the US legal domain, excelling in Similar Case Retrieval (SCR), Precedent Case Recommendation (PCR), and Legal Judgment Prediction (LJP). LawLLM distinguishes between similar and precedent cases, providing clarity for future research. We apply customized data preprocessing, in-context learning, and advanced information retrieval methods to transform raw legal data into a trainable format. LawLLM outperforms existing baselines in zero-shot and few-shot scenarios, addressing key challenges in legal analytics.

LawLLM: Law Large Language Model for the US Legal System

Download here

ToolNet: Connecting Large Language Models With Massive Tools

Xukun Liu, Zhiyuan Peng, DK Xu, 2024

We introduce ToolNet, a plug-and-play method to assists LLMs in handling massive tools. ToolNet organizes tools in a weighted directed graph (node represents tools and edges denote tool transition) based on the tool-use trajectories produced by LLMs. An LLM navigates in the graph by iteratively choosing the next one from its successors until the task is resolved. Graphs are updated online, enabling adjustment to accommodate the frequent updates of tools and new tasks.

ToolNet: Connecting Large Language Models With Massive Tools

Download here

Gentopia.AI: A Collaborative Platform for Tool-Augmented LLMs

Binfeng Xu, Xukun Liu, et al., 2023

Gentopia is a lightweight and extensible framework for LLM-driven Agents and ALM research. It provides essential components to build, test and evaluate agents. At its core, Gentopia aims to assemble an agent with a single config, thus minimizing your effort in building, tuning, and sharing agents.

Gentopia.AI: A Collaborative Platform for Tool-Augmented LLMs

Download here