-
HKUST, @HLTCHKUST, @IndoNLP
- Hong Kong SAR
-
04:06
(UTC +08:00) - https://scholar.google.com/citations?user=0pnaXO8AAAAJ&hl=en
- @bryanwilie92
- https://bryanwilie.github.io/
Highlights
- Pro
Stars
A framework for efficient model inference with omni-modality models
Belief-R test LMs' belief revision ability when presented with new evidence. Inspired by how humans suppress prior inferences, this task assesses LMs within delta reasoning (ΔR) framework. Belief-R…
Official repository of the work titled "High-Dimensional Interlingual Representations of Large Language Models"
Resources and Implementations (PyTorch) for Information Theoretical concepts in Deep Learning
WorldCuisines is an extensive multilingual and multicultural benchmark that spans 30 languages, covering a wide array of global cuisines. Best Theme Paper 🏆 NAACL 2025
Indonesian T0 | Instruction-tuning for low-resource and extremely low-resource Austronesian languages
Official reposity for paper "High-Dimension Human Value Representation in Large Language Models" (NAACL'25 Main)
A curated list of Human Preference Datasets for LLM fine-tuning, RLHF, and eval.
The project page for "LOGIC-LM: Empowering Large Language Models with Symbolic Solvers for Faithful Logical Reasoning"
NusaWrites is an in-depth analysis of corpora collection strategy and a comprehensive language modeling benchmark for underrepresented and extremely low-resource Indonesian local languages.
Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks
😎 A curated list of awesome GitHub Profile which updates in real time
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts…
Code for paper RHO (ρ): Reducing Hallucination in Open-domain Dialogues with Knowledge Grounding
[EMNLP 2023] The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning
A playbook for systematically maximizing the performance of deep learning models.
How Long Is Enough? Exploring the Optimal Intervals of Long-Range Clinical Note Language Modeling
This repository provides details and links to the ACL anthology corpus/collection including .bib, .pdf and grobid extractions of the pdfs
A GPT-J API to use with python3 to generate text, blogs, code, and more
BioBART: Pretraining and Evaluation of A Biomedical Generative Language Model [ACL-BioNLP 2022]
Tools for curating biomedical training data for large-scale language modeling
A collection of (mostly) technical things every software developer should know about
The implementation of the paper "Retrieval-Free Knowledge-Grounded Dialogue Response Generation with Adapters".