Stars
[CVPR 2025 HIghlight] XLRS-Bench: ould Your Multimodal LLMs Understand Extremely Large Ultra-High-Resolution Remote Sensing Imagery?
(ACL 2025 main) FR-Spec: Frequency-Ranked Speculative Sampling
Scalable RL solution for advanced reasoning of language models
An open-source framework for collaborative AI agents, enabling diverse, distributed agents to team up and tackle complex tasks through internet-like connectivity.
The code of our paper "InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory"
MiniCPM3-4B: An edge-side LLM that surpasses GPT-3.5-Turbo.
[ACL 2024]Official GitHub repo for OlympiadBench: A Challenging Benchmark for Promoting AGI with Olympiad-Level Bilingual Multimodal Scientific Problems.
Ouroboros: Speculative Decoding with Large Model Enhanced Drafting (EMNLP 2024 main)
MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone
Codes for the paper "∞Bench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718
The most comprehensive database of Chinese poetry 🧶最全中华古诗词数据库, 唐宋两朝近一万四千古诗人, 接近5.5万首唐诗加26万宋诗. 两宋时期1564位词人,21050首词。
Full description can be found here: https://discuss.huggingface.co/t/pretrain-gpt-neo-for-open-source-github-copilot-model/7678?u=ncoop57
Efficient Training (including pre-training and fine-tuning) for Big Models
Efficient, Low-Resource, Distributed transformer implementation based on BMTrain
Few-Shot-Intent-Detection includes popular challenging intent detection datasets with/without OOS queries and state-of-the-art baselines and results.
The dataset for the 2019 Cuneiform Language Identification (CLI) shared task
PyTorch implementation of various methods for continual learning (XdG, EWC, SI, LwF, FROMP, DGR, BI-R, ER, A-GEM, iCaRL, Generative Classifier) in three different scenarios.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Tools for state of the art Knowledge Base Completion.
S2ORC: The Semantic Scholar Open Research Corpus: https://www.aclweb.org/anthology/2020.acl-main.447/
Code associated with the Don't Stop Pretraining ACL 2020 paper