Stars
Based on paper Stock Embeddings Acquired from News Articles and Price History, and an Application to Portfolio Optimization by X Du et al. (2020)
Code relating to the paper - Stock Embeddings: Learning Distributed Representations for Financial Assets
Distances and representations of persistence diagrams
Material for QuantUniversity talk on Sythetic Data Generation for Finance.
A simple and flexible code for Reservoir Computing architectures like Echo State Networks
Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax.
The Electricity Transformer dataset is collected to support the further investigation on the long sequence forecasting problem.
KErnel OPerationS, on CPUs and GPUs, with autodiff and without memory overflows
PyTorch implementation of Structured State Space for Sequence Modeling (S4), based on Annotated S4.
An extremely fast Python package and project manager, written in Rust.
Code for the paper "A Theoretical Analysis of the Repetition Problem in Text Generation" in AAAI 2021.
[ICLR 2025] LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
HelloBench: Evaluating Long Text Generation Capabilities of Large Language Models
Python code for paper - Variational Deep Embedding : A Generative Approach to Clustering
MTEB: Massive Text Embedding Benchmark
This is the all of the pdf of Head First series
Pytorch implementation of various Knowledge Distillation (KD) methods.
A coding-free framework built on PyTorch for reproducible deep learning studies. PyTorch Ecosystem. 🏆26 knowledge distillation methods presented at CVPR, ICLR, ECCV, NeurIPS, ICCV, etc are implemen…
A Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quantization.
[ACL 2023]: Training Trajectories of Language Models Across Scales https://arxiv.org/pdf/2212.09803.pdf
Understanding how features learned by neural networks evolve throughout training
A small library for automatically adjustment of text position in matplotlib plots to minimize overlaps.
Official code repository for article Intrinsic Dimension Estimation for Robust Detection of AI-Generated Texts
The hub for EleutherAI's work on interpretability and learning dynamics