Open Source Deep Research Alternative to Reason and Search on Private Data. Written in Python.
-
Updated
May 24, 2025 - Python
8000
Open Source Deep Research Alternative to Reason and Search on Private Data. Written in Python.
MedReason: Eliciting Factual Medical Reasoning Steps in LLMs via Knowledge Graphs
Simple extension on vLLM to help you speed up reasoning model without training.
xVerify: Efficient Answer Verifier for Reasoning Model Evaluations
Pivotal Token Search
Official implementation of the paper "Soft Thinking: Unlocking the Reasoning Potential of LLMs in Continuous Concept Space"
Lightweight replication study of DeepSeek-R1-Zero. Interesting findings include "No Aha Moment", "Longer CoT ≠ Accuracy", and "Language Mixing in Instruct Models".
VeriThinker: Learning to Verify Makes Reasoning Model Efficient
Pure RL to post-train base models for social reasoning capabilities. Lightweight replication of DeepSeek-R1-Zero with Social IQa dataset.
Code for the 2025 ACL publication "Fine-Tuning on Diverse Reasoning Chains Drives Within-Inference CoT Refinement in LLMs"
[arXiv 2025] Can MLLMs Guide Me Home? A Benchmark Study on Fine-Grained Visual Reasoning from Transit Maps
This repository hosts the instructions and workshop materials for Lab 333 - Evaluate Reasoning Models for Your Generative AI Solutions
Agentic Deep Graph Reasoning Implementation
AI Lawyer is an intelligent reasoning legal assistant powered by DeepSeek , Ollama RAG and LangChain, designed to streamline legal research and document analysis. By leveraging retrieval-augmented generation (RAG), it provides precise legal insights, and contract summarization. With an intuitive Streamlit-based UI, analyze legal documents.
State Sandbox is an experimental game for socioeconomic simulation. It uses Large Language Models (o3-mini) to simulate the world and complex policy impacts.
Sudoku4LLM is a Sudoku dataset generator for training and evaluating reasoning in Large Language Models (LLMs). It offers customizable puzzles, difficulty levels, and 11 serialization formats to support structured data reasoning and Chain of Thought (CoT) experiments.
R1-Code-Interpreter: Training LLMs to Reason with Code via Supervised and Reinforcement Learning
LLM reasoning method: combining reflexive cueing with real-time web search and multi-stage analysis for more accurate and explainable answers.
Replication Materials for, "Implicit Bias-Like Patterns in Reasoning Models"
Add a description, image, and links to the reasoning-models topic page so that developers can more easily learn about it.
To associate your repository with the reasoning-models topic, visit your repo's landing page and select "manage topics."