-
Industrial University of Ho Chi Minh City (IUH)
- Go Vap, Ho Chi Minh, Viet Nam
- https://www.kaggle.com/danghuutri
- https://fuutoru.github.io/
Highlights
- Pro
Stars
Implementation of all RAG techniques in a simpler way
Robin: A multi-agent system for automating scientific discovery
Sample code and notebooks for Generative AI on Google Cloud, with Gemini on Vertex AI
Awesome list of 300+ agentic AI resources
linhkid / ArxivDigest-extra
Forked from AutoLLM/ArxivDigestArXiv Digest and Personalized Recommendations using Large Language Models
A collection of sample agents built with Agent Development (ADK)
A library for federated learning (a distributed machine learning process) in an enterprise environment.
VIP cheatsheet for Stanford's CME 295 Transformers and Large Language Models
This repository is created to share current progress of transformer based optical character recognition(OCR). Welcome to share~
Ichigo Whisper is a compact (22M parameters), open-source speech tokenizer for the Whisper-medium, designed to enhance performance on multilingual with minimal impact on its original English capabi…
Collection of awesome LLM apps with AI Agents and RAG using OpenAI, Anthropic, Gemini and opensource models.
Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audio codec.
PDFwhisper allows you to have a conversation with your PDF docs. Finding info on PDF files is now easier than ever. 🚀🔥 Most secure authentication measures using Passkeys🔑
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
Large-scale, Informative, and Diverse Multi-round Chat Data (and Models)
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
A Web Interface for chatting with your local LLMs via the ollama API
streamline the fine-tuning process for multimodal models: PaliGemma 2, Florence-2, and Qwen2.5-VL
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
🏆 A ranked list of awesome machine learning Python libraries. Updated weekly.
hllj / Vistral-V
Forked from haotian-liu/LLaVAVistral-V: Visual Instruction Tuning for Vistral - Vietnamese Large Vision-Language Model.
Aligning pretrained language models with instruction data generated by themselves.
Instruction Fine-Tuning of Meta Llama 3.2-3B Instruct on Kannada Conversations. Tailoring the model to follow specific instructions in Kannada, enhancing its ability to generate relevant, context-a…
Qwen2.5-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.