Stars
An repository of 2024-2025 AI Safety and Alignment programs, camps, and workshops.
A library for mechanistic interpretability of GPT-style language models
The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery 🧑🔬
The Official Repo for "Quick Start Guide to Large Language Models"
[NeurIPS 2024] Knowledge Circuits in Pretrained Transformers
Transformer models from BERT to GPT-4, environments from Hugging Face to OpenAI. Fine-tuning, training, and prompt engineering examples. A bonus section with ChatGPT, GPT-3.5-turbo, GPT-4, and DALL…
An self-improving embodied conversational agent seamlessly integrated into the operating system to automate our daily tasks.
Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
A curated list of Large Language Model (LLM) Interpretability resources.
The most comprehensive database of Chinese poetry 🧶最全中华古诗词数据库, 唐宋两朝近一万四千古诗人, 接近5.5万首唐诗加26万宋诗. 两宋时期1564位词人,21050首词。
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
Data and tools for generating and inspecting OLMo pre-training data.
Collection of links, tutorials and best practices of how to collect the data and build end-to-end RLHF system to finetune Generative AI models
GPT4All: Run Local LLMs on Any Device. Open-source and available for commercial use.
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Implementing a ChatGPT-like LLM in PyTorch from scratch, step by step
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
Collection of recent methods on (deep) neural network compression and acceleration.
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
MOSAIC: A Prune-and-Assemble Approach for Efficient Model Pruning in Privacy-Preserving Deep Learning.
Retrieval and Retrieval-augmented LLMs
Code for "LoFTR: Detector-Free Local Feature Matching with Transformers", CVPR 2021, T-PAMI 2022
StableLM: Stability AI Language Models