Python client library for improving your LLM app accuracy
-
Updated
Jun 11, 2024 - Python
Python client library for improving your LLM app accuracy
Argilla is a collaboration platform for AI engineers and domain experts that require high-quality outputs, full data ownership, and overall efficiency.
⚗️ distilabel is a framework for synthetic data and AI feedback for AI engineers that require high-quality outputs, full data ownership, and overall efficiency.
Official release of InternLM2 7B and 20B base and chat models. 200K context support
Unify Efficient Fine-Tuning of 100+ LLMs
Official Codebase for Rating-Based Reinforcement Learning.
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
Improving LLM truthfulness via reporting confidence
RewardBench: the first evaluation tool for reward models.
AI research lab🔬: implementations of AI papers and theoretical research: InstructGPT, llama, transformers, diffusion models, RLHF, etc...
Language Models Resist Alignment
Achieving Efficient Alignment through Learned Correction
Tracking instruction-tuned LLM openness. Paper: Liesenfeld, Andreas, Alianda Lopez, and Mark Dingemanse. 2023. “Opening up ChatGPT: Tracking Openness, Transparency, and Accountability in Instruction-Tuned Text Generators.” In Proceedings of the 5th International Conference on Conversational User Interfaces. doi:10.1145/3571884.3604316.
Language Modeling Research Hub, a comprehensive compendium for enthusiasts and scholars delving into the fascinating realm of language models (LMs), with a particular focus on large language models (LLMs)
Quality Diversity through Human Feedback: Towards Open-Ended Diversity-Driven Optimization (ICML 2024)
Add a description, image, and links to the rlhf topic page so that developers can more easily learn about it.
To associate your repository with the rlhf topic, visit your repo's landing page and select "manage topics."