Open Source LLM toolkit to build trustworthy LLM applications. TigerArmor (AI safety), TigerRAG (embedding, RAG), TigerTune (fine-tuning)
-
Updated
Dec 2, 2023 - Jupyter Notebook
Open Source LLM toolkit to build trustworthy LLM applications. TigerArmor (AI safety), TigerRAG (embedding, RAG), TigerTune (fine-tuning)
[CoRL'23] Adversarial Training for Safe End-to-End Driving
Website to track people, organizations, and products (tools, websites, etc.) in AI safety
Can Large Language Models Solve Security Challenges? We test LLMs' ability to interact and break out of shell environments using the OverTheWire wargames environment, showing the models' surprising ability to do action-oriented cyberexploits in shell environments
Common repository for our readings and discussions
Safe Option Critic: Learning Safe Options in the A2OC Architecture
where I learn and explore mechanistic interpretability of transformers
Add a description, image, and links to the aisafety topic page so that developers can more easily learn about it.
To associate your repository with the aisafety topic, visit your repo's landing page and select "manage topics."