This is the repository holding code and data for "FrugalML: How to Use ML Prediction APIs More Accurately and Cheaply".
-
Updated
Jan 12, 2021 - Python
This is the repository holding code and data for "FrugalML: How to Use ML Prediction APIs More Accurately and Cheaply".
Bringing local LLMs to a Minecraft front-end through commands.
AccIo - Enterprise LLM : Unifying intelligence at your command!
Python-based WebSocket for CLI LLaVA inference.
Mamba for Vision, Perception and Action
Detailed code explanation of google LLM gemini
HugNLP is a unified and comprehensive NLP library based on HuggingFace Transformer. Please hugging for NLP now!😊 HugNLP will released to @HugAILab
Logical verification of probabilistic/language model 'intuitions'.
Inference Llama 2 in one file of pure C
Specify what you want it to build, the AI asks for clarification, and then builds it.
Automating the deployment of the Takeoff Server on AWS for LLMs
creating a workflow to train t5 language models
A platform to test multiple LLM models inside a RAG workflow to choose the best model for embedding and retrieval and the best prompt according to the use case
Experimental autonomous AI LLM & RAG IETF reviewer
A custom framework for easy use of LLMs, VLMs, etc. supporting various modes and settings via web-ui
Plug in and Play Implementation of Tree of Thoughts: Deliberate Problem Solving with Large Language Models that Elevates Model Reasoning by atleast 70%
This repository contains question-answers model as an interface which retrieves answers from vector database for a question. Embeddings or tokenised vector being computed using OpenAI API call which gets inserted into ChromaDB as a RAG. OpenAI API key would be required to run this service.
Add a description, image, and links to the llm-inference topic page so that developers can more easily learn about it.
To associate your repository with the llm-inference topic, visit your repo's landing page and select "manage topics."