{"payload":{"header_redesign_enabled":false,"results":[{"id":"690511522","archived":false,"color":"#f1e05a","followers":677,"has_funding_file":false,"hl_name":"RahulSChand/gpu_poor","hl_trunc_description":"Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization","language":"JavaScript","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":690511522,"name":"gpu_poor","owner_id":16897807,"owner_login":"RahulSChand","updated_at":"2023-11-04T23:17:31.643Z","has_issues":true}},"sponsorable":false,"topics":["gpu","pytorch","llama","quantization","language-model","huggingface","llm","llamacpp","ggml","llama2"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":95,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253ARahulSChand%252Fgpu_poor%2B%2Blanguage%253AJavaScript","metadata":null,"csrf_tokens":{"/RahulSChand/gpu_poor/star":{"post":"Ev-MgDzkQ8OMw_4MeoaHM45yTGvka5K4VZJWnQt8skIeJaDvLiyk-ZQ0VxIkyTSqhRTeWf_lv5LXOUHLvfaZAA"},"/RahulSChand/gpu_poor/unstar":{"post":"z5VpMjF-jLsfFIvWSoKVtCFzlH6TRz1O2wfQPmJHMBsztD46NNpDHUyuYnhBlSyy_s-cfnN01wmS_pco5XyrrQ"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"vLB6ZjpYAv-YoSQRBBv6XWpKW1sVBrZz2HqsFGQ_pzGUnAfLNxIxuHa1wkhUwjdaZ5agH23Qenegsl9H3DljyA"}}},"title":"Repository search results"}