{"payload":{"header_redesign_enabled":false,"results":[{"id":"682775635","archived":false,"color":"#3572A5","followers":189,"has_funding_file":false,"hl_name":"TUDB-Labs/multi-lora-fine-tune","hl_trunc_description":"Provide Efficient LLM Fine-Tune via Multi-LoRA Optimization ","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":682775635,"name":"multi-lora-fine-tune","owner_id":101306763,"owner_login":"TUDB-Labs","updated_at":"2024-05-12T08:18:55.138Z","has_issues":true}},"sponsorable":false,"topics":["gpu","llama","lora","finetune","peft","baichuan","llm","chatglm","llama2"],"type":"Public template","help_wanted_issues_count":0,"good_first_issue_issues_count":4,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":85,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253ATUDB-Labs%252Fmulti-lora-fine-tune%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/TUDB-Labs/multi-lora-fine-tune/star":{"post":"fVPCgUPX6FnBHDs7UvwNcgwTRgArj34FnbNxQU9k1rW7feWT31tIggj9yGmYJXDnK-mT_njUwknT9iSllJpNHg"},"/TUDB-Labs/multi-lora-fine-tune/unstar":{"post":"fPFC-D0E8v2_R8hcHQZJY4OytTcnQYW9q27VdpELeN1AyHFt0tMm9Fhxy1SBIJ0NOWEUy3vUGj_kyW7scxNsmA"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"we1hvjzUhdCw711YHWyQt3HOWCNS1y1-E4i08_B_OH6qPW94HEwcIcFY5JLni5ctPn3Da5pr_Pnab19-M5QECA"}}},"title":"Repository search results"}