{"payload":{"header_redesign_enabled":false,"results":[{"id":"351783073","archived":false,"color":"#3572A5","followers":421,"has_funding_file":false,"hl_name":"Xirider/finetune-gpt2xl","hl_trunc_description":"Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed","language":"Python","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":351783073,"name":"finetune-gpt2xl","owner_id":37597043,"owner_login":"Xirider","updated_at":"2023-06-14T02:34:46.365Z","has_issues":true}},"sponsorable":false,"topics":["finetuning","gpt2","huggingface","huggingface-transformers","gpt3","deepspeed","gpt-neo","gpt-neo-fine-tuning"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":53,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253AXirider%252Ffinetune-gpt2xl%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/Xirider/finetune-gpt2xl/star":{"post":"c5enI9G3OJMVmdDCXpBqCondv0EUFWLVvDJX14aiOkJ4_nfCMe1aXD4Ts_bbP-QVUF0KOuEHqYYr9W2wScKadQ"},"/Xirider/finetune-gpt2xl/unstar":{"post":"2Aian3aFcF-tJWq5dQnc6YymCLXKpnCY9bOAh-aUXPxRhi1RaVCB3UPG-yr1DZ52-SqfmmFwjtV7IRnmj4OcdQ"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"bgsz66eZR5ReXPw0MLf2HoO2K0tGcUzVrhcDdAmyI55h1f3JRO0yAlAcQd2goNjGHV3jA1kaD8Yh4WS30rNfuQ"}}},"title":"Repository search results"}