{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"node-llama-cpp","owner":"withcatai","isFork":false,"description":"Run AI models locally on your machine with node.js bindings for llama.cpp. Force a JSON schema on the model output on the generation level","allTopics":["nodejs","cmake","ai","metal","json-schema","grammar","cuda","self-hosted","bindings","llama","cmake-js","prebuilt-binaries","llm","llama-cpp","catai","gguf"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":1,"issueCount":4,"starsCount":688,"forksCount":58,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T13:13:03.130Z"}},{"type":"Public","name":"catai","owner":"withcatai","isFork":false,"description":"Run AI ✨ assistant locally! with simple API for Node.js 🚀","allTopics":["nodejs","ai","chatbot","openai","chatui","vicuna","ai-assistant","llm","chatgpt","dalai","llama-cpp","vicuna-installation-guide","localai","wizardlm","local-llm","catai","ggmlv3","gguf","node-llama-cpp"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":0,"issueCount":1,"starsCount":416,"forksCount":27,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-13T12:00:42.110Z"}}],"repositoryCount":2,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}