OneDiff: An out-of-the-box acceleration library for diffusion models.
-
Updated
Jun 1, 2024 - Python
OneDiff: An out-of-the-box acceleration library for diffusion models.
🍺 Obrew Server: A local & private Ai inference engine.
A robust and efficient TinyML inference engine.
PyTorch library for cost-effective, fast and easy serving of MoE models.
docs for search system and ai infra
MIVisionX toolkit is a set of comprehensive computer vision and machine intelligence libraries, utilities, and applications bundled into a single toolkit. AMD MIVisionX also delivers a highly optimized open-source implementation of the Khronos OpenVX™ and OpenVX™ Extensions.
带你从零实现一个高性能的深度学习推理库,支持大模型 llama2 、Unet、Yolov5、Resnet等模型的推理。Implement a high-performance deep learning inference library step by step
Python Computer Vision & Video Analytics Framework With Batteries Included
Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM). Powers 👋 Jan
What Happen Next ? Live Inference
FEDML - The unified and scalable ML library for large-scale distributed training, model serving, and federated learning. FEDML Launch, a cross-cloud scheduler, further enables running any AI jobs on any GPU cloud or on-premise cluster. Built on this library, TensorOpera AI (https://TensorOpera.ai) is your generative AI platform at scale.
PygmalionAI's large-scale inference engine
The Qualcomm® AI Hub Models are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
Wingman is the fastest and easiest way to run Llama models on your PC or Mac.
An optimized neural network operator library for chips base on Xuantie CPU.
A common base representation of python source code for pylint and other projects
Repository for OpenVINO's extra modules
Simple first-order logic implementation for .NET.
Friendli: the fastest serving engine for generative AI
The collections of tools for testing and dumping LLMs
Add a description, image, and links to the inference-engine topic page so that developers can more easily learn about it.
To associate your repository with the inference-engine topic, visit your repo's landing page and select "manage topics."