A high-throughput and memory-efficient inference and serving engine for LLMs
-
Updated
Jun 2, 2024 - Python
A high-throughput and memory-efficient inference and serving engine for LLMs
An orchestration platform for the development, production, and observation of data assets.
Run any open-source LLMs, such as Llama 2, Mistral, as OpenAI compatible API endpoint in the cloud.
🦋 A personal research and development (R&D) lab that facilitates the sharing of knowledge.
Apache Airflow - A platform to programmatically author, schedule, and monitor workflows
MLRun is an open source MLOps platform for quickly building and managing continuous ML applications across their lifecycle. MLRun integrates into your development and CI/CD environment and automates the delivery of production data, ML pipelines, and online applications.
Standardized Serverless ML Inference Platform on Kubernetes
ML/AI meta-model, used in MLRun/Iguazio/Nuclio, see qgate-sln-<solution>
Machine Learning Operations with a denoising diffusion model using a butterfly dataset
Qdrant - High-performance, massive-scale Vector Database for the next generation of AI. Also available in the cloud https://cloud.qdrant.io/
Self-Driven Autonomous Python Libraries
Turns Data and AI algorithms into production-ready web applications in no time.
Serve, optimize and scale PyTorch models in production
A curated list of awesome open source libraries to deploy, monitor, version and scale your machine learning
Build applications that make decisions (chatbots, agents, simulations, etc...). Monitor, persist, and execute on your own infrastructure.
RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry
A resuable template for Machine Learning projects
Add a description, image, and links to the mlops topic page so that developers can more easily learn about it.
To associate your repository with the mlops topic, visit your repo's landing page and select "manage topics."