This is suite of the hands-on training materials that shows how to scale CV, NLP, time-series forecasting workloads with Ray.
-
Updated
Feb 13, 2024 - Jupyter Notebook
This is suite of the hands-on training materials that shows how to scale CV, NLP, time-series forecasting workloads with Ray.
AI-based search engine done right
Building Real-Time Inference Pipelines with Ray Serve
A Production-Ready, Scalable RAG-powered LLM-based Context-Aware QA App
Custom AI Generator -- Pretrain your LLM Models with this Automated Embedding Generator and model Q&A Interface. Uses Retrieval Augmented Generation (RAG) to reduce hallucinations and ground the LLM on a source of truth
A drop-in replacement of fastapi to enable scalable and fault tolerant deployments with ray serve
contains the basic structure that a model serving application should have. This implementation is based on the Ray Serve framework.
This MLOps repository contains python modules intended for distributed model training, tuning, and serving using PyTorch and Ray, a distributed computing framework.
Add a description, image, and links to the ray-serve topic page so that developers can more easily learn about it.
To associate your repository with the ray-serve topic, visit your repo's landing page and select "manage topics."