hitomi-team/sukima
A ready-to-deploy container for implementing an easy to use REST API to access Language Models.
Sukima helps developers easily deploy and manage Language Models as a service. It takes requests to load models and generate or classify text, providing model outputs in return. This is useful for backend developers and MLOps engineers who need to integrate AI language capabilities into their applications.
No commits in the last 6 months.
Use this if you need a straightforward way to expose Language Models via a REST API without complex infrastructure setup.
Not ideal if you require advanced natural language processing tasks beyond basic generation and classification, or need built-in features like autoscaling and HTTPS support out-of-the-box.
Stars
66
Forks
14
Language
Python
License
GPL-2.0
Category
Last pushed
Feb 14, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/hitomi-team/sukima"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
containers/ramalama
RamaLama is an open-source developer tool that simplifies the local serving of AI models from...
av/harbor
One command brings a complete pre-wired LLM stack with hundreds of services to explore.
RunanywhereAI/runanywhere-sdks
Production ready toolkit to run AI locally
runpod-workers/worker-vllm
The RunPod worker template for serving our large language model endpoints. Powered by vLLM.
foldl/chatllm.cpp
Pure C++ implementation of several models for real-time chatting on your computer (CPU & GPU)