tjake/Jlama
Jlama is a modern LLM inference engine for Java
Jlama helps Java developers integrate large language models (LLMs) directly into their Java applications. It takes various LLM models (like Llama, Gemma, Mistral) as input and allows the application to generate text, complete prompts, or create embeddings. This is ideal for Java application developers who want to add AI capabilities without relying on external services.
1,259 stars. No commits in the last 6 months.
Use this if you are a Java developer building applications and need to embed LLM inference directly within your Java environment for tasks like text generation or content classification.
Not ideal if you are not a Java developer or if you prefer to use pre-built LLM APIs without needing to manage the inference engine yourself.
Stars
1,259
Forks
151
Language
Java
License
Apache-2.0
Category
Last pushed
Oct 12, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/tjake/Jlama"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ludwig-ai/ludwig
Low-code framework for building custom LLMs, neural networks, and other AI models
withcatai/node-llama-cpp
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema...
mudler/LocalAI
:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and...
zhudotexe/kani
kani (カニ) is a highly hackable microframework for tool-calling language models. (NLP-OSS @ EMNLP 2023)
SciSharp/LLamaSharp
A C#/.NET library to run LLM (🦙LLaMA/LLaVA) on your local device efficiently.