JSPM

Found 23 results for llama.cpp

node-llama-cpp

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

  • v3.8.1
  • 75.37
  • Published

llama.rn

React Native binding of llama.cpp

  • v0.5.11
  • 55.93
  • Published

hyllama

llama.cpp gguf file parser for javascript

  • v0.2.2
  • 46.58
  • Published

cui-llama.rn

Fork of llama.rn for ChatterUI

  • v1.7.0
  • 40.07
  • Published

@electron/llm

Load and use an LLM model directly in Electron. Experimental.

  • v1.1.1
  • 37.30
  • Published

@aibrow/node-llama-cpp

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

  • v1.3.3
  • 35.57
  • Published

inference-server

Libraries and server to build AI applications. Adapters to various native bindings allowing local inference. Integrate it with your application, or use as a microservice.

  • v1.0.0-beta.31
  • 29.60
  • Published

inferra-llama.rn

React Native binding of llama.cpp for Inferra

  • v1.8.0
  • 22.44
  • Published

grammar-builder

A simple grammar builder compatible with GBNF (llama.cpp)

  • v0.0.5
  • 18.08
  • Published

llama.cpp-ts

Node.js bindings for LlamaCPP, a C++ library for running language models.

  • v1.2.0
  • 17.78
  • Published

pllama.rn

React Native binding of llama.cpp

  • v0.4.4
  • 17.28
  • Published

quiad

Run AI models locally on your machine with node.js bindings for llama.cpp. Force a JSON schema on the model output on the generation level

  • v1.3.1
  • 16.73
  • Published

custom-koya-node-llama-cpp

Run AI models locally on your machine with node.js bindings for llama.cpp. Force a JSON schema on the model output on the generation level

  • v0.1.0
  • 15.51
  • Published

llama.native.js

use `npm i --save llama.native.js` to run lama.cpp models on your local machine. features a socket.io server and client that can do inference with the host of the model.

    • v1.1.0
    • 14.35
    • Published

    llama-ggml.js

    serve websocket GGML 4/5bit Quantized LLM's based on Meta's LLaMa model with llama.ccp

      • v0.1.0
      • 11.33
      • Published