llama-cpp-capacitor@0.0.22 vulnerabilities

A native Capacitor plugin that embeds llama.cpp directly into mobile apps, enabling offline AI inference with chat-first API design. Supports both simple text generation and advanced chat conversations with system prompts, multimodal processing, TTS, LoRA

  • latest version

    0.0.22

  • latest non vulnerable version

  • first published

    3 months ago

  • latest version published

    2 months ago

  • licenses detected

  • Direct Vulnerabilities

    No direct vulnerabilities have been found for this package in Snyk’s vulnerability database. This does not include vulnerabilities belonging to this package’s dependencies.

    Does your project rely on vulnerable package dependencies?

    Automatically find and fix vulnerabilities affecting your projects. Snyk scans for vulnerabilities (in both your packages & their dependencies) and provides automated fixes for free.

    Scan for indirect vulnerabilities

    No vulnerabilities found in the latest version

    No vulnerabilities found