Native LLM engine
Optimized pipelines for Android and iOS with predictable startup.
React Native Amaryllis
Native LLMs for mobile apps
Amaryllis is a modern React Native AI module that powers on-device inference, supports text + images, and streams tokens through hooks and observables.
npm install react-native-amaryllis
# or
yarn add react-native-amaryllis
# or
pnpm add react-native-amaryllis
Native performance with a developer friendly API.
Optimized pipelines for Android and iOS with predictable startup.
Send prompts with images for grounded, visual responses.
Use hooks and observables to render partial tokens fast.
Centralize configuration for the entire React Native app.
Fine tune behavior with LoRA adapters on GPU devices.
Cancel and manage sessions with explicit APIs.
Keep performance, memory, and safety predictable.
Show partial tokens early for a responsive UX.
Always cancel async inference in cleanup handlers.
Limit image count and size to protect memory.
Prevent invalid or unsafe native file access.
Surface custom error types and graceful fallbacks.
Track model versions and update strategies.
Wrap your app, then generate from hooks.
import { LLMProvider } from 'react-native-amaryllis';
<LLMProvider
config={{
modelPath: 'gemma3-1b-it-int4.task',
visionEncoderPath: 'mobilenet_v3_small.tflite',
visionAdapterPath: 'mobilenet_v3_small.tflite',
maxTopK: 32,
maxNumImages: 2,
maxTokens: 512,
}}
>
{/* App components */}
</LLMProvider>
import { useInferenceAsync } from 'react-native-amaryllis';
const generate = useInferenceAsync({
onResult: (chunk, isFinal) => {
// Update UI with tokens
},
onError: (err) => setError(err),
});
await generate({ prompt, images });
Sample UI built on top of Amaryllis streaming hooks.