A production-ready reference app demonstrating the RunAnywhere React Native SDK capabilities for on-device AI. This cross-platform app showcases how to build privacy-first, offline-capable AI features with LLM chat, speech-to-text, text-to-speech, and a complete voice assistant pipeline—all running locally on your device.
Important: This sample app consumes the RunAnywhere React Native SDK as local workspace dependencies. Before opening this project, you must first build the SDK's native libraries.
# 1. Navigate to the React Native SDK directory
cd runanywhere-sdks/sdk/runanywhere-react-native
# 2. Run the setup script (~15-20 minutes on first run)
# This builds the native C++ frameworks/libraries and enables local mode
./scripts/build-react-native.sh --setup
# 3. Navigate to this sample app
cd ../../examples/react-native/RunAnywhereAI
# 4. Install dependencies
npm install
# 5. For iOS: Install pods
cd ios && pod install && cd ..
# 6a. Run on iOS
npx react-native run-ios
# 6b. Or run on Android
npx react-native run-android
# Or open in VS Code / Cursor and run from thereThis sample app's package.json uses workspace dependencies to reference the local React Native SDK packages:
This Sample App → Local RN SDK packages (sdk/runanywhere-react-native/packages/)
↓
Local XCFrameworks/JNI libs (in each package's ios/ and android/ directories)
↑
Built by: ./scripts/build-react-native.sh --setup
The build-react-native.sh --setup script:
- Downloads dependencies (ONNX Runtime, Sherpa-ONNX)
- Builds the native C++ libraries from
runanywhere-commons - Copies XCFrameworks to
packages/*/ios/Binaries/andpackages/*/ios/Frameworks/ - Copies JNI
.sofiles topackages/*/android/src/main/jniLibs/ - Creates
.testlocalmarker files (enables local library consumption)
- TypeScript SDK code changes: Metro bundler picks them up automatically (Fast Refresh)
- C++ code changes (in
runanywhere-commons):cd sdk/runanywhere-react-native ./scripts/build-react-native.sh --local --rebuild-commons
Download the app from the App Store or Google Play Store to try it out.
This sample app demonstrates the full power of the RunAnywhere React Native SDK:
| Feature | Description | SDK Integration |
|---|---|---|
| AI Chat | Interactive LLM conversations with streaming responses | RunAnywhere.generateStream() |
| Conversation Management | Create, switch, and delete chat conversations | ConversationStore |
| Real-time Analytics | Token speed, generation time, inference metrics | Message analytics display |
| Speech-to-Text | Voice transcription with batch & live modes | RunAnywhere.transcribeFile() |
| Text-to-Speech | Neural voice synthesis with Piper TTS | RunAnywhere.synthesize() |
| Voice Assistant | Full STT → LLM → TTS pipeline | Voice pipeline orchestration |
| Model Management | Download, load, and manage multiple AI models | RunAnywhere.downloadModel() |
| Storage Management | View storage usage and delete models | RunAnywhere.getStorageInfo() |
| Offline Support | All features work without internet | On-device inference |
| Cross-Platform | Single codebase for iOS and Android | React Native + Nitrogen/Nitro |
The app follows modern React Native architecture patterns with a multi-package SDK structure:
┌─────────────────────────────────────────────────────────────────────────┐
│ React Native UI Layer │
│ ┌──────────┐ ┌──────────┐ ┌──────────┐ ┌──────────┐ ┌────────────────┐ │
│ │ Chat │ │ STT │ │ TTS │ │ Voice │ │ Settings │ │
│ │ Screen │ │ Screen │ │ Screen │ │ Assistant│ │ Screen │ │
│ └────┬─────┘ └────┬─────┘ └────┬─────┘ └────┬─────┘ └───────┬────────┘ │
├───────┼────────────┼────────────┼────────────┼───────────────┼──────────┤
│ │ │ │ │ │ │
│ ┌────▼────────────▼────────────▼────────────▼───────────────▼────────┐ │
│ │ @runanywhere/core (TypeScript API) │ │
│ │ RunAnywhere.initialize(), loadModel(), generate(), etc. │ │
│ └──────────────────────────────┬──────────────────────────────────────┘ │
│ │ │
│ ┌───────────────────────┼───────────────────────┐ │
│ │ │ │ │
│ ┌──────▼──────┐ ┌──────▼──────┐ ┌──────▼──────┐ │
│ │@runanywhere │ │@runanywhere │ │ Native │ │
│ │ /llamacpp │ │ /onnx │ │ Bridges │ │
│ │ (LLM/GGUF) │ │ (STT/TTS) │ │ (JSI/Nitro)│ │
│ └──────┬──────┘ └──────┬──────┘ └──────┬──────┘ │
├─────────┼───────────────────────┼───────────────────────┼───────────────┤
│ │ │ │ │
│ ┌──────▼───────────────────────▼───────────────────────▼──────────────┐│
│ │ runanywhere-commons (C++) ││
│ │ Core inference engine, model management ││
│ └──────────────────────────────────────────────────────────────────────┘│
└─────────────────────────────────────────────────────────────────────────┘
- Multi-Package SDK — Core API, LlamaCPP, and ONNX as separate packages for modularity
- TypeScript First — Full type safety across the entire SDK API surface
- JSI/Nitro Bridges — Direct native module communication for performance
- Zustand State Management — Lightweight, performant state for conversations
- Tab-Based Navigation — React Navigation bottom tabs matching iOS/Android patterns
- Theme System — Consistent design tokens across all components
RunAnywhereAI/
├── App.tsx # App entry, SDK initialization, model registration
├── index.js # React Native entry point
├── package.json # Dependencies and scripts
├── tsconfig.json # TypeScript configuration
│
├── src/
│ ├── screens/
│ │ ├── ChatScreen.tsx # LLM chat with streaming & conversation management
│ │ ├── ChatAnalyticsScreen.tsx # Message analytics and performance metrics
│ │ ├── ConversationListScreen.tsx # Conversation history management
│ │ ├── STTScreen.tsx # Speech-to-text with batch/live modes
│ │ ├── TTSScreen.tsx # Text-to-speech synthesis & playback
│ │ ├── VoiceAssistantScreen.tsx # Full STT → LLM → TTS pipeline
│ │ └── SettingsScreen.tsx # Model & storage management
│ │
│ ├── components/
│ │ ├── chat/
│ │ │ ├── ChatInput.tsx # Message input with send button
│ │ │ ├── MessageBubble.tsx # Message display with analytics
│ │ │ ├── TypingIndicator.tsx # AI thinking animation
│ │ │ └── index.ts # Component exports
│ │ ├── common/
│ │ │ ├── ModelStatusBanner.tsx # Shows loaded model and framework
│ │ │ ├── ModelRequiredOverlay.tsx # Prompts model selection
│ │ │ └── index.ts
│ │ └── model/
│ │ ├── ModelSelectionSheet.tsx # Model picker with download progress
│ │ └── index.ts
│ │
│ ├── navigation/
│ │ └── TabNavigator.tsx # Bottom tab navigation (5 tabs)
│ │
│ ├── stores/
│ │ └── conversationStore.ts # Zustand store for chat persistence
│ │
│ ├── theme/
│ │ ├── colors.ts # Color palette matching iOS design
│ │ ├── typography.ts # Font styles and text variants
│ │ └── spacing.ts # Layout constants and dimensions
│ │
│ ├── types/
│ │ ├── chat.ts # Message and conversation types
│ │ ├── model.ts # Model info and framework types
│ │ ├── settings.ts # Settings and storage types
│ │ ├── voice.ts # Voice pipeline types
│ │ └── index.ts # Root navigation types
│ │
│ └── utils/
│ └── AudioService.ts # Native audio recording abstraction
│
├── ios/
│ ├── RunAnywhereAI/
│ │ ├── AppDelegate.swift # iOS app delegate
│ │ ├── NativeAudioModule.swift # Native audio recording/playback
│ │ └── Images.xcassets/ # iOS app icons and images
│ ├── Podfile # CocoaPods dependencies
│ └── RunAnywhereAI.xcworkspace/ # Xcode workspace
│
└── android/
├── app/
│ ├── src/main/
│ │ ├── java/.../MainActivity.kt
│ │ ├── res/ # Android resources
│ │ └── AndroidManifest.xml
│ └── build.gradle
└── settings.gradle
- Node.js 18+
- React Native CLI or npx
- Xcode 15+ (iOS development)
- Android Studio Hedgehog+ (Android development)
- CocoaPods (iOS)
- ~2GB free storage for AI models
# Clone the repository
git clone https://github.com/RunanywhereAI/runanywhere-sdks.git
cd runanywhere-sdks/examples/react-native/RunAnywhereAI
# Install JavaScript dependencies
npm install
# Install iOS dependencies
cd ios && pod install && cd ..# Start Metro bundler
npm start
# In another terminal, run on iOS
npx react-native run-ios
# Or run on a specific simulator
npx react-native run-ios --simulator="iPhone 15 Pro"# Start Metro bundler
npm start
# In another terminal, run on Android
npx react-native run-android# iOS - Build and run
npx react-native run-ios --mode Release
# Android - Build and run
npx react-native run-android --mode releaseThe SDK is initialized in App.tsx with a two-phase initialization pattern:
import { RunAnywhere, SDKEnvironment, ModelCategory } from '@runanywhere/core';
import { LlamaCPP } from '@runanywhere/llamacpp';
import { ONNX, ModelArtifactType } from '@runanywhere/onnx';
// Phase 1: Initialize SDK
await RunAnywhere.initialize({
apiKey: '', // Empty in development mode
baseURL: 'https://api.runanywhere.ai',
environment: SDKEnvironment.Development,
});
// Phase 2: Register backends and models
LlamaCPP.register();
await LlamaCPP.addModel({
id: 'smollm2-360m-q8_0',
name: 'SmolLM2 360M Q8_0',
url: 'https://huggingface.co/prithivMLmods/SmolLM2-360M-GGUF/...',
memoryRequirement: 500_000_000,
});
ONNX.register();
await ONNX.addModel({
id: 'sherpa-onnx-whisper-tiny.en',
name: 'Sherpa Whisper Tiny (ONNX)',
url: 'https://github.com/RunanywhereAI/sherpa-onnx/releases/...',
modality: ModelCategory.SpeechRecognition,
artifactType: ModelArtifactType.TarGzArchive,
memoryRequirement: 75_000_000,
});// Download with progress tracking
await RunAnywhere.downloadModel(modelId, (progress) => {
console.log(`Download: ${(progress.progress * 100).toFixed(1)}%`);
});
// Load LLM model into memory
const success = await RunAnywhere.loadModel(modelPath);
// Check if model is loaded
const isLoaded = await RunAnywhere.isModelLoaded();// Generate with streaming
const streamResult = await RunAnywhere.generateStream(prompt, {
maxTokens: 1000,
temperature: 0.7,
});
let fullResponse = '';
for await (const token of streamResult.stream) {
fullResponse += token;
// Update UI in real-time
updateMessage(fullResponse);
}
// Get final metrics
const result = await streamResult.result;
console.log(`Speed: ${result.tokensPerSecond} tok/s`);
console.log(`Latency: ${result.latencyMs}ms`);const result = await RunAnywhere.generate(prompt, {
maxTokens: 256,
temperature: 0.7,
});
console.log('Response:', result.text);
console.log('Tokens:', result.tokensUsed);
console.log('Model:', result.modelUsed);// Load STT model
await RunAnywhere.loadSTTModel(modelPath, 'whisper');
// Check if loaded
const isLoaded = await RunAnywhere.isSTTModelLoaded();
// Transcribe audio file
const result = await RunAnywhere.transcribeFile(audioPath, {
language: 'en',
});
console.log('Transcription:', result.text);
console.log('Confidence:', result.confidence);// Load TTS voice model
await RunAnywhere.loadTTSModel(modelPath, 'piper');
// Synthesize speech
const result = await RunAnywhere.synthesize(text, {
voice: 'default',
rate: 1.0,
pitch: 1.0,
volume: 1.0,
});
// result.audio contains base64-encoded float32 PCM
// result.sampleRate, result.numSamples, result.duration// 1. Record audio using AudioService
const audioPath = await AudioService.startRecording();
// 2. Stop and get audio
const { uri } = await AudioService.stopRecording();
// 3. Transcribe
const sttResult = await RunAnywhere.transcribeFile(uri, { language: 'en' });
// 4. Generate LLM response
const llmResult = await RunAnywhere.generate(sttResult.text, {
maxTokens: 500,
temperature: 0.7,
});
// 5. Synthesize speech
const ttsResult = await RunAnywhere.synthesize(llmResult.text);
// 6. Play audio (using native audio module)// Get available models
const models = await RunAnywhere.getAvailableModels();
const downloaded = await RunAnywhere.getDownloadedModels();
// Get storage info
const storage = await RunAnywhere.getStorageInfo();
console.log('Used:', storage.usedSpace);
console.log('Free:', storage.freeSpace);
console.log('Models:', storage.modelsSize);
// Delete a model
await RunAnywhere.deleteModel(modelId);
// Clear cache
await RunAnywhere.clearCache();
await RunAnywhere.cleanTempFiles();What it demonstrates:
- Streaming text generation with real-time token display
- Conversation management (create, switch, delete)
- Message analytics (tokens/sec, generation time, time to first token)
- Model selection bottom sheet integration
- Model status banner showing loaded model
Key SDK APIs:
RunAnywhere.generateStream()— Streaming generationRunAnywhere.loadModel()— Load LLM modelRunAnywhere.isModelLoaded()— Check model statusRunAnywhere.getAvailableModels()— List models
What it demonstrates:
- Batch mode: Record full audio, then transcribe
- Live mode: Pseudo-streaming with interval-based transcription
- Audio level visualization during recording
- Transcription metrics (confidence percentage)
- Microphone permission handling
Key SDK APIs:
RunAnywhere.loadSTTModel()— Load Whisper modelRunAnywhere.isSTTModelLoaded()— Check STT model statusRunAnywhere.transcribeFile()— Transcribe audio file- Native audio recording via
AudioService
What it demonstrates:
- Neural voice synthesis with Piper TTS models
- Speed, pitch, and volume controls
- Audio playback with progress tracking
- System TTS fallback support
- WAV file generation from float32 PCM
Key SDK APIs:
RunAnywhere.loadTTSModel()— Load TTS modelRunAnywhere.isTTSModelLoaded()— Check TTS model statusRunAnywhere.synthesize()— Generate speech audio- Native audio playback via
NativeAudioModule(iOS)
What it demonstrates:
- Complete voice AI pipeline (STT → LLM → TTS)
- Push-to-talk interaction with visual feedback
- Model status tracking for all 3 components
- Pipeline state machine (Idle, Listening, Processing, Thinking, Speaking)
- Conversation history display
Key SDK APIs:
- Full integration of STT, LLM, and TTS APIs
AudioService.startRecording()/stopRecording()- Sequential pipeline execution with error handling
What it demonstrates:
- Model catalog with download/delete functionality
- Download progress tracking
- Storage usage overview (total, models, cache, free)
- Generation settings (temperature, max tokens)
- SDK version and backend information
Key SDK APIs:
RunAnywhere.getAvailableModels()— List all modelsRunAnywhere.getDownloadedModels()— List downloaded modelsRunAnywhere.downloadModel()— Download with progressRunAnywhere.deleteModel()— Remove modelRunAnywhere.getStorageInfo()— Storage metricsRunAnywhere.clearCache()— Clear temporary files
# ESLint check
npm run lint
# ESLint with auto-fix
npm run lint:fixnpm run typecheck# Check formatting
npm run format
# Auto-fix formatting
npm run format:fixnpm run unused# Full clean (removes node_modules and Pods)
npm run clean
# Just reinstall pods
npm run pod-installThe app uses console.warn with tags for debugging:
# iOS: View logs in Xcode console or use:
npx react-native log-ios
# Android: View logs with:
npx react-native log-android
# Or filter with adb:
adb logcat -s ReactNative:D| Tag | Description |
|---|---|
[App] |
SDK initialization, model registration |
[ChatScreen] |
LLM generation, model loading |
[STTScreen] |
Speech transcription, audio recording |
[TTSScreen] |
Speech synthesis, audio playback |
[VoiceAssistant] |
Voice pipeline orchestration |
[Settings] |
Storage info, model management |
# Reset Metro cache
npx react-native start --reset-cache
# Clear watchman
watchman watch-del-allFor production builds, configure via environment variables:
# Create .env file (git-ignored)
RUNANYWHERE_API_KEY=your-api-key
RUNANYWHERE_BASE_URL=https://api.runanywhere.ai- Minimum iOS: 15.1
- Bridgeless Mode: Disabled (for Nitrogen compatibility)
- Architectures: arm64 (device), x86_64/arm64 (simulator)
- Minimum SDK: 24 (Android 7.0)
- Target SDK: 36
- Architectures: arm64-v8a, armeabi-v7a
| Model | Size | Memory | Description |
|---|---|---|---|
| SmolLM2 360M Q8_0 | ~400MB | 500MB | Fast, lightweight chat |
| Qwen 2.5 0.5B Q6_K | ~500MB | 600MB | Multilingual, efficient |
| LFM2 350M Q4_K_M | ~200MB | 250MB | LiquidAI, ultra-compact |
| LFM2 350M Q8_0 | ~350MB | 400MB | LiquidAI, higher quality |
| Llama 2 7B Chat Q4_K_M | ~4GB | 4GB | Powerful, larger model |
| Mistral 7B Instruct Q4_K_M | ~4GB | 4GB | High quality responses |
| Model | Size | Description |
|---|---|---|
| Sherpa Whisper Tiny (EN) | ~75MB | English transcription |
| Model | Size | Description |
|---|---|---|
| Piper US English (Medium) | ~65MB | Natural American voice |
| Piper British English (Medium) | ~65MB | British accent |
- ARM64 Preferred — Native libraries optimized for arm64; x86 emulators may have issues
- Memory Usage — Large models (7B+) require devices with 6GB+ RAM
- First Load — Initial model loading takes 1-3 seconds
- iOS Bridgeless — Disabled for Nitrogen/Nitro module compatibility
- Live STT — Uses pseudo-streaming (interval-based) since Whisper is batch-only
See CONTRIBUTING.md for guidelines.
# Fork and clone
git clone https://github.com/YOUR_USERNAME/runanywhere-sdks.git
cd runanywhere-sdks/examples/react-native/RunAnywhereAI
# Install dependencies
npm install
cd ios && pod install && cd ..
# Create feature branch
git checkout -b feature/your-feature
# Make changes and test
npm run lint
npm run typecheck
npm run ios # or npm run android
# Commit and push
git commit -m "feat: your feature description"
git push origin feature/your-feature
# Open Pull RequestThis project is licensed under the Apache License 2.0 - see LICENSE for details.
- Discord: Join our community
- GitHub Issues: Report bugs
- Email: san@runanywhere.ai
- Twitter: @RunanywhereAI
- RunAnywhere React Native SDK — Full SDK documentation
- iOS Example App — iOS native counterpart
- Android Example App — Android native counterpart
- Main README — Project overview
