Skip to main content
Support Us

AI Features & Privacy

Learn how AssisT handles AI processing with local LLMs and optional cloud APIs while keeping your data private.

Overview

AssisT uses a privacy-first hybrid AI system that prioritizes local processing while offering optional cloud capabilities. Your data stays on your device by default.

Key Principles

  • Local First: AI runs on your computer using Ollama—no internet required
  • No Data Collection: We never see, store, or transmit your data
  • Optional Cloud: Use your own API key if you want cloud AI capabilities
  • Graceful Fallback: Features work even without AI (with reduced functionality)

Local AI with Ollama

AssisT integrates with Ollama, a free, open-source tool that runs AI models directly on your computer.

Why Local AI?

BenefitDescription
PrivacyData never leaves your device
ComplianceSafe for GDPR, FERPA, and HIPAA environments
No CostNo API fees or subscriptions
OfflineWorks without internet connection
SpeedNo network latency for requests

Supported Models

AssisT automatically detects and uses available Ollama models:

ModelSizeBest For
phi3:mini2GBFast responses, basic tasks
llama3.25GBBalanced performance
mistral4GBComplex analysis, detailed responses
llava4GBImage understanding (vision)

Installing Ollama

  1. Download Ollama from ollama.ai
  2. Install and run Ollama on your computer
  3. AssisT will automatically detect it

Installing Models

Once Ollama is running, you can install models directly from AssisT:

  1. Open AssisT settings
  2. Go to AI Settings > Local Models
  3. Click Install next to your preferred model
  4. Wait for the download to complete

Recommended Model Sets:

  • Minimal (2GB): phi3:mini - Fast responses for basic tasks
  • Balanced (5GB): phi3:mini + llama3.2 - Good for most users
  • Full (10GB): All models including vision - Complete AI capabilities

AssisT will recommend a model set based on your system’s available memory.

How Local AI Works

Your Browser (AssisT)

    Message Bridge

Ollama (localhost:11434)

    AI Response

Back to AssisT

All communication happens locally on your machine. Nothing is sent to external servers.

Cloud Providers (Optional)

For users who want more powerful AI capabilities, AssisT supports multiple cloud providers through API keys you provide.

Supported Providers

ProviderStrengthsBest For
Anthropic (Claude)Coding, academic writing, analysisText simplification, tutoring
OpenAI (ChatGPT)Creative, conversationalBrainstorming, general tasks
Google (Gemini)Multimodal, visual, factualImage understanding
PerplexityReal-time web, citationsResearch, fact-checking

Bringing Your Own API Key

  1. Get an API key from your preferred provider:
  2. Open AssisT settings
  3. Go to AI Settings > Cloud Providers
  4. Select your provider and enter your API key
  5. Choose your preferred model

Cost vs Quality

Model TypeCostBest For
Fast (Haiku, GPT-4o-mini, Flash)Cheaper per tokenSimple tasks, high volume
Balanced (Sonnet, GPT-4o, Pro)ModerateMost use cases
Quality (Opus, GPT-4)Higher per tokenComplex tasks, accuracy critical

Tip: Start with faster models for simple tasks. Use larger models when you need more nuanced or accurate responses.

API Key Security

  • Your API keys are stored locally in Chrome’s secure storage
  • They are never sent to Fiavaion servers
  • Only transmitted directly to the provider when you use cloud features
  • You can remove them anytime from settings

How the Hybrid System Works

AssisT intelligently routes requests based on availability:

Feature Request

Is Ollama available?
   ├─ YES → Use local AI (privacy preserved)
   └─ NO → Is cloud API configured?
            ├─ YES → Use cloud AI (with your key)
            └─ NO → Use fallback behavior

Task-Based Model Selection

Different features use the best available model for optimal results:

FeatureLocal ModelRecommended Cloud
Summarizationphi3:mini, llama3.2Any fast model
Text Simplificationllama3.2, mistralAnthropic (clarity)
Socratic TutormistralAnthropic (reasoning)
Image UnderstandingllavaGemini or GPT-4o
Research & CitationsPerplexity (web access)

Fallback Behaviors

When AI isn’t available, features gracefully degrade:

FeatureFallback Behavior
SummarizeShows first paragraph
SimplifyFeature disabled with message
Image DescribeRequires vision model
TTS ProsodyUses neutral tone

Privacy Guarantees

What We Never Do

  • Collect or store your data
  • Send data to our servers
  • Track your AI usage
  • Share information with third parties

What Stays Local

  • All text you process
  • Documents you summarize
  • Images you analyze
  • Conversation history

GDPR/FERPA/HIPAA Compliance

Because AssisT processes everything locally:

  • GDPR: No personal data is transmitted
  • FERPA: Student data stays on the device
  • HIPAA: Patient information never leaves the browser

This makes AssisT safe for educational institutions and healthcare settings.

Performance Tips

For Best Local AI Performance

  1. Use an SSD: Faster model loading
  2. 8GB+ RAM/VRAM: Required for larger models
  3. Keep Ollama Running: Faster first response
  4. Choose Appropriate Models: Match model size to your hardware

Why Memory Matters

  • More VRAM = Better Models: With more video memory (or unified memory on Apple Silicon), you can run larger, more capable models
  • More Memory = Longer Context: Additional memory allows longer context windows—the AI can “remember” more of your document
  • Longer Context = Fewer Hallucinations: When AI sees more context, it makes fewer mistakes because it has more information to work with

Memory Types

TypeWhat MattersNotes
Dedicated GPUVRAM (8GB good, 12GB+ great)NVIDIA/AMD graphics cards
Apple SiliconUnified memory (16GB good, 32GB+ excellent)M1/M2/M3/M4 Macs
CPU-onlySystem RAM (16GB min, 32GB recommended)Slower but works
SetupRAM/VRAMStorageModels
Minimal8GB4GB freephi3:mini
Standard16GB8GB freephi3:mini + llama3.2
Full32GB+15GB freeAll models + longer context

Troubleshooting

Ollama Not Detected

  1. Ensure Ollama is installed and running
  2. Check that it’s accessible at localhost:11434
  3. Restart Ollama if needed
  4. Refresh the AssisT extension

Slow Responses

  1. Try a smaller model (phi3:mini is fastest)
  2. Ensure Ollama isn’t processing other requests
  3. Check your system’s available memory
  4. Close other resource-intensive applications

Model Download Failed

  1. Check your internet connection
  2. Ensure enough disk space is available
  3. Try downloading a smaller model first
  4. Restart Ollama and try again