GPU-accelerated speech recognition engine providing industry-leading accuracy with minimal latency.
audio/record.py • Wake word: "Aura"
High-quality neural TTS running entirely on local GPU, providing natural and expressive speech.
voices/backend_voice.py • Model: VCTK/vits
State-of-the-art multimodal model for understanding game scenes and providing visual context.
llm/query.py • Ollama model: qwen2-vl
Fast and efficient LLM providing intelligent game assistance with personality-aware responses.
llm/query.py • Personality: personality_manager.py
Battle-tested OCR engine for extracting game text, UI elements, and puzzle clues.
computer_vision/ocr.py • Preprocessing included
Lightning-fast cross-platform screenshot library with minimal performance impact.
computer_vision/capture.py • Multi-monitor support
# Create virtual environment python -m venv aura-env aura-env\Scripts\activate.bat # Install requirements pip install -r requirements.txt
# Install Ollama from https://ollama.ai # Then pull required models ollama pull mistral ollama pull qwen2-vl # Coqui TTS models download automatically on first run
# Follow NVIDIA RIVA setup guide # Ensure RIVA server is running on localhost:50051
# Start with Blue Prince demo python main_new.py --game blue_prince --input-mode riva --whisper-mode riva # Open admin console in separate terminal python utils/admin_console_system_shock.py # Voice activation: Say "Aura" followed by your command
# Test wake word detection "Aura, can you hear me?" # Test vision capabilities "Aura, what do you see on screen?" # Check admin console for real-time logs