A Large Language Model is an artificial intelligence system trained on vast amounts of text data to understand and generate human-like text. Think of it as a sophisticated pattern recognition system that has "read" billions of pages of text and learned the statistical relationships between words and concepts.
Key Concepts:
Privacy: Your data never leaves your computer - no cloud servers, no logging
Cost: No API fees or subscription costs after initial setup
Customization: Full control over model behavior and system prompts
Offline: Works without internet connection
Learning: Understand how AI systems actually work
The "size" of a model refers to the number of parameters (weights) it contains:
| Model Size | Parameters | RAM Required | Example Models |
|---|---|---|---|
| Tiny | 1-2B | 4-8 GB | tinyllama:1.1b, gemma:2b |
| Small | 3-7B | 8-16 GB | phi3:mini, llama3.2:3b, mistral:7b |
| Medium | 13-34B | 24-48 GB | llama2:13b, codellama:34b |
| Large | 70B+ | 80+ GB | llama3.1:70b, mixtral:8x7b |
More parameters ≠ Always better: Smaller, well-trained models can outperform larger ones for specific tasks.
Example:
Input: "The capital of France is"
Model thinks: "Paris" (95%), "Lyon" (2%), "Marseille" (1%)...
Output: "Paris"
Temperature controls the randomness of responses:
Top-K & Top-P Sampling: Additional techniques to control output quality by limiting which tokens can be selected.
Why GPUs are faster:
CPU Processing:
Practical Impact:
❌ "AI understands like humans do"
✅ AI recognizes patterns in text but doesn't "understand" meaning in a human sense
❌ "Bigger models are always better"
✅ Smaller specialized models often perform better for specific tasks
❌ "AI needs GPU to work"
✅ CPU-only operation is perfectly viable, just slower
❌ "AI is always accurate"
✅ Models can "hallucinate" (generate plausible-sounding but incorrect information)
Ready to get started? Check the next tabs for installation and setup instructions!
This guide provides an introduction to running AI models locally on your computer. You'll learn how to work with Large Language Models (LLMs) using tools like Ollama, Jupyter Notebooks, and Python virtual environments.
Running AI models locally offers several advantages: full control over your data, enhanced privacy, no dependency on internet connectivity, and no recurring cloud service costs. This approach is particularly valuable for academic work, research projects, and learning the fundamentals of AI implementation.
| Component | Minimal | Basic | Enthusiast |
|---|---|---|---|
|
RAM
|
4-8 GB | 16 GB | 32 GB+ |
|
GPU VRAM
|
Optional / CPU only | 8 GB | 16 GB+ |
|
Storage
|
10 GB free | 50 GB SSD | 1 TB+ SSD |
|
CPU
|
2+ Cores | 4+ Cores | 8+ Cores |
|
Suitable Models
|
tinyllama:1.1b gemma:2b phi3:mini |
llama3.2:3b mistral:7b codellama:7b |
llama3.1:70b mixtral:8x7b command-r:35b |
Ollama serves as the primary interface for managing and running AI models locally. It simplifies the process of installing, configuring, and executing various language models. This guide demonstrates how to configure a development environment using Visual Studio Code (VS Code) and Jupyter Notebooks for interactive AI experimentation.
The following sections cover:
Each step includes detailed instructions and code examples. Prior programming experience is helpful but not required, as all necessary commands and configurations are provided.
The hardware requirements table provides recommended specifications for different use cases. Performance of AI models is primarily determined by available Random Access Memory (RAM) and, optionally, Graphics Processing Unit (GPU) capabilities.
tinyllama:1.1b, phi3:mini, gemma:2b) operate efficiently on systems with 4-8GB RAM. Response generation is slower compared to GPU-accelerated setups but remains practical for learning and development purposes.These specifications serve as guidelines rather than strict requirements. Entry-level hardware configurations are sufficient for experimentation with compact models and learning fundamental concepts. CUDA installation is only necessary for leveraging NVIDIA GPU acceleration and can be omitted for CPU-based workflows.
Want to skip manual setup? The A1-Terminal project includes automatic installation scripts that handle everything for you!
The installation scripts automatically install:
Perfect for: Beginners who want a working setup immediately, or anyone who prefers using a modern GUI instead of command-line tools.
📖 See the "A1-Terminal" tab above for complete installation instructions and features.
python --versionlocalhost:11434ollama --versionollama serve (automatic on Windows/macOS)Ctrl+Shift+X (Extensions)Ctrl+Shift+P (Command Palette)Python: Create EnvironmentWhy virtual environments? Isolates project dependencies, prevents version conflicts, and keeps your system Python clean.
Open integrated terminal (`Ctrl+``) and run:
pip install ipykernel
This enables Jupyter notebooks to use your virtual environment.
Check if Ollama is running:
ollama list
If not running, start it:
ollama serve
ollama pull tinyllama:1.1b
tinyllama:1.1b (included) or phi3:mini for best balance. Download models in A1-Terminal's "Models" tab or via CLI: ollama pull model-namepython --versionollama listollama listpip list | grep ipykernel🎯 Next Steps: Your environment is ready! Start experimenting with Jupyter notebooks or launch A1-Terminal for a full-featured chat interface.
The installation script handles everything automatically - perfect for beginners or blank systems!
Windows Installation
# 1. Clone repository
git clone https://github.com/Nr44suessauer/A1-Terminal.git
cd A1-Terminal
# 2. Run as Administrator (Right-click → "Run as Administrator")
.\scripts\install.bat
What the install.bat script does:
requirements.txttinyllama:1.1b (~600 MB) for immediate useAfter installation:
cd a1_terminal_modular
.\start.bat
Linux/macOS Installation
# 1. Clone repository
git clone https://github.com/Nr44suessauer/A1-Terminal.git
cd A1-Terminal
# 2. Make executable and run
chmod +x scripts/install.sh
./scripts/install.sh
What the install.sh script does:
After installation:
cd a1_terminal_modular
./start.sh
⏱️ Installation takes 5-10 minutes. Everything is automatic!
For full control over each installation step:
Windows Manual Installation
# 1. Install Ollama manually
# Visit https://ollama.com/download
# 2. Clone repository
git clone https://github.com/Nr44suessauer/A1-Terminal.git
cd A1-Terminal/a1_terminal_modular
# 3. Install Python dependencies
pip install -r requirements.txt
# 4. Download a model (optional)
ollama pull tinyllama:1.1b
# 5. Start application
python main.py
# Or use start script:
# .\start.bat
Linux/macOS Manual Installation
# 1. Install Ollama manually
# Visit https://ollama.com/download
# Or use: curl -fsSL https://ollama.com/install.sh | sh
# 2. Clone repository
git clone https://github.com/Nr44suessauer/A1-Terminal.git
cd A1-Terminal/a1_terminal_modular
# 3. Install Python dependencies
pip install -r requirements.txt
# 4. Download a model (optional)
ollama pull tinyllama:1.1b
# 5. Start application
python3 main.py
# Or use start script:
# ./start.sh
After installation, tinyllama:1.1b is ready. Browse models by category:
Perfect for quick tests, learning, and low-resource systems. Runs smoothly on any CPU with 4-6 GB RAM.
Best quality-to-size ratio. Great for most tasks with 8-12 GB RAM. Industry standard performance.
Optimized for programming tasks: code generation, debugging, completion, and refactoring.
Models with excellent German language support for professional and casual German text generation.
Specialized models for system administration, CLI workflows, scripting, and technical problem-solving.
tinyllama:1.1b (included) or phi3:mini for best balance. Download models in A1-Terminal's "Models" tab or via CLI: ollama pull model-nameA1-Terminal/
scripts/
install.bat # Windows auto-installer
install.sh # Linux/macOS auto-installer
start.bat # Quick start (from root)
a1_terminal_modular/
main.py # Entry point
start.bat # Windows start
requirements.txt # Dependencies
a1_terminal_config.yaml # Config
sessions/ # Saved chats
src/
core/
a1_terminal.py # Main app
ollama_manager.py # API client
ui/
ultimate_ui.py # Modern UI
chat_bubble.py # Messages
session_card.py # Session list
model_selector.py # Model picker
color_wheel.py # Color picker
Auto-created a1_terminal_config.yaml:
# Colors
user_bg_color: "#003300"
user_text_color: "#00FF00"
ai_bg_color: "#1E3A5F"
ai_text_color: "white"
# Fonts
user_font: "Courier New"
ai_font: "Consolas"
# UI
ui_window_width: 1400
ui_window_height: 900
# Options
show_system_messages: true
auto_scroll_chat: true
Ollama Not Running:
ollama list # Check status
ollama serve # Start manually
App Won't Start:
pip install -r requirements.txt --upgrade
python --version # Needs 3.8+
Model Download Failed:
ollama pull <model_name>Ready to start? Run the installation script!
The console provides real-time process monitoring through log outputs, showing the ideal workflow.
Switch between different AI models mid-conversation to compare responses and capabilities.
Configure system prompts (BIAS) to define AI behavior and personality for each session.
Customize your A1-Terminal experience with colors, fonts, and UI preferences in the configuration screen.
The model download interface allows you to browse and install AI models directly from the application.
Access this customization window by clicking the gear icon next to any session in the session list.