Skip to content

linode/ai-quickstart-mistral

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

AI Quickstart - Mistral 7B LLM

Deploy a complete, pre-configured AI inference stack with both a chat UI and an OpenAI-compatible API endpoint in minutes on Linode GPU instances.

🚀 Quick Start

Get started quickly by deploying to a clean Linode GPU instance. See the Quick Start Guide for step-by-step instructions.

✨ Features

  • One-Click Deployment: Fully automated setup via cloud-init
  • Complete AI Stack: Includes both a web-based chat interface and an OpenAI-compatible API
  • Pre-Configured: NVIDIA drivers, Docker, and all dependencies pre-installed
  • Fast Time-to-Value: From instance boot to working AI in under 5 minutes
  • Model Flexibility: Choose any7model from Hugging Face at deployment
  • OpenAI-Compatible API: Drop-in replacement for OpenAI endpoints—just change your BASE_URL

🏗️ Architecture

AI Quickstart - Mistral LLM consists of two containerized services working together to provide a complete AI inference stack. See the Architecture Documentation for detailed information.

📋 Requirements

  • A Linode GPU instance (any supported GPU instance type)
    • Note: GPU instances require GPU access to be enabled on your account. If you don't see GPU instance types available, please contact Linode Support to enable GPU access.
  • Linode CLI configured with API token (for script-based deployment)

🎯 Use Cases

For AI Explorers

Try the latest open-source models (like Llama 3) in a chat interface without writing code or paying per-token API fees.

For Backend Engineers

Get a stable, OpenAI-compatible API endpoint. Point your existing application to your own endpoint by simply changing the BASE_URL.

For Full-Stack Developers

Use the chat UI to experiment with prompts, then use the same underlying API in your application for consistent results.

🚦 Getting Started

See the Quick Start Guide for detailed deployment instructions. The guide covers:

  • Prerequisites and setup
  • Deploying to a clean Linode GPU instance
  • Accessing your services after deployment
  • Troubleshooting common issues

🔒 Security

⚠️ IMPORTANT: By default, both services are exposed to the internet without authentication. You must configure a Linode Cloud Firewall to protect your services.

See the Security Guide for detailed firewall setup instructions and security best practices.

🛠️ Maintenance

Common maintenance tasks including updating services, changing models, viewing logs, and troubleshooting are covered in the Maintenance Guide.

📝 Limitations (V1)

  • No automatic API authentication (use firewall)
  • No user accounts for the UI (open by default)
  • No automatic HTTPS/SSL
  • Inference only (no fine-tuning support)

📚 Documentation

🤝 Contributing

For issues or feature requests, please open an issue in this repository.

📄 License

This project is licensed under the MIT License. See the LICENSE file for details.


Status: Draft v1.0

About

Quickstart script to start a GPU instance running the Mistral LLM model

Topics

Resources

License

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages