Skip to main content

Primary links

Show — Primary links Hide — Primary links
  • Home
  • Docker
  • Kubernetes
  • Incus
  • Ansible
  • Terraform
  • OpenStack
  • Virtualization
  • Linux
  • SmartHome
  • HowTo

llm

Large language models (LLM)

Ollama
https://ollama.com/download

curl -fsSL https://ollama.com/install.sh | sh
ollama run llama3

Nvidia GPUs / CUDA capability
https://github.com/ollama/ollama/blob/main/docs/gpu.md

Docker
https://hub.docker.com/r/ollama/ollama

Links
https://github.com/ollama/ollama

  • Read more about Large language models (LLM)
Subscribe to llm

Misc

Show — Misc Hide — Misc
  • Installation
  • Grafana
  • Linux
  • Hardware
  • Programming
  • Databases
  • Ansible
  • CLI
  • Multimedia
  • Plugins
  • Windows

Cloud

Show — Cloud Hide — Cloud
  • OpenStack
  • cloud-config
  • nextcloud

Virtualization

Show — Virtualization Hide — Virtualization
  • Virtualization
  • Docker
  • KVM
  • Kubernetes
  • LXC
  • LXD
  • QEMU
  • VMware
  • VirtualBox
  • multipass
  • podman
  • vagrant
  • XEN
  • Incus

Network

Show — Network Hide — Network
  • DNS
  • Firewall
  • Linux
  • OpenvSwitch
  • SSL
  • VLAN
  • VPN
  • iPXE
  • namespaces
  • nmcli
  • tcpdump

Storage

Show — Storage Hide — Storage
  • CEPH
  • DRBD
  • LVM
  • S3
  • ZFS
  • btrfs

Automation / CI/CD

Show — Automation / CI/CD Hide — Automation / CI/CD
  • Install
  • Ansible
  • GitLab
  • Preseed
  • Puppet
  • Terraform
  • Ubuntu autoinstall

Monitoring / Visualisation

Show — Monitoring / Visualisation Hide — Monitoring / Visualisation
  • Icinga
  • Prometheus
  • Monitoring
  • ELK
  • mermaid

Page QR Code

QR code for this page URL

Profiles

  

© panticz 2025

Footer menu

Show — Footer menu Hide — Footer menu
  • Contact
  • Imprint