Skip to content

nitinab/terraform-provider-lxd

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

59 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

terraform-provider-lxd

LXD Resource provider for Terraform

Build Status

Prerequisites

Usage

This provider connects to the LXD daemon over local Unix socket or HTTPS.

It makes use of the LXD client library, which currently looks in ~/.config/lxc/ for client.crt and client.key files that must exist to be able to communicate with the LXD daemon.

To generate these files and store them in the LXD daemon, follow these steps.

Example Configurations

Provider (HTTPS)

provider "lxd" {
  scheme  = "https"
  address = "10.1.1.8"
}

Basic Example

This assumes the LXD server has been configured per the LXD documentation, including running lxd init to create a default network configuration.

This example also assumes an image called ubuntu has been cached locally on the LXD server. This can be done by running:

$ lxc image copy images:ubuntu/xenial/amd64 local: --alias=ubuntu

With those pieces in place, you can launch a basic container with:

resource "lxd_container" "test1" {
  name      = "test1"
  image     = "ubuntu"
  ephemeral = false
}

A container can also take a number of configuration and device options. A full reference can be found here. For example, to create a container with 2 CPUs and to share the /tmp directory with the LXD host:

resource "lxd_container" "test1" {
  name      = "test1"
  image     = "ubuntu"
  ephemeral = false

  config {
    limits.cpu = 2
  }

  device {
    name = "shared"
    type = "disk"

    properties {
      source = "/tmp"
      path   = "/tmp"
    }
  }
}

Profiles

Profiles can be used to share common configurations between containers. Profiles accept the same configuration and device options that containers can use.

The order which profiles are specified is important. LXD applies profiles from "left to right", so profile options may be overridden by other profiles.

resource "lxd_profile" "profile1" {
  name = "profile1"

  config {
    limits.cpu = 2
  }

  device {
    name = "shared"
    type = "disk"

    properties {
      source = "/tmp"
      path   = "/tmp"
    }
  }
}

resource "lxd_container" "test1" {
  name      = "test1"
  image     = "ubuntu"
  ephemeral = false
  profiles  = ["default", "${lxd_profile.profile1.name}"]
}

Networks

If you're using LXD 2.3 or later, you can create networks with the lxd_network resource. See this blog post for details about LXD networking and the configuration reference for all network details.

This example creates a standard NAT network similar to what lxd init creates. Containers will access this network via their eth0 interface:

resource "lxd_network" "new_default" {
  name = "new_default"

  config {
    ipv4.address = "10.150.19.1/24"
    ipv4.nat     = "true"
    ipv6.address = "fd42:474b:622d:259d::1/64"
    ipv6.nat     = "true"
  }
}

resource "lxd_profile" "profile1" {
  name = "profile1"

  device {
    name = "eth0"
    type = "nic"

    properties {
      nictype = "bridged"
      parent  = "${lxd_network.new_default.name}"
    }
  }
}

resource "lxd_container" "test1" {
  name      = "test1"
  image     = "ubuntu"
  ephemeral = false
  profiles  = ["${lxd_profile.profile1.name}"]
}

This example creates a second internal network that containers will access via eth1. Containers will use the default profile to gain access to the default network on eth0.

resource "lxd_network" "internal" {
  name = "internal"

  config {
    ipv4.address = "192.168.255.1/24"
  }
}

resource "lxd_profile" "profile1" {
  name = "profile1"

  device {
    name = "eth1"
    type = "nic"

    properties {
      nictype = "bridged"
      parent  = "${lxd_network.internal.name}"
    }
  }
}

resource "lxd_container" "test1" {
  name      = "test1"
  image     = "ubuntu"
  ephemeral = false
  profiles  = ["default", "${lxd_profile.profile1.name}"]

	provisioner "local-exec" {
		command = "lxc exec local:${self.name} dhclient eth1"
	}
}

Finally, LXD networks can be used to create tunnels to other LXD servers. In order to create a tunnel, designate one LXD server as the tunnel "server". This server will offer DHCP leases to the tunnel "client". For example:

resource "lxd_network" "vxtun" {
  name = "vxtun"

  config {
    tunnel.vxtun.protocol = "vxlan"
    tunnel.vxtun.id       = 9999
    tunnel.vxtun.local    = "10.1.1.1"
    tunnel.vxtun.remote   = "10.255.1.1"
    ipv4.address          = "192.168.255.1/24"
    ipv6.address          = "none"
  }
}

For the tunnel client:

resource "lxd_network" "vxtun" {
  name = "vxtun"

  config {
    tunnel.vxtun.protocol = "vxlan"
    tunnel.vxtun.id       = 9999
    tunnel.vxtun.local    = "10.255.1.1"
    tunnel.vxtun.remote   = "10.1.1.1"
    ipv4.address          = "none"
    ipv6.address          = "none"
  }
}

Note how the local and remote addresses are swapped between the two. Also note how the client does not provide an IP address range.

With these resources in place, attach them to a profile in the exact same way described in the other examples.

note: local and remote accept both IPv4 and IPv6 addresses.

Reference

Provider

lxd

Parameters
  • address - Optional - Unix socket file path or IP / FQDN where LXD daemon can be reached. Defaults to /var/lib/lxd/unix.socket
  • scheme - Optional - https or unix. Defaults to unix.
  • port - Optional - https scheme only - The port on which the LXD daemon is listening. Defaults to 8443.
  • remote - Optional - Name of the remote LXD as it exists in the local lxc config. Defaults to local.

Resources

The following resources currently exist:

  • lxd_container - Creates and manages a Container
  • lxd_network - Creates and manages a Network
  • lxd_profile - Creates and manages a Profile

lxd_container

Parameters
  • name - Required -Name of the container.
  • image - Required -Base image from which the container will be created.
  • profiles - Optional -Array of LXD config profiles to apply to the new container.
  • ephemeral - Optional -Boolean indicating if this container is ephemeral. Default = false.
  • privileged- Optional -Boolean indicating if this container will run in privileged mode. Default = false.
  • config - Optional -Map of key/value pairs of container config settings.
  • device - Optional -Device definition. See reference below.
Device Block
  • name - Required -Name of the device.
  • type - Required -Type of the device Must be one of none, disk, nic, unix-char, unix-block, usb, gpu.
  • properties- Required -Map of key/value pairs of device properties.

lxd_network

Parameters
  • name - Required -Name of the network. This is usually the device the network will appear as to containers.
  • config - Optional -Map of key/value pairs of network config settings.
Exported Attributes
  • type - The type of network. This will be either bridged or physical.
  • managed - Whether or not the network is managed.

lxd_profile

Parameters
  • name - Required -Name of the container.
  • config - Optional -Map of key/value pairs of container config settings.
  • device - Optional -Device definition. See reference below.
Device Block
  • name - Required -Name of the device.
  • type - Required -Type of the device Must be one of none, disk, nic, unix-char, unix-block, usb, gpu.
  • properties- Required -Map of key/value pairs of device properties.

Known Limitations

All the base LXD images do not include an SSH server, therefore terraform will be unable to execute any provisioners. A basic base image must be prepared in advance, that includes the SSH server.

To Do

  • Support for using client cert / key from other paths
  • Ability to update container config
  • Ability to exec commands via LXD WebSocket channel
  • Ability to upload files via LXD WebSocket channel
  • Volumes support
  • Add LXD image resource

About

LXD Resource provider for Terraform

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Go 98.3%
  • HCL 1.1%
  • Makefile 0.6%