Release Notes History for 1.x series

This section details all the available release notes for the 1.x stable series of Juju

The versions covered here are:

juju-core 1.25.5

A stable release of Juju, juju-core 1.25.5, is now available. This release replaces version 1.25.3

Getting Juju

juju-core 1.25.5 is available for Xenial and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows, Centos, and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.25.5

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Jujud offers poodle vulnerable sslv3 on 17070 Lp 1536269

  • Leadership dependency failures in ci deploy tests Lp 1539656

  • Upgrade 1.24.7 -> 1.25.3 fails Lp 1546100

  • Destroyed leader, new leader not elected. Lp 1511659

  • New ec2 korea region Lp 1530957

  • Maas bridge script handles vlan nics incorrectly Lp 1532167

  • 1.25.0: bootstrap failure - warning discarding api open error: eof Lp 1538303

  • Bootstrap node does not use the proxy to fetch tools from streams.c.c Lp 1515289

  • Payload/persistence intermittent failure Lp 1519061

  • 1.25.3 can't bootstrap xenial environments Lp 1550306

  • Enable daily image stream in juju local provider Lp 1554073

  • Juju's maas bridge script is echoed to the console by cloud-init during bootstrap Lp 1536587

  • Cookies file locked for too long Lp 1534643

  • 1.25.4: units attempt to go through the proxy to download charm from state server Lp 1556207

  • Xenial juju 1.25.3 unable to deploy to lxc containers Lp 1557345

  • Provider/maas bridge script is not idempotent Lp 1553915

  • Unable to bootstrap lxd provider on s390x Lp 1554675

  • Gce invalid volume id destroying environment Lp 1556293

  • Juju_availability_zone not set in maas Lp 1559099

  • Handle multi-series charms in 1.25 Lp 1563607

  • Apt-mirror is not used in containers with maas provider Lp 1560391

  • Proxy updater fails with "permission denied" Lp 1564694

  • Can-upgrade-to suggests a downgrade Lp 1319890

juju-core 1.25.3

A new stable release of Juju, juju-core 1.25.3, is now available. This release replaces version 1.25.0.

Getting Juju

juju-core 1.25.3 is available for Xenial and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows, Centos, and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.25.3

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Unit loses network connectivity during bootstrap: juju 1.25.2 + maas 1.9 Lp 1534795

  • "cannot allocate memory" when running "juju run" Lp 1382556

  • Bootstrap with the vsphere provider fails to log into the virtual machine Lp 1511138

  • Add-machine with vsphere triggers machine-0: panic: juju home hasn't been initialized Lp 1513492

  • Using maas 1.9 as provider using dhcp nic will prevent juju bootstrap Lp 1512371

  • Worker/storageprovisioner: machine agents attempting to attach environ-scoped volumes Lp 1483492

  • Restore: agent old password not found in configuration Lp 1452082

  • "ignore-machine-addresses" broken for containers Lp 1509292

  • Deploying a service to a space which has no subnets causes the agent to panic Lp 1499426

  • /var/lib/juju gone after 1.18->1.20 upgrade and manual edit of agent.conf Lp 1444912

  • Juju bootstrap fails to successfully configure the bridge juju-br0 when deploying with wily 4.2 kernel Lp 1496972

  • Incompatible cookie format change Lp 1511717

  • Error environment destruction failed: destroying storage: listing volumes: get https://x.x.x.x:8776/v2//volumes/detail: local error: record overflow Lp 1512399

  • Replica set emptyconfig maas bootstrap Lp 1412621

  • Juju can't find daily image streams from cloud- images.ubuntu.com/daily Lp 1513982

  • Improper address:port joining Lp 1518128

  • Juju status broken Lp 1516989

  • 1.25.1 with maas 1.8: devices dns allocation uses non-unique hostname Lp 1525280

  • Increment minimum juju version for 2.0 upgrade to 1.25.3 Lp 1533751

  • Make assignment of units to machines use a worker Lp 1497312

  • juju environments fails due to missing ~/.juju/current- environment Lp 1506680

  • Juju 1.25 misconfigures juju-br0 when using maas 1.9 bonded interface Lp 1516891

  • Destroy-environment on an unbootstrapped maas environment can release all my nodes Lp 1490865

  • On juju upgrade the security group lost ports for the exposed services Lp 1506649

  • Support centos and windows image metadata Lp 1523693

  • Upgrade-juju shows available tools and best version but did not output what it decided to do Lp 1403655

  • Invalid binary version, version "1.23.3--amd64" or "1.23.3--armhf" Lp 1459033

  • Add xenial to supported series Lp 1533262

juju-core 1.25.0

A new stable release of Juju, juju-core 1.25.0, is now available. This release replaces version 1.24.7.

Getting Juju

juju-core 1.25.0 is available for Wily and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows Centos, and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.25.0

Notable Changes

  • (Experimental) Improved networking features for AWS
    • New 'spaces' and 'subnet' commands
    • New 'spaces' constraints support
  • Support for "devices" on MAAS 1.8+
  • Storage support for GCE and Azure providers
  • Payload management for charmers

(Experimental) Improved networking features for AWS

Juju has experimental support for deploying services on AWS in isolation, taking advantage of the enhanced AWS VPC networking features. This is just a first step towards a full-fledged networking model support in Juju.

New 'spaces' and 'subnet' commands

Juju introduces the 'space' and 'subnet' commands to manage the networks available to services. A Juju network space is a collection of related subnets that have no firewalls between each other and have the same ingress and egress policies.

You can create a new Juju space, and optionally associate existing subnets with it by specifying their CIDRs using the 'create' subcommand. The command looks like this:

juju space create <name> [<CIDR1> <CIDR2> …]

The list of registered spaces can be seen using the 'list' subcommand:

juju space list [--short] [--format yaml|json] [--output <path>]

You can add an existing AWS subnet to a space by CIDR or AWS ID, for example:

juju subnet add <CIDR>|<AWS-subnet-id> <space-name>

You can list all the subnets known by Juju and optionally filtering them by space and/or availability zone like so:

juju subnet list [--zone <name>] [--space <name>] [<zone1> <zone2> …]

For more information about these commands, type:

juju <command> --help

New 'spaces' constraints support

The new 'spaces' constraint instructs Juju to deploy a service's units in subnets of the given space.

Similar to the 'tags' constraint, the 'spaces' constraint takes a comma- delimited list of existing Juju network spaces. Both positive (without prefix) and negative (prefixed by "^") spaces are allowed in the list. For this alpha release, the first positive space name is used, the rest is ignored.

You can constrain a service to a space like this:

juju deploy <charm-URL> [<service-name>] [--constraints "spaces=<list>"]

For more information, run

juju help glossary

and

juju help constraints

Known issues

Deploying a service to a space that has no subnets will cause the agent to panic and is a known issue (https://launchpad.net/bugs/1499426). This issue can be mitigated by always adding at least one subnet to the space.

Support "devices" on MAAS 1.8+

MAAS 1.8 introduced a new feature called "devices". This allows the association of a "device", that requires an IP address, with a parent machine managed by MAAS. There is a view in the MAAS UI showing all devices.

Storage support for GCE and Azure providers

Juju's storage feature can now be used with the Google Compute Engine and Azure providers. Storage is now supported for local, AWS, Openstack, MAAS, GCE and Azure. See https://jujucharms.com/docs/devel/storage for more information.

Status for storage volumes

Volumes now have status associated, so provisioning progress can be observed. List the volumes to see their status:

juju storage volume list

Payload management for charmers

The new payload management feature allows charmers to more accurately define large and complex deployments by registering different payloads, such as LXC, KVM, and docker, with Juju. This lets the operator better understand the purpose and function of these payloads on a given machine.

You define these payloads in the metadata.yaml of the charm under the payloads section. You create a class for the payload, "monitoring" or "kvm- guest", and assign the type.

payloads:
    monitoring:
        type: docker
    kvm-guest:
        type: kvm

From your charm hook you can manage your payload with the following commands:

payload-register <type> <class> <ID> [tags]
payload-unregister <type> <class> <ID>
payload-status-set <type> <class> <ID> <starting, running, stopping, stopped>

From the Juju command line you can view your payloads like this:

  juju payloads <filter>

For more information run:

juju help payloads

Resolved issues

  • Failed configuring a static ip for container "1/lxc/0": cannot allocate addresses: instid not supported Lp 1498982

  • Addresses logged by apiaddressupdater aren't accurate Lp 1497098

  • Server should handle tools of unknown or unsupported series Lp 1403689

  • Provider/gce: devicenames populated in volume attachment info are invalid Lp 1500703

  • Provider/ec2: volumes should be attached with name "xvdf" instead of "xvdf1" by default Lp 1500721

  • Provider/ec2: "iops" should be per-gib in ebs pool config Lp 1501637

  • Provider/ec2: min/max ebs volume sizes are wrong for ssd/iops Lp 1501642

  • Worker/storageprovisioner: worker bounces upon finding invalid volume/filesystem params Lp 1501710

  • Config-get returns scientific notation for large ints Lp 1302118

  • Sending a sigabrt to jujud process causes jujud to uninstall (wiping /var/lib/juju) Lp 1464304

  • Juju-core 1.20.11 crashes (sigsegv) if apiaddresses key is missing from agent.conf Lp 1392814

  • Session closed in data source Lp 1497829

  • Support use of cloud-images mirror for container images Lp 1357045

juju-core 1.24.7

A new stable release of Juju, juju-core 1.24.7, is now available. This release replaces version 1.24.6.

Getting Juju

juju-core 1.24.7 is available for Wily and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows, Centos7, and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.24.7

Notable Changes

The default EC2 instance is m3.medium

AWS deprecated the m1.* instance type. Juju now uses the m3.medium as the default instance type when provisioning machines in EC2.

  • m1.* instances will only be selected when specified using the "instance-type" constraint.

  • When no constraints are specified, the m3.medium type will be selected.

  • When constraints are specified, m1 instance types which satisfy the constraints will be ignored.

Use the "storage" feature flag to try charms that support storage.

Storage was classified as experimental in 1.24, and has been found to be problematic. Since storage is used implicitly by charms with default stores, the feature was placed behind a feature flag. You can test charms that support storage by bootstrapping a new environment with "storage" enabled like so:

JUJU_DEV_FEATURE_FLAG=storage juju bootstrap

Resolved issues

  • Maas provider with maas 1.9 - /etc/network/interfaces "auto eth0" gets removed and bridge is not setup Lp 1494476

  • Switch default instance type from m1.small to t2.small/m3.medium for ec2 provider Lp 1373516

  • Destroy-environment reports warning cannot delete security group Lp 1335885

  • Storage should be behind a feature flag in 1.24 Lp 1503740

  • Configstore should break fslock if time > few seconds Lp 1500613

  • Config-changed error does not cause error state Lp 1494542

  • Setapihostports shouldn't record just one address per api server Lp 1497094

  • Failed worker can result in large number of goroutines and open socket connections and eventually gets picked on by the oom killer Lp 1496750

  • Upgrade in progress reported, but panic happening behind scenes Lp 1493123

  • juju occasionally switches a units public-address if an additional interface is added post-deployment Lp 1435283

juju-core 1.24.6

A new stable release of Juju, juju-core 1.24.6, is now available. This releases replaces 1.24.5.

Getting Juju

juju-core 1.24.6 is available for Wily and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows, Centos7, and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.24.6

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Regression: juju ssh dies with (publickey) Lp 1472632

  • Juju bootstrap fails to work due to incorrect identities restriction. Lp 1486074

  • Cpu-power constraint conflicts with with instance-type when trying to launch a t2.medium Lp 1489142

  • I/o timeout errors can cause non-atomic service deploys Lp 1486553

  • Juju status should show if an upgrade is available Lp 1464633

  • Juju bootstrap failed - cannot dial mongo to initiate replicaset: no reachable servers Lp 1468579

  • Storage/provider: managedfs fails to create fs on full disks Lp 1483086

  • Provider/maas: volume source won't work for physical machines/disks Lp 1483082

  • Upgrade fails if no explicit version is specified Lp 1447899

  • Juju upgrade from 1.24-beta2 to 1.24.5 broken Lp 1493444

  • Cannot remove an environment user with upper case characters Lp 1467037

juju-core 1.24.5

A new stable release of Juju, juju-core 1.24.5, is now available. This release replaces version 1.24.4.

Getting Juju

juju-core 1.24.5 is available for Wily and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows, Centos, and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.24.5

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Deploy hangs: jujud-machine-0 upstart process running but juju status shows it down Lp 1481368

  • Juju uses proxy to access bootstrap node Lp 1478660

juju-core 1.24.4

A new stable release of Juju, juju-core 1.24.4, is now available. This release replaces version 1.24.3.

Getting Juju

juju-core 1.24.4 is available for Wily and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.24.4

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Looping config-changed hooks in fresh juju-core 1.24.3 openstack deployment Lp 1478024

  • Cannot boostrap from custom image-metadata-url or by specifying metadata-source Lp 1452422

  • No output from ssh problems during bootstrap Lp 1477358

  • Leadership unreliable in HA Lp 1461890

  • Juju 1.22.6 cannot upgrade to 1.24.3/1.24.4 Lp 1479931

juju-core 1.24.3

A new stable release of Juju, juju-core 1.24.3, is now available. This release replaces version 1.24.2.

Getting Juju

juju-core 1.24.3 is available for Wily and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows, OS X, and Centos users will find installers at:

https://launchpad.net/juju-core/+milestone/1.24.3

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Upgrade fails on windows machines Lp 1471332

  • Juju 1.24.0: wget cert issues causing failure to create containers on 14.04.2 with lxc 1.07 Lp 1472014

  • Juju environment not usable after the upgrade Lp 1473517

  • Rsyslog connections fail with certificate verification errors after upgrade to 1.24.2 Lp 1474614

  • Rebooting the virtual machines breaks juju networking Lp 1474508

  • Failed upgrade, mixed up ha addresses Lp 1463480

  • Debug-log does not work with local provider Lp 1464335

  • Jujud hanging after upgrading from 1.24.0 to 1.24.1(and 1.24.2) Lp 1468653

  • Juju bootstrap fails - waiting for api to become available error cannot get all blocks: eof Lp 1468581

  • Tools migration fails when upgrading 1.20.14 to 1.24.1 on ec2 Lp 1469130

  • State server seems to have died Lp 1469199

  • Github.com/juju/utils has contradictory licences Lp 1472749

  • Transient error with juju-unit-agent on windows hosts Lp 1430245

  • Agent shutdown can cause cert updater channel already closed panic Lp 1472729

  • Juju 1.24 memory leakage Lp 1474195

  • $set updates may clear out the env-uuid field Lp 1474606

  • Ec2: provisioning machines sometimes fails with "tagging instance: the instance id does not exist" Lp 1474788

  • When the uniter fails to run an operation due to an error, the agent state is not set to "failed" Lp 1475163

juju-core 1.24.2

A new stable release of Juju, juju-core 1.24.2, is now available. This release replaces version 1.24.0.

Getting Juju

juju-core 1.24.2 is available for Wily and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows, Centos, and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.24.2

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Destroy-machine --force no longer forceably destroys machine Lp 1464616

  • Inconsistent juju status from cli vs api Lp 1467690

  • Dhcp addresses for containers should be released on teardown Lp 1348663

  • "relation-set --file -" doesn't seem to work Lp 1454678

  • Debug-hooks no longer works with 1.24-beta6 Lp 1466167

  • Panic if jujud restarts while action is running Lp 1447895

  • Mysterious env.lock held message Lp 1454323

  • Upgrade fails moving syslog config files "invalid argument" Lp 1468815

  • Leadership settings documents written out without env-uuid field Lp 1468994

  • Debug-log does not work with local provider Lp 1464335

  • A subordinate charm hook scheduled to run(but waiting for the principal charm hook to release the lock) goes to an error state after the principal charm hook triggers a reboot. Lp 1464470

  • Upgrading 1.20.14 -> 1.24.0 fails Lp 1466969

  • Subordinate charm action data not reported by api Lp 1457205

  • Centos fails to bootstrap using juju 1.24 Lp 1470150

juju-core 1.24.0

A new stable release of Juju, juju-core 1.24.0, is now available. This release replaces version 1.23.3.

Getting Juju

juju-core 1.24.0 is available for vivid and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.24.0

Notable Changes

  • VMWare (vSphere) Provider
  • Resource Tagging (EC2, OpenStack)
  • MAAS root-disk Constraint
  • Service Status
  • CentOS 7 Preview
  • Storage (experimental)

VMWare (vSphere) Provider

Juju now supports VMWare's vSphere ("Software-Defined Data Center") installations as a targetable cloud. It uses the vSphere API to interact with the vCenter server. The vSphere provider uses the OVA images provided by Ubuntu's official repository. API authentication credentials, as well as other config options, must be added to your environments.yaml file before running 'juju bootstrap'. The different options are described below.

The basic config options in your environments.yaml will look like this:

my-vsphere:
  type: vsphere
  host: <192.168.1.10>
  user: <some-user>
  password: <some-password>
  datacenter: <datacenter-name>
  external-network: <external-network-name>

The values in angle brackets need to be replaced with your vSphere information. 'host' must contain the IP address or DNS name of vSphere API endpoint. 'user' and 'password' are fields that must contain your vSphere user credentials. 'datacenter' field must contain the name of your vSphere virtual datacenter. 'external-network' is an optional field. If set, it must contain name of the network that will be used to obtain public IP addresses for each virtual machine provisioned by juju. An IP pool must be configured in this network and all available public IP addresses must be added to this pool. For more information on IP pools, see official documentation:

https://pubs.vmware.com/vsphere-51/index.jsp?topic=2Fcom.vmware.vsphere.vm_admin.doc%2FGUID-5B3AF10D-8E4A-403C-B6D1-91D9171A3371.html

NOTE that using the vSphere provider requires an existing vSphere installation. Juju does not set up vSphere for you. The OVA images we use support VMWare's Hardware Version 8 (or newer). This should not be a problem for most vSphere installations.

Resource Tagging (EC2, OpenStack)

Juju now tags instances and volumes created by the EC2 and OpenStack providers with the Juju environment UUID. Juju also adds any user- specified tags set via the "resource-tags" environment setting. The format of this setting is a space-separated list of key=value pairs:

resource-tags: key1=value1 [key2=value2 ...]

These tags may be used, for example, to set up chargeback accounting. Any tags that Juju manages will be prefixed with "juju-"; users must avoid modifying these.

Instances and volumes are now named consistently across EC2 and OpenStack, using the scheme "juju---", where is the human-readable name of the environment as specified in environments.yaml; is the type of the resource ("machine" or "volume") and is the numeric ID of the Juju machine or volume corresponding to the IaaS resource.

MAAS root-disk Constraint

The MAAS provider now honours the root-disk constraint, if the targeted MAAS supports disk constraints. Support for disk constraints was added to MAAS 1.8.

Service Status

Juju provides new hooks for charm authors to report service status, and 'juju status' now includes the service status. This new functionality allows charms to explicitly inform Juju of their status, rather than Juju guessing. Charm authors have access to 2 new hook tools, and the status report includes more information.

The 'status-set' hook tool allows a charm to report its status to Juju. This is known as the workload status and is meant to reflect the state of the software deployed by the charm. Charm authors are responsible for setting the workload's status to "Active" when the charm is ready to run its workload, and "Blocked" when it needs user intervention to resolve a problem.

status-set:
    status-set <maintenance | blocked | waiting | active> "message"

The 'status-get' hook tool allows a charm to query the current workload status recorded in Juju. Without arguments, it just prints the workload status value eg maintenance. With '--include-data' specified, it prints YAML which contains the status value plus any data associated with the status.

status-get:
    status-get [--include-data]

Charms that do not make use of these hook tools will still work as before, but Juju will not provide details about the workload status.

The above commands set the status of the individual units. Unit leaders may also set and get the status of the service to which they belong:

print the status of all units of the service and the service itself: status-get --service

set the status of the service: status-set --service "message"

The 'juju status' command includes the 'workload-status' and 'service-status' in the report. for example:

services:
 ...
  wordpress:
    charm: local:trusty/wordpress-93
    exposed: false
    service-status:                    <-- new service status
      current: blocked
      message: Waiting for database
      since: 01 May 2015 17:39:38 AEST
    relations:
      loadbalancer:
      - wordpress
    units:
      wordpress/0:
        workload-status:                <-- new workload status
          current: blocked
          message: Waiting for database
          since: 01 May 2015 17:39:38 AEST
        agent-status:                   <-- new agent status
          current: idle
          since: 01 May 2015 17:39:44 AEST
          version: 1.24-alpha1.1
        agent-state: started            <-- legacy Juju agent state
        agent-version: 1.24-alpha1.1
        machine: "1"
        open-ports:
        - 80/tcp
        public-address: 23.20.250.14

Juju aggregates all the unit 'workload-status' values to represent the 'service-status'. The 'service-status' value is derived from the worst case status of all the units; eg. if any unit is in error, then the service is in error.

The 'status' command will use a table layout in the future, and you can set the environmental variable 'JUJU_CLI_VERSION' to "2" to see it like so:

export JUJU_CLI_VERSION=2
juju status

NAME       STATUS  EXPOSED CHARM
mysql      unknown false   local:trusty/mysql-326
wordpress  blocked false   local:trusty/wordpress-93

The legacy status values are omitted from output. You can use the '--yaml' option to see status in the Juju 1.x layout.

Juju also records a history of status changes for a unit, and tracks the time when the status was last updated. The 'juju status-history' command allows you to inspect a charm's status changes over time

juju status-history [options] [-n N] <unit>

options:
-e, --environment (= "")
   juju environment to operate in
-n  (= 20)
   size of logs backlog.
--type (= "combined")
   type of statuses to be displayed [agent|workload|combined].
--utc  (= false)
   display time as UTC in RFC3339 format

This command will report the history of status changes for
a given unit.
The statuses for the unit workload and/or agent are available.
-type supports:
   agent: will show statuses for the unit's agent
   workload: will show statuses for the unit's workload
   combined: will show agent and workload statuses combined
and sorted by time of occurrence.

For example, to see the history of the unit wordpress/0

juju status-history wordpress/0

TIME                       TYPE     STATUS       MESSAGE
01 May 2015 17:33:20+06:00  workload  unknown      Waiting for agent initialization to finish
01 May 2015 17:33:20+06:00  agent     allocating
01 May 2015 17:36:37+06:00  agent     executing    running install hook
01 May 2015 17:36:37+06:00  workload  maintenance  installing charm software
01 May 2015 17:36:38+06:00  workload  maintenance  installing dependencies
01 May 2015 17:39:11+06:00  workload  maintenance  installing components
01 May 2015 17:39:18+06:00  agent     executing    running leader-elected hook
01 May 2015 17:39:18+06:00  agent     executing    running config-changed hook
01 May 2015 17:39:19+06:00  workload  maintenance  configuring nginx
01 May 2015 17:39:34+06:00  workload  maintenance  restarting services
01 May 2015 17:39:38+06:00  workload  blocked      Waiting for database
01 May 2015 17:39:39+06:00  agent     executing    running start hook
01 May 2015 17:39:44+06:00  agent     idle

CentOS 7 Preview

Juju 1.24.0 has initial CentOS support. This is experimental and has a number of known issues. However, most of the functionality of Juju should be there and ready to be used and tested. CentOS should be deployable on any cloud that supports cloud-init in it's CentOS images. It is possible to use CentOS as both a state machine (taking the limitations at the bottom into account) and as a normal machine.

Deploying a charm on CentOS is no different than deploying one on Ubuntu or Windows. The only thing that needs to change is the series which is "centos7". For example, from Launchpad:

juju deploy lp:~me/centos7/charm

or Locally:

juju deploy --repository=/home/user/charms local:centos7/charm

However there are no charms currently available for CentOS. The process or writing one should be no different from the Ubuntu charms besides keeping in mind the fact that one shouldn't use Ubuntu specific calls (such as apt).

There is a guide for setting up a MaaS environment using CentOS at

http://wiki.cloudbase.it/juju-centos

Note that Centos 7 agents are already in streams. There is no need install Go, compile, tar, and running juju metadata. You can sync the streams to a web server visible to your Juju environment.

mkdir local
juju sync-tools --local-dir local
cp -r local/tools <path/to/webserver>

Some of the known issues are:

  • Containers are not yet supported

  • There is a lack of mongo tools at the moment so any functionality depending on those is not available(for example backups)

  • There is no way to currently specify a proxy or mirror for yum in the environment configuration. The values that you specific for apt packages will be used for yum packages as well. This limitation will be fixed as soon as possible.

Storage (experimental)

Juju now models storage, charms can request storage (volumes and filesystems), and you can specify constraints on how to satisfy those requests (which provider, what options, how large, how many).

Initially, Juju supports native volumes for the EC2 (EBS) and OpenStack (Cinder), and MAAS providers. Juju also supports several cloud-independent storage providers: tmpfs, loop (loop devices), root filesystem. Future versions of Juju will extend this set with providers for Ceph, NFS, and others.

The storage feature is experimental: it has some known caveats, and has not yet been battle hardened. Instructions on use and caveats are documented at https://jujucharms.com/docs/devel/wip-storage.

Storage (experimental) MAAS Provider Support

The MAAS provider now supports storage. Storage directives are used to select machines which have the requisite number and size of volumes available on the machine (usually physical volumes such as SSD or magnetic drives).

Storage pools may be created to select volumes with specified tags.

juju storage create pool maas-ssd maas tags=ssd

The above creates a pool called "maas-ssd" and when used, will select volumes tagged in MAAS with the "ssd" tag. Tags may be a comma separated list.

Then to deploy a charm:

juju deploy mycharm --storage data=maas-ssd,50G

The above deploys a charm to a MAAS machine with the data store mounted on a volume at least 50GiB in size with the tag "ssd".

It is also possible to specify the size of the root-disk using the root disk constraint. This works the same way as for the AWS provider

juju deploy mysql --constraints root-disk=50G

Storage directives and root disk constraints may be combined.

juju deploy mysql --constraints root-disk=50G --storage data=maas-ssd,500G

NOTE: the root disk support has only just landed in MAAS trunk. the Juju/MAAS storage support has been smoke tested using the NEC MAAS test lab. It needs much more extensive testing!

NOTE: when using MAAS which does not support storage, if MAAS storage is requested, an error is returned and the node is cleaned up.

The storage feature is experimental: it has some known caveats, and has not yet been battle hardened. Instructions on use and caveats are documented at https://jujucharms.com/docs/devel/wip-storage.

Storage (experimental) Unit Placement

It is now possible to deploy units with storage to existing machines. This applies when using storage that is dynamically created, such as EBS/Cinder volumes, loop devices, TMPFS, rootfs. It can't be used with machine volumes on MAAS, but can be used to deploy charms to an existing MAAS machine if a dynamic storage source is specified. eg.

juju deploy charm --to 2 --storage data=loop,2G

An Openstack deployment example:

juju deploy charm --to 2 --storage data=cinder,2G

Resolved issues

  • Deploying into kvm with local provider, hostnames are not unique Lp 1326091

  • Juju status complains "config has no uuid" when no .jenv is present Lp 1436925

  • Agent panic on maas network with uppercase characters Lp 1446608

  • Debug-log spammed with leader-election messages Lp 1437015

  • Eu-central-1 aws region v4 signing required and not supported Lp 1447841

  • Environment variables are not propagated to jujud on vivid Lp 1449436

  • Rsyslog-gnutls is not installed when enable-os-refresh-update is false Lp 1424892

  • Juju stop responding after juju-upgrade Lp 1438489

  • Jujud won't start if apt-get of juju-mongodb package fails Lp 1441904

  • Juju cli compatibility option Lp 1450701

  • Upgrade from 1.18 to 1.23 fails: password for machine agent can't be set Lp 1451297

  • Broken db field ordering when upgrading to juju compiled with go 1.3+ Lp 1451674

  • Log files are lost when agents are restarted under systemd Lp 1452113

  • Worker/uniter: charm does not install properly if storage isn't provisioned before uniter starts Lp 1452207

  • Jujud does not restart after upgrade-juju on systemd hosts Lp 1452511

  • Maas provider chokes on unexpected device labels Lp 1452725

  • Juju 1.23-beta4 introduces ssh key bug when used w/ dhx Lp 1444861

  • Retry-provisioning launches instances for containers; cannot retry containers at all Lp 1428439

  • Debug-hooks exit 1 , doesn't mark hook as failed Lp 1415176

  • Default storage constraints are not quite correct Lp 1452535

  • Maas: implement root-disk constraint Lp 1212689

  • Juju upgrade-juju failed to configure mongodb replicasets Lp 1441913

  • Provider/openstack: volumes are recorded with 0 size Lp 1450740

  • Persist iptables rules / routes for addressable containers across host reboots Lp 1442012

  • Deployer sometimes fails with a unit status not found error Lp 1451283

  • Logs don't rotate Lp 1452285

  • Collect-metric hook failure, metric batch uuid already exists Lp 1452487

  • Juju machine service for windows creates incorrect tools symlinks Lp 1453280

  • /var/spool/rsyslog grows without bound Lp 1453801

  • Instancepoller compares wrong address list and always requests updated state addresses Lp 1454043

    • Agents see "too many open files" errors after many failed api attempts Lp 1420057
  • Container destruction doesn't mark ip addresses as dead Lp 1441206

  • Debug-hooks not working with manually provisioned machines Lp 1429790

  • Joyent machines get stuck in provisioning Lp 1446264

  • Cannot process charms: finding charm revision info: cannot get metadata from the charm store: verification failed: no macaroon cookies in request Lp 1453238

  • Juju log spams error juju.worker.diskmanager lsblk.go:111 error checking if "sr0" is in use: open /dev/sr0: no medium found Lp 1454481

  • Failed to retrieve the template to clone - 500 internal server error - error creating container juju-trusty-lxc-template - Lp 1454676

  • 1.20.x client cannot communicate with 1.22.x env Lp 1454829

  • Received disconnect from ...: 2: too many authentication failures for ubuntu Lp 1456315

    • Lxc network.mtu setting not set consistently across hosts Lp 1442257
  • Transaction collection (txns) grows without bound Lp 1453785

  • Firewaller gets an exception if a machine is not provisioned Lp 1454599

  • Jujud leaking file handles Lp 1454697

  • Destroying a machine with a placed unit errors then destroys anyway. Lp 1455158

  • Intermittent panic: rescanned document Lp 1449054

  • juju upgrade-juju --upload-tools leaves local environment unusable Lp 1457728

  • warning: log line attempted over max size - leadership related Lp 1457645

  • Debug-log eof: invalid tag and panic seen in state server logs Lp 1461354

  • Juju has conf files in /var/log/juju on instances Lp 1370896

  • Juju restore fails with "/var/lib/juju/agents: no such file or directory" Lp 1431372

  • Cannot destroy-environment with 'juju user add' .jenv file Lp 1403165

  • Juju cannot create vivid containers Lp 1442308

  • State: availability zone upgrade fails if containers are present Lp 1441478

  • 1.23.2.1, mongo: document is larger than capped size Lp 1454891

  • "relation-set --file -" doesn't seem to work Lp 1454678

  • Maas provider doesn't know about "failed deployment" instance status Lp 1376246

  • Juju status --utc does not display utc and is confused Lp 1459611

  • Juju agent opens api when upgrade is pending Lp 1459912

  • 'juju status' wall clock timestamps should display offset not tz abbrev Lp 1459616

  • Harvest mode setting not used by container provisioner Lp 1459885

  • Allow status-set/get to a service by its leader unit Lp 1461111

  • Unit storage add should only accept count as constraint Lp 1461342

  • Golang.org/x/sys/windows requires go 1.4 Lp 1463439

  • Erroneous juju user data on windows for juju version 1.23 Lp 1451626

  • Cmd/juju/storage: "add" fails to dynamically add filesystem for storage Lp 1462146

  • Worker/diskmanager sometimes goes into a restart loop due to failing to update state Lp 1461871

  • Juju 1.24-beta6.1 unit commands in debug-hooks hang indefinitely Lp 1463117

juju-core 1.23.3

A new stable stable release of Juju, juju-core 1.23.3, is now available. This release replaces version 1.23.2.

Getting Juju

juju-core 1.23.3 is available for vivid and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.23.3

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Broken db field ordering when upgrading to juju compiled with go 1.3+ Lp 1451674

  • Log files are lost when agents are restarted under systemd Lp 1452113

  • Jujud does not restart after upgrade-juju on systemd hosts Lp 1452511

  • Environment variables are not propagated to jujud on vivid Lp 1449436

  • Local charms are not added to storage on upgrade to 1.22.x Lp 1447853

  • Jujud won't start if apt-get of juju-mongodb package fails Lp 1441904

  • Upgrade from 1.18 to 1.23 fails: password for machine agent can't be set Lp 1451297

  • Juju 1.23-beta4 introduces ssh key bug when used w/ dhx Lp 1444861

  • Debug-hooks exit 1 , doesn't mark hook as failed Lp 1415176

  • Unnecessary errors emitted during init system discovery Lp 1452114

juju-core 1.23.2

A new stable release of Juju, juju-core 1.23.2, is now available. This release replaces version 1.22.1.

Getting Juju

juju-core 1.23.2 is available for vivid and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.23.2

Notable Changes

  • New Support for Google Compute Engine (GCE)
  • Support for systemd (and Vivid)
  • New Style Restore
  • Improved Proxy Support for Restrictive Networks
  • New Charm Actions
  • New blocks and Messages.
  • Experimental: Service Leader Elections
  • Experimental: Addressable LXC Containers and KVM Instances on AWS and MAAS

New Support for Google Compute Engine (GCE)

A new provider has been added that supports hosting a Juju environment in GCE. This feature leverages the support for Ubuntu cloud-images that GCE added late 2014. It uses Google's GCE API to interact with your account there. API authentication credentials, as well as other config options, must be added to your environments.yaml file before running 'juju bootstrap'. The different options are described below.

The basic config options in your environments.yaml will look like this:

yaml my-gce: type: gce project-id: <your-project-id> auth-file: <path-to-json-file-from-google> private-key: <your-private-key> client-email: <your-client-email> client-id: <your-client-id>

The values in angle brackets need to be replaced with your GCE information.

'project-id' must identify a GCE project that already exists before you run "juju bootstrap". This means creating a new one through the developer console (https://console.developers.google.com/project) before bootstrapping Juju. To make it easier to quickly identify in your GCE console, we recommend that the name start with 'juju-' and that it include the environment name you are planning to use. You could also use an existing project but we recommend against that if possible. Using a new project will make it easier for you to manage the environment's resources as well as to track the environment's cost and resource usage.

'auth-file' is the path to a JSON file you get from your GCE console that contains all the authentication credentials it needs. After you have created your project, download the file by pressing the 'Generate new JSON key' button in the 'Service Account' section of

https://console.developers.google.com/project//apiui/credential

If necessary, instead of 'auth-file' you may set the 'private-key', 'client-email', and 'client-id' fields in your environments.yaml. They are your GCE OAuth credentials (matching those in the JSON file). You may need to change "\u003d" in the private key to "=".

Either way, these details are associated with the 'service account' of the GCE project you will use for your Juju environment. For each GCE project, a service account is set up automatically when you create the project. Juju uses that service account to connect to the GCE API and does so with the proper authentication scope.

For more information please refer to

https://developers.google.com/accounts/docs/OAuth2ServiceAccount#creatinganaccount

and

https://developers.google.com/accounts/docs/OAuth2#serviceaccount

If the project's service account has any permissions problems go to the following page to fix them:

https://console.developers.google.com/project//permissions

(remember to insert the project id)

The GCE API should already be activated for the project. It it isn't, go to the following URL in your console:

https://console.developers.google.com/project//apiui/api

(insert the project name)

Also see step 2 on

https://cloud.google.com/compute/docs/api/how-tos/authorization.

The following config options in your environments.yaml file are optional:

region - (default us-central1) The location to place the
       environment.
image-endpoint - (default https://www.googleapis.com) This is
       where Juju will look for disk images when provisioning a
       new instance on GCE.

All Juju 1.23 provider capabilities are available for GCE except for networking.

Support for systemd (and Vivid)

In addition to upstart, Juju now supports Ubuntu hosts using systemd as their init system.

Support for systemd allows Juju to run on Ubuntu 15.04 (Vivid Vervet), which is the first Ubuntu release to boot with systemd. This means you can bootstrap Juju on a Vivid host. Note that the charm store (jujucharms.com) only support LTS releases. You can develop and test vivid charms in a local charm repository.

New Style Restore

You can now restore a backup with the new 'backups restore' command, which is more reliable and fast. New restore supports backups generated with the deprecated Juju backup plugin and with the recently added 'juju backups create' command. You can restore from a local backup file like so:

bash juju backups restore [-b] --file <backup file>

Which will optionally bootstrap a new state server, upload a backup file and restore it. The -b option will fail if there is a running state server.

You can also restore from a backup stored on the state-server:

bash juju backups restore --id <on server backup id>

To obtain a list of the existing backups in the state-server you can use:

bash juju backups list

Improved Proxy Support for Restrictive Networks

A few of issues around HTTP/HTTPS and apt proxy support were fixed (Lp 1403225, Lp 1417617). Charm downloads from the charm store which could not be completed due to connectivity issues are now retried every few minutes rather than once every 24 hours. Proxy settings from the environment (http-proxy, https-proxy, ftp-proxy, apt-http-proxy, apt-https-proxy, apt-ftp-proxy, and no-proxy) are properly propagated to all machines, and Juju agents use them for all external connectivity. The juju run command also uses proxy settings when defined, as well as debug-hooks and all hooks the a charm runs. You can specify one or more proxy settings via environment variables (http_proxy, https_proxy, etc.) or inside your environments.yaml. Other related proxies are configured as needed (e.g. you can specify just http-proxy, and that will also be used for https, ftp, and apt proxies).

New Charm Actions

Juju charms can describe actions that users can take on deployed services. These actions are scripts that can be triggered on a unit by the via the Juju CLI (support for triggering actions from the Juju GUI will follow soon). Schemas for each action are defined in an actions.yaml file in the charm root, and conform to JSON-Schema. When an action is invoked, passed parameters are validated against the respective schema as explained in "Actions for the Charm author" at both the API and the unit level:

https://jujucharms.com/docs/1.20/authors-charm-actions

CLI Actions are sub-commands of the 'juju action' command. For more details on their usage, 'juju action help' has examples and further material.

The following subcommands are currently specified:

  • defined - show actions defined for a service
  • do - queue an action for execution
  • fetch - show results of an action by ID
  • status - show results of actions filtered by optional ID prefix

New Blocks and Messages

You can now specify block message when you enable a block. For example, you can add a message to 'destroy-environment':

bash juju block destroy-environment "Don't destroy this environment" juju destroy-environment ERROR Don't destroy this environment

You can list the blocks enabled in the environment like so:

no-highlight juju block list destroy-environment=on, Don't destroy this environment remove-object=off all-changes=off

The Multiwatcher now has information about blocks. There is now block client capable of switching blocks on/off as well as listing all enabled blocks.

Experimental: Service Leader Elections

Services can now coordinate leadership among the deployed units using Juju's service leader election support.

Charms now have access to three new hook tools:

no-highlight is-leader - returns true only if the executing unit is guaranteed service leadership for the next 30s leader-get - as relation-get; accessible only within the service leader-set - as relation-set; will fail if not executed on leader

...and two new hooks:

no-highlight leader-elected - runs when the unit takes service leadership leader-settings-changed - runs when another unit runs leader-set

When a unit starts up, it will always run either leader-elected or leader-settings-changed as soon as possible, delaying only doing so only to run the install hook; complete any queued or in-flight operation; or resolve a hook or upgrade error.

Known limitations:

  • Amazon limits the number of addresses the containers an instance can have based on its size.
  • Statically allocated addresses are not released on container shutdown.
  • Container addressability does not survive a host reboot.
  • No public IP address is added on AWS, and we don’t yet support dynamic port mapping -- so you can not yet expose containerized services on Amazon.

Resolved issues

  • Allow annotations to be set on charms Lp 1313016

  • Juju-backup is not a valid plugin Lp 1389326

  • Juju needs to support systemd for >= vivid Lp 1409639

  • Joyent provider uploads user's private ssh key by default Lp 1415671

  • Unable to bootstrap on cn-north-1 Lp 1415693

  • Debug messages show when only info was asked for Lp 1421237

  • Juju default logging leaks credentials Lp 1423272

  • Juju resolve doesn't recognize error state Lp 1424069

  • Juju status --format=tabular Lp 1424590

  • Ec2 provider unaware of c3 types in sa-east-1 Lp 1427840

  • Ec2 eu-central-1 region not in provider Lp 1428117

  • Ec2 provider does not include c4 instance family Lp 1428119

  • Allwatcher does not remove last closed port for a unit, last removed service config Lp 1428430

  • Make kvm containers addressable (esp. on maas) Lp 1431130

  • Fix container addressability issues with cloud-init, precise, when lxc-clone is true Lp 1431134

  • Dhcp's "option interface-mtu 9000" is being ignored on bridge interface br0 Lp 1403955

  • Juju-1.23beta3 breaks glance <-> mysql relation when glance is hosted in a container Lp 1441811

  • Juju ensure-availability should be able to target existing machines Lp 1394755

  • Juju 1.23b4 vivid panic: runtime error: invalid memory address or nil pointer dereference Lp 1443541

  • 1.23-beta4 sporadically fails autotests Lp 1443440

  • 1.23.1: bootstrap failure, vivid, local provider Lp 1447446

  • Hooks don't fire after upgrade 1.23.0 Lp 1447846

juju-core 1.22.8

A new supported release of Juju, juju-core 1.22.8, is now available. This release replaces version 1.22.6.

Getting Juju

juju-core 1.22.8 is available for Wily and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/1.22

Windows and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.22.8

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Juju uses proxy to access bootstrap node Lp 1478660

  • Juju environment not usable after the upgrade Lp 1473517

juju-core 1.22.6

A new supported release of Juju, juju-core 1.22.6, is now available. This release replaces version 1.22.5.

Juju 1.22.x is an older supported version of Juju. The current Juju version is 1.24.0.

Getting Juju

juju-core 1.22.6 is available for vivid and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/1.22

Windows and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.22.6

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • State: availability zone upgrade fails if containers are present Lp 1441478

  • LXC provisioning fails on joyent Lp 1461150

  • Worker/diskmanager sometimes goes into a restart loop due to failing to update state Lp 1461871

  • Package github.com/juju/txn has conflicting licences Lp 1463455

juju-core 1.22.5

A new stable release of Juju, juju-core 1.22.5, is now available. This release replaces version 1.22.1.

Juju 1.22.x is an older supported version of juju. The current Juju version is 1.23.3

Getting Juju

juju-core 1.22.5 is available for vivid and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/1.22

Windows and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.22.5

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Transaction collection (txns) grows without bound Lp 1453785

  • /var/spool/rsyslog grows without bound Lp 1453801

  • Instancepoller compares wrong address list and always requests updated state addresses Lp 1454043

  • Firewaller gets an exception if a machine is not provisioned Lp 1454599

  • Jujud leaking file handles Lp 1454697

  • Agents see "too many open files" errors after many failed api attempts Lp 1420057

  • 1.20.x client cannot communicate with 1.22.x env Lp 1454829

  • Intermittent panic: rescanned document Lp 1449054

  • Harvest mode setting not used by container provisioner Lp 1459885

juju-core 1.22.1

A new stable release of Juju, juju-core 1.22.1, is now available. This release replaces 1.22.0.

Getting Juju

juju-core 1.22.1 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Notable Changes

This releases addresses stability and performance issues.

Vivid local-provider limitations

Juju 1.22.x on Vivid is suitable for creating and maintaining Juju environments in public and private clouds, including MAAS. Local-provider support for Vivid's default systemd installation will land in 1.23. To use Juju 1.22 on Vivid for local testing and charm development you must take some additional steps.

You can switch back to upstart, as explained at:

https://wiki.ubuntu.com/SystemdForUpstartUsers

Alternatively you can run Ubuntu Trusty inside a KVM or Docker running on your localhost. Many charm authors prefer to use Vagrant for development to separate the Juju from their localhost. We provide Juju Vagrant boxes:

https://jujucharms.com/docs/stable/config-vagrant

The Juju (1.23.x) will support systemd on Vivid. You can test and develop using the local-provider. 1.23.x will first be available via the official stable PPA:

https://launchpad.net/~juju/+archive/stable

And will appear in the release shortly thereafter.

Resolved issues

  • Juju cannot deploy when distro-info-data gets a new series Lp 1427879

  • Juju backup fails when journal files are present Lp 1423936

  • Copyright information is not available for some files Lp 1435974

  • Juju restore failed with "error: cannot update machines: machine update failed: ssh command failed:" Lp 1434437

  • Unit test failure: testnewdefaultserver.n40_github_com_juju_juju_cert_test.certsuite Lp 1437040

  • Certsuite.testnewdefaultserver failure Lp 1435860

  • Apt-http-proxy being reset to bridge address Lp 1437296

juju-core 1.22.0

A new stable release of Juju, juju-core 1.22.0, is now available. This release replaces 1.21.3.

Getting Juju

juju-core 1.22.0 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.22.0

Notable Changes

  • Blocking changes to the environment
  • LXC image caching and the cached-images command
  • Juju state backup improvements
  • Static internal route for Joyent instances
  • The instance availability zone is exposed to charms
  • Relaxed checked for container scoped relations
  • Windows workloads preview
  • New annotations client includes charms and bulk calls
  • Upgrade robustness
  • Multi-environment infrastructure

Blocking changes to the environment

The 'block' and 'unblock' commands control which changes can be made to a running environment. Use the 'block' command to lockdown the environment to prevent accidental changes. Three levels of restrictions can be applied, 'destroy-environment', 'remove-object', and 'all-changes'.

The 'destroy-environment' level blocks just the 'destroy-environment' command. You can prevent an environment from being destroyed like this:

bash juju block destroy-environment

The 'remove-object' level blocks the 'destroy-environment', 'remove-machine', 'remove-relation', 'remove-service', and 'remove-unit' commands from being run. You can prevent the machines, services, units and relations from being removed thusly:

bash juju block remove-object

The 'all-changes' level blocks commands that add, alter, or remove things in the running environment. These commands are blocked: 'add-machine', 'add-relation', 'add-unit', 'authorised-keys add', 'authorised-keys delete', 'authorised-keys import', 'deploy', 'destroy-environment', 'ensure-availability', 'expose', 'resolved', 'remove-machine', 'remove-relation', 'remove-service', 'remove-unit', 'retry-provisioning', 'run', 'set', 'set-constraints', 'set-env', 'unexpose', 'unset', 'unset-env', 'upgrade-charm', 'upgrade-juju', 'user add', 'user change-password', 'user disable', and 'user enable' You can lockdown the environment to prevent all changes like this:

bash juju block all-changes

When a command is blocked, and you are certain you need to alter the environment, you can remove the block using the 'unblock' command. For example, to permit the remove-relation command blocked by the 'remove-object' level, run:

bash juju unblock remove-object

You can restore the block after you complete your changes.

Note that some commands have a '--force' option bypasses the restrictions set by 'block'. Only use the '--force' option after the command was run without it to ensure you are aware of the restrictions set on the environment by the 'block' command.

For more information run 'juju help block' and 'juju help unblock'.

LXC image caching and the cached-images command

From Juju 1.22 onwards, LXC images are cached in the Juju environment when they are retrieved to instantiate a new LXC container. This applies to the local provider and all other cloud providers. This caching is done independently of whether image cloning is enabled.

Note: Due to current upgrade limitations, image caching is currently not available for machines upgraded to 1.22. Only machines deployed with 1.22 will cache the images.

In Juju 1.22, lxc-create is configured to fetch images from the Juju state server. If no image is available, the state server will fetch the image from http://cloud-images.ubuntu.com and then cache it. This means that the retrieval of images from the external site is only done once per environment, not once per new machine which is the default behaviour of lxc. The next time lxc-create needs to fetch an image, it comes directly from the Juju environment cache.

The 'cached-images' command can list and delete cached LXC images stored in the Juju environment. The 'list' and 'delete' subcommands support '--arch' and '--series' options to filter the result.

To see all cached images, run:

bash juju cached-images list

Or to see just the amd64 trusty images run:

bash juju cached-images list --series trusty --arch amd64

To delete the amd64 trusty cached images run:

bash juju cache-images delete --series trusty --arch amd64

Future development work will allow Juju to automatically download new LXC images when they becomes available, but for now, the only way update a cached image is to remove the old one from the Juju environment. Juju will also support KVM image caching in the future.

See 'juju cached-images list --help' and 'juju cached-images delete --help' for more details.

Juju state backup improvements

The backup CLI plugin is replaced by the native 'backups' command. The command can create, manage, and restore state backups. The new backup features are also available using the API endpoints too.

You can create a backup like this:

bash juju backups create

Backups are kept on the state-server by default, but you can use the 'download' and 'upload' subcommands to move backups between locations.

bash juju backups download <id> juju backups upload <filename>

You can list the backups and see info about a specific backup using these subcommands:

bash juju backups list juju backups info <id>

You can remove unwanted backups like this:

bash juju backups remove <id>

The juju 'restore' command is unchanged. You can learn more about backups by reading the details of 'juju help backups'.

The instance availability zone is exposed to charms

Charms can query the 'JUJU_AVAILABILITY_ZONE' environmental variable to learn the instance's availability zone. Not all clouds/providers support availability zones, but this information is available when the charm is deployed in EC2 and HP Cloud for example.

Relaxed checked for container scoped relations

Container scoped relations are now allowed as long a there is at least one subordinate service involved. Previously there had to be exactly one subordinate service in the relation.

Static internal route for Joyent instances

When running in Joyent, Juju creates a static route to ensure that it is always possible for instance to communicate over the internal network interface.

Windows workloads preview

Juju 1.22.0 has initial Windows workload support. This is BETA, and will be improved in the following releases.

Deploying Windows charms is just the same as it is for Ubuntu:

bash juju deploy lp:~me/win2012r2/my-charm

or from a local repository

bash juju deploy --repository=/home/me/charms local:win2012r2/my-charm

A Juju charm boilerplate is available:

https://github.com/cloudbase/windows-charms-boilerplate

Powershell is the language of choice when writing Juju charms for Windows, as it is tightly coupled to Windows automation. You may use any language you see fit.

There are a few notable differences between charms running on Ubuntu and charms running on Windows:

  • All charm hooks must have one of the following extensions: .ps1, .cmd, .bat, .exe. The extensions will be searched in order, with powershell being default.
  • All juju files/charms are located in: C:\juju . The folder structure should be the same as on ubuntu, with C:\juju as the root.
  • Cloudbase-init installs a local copy of python 2.7 (soon to switch to 3.X). If you wish, you can write python code to deploy your applications, but keep in mind you would have to wrap the execution of your script with one of the supported filetypes (.ps1, .cmd, .bat, .exe).
  • If the machine reboots, Juju machine agents and unit agents will have a delayed start. This means that Juju services will start 2 minutes after the last Windows service is started. So if the Juju services are listed as "Stopped" immediately after reboot, don't panic.
  • All Juju services (and therefor charms) will run under an administrator account called: jujud. This will allow charms to use installers that require a real user, and not just the "system" user.

Juju requires Windows images with cloudbase-init to deploy Juju services. For public clouds which run KVM/Hyper-V and have one of the supported metadata services in cloudbase-init, there is a downloadable:

http://www.cloudbase.it/ws2012r2/

You need to upload custom images. So roughly, the steps would be:

  1. Download image from http://www.cloudbase.it/ws2012r2/
  2. Add it to your cloud provider
  3. Generate necessary metadata using juju-metadata generate-image -s win2012r2 -i
  4. Make the generated files available. A simple webserver or simply uploading the files to the provider's object storage should do. As long as they are accessible to the newly booted up instances.
  5. Change agent-metadata-url in environments.yaml to point to the "tools" directory that 'generate-image' created.
  6. Bootstrap your environment

If you want to use other Windows version in your cloud, or if you need image formats other then vhdx or qcow2, you can generate your own image using the following tools:

https://github.com/cloudbase/windows-openstack-imaging-tools

Current Windows limitations:

  • Windows cannot be used as a bootstrap machine for the state-server. Only machine agents are supported.
  • 'juju ssh' currently does not work with Windows. The implementation is tightly coupled with *nix systems, and having OpenSSH deployed on Windows machines might later become a security risk due to the fact that it is not part of normal Windows update systems. Work is being done to integrate WinRM support into Juju. This will allow us to offer the same level of integration using powershell remoting (http://msdn.microsoft.com/en-us/library/dd357801.aspx)
  • Manual provider is not supported. Manual provider requires SSH to deploy. See above mention.
  • Local provider is not supported. Local provider relies on LXC or KVM and it deploys a state-server to the host, which is not supported on Windows.
  • Debug hooks are not yet supported yet. The current implementation requires SSH and tmux to work, which is very platform specific. This will be implemented for Windows soon.

New annotations client includes charms and bulk calls

The new api.annotations.client supports annotating charms, the environment, machines, services, and units. It supports bulk calls. The old api.client annotation functions are deprecated.

The SET annotations call takes this form:

json { "Type": "Annotations", "Request": "Set", "Params": { "Annotations": { { "EntityTag": a, "Annotations": pairs1 },{ "EntityTag": b, "Annotations": pairs2 } } }

The corresponding GET annotations call takes this form:

json { "Type": "Annotations", "Request": "Get", "Params": { "Entities": { {Entity {"Tag": a}}, {Entity {"Tag": b}, } } }

The GET response looks like this:

json { "Results": { {"EntityTag": a, "Annotations": pairs1, "Error": nil}, {"EntityTag": b, "Annotations": pairs2, "Error": nil}, } }

Juju upgrade robustness

During Juju upgrades (the 'upgrade-juju' command), steps specific to Juju state-servers are now run before other upgrade steps. Database migrations on the state-server are completed before machine updates and updates of the Juju agents on deployed units.

Multi-environment infrastructure

Juju will soon be able to support multiple environments managed by a single state server (or state server cluster). Although this is not yet visible, version 1.22 includes much of the infrastructure changes for multi-environment support.

Resolved issues

  • Interface mtu management across maas/juju Lp 1355813

  • State server certificate has empty san list Lp 1397909

  • Api server restarts soon after state server comes up Lp 1403721

  • Cmdline: implement juju environment jenv Lp 1409784

  • Juju should wait until a node's status is 'deployed' to attempt ssh'ing into it Lp 1394680

  • Lock contention running apt in bootstrap Lp 1384259

  • Juju status --format summary panics with unresolvable ips Lp 1410320

  • Machine hasvote/wantsvote should be included in allwatcher deltas Lp 1402965

  • Open-stack provider breaks swift with standard config Lp 1312217

  • 'juju switch' should not require an entry in environments.yaml Lp 1334768

  • Provider/maas: stopinstances may fail if nodes are removed out of band Lp 1319016

  • Unit "(anycharm)" is not assigned to a machine when deploying with juju 1.22-beta5 Lp 1430049

juju-core 1.21.3

A new stable release of Juju, juju-core 1.21.3, is released. This release may replaces 1.21.1.

Getting Juju

juju-core 1.21.3 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.21.3

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Unable to override network-bridge if container type is kvm (local provider) Lp 1416134

  • Src/bitbucket.org/kardianos/osext/license is wrong Lp 1416425

  • Github.com/juju/syslog has contradicting licensing info Lp 1416433

  • Some files refer to an include license file that is not included Lp 1416430

  • Github.com/juju/utils has contradictory licence info Lp 1416436

  • Juju restore no longer works with azure: error: cannot re-bootstrap environment: cannot determine state server instances: environment is not bootstrapped Lp 1417178

  • Unit ports not populated by api megawatcher Lp 1418433

  • Apt-proxy can be incorrectly set when the fallback from http-proxy is used Lp 1417617

  • Juju agent using lxcbr0 address as apiaddress instead of juju-br0 breaks agents Lp 1416928

  • Juju-quickstart: bad api server response: 'nonetype' object is not iterable Lp 1420403

  • Error juju.worker runner.go:219 exited "rsyslog": x509: certificate signed by unknown authority Lp 1417875

  • Juju references old dns-name and ip addresses after restore. (failed juju ssh) Lp 1420316

  • Ensure-availability fails with 'failed to find any voting machines' after backup/restore. Lp 1420306

juju-core 1.21.1

A new stable release of Juju, juju-core 1.21.1, is now available. This release replaces 1.20.14.

Getting Juju

juju-core 1.21.1 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Windows and OS X users will find installers at:

https://launchpad.net/juju-core/+milestone/1.21.1

Notable Changes

  • Selecting provisioner harvest modes
  • Using apt mirrors
  • Configuring OS update and upgrade for faster provisioning
  • Using daily image streams for faster provisioning
  • Selecting the agent stream to get alternate version of Juju
  • Tuning Juju to take advantage of NUMA
  • Configuring the MAAS network rules
  • Improved availability zone selection in OpenStack and MAAS
  • Juju now prefers SSD storage in AWS
  • Adding many machines
  • Choosing the nodes used to ensure high availability
  • Inspecting the API connection settings
  • Managing who can connect to the Juju environment
  • Upgrade robustness
  • Rebooting units from charm hooks
  • Improvements to ports management for charms
  • Developing Juju providers for clouds without storage
  • More mirrors for faster bootstraps

Selecting provisioner harvest modes

Juju keeps a model of what it thinks the environment looks like, and based on that model, can harvest machines which it deems are no longer required. This can help keep your costs low, and keep you out of web consoles. Juju supports several harvesting modes to suit your needs.

Destroyed: Juju will harvest only machine instances that are marked as dead, and that Juju knows about. Running instances unknown to Juju will not be harvested. This is the default mode.

Unknown: Juju will harvest only instances that Juju does not know about.

All: Juju will terminate all instances – destroyed or unknown – that it finds. This is a good option if you are only utilizing Juju for your environment.

None: Juju won't harvest any machines. This is the most conservative mode, and a good choice if you manage your machines utilizing a separate process outside of Juju.

Juju's harvesting behaviour is set through the environments.yaml file.

yaml provisioner-harvest-mode: <MODE>

provisioner-harvest-mode replaces safe-mode. Environments with safe-mode set will be converted to provisioner-harvest-mode when upgraded.

Using apt mirrors

You can now configure apt-mirror in environments.yaml to specify the mirror used by all the machines provisioned in the environment:

apt-mirror: http://my.archive.ubuntu.com

On precise, the cloud tools archive is now pinned before calling apt upgrade to ensure its packages are a lower priority than the default precise archives. Charms developed on precise will see the same packages when deployed into a Juju provisioned machine. If your precise charm requires packages from the cloud tool's archive, you can use the target-release option to specify the archive to select:

bash apt-get --target-release precise-updates/cloud-tools my-package

Configuring OS update and upgrade for faster provisioning

When Juju provisions a machine, its default behaviour is to update the list of available packages and upgrade the existing packages to the latest version. If your OS images are fresh or the services you deploy do not require updated packages, you can disable updates and upgrades to provision the machine faster.

Two configuration options are available to disable apt updates and upgrades. When your OS images are fresh, you can set both enable-os-refresh-update, and enable-os-upgrade to false. When you know that some charms want the latest packages to set up services, you will want to keep enable-os-refresh-update set to "true".

You can configure the options in environments.yaml for fast provisioning like so:

yaml enable-os-upgrade: false enable-os-refresh-update: false The local provider skips apt upgrades by default for faster provisioning. If you wish to enable upgrades in your local development, set enable-os-upgrade to "true" in your environments.yaml:

yaml enable-os-upgrade: true

If you are using the local-provider to develop charms or test, you will want to regularly purge the Juju template and LXC caches to be certain you are using fresh images. For example, before you start testing a new trusty charm, you can remove the template and cloud image like this:

bash sudo lxc-destroy -n juju-trusty-lxc-template sudo rm -r /var/cache/lxc/cloud-trusty

Using daily image streams for faster provisioning

Juju prefers to use the slow-changing "released" images when provisioning machines. The image-stream option in environments.yaml can be set to "daily" use more up-to-date images, thus shortening the time it takes to perform apt-get update/upgrade. While this feature has existed since 1.18.0, it was not applied consistently to KVM containers. KVM containers will now use "daily" when environments.yaml is set to:

yaml image-stream: daily

Selecting the agent stream to get alternate version of Juju

The agent-stream config option selects the versions of Juju that an environment can deploy and upgrade to. The default behaviour of Juju is to select agents from the "released" stream. These are the most stable versions of Juju. You can set agent-stream to select "devel" streams now to test the unstable versions of Juju:

yaml agent-stream: devel

You can evaluate the next stable version of Juju before it is the default version by selecting the "proposed" stream like this:

yaml agent-stream: proposed

The tools-metadata-url was renamed to agent-metadata-url and it does not need to be set to get "devel" or "proposed". You can remove it from environments.yaml if you have set it. agent-metadata-url is only needed to select a private stream.

If you have an existing test environment using tools-metadata-url or agent-metadata-url to test proposed versions, you can still upgrade to 1.21.0. After you upgrade, you can update the environment to use the devel streams at the default stream location:

bash juju unset-env agent-metadata-url juju set-env agent-stream=proposed

Subsequent upgrades will "just work".

Tuning Juju to take advantage of NUMA

Juju can be tuned to take advantage of NUMA machines. If your state-server will be on a machine with NUMA support, you can set set-numa-control-policy to true in environments.yaml like this:

yaml set-numa-control-policy: true

The default value is false.

Configuring the MAAS network rules

Juju and MAAS cannot both be in control of the network. When MAAS is managing the bridge and bringing networks up and down, set the disable-network-management option in environments.yaml to "true":

yaml disable-network-management: true

This tells Juju not to create a network bridge or to bring eth0 up and down during cloud-init. Juju will not make changes to the network config when its agents start.

Improved availability zone selection in OpenStack and MAAS

The OpenStack and MAAS providers now attempt to start instances in all available zones until they find one which succeeds, rather than trying just the first zone and failing. This aligns OpenStack and MAAS behaviour with that of AWS.

Juju now prefers SSD storage in AWS

When deploying workloads onto AWS, images with SSD volumes are now preferred. If no such images are available, an image using EBS storage will be used as a fallback.

Adding many machines

Juju's add-machine command now accepts the "-n" option to add many machines. For example, to add two machines:

bash juju add-machine -n 2

The "-n" option can be combined with placement. You can add two LXC containers to machine 1 thusly:

bash juju add-machine lxc:1 -n 2

Choosing the nodes used to ensure high availability

Just as juju bootstrap supports the ability to specify a particular node using "--to" placement directives, so too can juju ensure-availability specify a comma separated list of machines to use for any newly required state servers. For example:

bash juju ensure-availability -n 3 --to name1,name2

Inspecting the API connection settings

The juju api-info command shows the settings used to connect to the Juju state-server's API. You can see the settings for all the fields (except for password) like so:

bash juju api-info

If you want to see the password being used, you need to either use the "--password" option:

bash juju api-info --password

or specify the password field as the only field to show:

bash juju api-info password

You can learn the value of any field by including it in the command line. For example, to learn the name of user created during bootstrap, type:

bash juju api-info user

Managing who can connect to the Juju environment

Juju now supports multiple people connecting to the environment with their own identity and credentials.

When an environment is bootstrapped the "admin" user is created (this will change with 1.22 to reflect the name of the logged in user).

Even though there is support for multiple users, there is not yet support for fine grain permissions. These will come in time. The only permission checked at this stage is that only the "admin" user can create or disable other users. Any user is now able to change their own password.

The user commands are grouped under the juju user command:

bash juju user usage: juju user <command> ... purpose: manage user accounts and access control

"juju user" is used to manage the user accounts and access control
in the Juju environment.

commands:
    add - adds a user
    change-password - changes the password of the current user
    disable - disable a user to stop the user logging in
    enable - reenables a disabled user to allow the user
                      to log in
    help - show help on a command or other topic
    info - shows information on a user
    list - shows all users

You can add a user like this:

bash juju user add test "Test User"

To generate a random strong password, use the --generate option.

bash password: type password again: user "Test User (test)" added environment file written to /home/tim/some-dir/test.jenv

The generated environment file still needs to be copied into the user's $JUJU_HOME/environments directory in order to be used. Future versions of Juju will make this more simple. The name of the environments file is the name that the user needs to use to talk to the environment, so the user will probably want to rename it too. For example, an environment named app-stack will have an env named:

no-highlight $JUJU_HOME/environments/app-stack.jenv

Juju will ask for a password to be typed in. If you'd prefer a strong random password, you can use the "--generate" option. You can also control the location and name of the environments file that is created.

You can see which users have been created using the juju user list command:

bash juju user list NAME DISPLAY NAME DATE CREATED LAST CONNECTION admin admin 23 minutes ago just now test Test User 5 minutes ago never connected thumper Tim 5 seconds ago never connected

The output of this command can also be in YAML or JSON using the usual "--format" options.

Any user that is created will be able to access the environment. To stop this, you can disable the user.

```bash juju user disable test User "test" disabled

juju api-info user -e local-test test

juju user info -e local-test WARNING discarding API open error: environment "local-test" not found ERROR invalid entity name or password ```

Unfortunately the warning there is due to legacy environment support that is checked as a fallback when the initial connection failed due to the user being disabled.

Disabled users are not shown by default with the listing:

bash juju user list NAME DISPLAY NAME DATE CREATED LAST CONNECTION admin admin 30 minutes ago just now thumper Tim 6 minutes ago never connected

But they can be included with the '--all' option:

bash juju user list --all NAME DISPLAY NAME DATE CREATED LAST CONNECTION admin admin 32 minutes ago just now test Test User 13 minutes ago 2 minutes ago (disabled) thumper Tim 8 minutes ago never connected

Disabled users can be enabled again using the enable command:

```bash juju user enable test User "test" enabled

juju user info -e local-test user-name: test display-name: Test User date-created: 14 minutes ago last-connection: just now ```

Upgrade robustness

Many improvements have been made to make Juju software upgrades more reliable.

The upgrade process is now synchronised which is especially important in HA environments where multiple state servers exist. The master Juju state server now upgrades first, then the other state servers, followed by the remaining machines. If one or more state servers fail to start the upgrade process within a certain time, the upgrade is aborted and a rollback to the previous tools version occurs.

Upgrade progress is now shown in the Juju status output. A machine will now report when it is upgrading and both transient and permanent upgrade errors will also be indicated.

If a machine has trouble with the steps it needs to run to complete an upgrade, it will now retry several times. This helps to deal with some transient failures.

Rebooting units from charm hooks

There are several cases where a charm needs to reboot a machine, such as after a kernel upgrade, or to upgrade the entire system. The charm may not be able to complete the hook until the machine is rebooted.

The juju-reboot command allows charm authors schedule a reboot from inside a charm hook. The reboot will only happen if the hook completes without error. You can schedule a reboot like so:

bash juju-reboot

The --now option can be passed to block hook execution. The juju-reboot command will hang until the unit agent stops the hook and re-queues it for next run. This will allow you to create multi-step install hooks.

Charm authors must wrap calls to juju-reboot to ensure it is actually necessary, otherwise the charm risks entering a reboot loop. The preferred work-flow is to check if the feature/charm is in the desired state, and reboot when needed. This bash example assumes that "$FEATURE_IS_INSTALLED" variable was defined by a check for the feature, then 'juju-reboot' is called if the variable is false:

bash if [[ $FEATURE_IS_INSTALLED == "false" ]] then install_feature juju-reboot --now fi

The juju-reboot command can be called from any hook. It can also be called using the juju run command.

Improvements to ports management for charms

Your charm hooks can call the new opened-ports shell command to get a listing of the open ports on the unit.

bash opened-ports

The open-port and close-port commands both support single ports and ranges, for example, to open ports 80 through 100 on tcp:

bash open-port 80-100/tcp

And you can close a range like so:

bash close-port 90-100/tcp

Juju now keeps track of what ports are opened by other units on the same machine and does not allow conflicting ports to be opened. The open-port and close-port commands can return a conflict error when the port was opened or closed by another charm. Additionally, both these commands work transactionally, in common with other hook commands; until the hook is committed no actual changes are made (opening or closing ports).

Developing Juju providers for clouds without storage

Storage associated with a particular cloud (S3 for AWS, Swift for Openstack etc) was a mandatory requirement when developing a provider to allow Juju to deploy workloads to any particular cloud platform. This requirement is no longer necessary. Although existing providers still use cloud storage (to some extent), new providers can be written without needing to provide a storage implementation.

More mirrors for faster bootstraps

Juju agents are mirrored in the certified public clouds (AWS, Azure, HP Cloud, and Joyent) to make bootstraps fast. You do not need to do anything to use them, they just work when we add and update them.

We added mirrors to Joyent. We registered the new regions recently added to AWS, Azure, and HP Cloud.

Resolved issues

  • We should remove direct db access for clients Lp 1253652

  • Allow specifying a key when doing manual provisioning Lp 1270466

  • Juju doesn't use maas' knowledge of system architecture when picking tools Lp 1303853

  • Local provider is very slow to transition from agent-status: pending Lp 1322302

  • Juju should wrap apt-get invocations with eatmydata when provisioning cloud instances Lp 1335822

  • Cloudinit does not use ssh client Lp 1339976

  • Provisioner-safe-mode is undocumented Lp 1342729

  • Networker restarts every 3 seconds with the local provider (missing /etc/network/interfaces) Lp 1343219

  • Describe harvesting strategy rather than using "safe mode" name Lp 1345553

  • Configstore: if the size of the serialised jenv decreases the .jenv file will be corrupt Lp 1348458

  • Juju ignores environments.yaml on failed bootstrap if $provider.jenv exists Lp 1361680

  • Saved addresses should omit link-local addresses Lp 1362453

  • Blobstore's hashing needs improvement Lp 1364750

  • Removing a unit on an unclean machine should remove that machine Lp 1206532

  • Juju log files should not be world readable Lp 1286518

  • Juju uses hard-coded regions Lp 1319474

  • Cmd/juju: deploy --to a non existent machine fails too late in the process Lp 1212538

  • Cmd/juju: add-machine should take a -n param Lp 1214209

  • Container provisioner may choose bad tools Lp 1347984

  • Juju set help is written but not shown Lp 1359187

  • Status panics if environment not running Lp 1372264

  • Rsyslog worker continuously restarts due to x509 error following upgrade Lp 1375507

  • Allow open-port to expose several ports Lp 1216644

  • Failed add-machine ssh: leaves behind garbage in state Lp 1356886

  • Azure fails with juju bootstrap --upload-tools --upload-series Lp 1357511

  • Support maas zones for automatic az placement Lp 1360605

  • Juju should support an apt alternate mirror for private clouds Lp 1364200

  • "juju ssh" doesn't work during tools upgrade Lp 1367009

  • Bootstrap failed: unexpected error type *errors.errorstring Lp 1367896

  • Ensure-availability command doesn't support placement directives Lp 1370332

  • Api logins fail with "invalid entity name or password" before db migrations have run Lp 1372752

  • Juju needs to support the maas api's not_tags constraint Lp 1373385

  • Error message when trying to deploy to node 0 on lxc needs to be more user friendly Lp 1378792

  • Use ssd image types on amazon ec2 Lp 1381009

  • Configuration item tools-stream deprecated in favour of agent-stream Lp 1383070

  • Azure bootstrap fails when 'location' and 'storage-account-name' are not in the same region Lp 1236136

  • Juju run doesn't work with subordinate units Lp 1286613

  • Need to reset user passwords (+ui) Lp 1288750

  • Bootstrap without a jenv destroys an existing environment Lp 1336843

  • Juju doesn't retry hard enough when destroying maas environments Lp 1384001

  • Bad apt proxy config file in deployed nodes Lp 1309207

  • Juju deploy hangs for a long time and then fails Lp 1386405

  • Cmd/juju: add help on placement directives (zones, maas host name) Lp 1387421

  • Juju scp help text is unclear on how to pass additional arguments Lp 1387640

  • Juju set-env/get-env work with arbitrary strings Lp 1304126

  • Cannot set user for juju scp Lp 1387766

  • Status --errors only to show only things with errors Lp 1309260

  • Provider/azure: boilerplate uses <> inconsistently around values that require replacing Lp 1381289

  • Juju set should give feedback that the value is already set Lp 1384622

  • Logs are not logrotated Lp 1078213

  • Juju db should use numactl when running mongo on multi-socket nodes Lp 1350337

  • Juju status panic if state conn is shutdown || closing. Lp 1361316

  • Container failed to start with lxc-clone-aufs=true Lp 1364939

  • No defined 'state-servers' on environment file after bootstrap, works after run 'juju status' Lp 1370149

  • Juju add-unit --to fails too late, leaving unit unassigned Lp 1384732

  • Ec2 says agent-state-info: 'cannot run instances: no default subnet for availability zone: ''us-east-1e''. (invalidinput)' Lp 1388860

  • Provider/ec2: try alternative az on insufficientinstancecapacity error Lp 1389037

  • Provider should test and verify credentials as first operation before bootstrap Lp 1362072

  • Non subordinate container scoped relations broken Lp 1382751

  • --debug dumps sensitive information to terminal Lp 1289038

  • Juju destroy-environment could fail on MaaS with disk erasing enabled, or manually released nodes Lp 1381619

  • Unit-get public-address on ec2 returns split horizon dns Lp 1308374

  • Juju tries to use lxcbr0 when local provider is configured with kvm containers Lp 1307677

  • Add juju_machine_id to the hooks environment Lp 1359714

  • Openstack provider, instance-state doesn't change on instance shutdown Lp 1382709

  • Config-get error inside config-changed: "settings not found" Lp 1301996

  • Open-stack provider breaks swift with standard config Lp 1312217

juju-core 1.20.14

A new stable release of Juju, juju-core 1.20.14, is now available. This release replaces 1.20.13.

Getting Juju

juju-core 1.20.14 is available for utopic (Ubuntu 14.10) and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Container scoped relations between 2 subordinates broken in 1.20.12 Lp 1396625

  • Azure instance-types and regions missing Lp 1389422

  • Ec2 says agent-state-info: 'cannot run instances: no default subnet for availability zone: ''us-east-1e''. (invalidinput)' Lp 1388860

  • Lxc containers in pending state due to juju-br0 misconfiguration Lp 1395908

juju-core 1.20.13

A new stable release of Juju, juju-core 1.20.13, is now available. This release replaces stable 1.20.12.

Getting Juju

juju-core 1.20.13 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • 1.20.12 breaks neutron-gateway, since all interfaces are brought up Lp 1395081
juju-core 1.20.12

A new stable release of Juju, juju-core 1.20.12, is now available. This release replaces stable 1.20.11.

Getting Juju

juju-core 1.20.12 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Juju run doesn't after upgrade to 1.20.11 Lp 1392745

  • Non subordinate container scoped relations broken Lp 1382751

  • Juju add-machine lxc:0 fails to start due to incorrect network name in trusty config Lp 1340261

  • Cloud-init network error when using maas/juju Lp 1345433

  • Juju deploy --to lxc:0 works ok but juju deploy --to lxc:1 fails Lp 1368976

  • Juju deploy to kvm:0 fails for br0 Lp 1386575

juju-core 1.20.11

A new stable release of Juju, juju-core 1.20.11, is now available. This release replaced stable 1.20.10.

Getting Juju

juju-core 1.20.11 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Local environment machine deployments fail after "juju destroy- machine" has been used Lp 1379802

  • Juju does not consider whether it has permission to an availability zone Lp 1380557

juju-core 1.20.10

A new stable release of Juju, juju-core 1.20.10, is now available. This release replaces stable 1.20.9.

Getting Juju

juju-core 1.20.10 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Notable Changes

This releases addresses packaging and documentation issues.

Resolved issues

  • make-release-tarball could check the packages with dependencies.tsv Lp 1368417
juju-core 1.20.9

A new stable release of Juju, juju-core 1.20.9, is now available. This release replaces stable 1.20.8.

Getting Juju

juju-core 1.20.9 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Not okforstorage error when deploying local charm Lp 1308146

  • Cloud-archive on precise not pinned when juju calls apt-get upgrade Lp 1370781

juju-core 1.20.8

A new stable release of Juju, juju-core 1.20.8, is now available. This release may replace stable 1.20.7.

Getting Juju

juju-core 1.20.8 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Notable Changes

This releases addresses stability and performance issues.

Resolved issues

  • Maas provider assumes machine uses dhcp for eth0 Lp 1361374

  • Relation-get with invalid relation name panics agent Lp 1365412

  • Bootstrap on maas fails trying to access cloud-images.ubuntu.com Lp 1365135

  • Not okforstorage error when deploying local charm Lp 1308146

  • Add-machine containers should default to latest lts Lp 1363971

  • Juju add-machine still assumes precise (maas) Lp 1315473

  • Juju-core client panics with juju set empty string Lp 1348829

  • --keep-broken option still allows instance to be stopped Lp 1365772

  • Some third party embedded sources in the source tarball are missing dependencies.tsv entries Lp 1368321

  • Licensing is inconsistent Lp 1368358

  • Sshstorage fails in non-english locale Lp 1367695

juju-core 1.20.7

A new stable release of Juju, juju-core 1.20.7, is now available. This release replaces 1.20.6.

Getting Juju

juju-core 1.20.7 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Noteworthy

This release contains various bug fixes.

Resolved issues

  • --keep-broken bootstrap option to keep failed environments Lp 1362923

  • LXC was not created, no errors, no logs -> pending state. Lp 1354027

  • Juju status still returns private IP in 'public-ip' field Lp 1364419

juju-core 1.20.6

A new stable release of Juju, juju-core 1.20.6, is now available. This release replaces 1.20.5.

Getting Juju

juju-core 1.20.6 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Noteworthy

This releases addresses stability issues, particularly with LXC.

Resolved issues

  • Arguments no longer passed to plugins if you don't have an environment set Lp 1359170

  • ERROR Nonce already used Lp 1190986

  • Juju status returns private IP in 'public-ip' field. Lp 1348287

  • Juju status nil pointer reference. Lp 1357482

  • LXC containers created, juju can't seen or communicate with them Lp 1357552

  • Cannot get tools from machine for lxc container Lp 1359800

  • Tools download can fail and is not always retried Lp 1360994

  • Containers not starting on machine Lp 1362360

  • Update the add machine api to accept both environment name and UUID Lp 1360063

juju-core 1.20.5

A new stable release of Juju, juju-core 1.20.5, is now available. This release replaces 1.20.1.

Getting Juju

juju-core 1.20.5 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Noteworthy

This releases addresses stability and performance issues.

Resolved issues

  • Juju is stripping underscore from options Lp 1243827

  • Juju api-endpoints cannot distinguish public and private addresses Lp 1311227

  • API server inaccessible for roughly 5 seconds after bootstrap Lp 1351083

  • Juju bootstrap in an existing environment destroys the environment Lp 1340893

  • The jujud on state server panic misses transaction in queue Lp 1318366

  • Juju machine agents suiciding Lp 1345014

  • Juju state server database is overly large Lp 1344940

  • Jujud rewrites /etc/init/juju-db.conf constantly Lp 1349969

  • Juju writes to mongo without an actual change occurring Lp 1345832

  • Talking to mongo can fail with "TCP i/o timeout" Lp 1307434

  • Bootstrap fails if /usr/lib/juju/bin/mongod doesn't already exist Lp 1350700

  • Support the new v4 signing format used by the new China region Lp 1319475

  • Azure: secondary state servers do not load balance API server port Lp 1347371

  • Network error causes tools download to fail Lp 1349989

  • Maas provider: allow users to specify network bridge interface. Lp 1337091

  • ppc64 architecture miss match for MAAS ppc64el Lp 1349771

  • MAAS provider uses dead instance address Lp 1353442

  • Juju bootstrap fails if kvm-ok not in path Lp 1342747

  • Juju-local needs new dep dbus -- specifically to work in lxc Lp 1343301

  • LXC clonetemplate lock can be left stale Lp 1311668

  • LXC deploys broken on precise Lp 1350011

  • LXC template fails to stop Lp 1348386

  • Distribution tarball has licensing problems that prevent redistribution Lp 1341589

juju-core 1.20.1

A new stable release of Juju, juju-core 1.20.1, is now available. This release replaces 1.20.0.

Getting Juju

juju-core 1.20.1 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

Noteworthy

This release fixes several issues seen in slower environments. The performance improvements also improved reliability. Juju CI saw a 35% speed improvement in the test suite. While we had planned to release 1.20.1 on 2014-07-17, the performance improvements were just too good to delay a whole week.

Resolved issues

  • Juju 1.20 consistently fails to bootstrap a MAAS environment Lp 1339240

  • Juju bootstrap fails because mongodb is unreachable Lp 1337340

  • Juju 1.20.x slow bootstrap Lp 1338179

  • API-endpoints fails if run just after bootstrap Lp 1338511

  • Machines are killed if mongo fails Lp 1339770

  • Restore doesn't Lp 1336967

juju-core 1.20.0

A new stable release of Juju, juju-core 1.20.0, is now available.

Getting Juju

juju-core 1.20.0 is available for utopic and backported to earlier series in the following PPA:

https://launchpad.net/~juju/+archive/stable

New and Notable

  • High Availability
  • Availability Zone Placement
  • Azure Availability Sets
  • Juju debug-log Command Supports Filtering and Works with LXC
  • Constraints Support instance-type
  • The lxc-use-clone Option Makes LXC Faster for Non-Local Providers
  • Support for Multiple NICs with the Same MAC
  • MaaS Network Constraints and Deploy Argument
  • MAAS Provider Supports Placement and add-machine
  • Server-Side API Versioning

Resolved issues

  • Juju client is not using the floating ip to connect to the state server Lp 1308767

  • Juju help promotes the 'wrong' versions of commands Lp 1299120

  • Juju backup command fails against trusty bootstrap node Lp 1305780

  • The root-disk constraint is broken on ec2 Lp 1324729

  • Bootstrapping juju from within a juju deployed unit fails Lp 1285256

  • Images are not found if endpoint and region are inherited from the top level in simplestreams metadata Lp 1329805

  • Missing @ syntax for reading config setting from file content Lp 1216967

  • Deploy --config assumes relative path if it starts with a tilde (~) Lp 1271819

  • local charm deployment fails with symlinks Lp 1330919

  • Git usage can trigger disk space/memory issues for charms with blobs Lp 1232304

  • Juju upgrade-charm fails because of git Lp 1297559

  • .pyc files caused upgrade-charm to fail with merge conflicts Lp 1191028

  • Juju debug-hooks should display a start message Lp 1270856

  • Can't determine which relation is in error from status Lp 1194481

  • Dying units departure is reported late Lp 1192433

  • Juju upgrade-juju needs a dry run mode Lp 1272544

  • Restarting API server with lots of agents gets hung Lp 1290843

  • Juju destroy-environment >=256 nodes fails Lp 1316272

  • Azure destroy-environment does not complete Lp 1324910

  • Azure bootstrap dies with xml schema validation error Lp 1259947

  • Azure provider stat output does not show machine hardware info Lp 1215177

  • Bootstrapping azure causes memory to fill Lp 1250007

  • Juju destroy-environment on openstack should remove sec groups Lp 1227574

  • Floating IPs are not recycled in OpenStack Havana Lp 1247500

  • LXC cloning should be default behaviour Lp 1318485

  • Nasty worrying output using local provider Lp 1304132

  • Intermittent error destroying local provider environments Lp 1307290

  • Default bootstrap timeout is too low for MAAS environments Lp 1314665

  • MAAS provider cannot provision named instance Lp 1237709

  • Manual provisioning requires target hostname to be directly resolvable Lp 1300264

  • Manual provider specify bash as shell for ubuntu user Lp 1303195

High Availability

The juju state-server (bootstrap node) can be placed into high availability mode. Juju will automatically recover when one or more the state-servers fail. You can use the 'ensure-availability' command to create the additional state-servers:

bash juju ensure-availability

The 'ensure-availability' command creates 3 state servers by default, but you may use the '-n' option to specify a larger number. The number of state servers must be odd. The command supports the 'series' and 'constraints' options like the 'bootstrap' command. You can learn more details by running 'juju ensure-availability --help'

Availability Zone Placement

Juju supports explicit placement of machines to availability zones (AZs), and implicitly spreads units across the available zones.

When bootstrapping or adding a machine, you can specify the availability zone explicitly as a placement directive. e.g.

bash juju bootstrap --to zone=us-east-1b juju add-machine zone=us-east-1c

If you don't specify a zone explicitly, Juju will automatically and uniformly distribute units across the available zones within the region. Assuming the charm and the charm's service are well written, you can rest assured that IaaS downtime will not affect your application. Commands you already use will ensure your services are always available. e.g.

bash juju deploy -n 10 <service>

When adding machines without an AZ explicitly specified, or when adding units to a service, the ec2 and openstack providers will now automatically spread instances across all available AZs in the region. The spread is based on density of instance "distribution groups".

State servers compose a distribution group: when running 'juju ensure-availability', state servers will be spread across AZs. Each deployed service (e.g. mysql, redis, whatever) composes a separate distribution group; the AZ spread of one service does not affect the AZ spread of another service.

Amazon's EC2 and OpenStack Havana-based clouds and newer are supported. This includes HP Cloud. Older versions of OpenStack are not supported.

Azure availability sets

Azure environments can be configured to use availability sets. This feature ensures services are distributed for high availability; as long as at least two units are deployed, Azure guarantees 99.95% availability of the service overall. Exposed ports will be automatically load balanced across all units within the service.

New Azure environments will have support for availability sets by default. To revert to the previous behaviour, the 'availability-sets-enabled' option must be set in environments.yaml like so:

yaml availability-sets-enabled: false

Placement is disabled when 'availability-sets-enabled' is true. The option cannot be disabled after the environment is bootstrapped.

Juju debug-log Command Supports Filtering and Works with LXC

The 'debug-log' command shows the consolidate logs of all juju agents running on all machines in the environment. The command operates like 'tail -f' to stream the logs to the your terminal. The feature now support local-provider LXC environments. Several options are available to select which log lines to display.

The 'lines' and 'limit' options allow you to select the starting log line and how many additional lines to display. The default behaviour is to show the last 10 lines of the log. The 'lines' option selects the starting line from the end of the log. The 'limit' option restricts the number of lines to show. For example, you can see just 20 lines from last 100 lines of the log like this:

bash juju debug-log --lines 100 --limit 20

There are many ways to filter the juju log to see just the pertinent information. A juju log line is written in this format:

no-highlight <entity> <timestamp> <log-level> <module>:<line-no> <message>

The 'include' and 'exclude' options select the entity that logged the message. An entity is a juju machine or unit agent. The entity names are similar to the names shown by 'juju status'. You can exclude all the log messages from the bootstrap machine that hosts the state-server like this:

bash juju debug-log --exclude machine-0

The options can be used multiple times to select the log messages. This example selects all the message from a unit and its machine as reported by status:

bash juju debug-log --include unit-mysql-0 --include machine-1

The 'level' option restricts messages to the specified log-level or greater. The levels from lowest to highest are TRACE, DEBUG, INFO, WARNING, and ERROR. The WARNING and ERROR messages from the log can seen thusly:

bash juju debug-log --level WARNING

The 'include-module' and 'exclude-module' are used to select the kind of message displayed. The module name is dotted. You can specify all or some of a module name to include or exclude messages from the log. This example progressively excludes more content from the logs

bash juju debug-log --exclude-module juju.state.apiserver juju debug-log --exclude-module juju.state juju debug-log --exclude-module juju

The 'include-module' and 'exclude-module' options can be used multiple times to select the modules you are interested in. For example, you can see the juju.cmd and juju.worker messages like this:

bash juju debug-log --include-module juju.cmd --include-module juju.worker

The 'debug-log' command output can be piped to grep to filter the message like this:

bash juju debug-log --lines 500 | grep amd64

You can learn more by running 'juju debug-log --help' and 'juju help logging'

Constraints Support instance-type

You can specify 'instance-type' with the 'constraints' option to select a specific image defined by the cloud provider. The 'instance-type' constraint can be used with Azure, EC2, HP Cloud, and all OpenStack-based clouds. For example, when creating an EC2 environment, you can specify 'm1.small':

bash juju bootstrap --constraints instance-type=m1.small

Constraints are validated by all providers to ensure values conflicts and unsupported options are rejected. Previously, juju would reconcile such problems and select an image, possibly one that didn't meet the needs of the service.

The lxc-use-clone Option Makes LXC Faster for Non-Local Providers

When 'lxc-use-clone' is set to true, the LXC provisioner will be configured to use cloning regardless of provider type. This option cannot be changed once it is set. You can set the option to true in environments.yaml like this:

yaml lxc-use-clone: true

This speeds up LXC provisioning when using placement with any provider. For example, deploying mysql to a new LXC container on machine 0 will start faster:

bash juju deploy --to lxc:0 mysql

Support for Multiple NICs with the Same MAC

Juju now supports multiple physical and virtual network interfaces with the same MAC address on the same machine. Juju takes care of this automatically, there is nothing you need to do.

Caution, network settings are not upgraded from 1.18.x to 1.20.x. If you used juju 1.18.x to deploy an environment with specified networks, you must redeploy your environment instead of upgrading to 1.20.0.

The output of 'juju status' will include information about networks when there is more than one. The networks will be presented in this manner:

no-highlight machines: ... services: ... networks: net1: provider-id: foo cidr: 0.1.2.0/24 vlan-tag: 42

MaaS Network Constraints and Deploy Argument

You can specify which networks to include or exclude as a constraint to the deploy command. The constraint is used to select a machine to deploy the service's units too. The value of 'networks' is a comma-delimited list of juju network names (provided by MaaS). Excluded networks are prefixed with a "^". For example, this command specify the service requires the "logging" and "storage" networks and conflicts with the "db" and "dmz" networks.

bash juju deploy mongodb --constraints networks=logging,storage,^db,^dmz

The network constraint does not enable the network for the service. It only defines what machine to pick.

Use the 'deploy' command's 'networks' option to specify service-specific network requirements. The 'networks' option takes a comma-delimited list of juju-specific network names. Juju will enable the networks on the machines that host service units.

Juju networking support is still experimental and under development, currently only supported with the MaaS provider.

bash juju deploy mongodb --networks=logging,storage

The 'exclude-network' option was removed from the deploy command as it is superseded by the constraint option.

There are plans to add support for network constraint and argument with Amazon EC2, Azure, and OpenStack Havana-based clouds like HP Cloud in the future.

MAAS Provider Supports Placement and add-machine

You can specify which MAAS host to place the juju state-server on with the 'to' option. To bootstrap on a host named 'fnord', run this:

bash juju bootstrap --to fnord

The MAAS provider support the add-machine command now. You can provision an existing host in the MAAS-based Juju environment. For example, you can add running machine named fnord like this:

bash juju add-machine fnord

Server Side API Versioning

The Juju API server now has support for a Version field in requests that are made. For this release, there are no RPC calls that require anything other than 'version=0' which is the default when no Version is supplied. This should have limited impact on existing CLI or API users, since it allows us to maintain exact compatibility with existing requests. New features and APIs should be exposed under versioned requests.

For details on the internals (for people writing API clients), see: this document

Finally

We encourage everyone to subscribe to the Juju Discourse forum or join us on #juju on freenode.

juju-core 1.18.3

A new stable release of Juju, juju-core 1.18.3, is now available. This release replaces 1.18.2.

Getting Juju

juju-core 1.18.3 is available in trusty and backported to earlier series in the following PPA

https://launchpad.net/~juju/+archive/stable

If you use the local provider, be sure to install the juju-local package if it is not already installed. Juju's local requirements have changed. Upgrading local juju environments without the juju-local package is not advised.

Resolved issues

  • juju sync-tools destroys the environment when given an invalid source Lp 1316869
  • Local provider behaves poorly when juju-mongodb is not installed on trusty Lp 1301538
  • Juju bootstrap defaults to i386 Lp 1304407
  • LXC local provider fails to provision precise instances from a trusty host Lp 1306537
  • LXC nested within kvm machines do not use lxc-clone Lp 1315216
juju-core 1.18.2

A new stable release of Juju, juju-core 1.18.2, is now available. This release replaces 1.18.1.

Getting Juju

juju-core 1.18.2 is available in trusty and backported to earlier series in the following PPA https://launchpad.net/~juju/+archive/stable

If you use the local provider, be sure to install the juju-local package if it is not already installed. Juju's local requirements have changed. Upgrading local juju environments without the juju-local package is not advised.

Resolved issues

  • Manual provisioned systems stuck in pending on arm64 Lp 1302205
  • Version reports "armhf" on arm64 Lp 1304742
  • Juju deploy --to lxc:0 cs:trusty/ubuntu creates precise container Lp 1302820
  • If the provisioner fails to find tools for one machine it fails to provision the others LP 1311676
  • Juju status panic() when unable to parse .jenv Lp 1312136
  • Juju authorised-keys should have an authorized-keys alias Lp 1312537
  • Race condition on bootstrap machine addresses with manual provider Lp 1314430
juju-core 1.18.1

A new stable release of Juju, juju-core 1.18.1, is now available. This release replaces 1.18.0.

Getting Juju

juju-core 1.18.1 is available in trusty and backported to earlier series in the following PPA https://launchpad.net/~juju/+archive/stable

If you use the local provider, be sure to install the juju-local package if it is not already installed. Juju's local requirements have changed. Upgrading local juju environments without the juju-local package is not advised.

Notable

  • Working with Ubuntu 12.04 Precise and 14.04 Trusty

Resolved issues

  • Juju upgrade-juju 1.16.6 -> 1.18 (tip) fails Lp 1299802
  • Peer relation disappears during upgrade of juju Lp 303697
  • The public-address changed to internal bridge after juju-upgrade Lp 1303735
  • juju cannot downgrade to same major.minor version with earlier patch number Lp 1306296
  • Juju 1.18.0, can not deploy local charms without series Lp 1303880
  • Juju scp no longer allows multiple extra arguments to pass through Lp 1306208
  • Juju tests fail with release version number 1.18.0 Lp 1302313
  • Charm store tests do not pass with juju-mongodb Lp 1304770

Working with Ubuntu 12.04 Precise and 14.04 Trusty

Juju 1.18.x is the first Juju that is aware of two Ubuntu LTS releases. In the past, when bootstrapping the state-server or deploying a charm, Juju selected Ubuntu 12.04 Precise when the serie was not specified. If you care about the version of Ubuntu used, then you want to specify the series

You can specify the series of the state-server and set the default charm series by adding "default-series" to environments.yaml. Setting "default- series" is the only way to specify the series of the state-server (bootstrap node). For example, if your local machine is Ubuntu 14.04 trusty, but your organisation only supports Ubuntu 12.04 Precise in the cloud, you can add this to your environment in environments.yaml:

yaml default-series: precise

There are many ways to specify the series to deploy with a charm. In most cases you don't need to. When you don't specify a series, Juju checks the environment's default-series, and if that isn't set, Juju asks the charm store to select the best series to deploy with the charm. The series of the state-server does not restrict the series of the charm. You can use the best series for a charm when deploying a service.

When working with local charms, Juju cannot fall back to the charm store, it falls back to the environment's default-series. You must specify the series in the environment or when deploying the charm. If your environment is running, you can add default-series like so:

bash juju set-environment default-series=precise

These commands choose Ubuntu 12.04 Precise when default-series is set to "precise" in the environment:

bash juju deploy cs:precise/mysql juju deploy precise/mysql juju deploy mysql juju deploy local:precise/mysql juju deploy loca:mysql

juju-core 1.18.0

A new stable release of Juju, juju-core 1.18.0, is now available. This release replaces 1.16.6.

Getting Juju

juju-core 1.18.0 is available in trusty and backported to earlier series in the following PPA https://launchpad.net/~juju/+archive/stable

If you use the local provider, be sure to install the juju-local package if it is not already installed. Juju’s local requirements have changed. Upgrading local juju environments without the juju-local package is not advised.

New and Notable

  • Support for proxies
  • Managing authorised ssh keys
  • Backup and restore the state-server (bootstrap node)
  • Run arbitrary commands on some or all instances
  • Use metadata generate-tools instead of sync-tools to generated simple streams metadata
  • The bootstrap --source option was replaced with --metadata-source
  • Timeouts for bootstrap are configurable for environments
  • Bootstrapping the local provider for lxc no longer requires sudo
  • Juju local provider can use clone for faster LXC
  • Juju now supports juju-mongodb, a mongodb tuned for juju’s needs
  • The "null" provider is now called "manual" in the juju config.
  • The destroy-environment command requires the environment name
  • Juju unset-env
  • Juju retry-provisioning
  • Bash completions include flags, machines, services, and units

Resolved issues

  • --upload-tools failure preventing bootstrap completing. Lp 1239558
  • Invalid SSL certificate after rebootstrapping. Lp 1130255
  • Killing instance outside of juju, doesn't get noticed. Lp 1205451
  • Juju bootstrap fails with openstack provider (failed unmarshaling the response body) Lp 1209003
  • Manual provider bootstrap fails with error about sftp scheme. Lp 1235717
  • Manual provider doesn't install mongo from the cloud archive. Lp 1238934
  • Manual provider uses reverse-lookup result instead of "bootstrap-host" Lp 1246905
  • Manual provider requires that machines have DNS records. Lp 1254629
  • Juju broken with OpenStack Havana for tenants with multiple networks. Lp 1241674
  • Juju destroy-environment no longer returns error when no environment exists Lp 1225238
  • Local provider isn't usable after an old environment has been destroyed Lp 1238541
  • Manual provider client cache prevents reuse of env by name Lp 1238948
  • destroy-environment no longer removes .jenv Lp 1246343
  • Manual bootstrap / provisioning does not add the ubuntu user Lp 1261343
  • Concurrent charm deployments corrupts deployments Lp 1067979
  • Juju status reports 'missing' instance-state when not run as root Lp 1237259
  • Juju uses tools for the wrong architecture when unable to find correct tools Lp 1227722
  • Call to relation-get failing with 'permission denied' Lp 1239681

Support for proxies

Proxies can now be configured for the providers in the environments.yaml file, or added to an existing environment using "juju set-env" The configuration options are:

- http-proxy
- https-proxy
- ftp-proxy
- no-proxy

The protocol-specific options accept a URL. The "no-proxy" option accepts a comma-separated list of host names or addresses.

The proxy options are exported in all hook execution contexts, and also available in the shell through "juju ssh" or "juju run".

There are three additional proxy options specific for apt. These are set to be the same as the non-apt proxy values, but can be overridden independently:

- apt-http-proxy
- apt-https-proxy
- apt-ftp-proxy

For example, with a squid-deb-proxy running on a laptop, you can specify the apt-http-proxy to use it for the containers by specifying the host machine’s network-bridge:

yaml apt-http-proxy: http://10.0.3.1:8000

Managing authorised ssh keys

Juju's ssh key management allows people other than the person who bootstrapped an environment to ssh into Juju machines/nodes. The authorised-keys command accepts 4 subcommands:

  • add - add ssh keys for a Juju user
  • delete - delete ssh keys for a Juju user
  • list - list ssh keys for a Juju user
  • import - import Launchpad or Github ssh keys

import can be used in clouds with open networks to pull ssh keys from Launchpad or Github. For example:

bash juju authorised-keys import lp:wallyworld

add can be used to import the provided key, which is necessary for clouds that do not have internet access.

Use the key fingerprint or comment to specify which key to delete. You can find the fingerprint for a key using ssh-keygen.

Juju cannot not manage existing keys on manually provisioned machines. Juju will prepend "Juju:" to the comments of all keys that it adds to a machine. These are the only keys it can "list" or "delete".

Note that keys are global and grant access to all machines. When a key is added, it is propagated to all machines in the environment. When a key is deleted, it is removed from all machines. You can learn more details by running juju authorised-keys --help.

Backup and restore state-server (bootstrap node)

Juju provides backup and restore commands to recover the juju state-server in case or failure. The juju backup command creates a tarball of the state- server’s configuration, keys, and environment data.

bash juju switch my-env juju backup

The juju restore command creates a new juju bootstrap instance using a backup file. It updates the existing instances in the environment to recognise the new state-server. The command ensures the state-server is not running before it creates the replacement. As with the normal bootstrap command, you can pass --constraints to setup the new state-server.

bash juju switch my-env juju restore juju-backup-2014-02-21.tgz

You can learn more details by running "juju restore --help".

Run arbitrary commands on some or all instances

The run command can be used by devops or scripts to inspect or do work on services, units, or machines. Commands for services or units are executed in a hook context. Charm authors can use the run command to develop and debug scripts that run in hooks.

For example, to run uname on every instance:

bash juju run "uname -a" --all

Or to run uptime on some instances:

bash juju run "uptime" --machine=2 juju run "uptime" --service=mysql

You can learn more details by running juju run --help.

Use metadata generate-tools instead of sync-tools to generate

simplestreams metadata

The sync-tools command previously generated simple streams metadata for local juju tools when provided with the --destination option. This is no longer the case. You can create the simple streams metadata for tools thusly:

bash mkdir -p $MY_DIR/tools/streams/v1 mkdir -p $MY_DIR/tools/releases cp $PUBLIC_TOOLS/*tgz $MY_DIR/tools/releases juju metadata generate-tools -d $MY_DIR

Upload the tools directory to your private cloud. Set the tools-metadata-url option in the environment’s yaml to point to the tools URL. For more details, run "juju metadata --help".

The bootstrap --source option was replaced with --metadata-source

The juju bootstrap command previously accepted the --source option which was the local path to a directory of juju tools. The bootstrap command now has a --metadata-source option that accepts the local path to simple streams metadata and tools. If your workflow previously was to download the juju tools to a local directory, then bootstrap with the --source option to upload the tools to your environment, you need to run juju metadata generate-tools per the previous section. For more details, run juju bootstrap --help.

Timeouts for bootstrap are configurable for environments.

Environments that need more time to provision an instance can configure 3 options the environments.yaml. MAAS environments often need to set bootstrap- timeout to 1800.

  • bootstrap-timeout (default: 600s)
  • bootstrap-retry-delay (default: 5s)
  • bootstrap-addresses-delay (default: 10s)

Bootstrapping the local-provider for lxc no longer requires sudo

The Juju bootstrap command cannot be run as root. Bootstrap will prompt for a password to use sudo as needed to set up the environment. This addresses several issues that arose because the owner and permissions of the local environment files were different from the owner of the process. The juju status command for example now reports the status of the machines without the need to run it with sudo.

If you have used the local provider before, you may need to manually clean up some files that are still owned by root. The environment’s jenv file commonly needs to be removed. If your local environment is named "local" then there may be a local.jenv owned by root with the JUJU_HOME directory (~/.juju). After the local environment is destroyed, you can remove the file like this

bash sudo rm ~/.juju/environments/local.jenv

Juju local provider can use clone for faster LXC

The local provider can use lxc-clone to create the containers used as machines. This feature is controlled by the "lxc-clone" option. The default is "true" for Trusty and above, and "false" for earlier Ubuntu releases. You can try to use lxc-clone on earlier releases, but it is not a supported. It may well work. You can enable lxc-clone in environments.yaml thusly:

yaml lxc-clone: true

The local provider is btrfs-aware. If your LXC directory is on a btrfs filesystem, the clones use snapshots and are much faster to create and take up much less space. There is also support for using aufs as a backing-store for the LXC clones, but there are some situations where aufs doesn’t entirely behave as intuitively as one might expect, so this must be turned on explicitly.

yaml lxc-clone-aufs: true

When using clone, the first machine to be created will create a "template" machine that is used as the basis for the clones. This will be called juju-[series]-template, so for a precise image, the name is "juju-precise- template". Do not modify or start this image while a local provider environment is running because you cannot clone a running lxc machine.

Juju now supports juju-mongodb, a mongodb tuned for juju’s needs

The Juju state-server (bootstrap node) prefers juju-mongodb. The package is available in Ubuntu Trusty, the new db will be used when a Trusty environment is bootstrapped. The juju-local package on Trusty will install juju-mongodb if it is not already installed. Upgrades of the juju-local package will continue to use mongodb-server to preserve continuity with existing local environments.

Destroying environments

The destroy-environment command requires you to specify the environment name:

bash juju destroy-environment my-press-site

Previously, destroy-environment would destroy the current environment, which might not be the one you want to destroy.

Juju unset-env

The unset-env command allows you to reset one or more the environment configuration attributes to its default value in a running Juju instance. Attributes without defaults are removed. Attempting to remove a required attribute with no default will result in an error. Multiple attributes may be removed at once; keys are space-separated.

bash juju unset-env

Juju retry-provisioning

You can use the retry-provisioning command in cases where deploying services, adding units, or adding machines fails. The command allows you to specify machines which should be retried to resolve errors reported in status.

For example, after you deployed 100 units and machines, status reports that machines 3, 27 and 57 could not be provisioned because of a rate limit exceeded error. You can ask juju to retry:

bash juju retry-provisioning 3 27 5

Bash completions include flags, machines, services, and units

Bash command line completions are improved. Pressing the TAB key will complete juju commands and their flags. Completion will also look up the machines, services, and units in an environment and suggest them.

We encourage everyone to subscribe to the Juju Discourse forum or join us on #juju on freenode.

juju-core 1.16.6

A new stable release of Juju, juju-core 1.16.6, is now available. This release replaces 1.16.5.

Getting Juju

juju-core 1.16.6 is available in trusty and backported to earlier series in the following PPA https://launchpad.net/~juju/+archive/stable

Resolved issues

  • Azure provider handles 307 temporary redirects poorly Lp 1269835
  • Loading state on bootstrap ignores ssl-hostname-verification setting Lp 1268913
  • Provide remove-* aliases for destroy-service and destroy-machine Lp 1261628

Notable Changes

This release addresses intermittent failures observed in recent weeks in Azure. Users may have seen “(request failed...307: Temporary Redirect)” when issuing juju commands.

The destroy-service and destroy-machine commands were renamed to remove- service and remove-machine respectively. The commands are still available under their old names; no scripts were harmed in the making of this release.

juju-core 1.16.5

A new stable release of Juju, juju-core 1.16.5, is now available. This release replaces 1.16.4.

Getting Juju

juju-core 1.16.5 is available in trusty and backported to earlier series in the following PPA https://launchpad.net/~juju/+archive/stable

Resolved issues

  • service with no units stuck in lifecycle dying Lp 1233457
  • destroy-machine --force Lp 089291
  • destroy-machine --force plays badly with provisioner-safe-mode Lp 1259473
  • API incompatibility: ERROR no such request "DestroyMachines" on Client Lp 1252469
  • juju destroy-machine is incompatible in trunk vs 1.16 Lp 1253643
  • juju get give a "panic: index out of range" error Lp 1227952
  • [maas] juju set-env fails after upgrading to 1.16.2+ Lp 1256179
  • juju destroy-environment destroys other environments Lp 1257481

Notable Changes

Juju may report the status of a service and machine that was terminated outside of Juju. Juju did not notice that the machine is gone. This release provides minor fixes to identify the case, and a feature to explicitly tell juju that the machine is gone.

bash juju destroy-machine --force

The destroy-machine command will destroy the machine as needed, and remove it from its list of available machines. Machines that are responsible for the environment cannot be destroyed. Machines running units or containers can only be destroyed with the --force flag; doing so will also destroy all those units and containers without giving them any opportunity to shut down cleanly.

Future version of Juju will notice when machines disappear without the need of destroy-machine --force.

juju-core 1.16.4

A new stable release of Juju, juju-core 1.16.4, is now available. This release replaces 1.16.3.

Getting Juju

juju-core 1.16.4 is available in trusty and backported to earlier series in the following PPA https://launchpad.net/~juju/+archive/stable

Resolved issues

  • Update Juju to make a "safe mode" for the provisioner Lp 1254729
  • juju set-env fails for boolean attributes Lp 1254938
  • juju-update-bootstrap-address plugin Lp 1254579
juju-core 1.16.3

A new stable release of Juju, juju-core 1.16.3, is now available. This release replaces 1.16.0.

Getting Juju

juju-core 1.16.3 is available in trusty and backported to earlier series in the following PPA https://launchpad.net/~juju/+archive/stable

Resolved issues

  • local provider deploys fail with 'install hook failed' Lp 247299
juju-core 1.16.2

A new stable release of Juju, juju-core 1.16.2, is now available. This release replaces 1.16.0.

Getting Juju

juju-core 1.16.2 is available in trusty and backported to earlier series in the following PPA https://launchpad.net/~juju/+archive/stable

Resolved issues

  • Azure bootstrap fails, versioning header is not specified. All versions of juju deploying to Azure experienced this bug when Azure updated its API on 2013-10-28. This release restores service. Lp 1246320.
  • Local provider fails to start; connection refused or machines stuck in pending. Local deploys (LXC) are now given storage. Container information is now provided for MaaS config. Lp 1240709 and Lp 1246556
  • Running destroy-environment with juju on MaaS must only destroy nodes that juju controls. Lp 1081247 and Lp 1081247
  • Ensure the cloud-archive key is installed so that downloads work without additional verification. Lp 1243861.
  • Relation names can contain underscores. Lp 1245004
juju core 1.16.1

Released:2013-10-31

juju-core 1.16.0

A new stable release of Juju, juju-core 1.16.0, is now available.

Getting Juju

juju-core 1.16.0 is available in Saucy and backported to earlier series in the following PPA https://launchpad.net/~juju/+archive/stable

New and Notable

  • bootstrap now matches on major and minor version numbers when locating tools
  • public-bucket-url is deprecated, use tools-url and/or image-metadata-url instead
  • users running from source no longer need to manually specify --upload-tools in order to have tools available for bootstrapping
  • tools retrieval is now handled using simplestreams metadata, similar to how image id lookup works
  • tools no longer have to be located in a publicly readable cloud storage instance; they can be served by any HTTP server eg apache
  • new constraint: tags. This constraint is currently only supported by MaaS environments, and allows you to specify nodes by tags applied to them in MaaS. Example: juju deploy mysql --constraints tags=foo,bar will deploy mysql to a node that is tagged with both the "foo" tag and the "bar" tag.
  • the "null" provider is now available, for creating environments with existing machines.
  • admin-secret is now chosen automatically if omitted from the configuration of a new environment.
  • control-bucket is now chosen automatically if omitted from the configuration for new ec2 and openstack environments.
  • Logging has changed. You can specify an environment variable "JUJU_LOGGING_CONFIG", or you can specify --logging-config on the command line. To show the log on the command line, now use --show-log. The --debug has been kept to be short-hand for --logging-config==DEBUG --show-log, and --verbose has been deprecated with its current meaning.

Resolved issues

  • "juju add-machine ssh:..." no longer fails when an IP address is specified instead of a hostname. LP #1225825
  • Bootstrap nodes will now only run one copy of Mongodb, this will result in lower resource utilisation and faster startup of the bootstrap node. Closes LP #1210407
  • EC2 accounts in default VPC mode now work with juju. LP #1221868
  • The unset command is now available from the command line. LP #1237929 Configuration changes
  • tools-url is a new configuration option used to specify the location where the tools tarballs may be found. If specified, this overrides the default tools repository.
  • image-metadata-url is a new configuration option used to specify the location where image metadata may be found. This is only normally required for private Openstack deployments.
  • public-bucket-url is deprecated - use tools-url to specify where tools are located (if not handled automatically by your cloud provider).

Testing on Canonistack and HP Cloud

Starting with this release, there is no longer any requirement to use a public-bucket-url configuration attribute when bootstrapping Canonistack or HP Cloud environments. The public-bucket-url attribute is deprecated and you are advised to remove it from your environment configuration.

Bootstrapping and tools

When Juju bootstraps an environment, it searches for a tools tarball matching the version of the Juju client used to run the bootstrap. Prior to this release, any tools with the same major version number as the Juju client were considered a match. Starting with this release, both the major and minor version numbers must match. This is to prevent any subtle and hard to find bugs due to version mismatch from manifesting themselves.

For users of Juju on EC2, HP Cloud, Azure, and Canonistack, where the tools are maintained by Canonical in the cloud itself, and MASS, with access to the web to download the tools, there will be no noticeable change. For private cloud setups, more care will be needed to ensure all the required tools tarballs are available.

Configuring tools

A focus of this release is to make bootstrapping a Juju environment work better out of the box, eliminating the need for manual configuration steps. To this end, the following improvements are delivered:

  1. HP Cloud and Canonistack users no longer need any special configuration over and above what goes into any generic Openstack setup for Juju.
  2. Users who compile and run Juju from source can bootstrap without manually having to tell Juju to upload a custom tools tarball. The upload-tools bootstrap option is still available to force a tools upgrade, but it is no longer required just to bootstrap the first time or when starting again after an environment has been destroyed.

Users who wish to run Juju in HP Cloud from scratch can now just follow the same steps as used for EC2:

  1. run juju init
  2. source a credentials file containing username, password, region etc, or edit the newly created ~/.juju/environments.yaml file to add username, password, region etc to the hpcloud environment block
  3. run juju bootstrap -e hpcloud

HP Cloud users can run juju switch hpcloud to make the default environment hpcloud and remove the need to use the -e parameter with each Juju command.

Private Openstack deployments

Tools can now be served by any http server, or from a shared mounted directory accessible from the cloud, removing the requirement to place the tools in a cloud storage instance. Using a cloud storage instance is still possible of course, but this option may be more desirable for users who wish to set up a private cloud without outgoing web access. The steps are as follows:

  1. Copy the tools tarballs for a given Juju version to a directory named like: /tools/releases
  2. run juju sync-tools --all --source= --destination=
  3. Configure your http server to serve to contents of under http://
  4. Add tools-url: http:// to your environments.yaml file.

An alternative to using http is to copy the tools directory itself, from , to a shared drive which is accessible to client machines and cloud instances via a common mounted directory. In that case, add an entry like this to your environments.yaml:

yaml tools-url: file://mounted_dir

Users who still wish to place the tools in a cloud instance should copy the tools directory itself to a publicly readable storage instance. Thenadd the tools-url entry to environments.yml:

yaml tools-url: https:///tools

A final alternative for locating tools is to copy the entire tools directory created above to the cloud’s control bucket, as specified in the environments.yaml file. No tools-url configuration is required, but note that these tools will only be visible to that one particular Juju environment.

Configuring image metadata

Just as private Openstack deployments need to configure where tools are located, so too does the location of the image metadata need to be specified. This is not applicable for supported openstack clouds like HP Cloud and Canonistack, only private deployments. Prior to this release, the public- bucket-url was used to point to the location of the metadata. Now, configure the image-metadata-url as follows:

yaml image-metadata-url: /juju-dist

Note that the image metadata no longer has to be kept in a cloud storage instance. The above example simply assumes the metadata will be kept in the same location as prior to this release. But as with tools, the image metadata may be server from any http url or even a directory using file://

juju-core 1.14.1

A new stable release of Juju, juju-core 1.14.1, is now available.

This release replaces juju-core 1.14.0. It reconciles differences between juju clients.

Getting Juju

juju-core 1.14.1 is available in Saucy and backported to earlier series in the following PPA

https://launchpad.net/~juju/+archive/stable

New and Notable

  • The Microsoft Azure provider is now considered ready for use.
  • lp # 1184786 juju cli now gives a clearer error message when the deprecated default-image-id key is found.
  • When debugging hooks, juju will set the $PS1 value to $JUJU_UNIT_NAME:$JUJU_HOOK_NAME %. This results in a shorter and more useful prompt. Closes ##1212903
  • Juju plugins no longer required the caller to escape all arguments passed to the plugin.
  • The new command “juju unset …” allows a configuration setting for a service to be removed. Removing the configuration setting will revert it to the charm default. This is the first step in solving #1194945.
  • The Windows Azure provider now defines a public storage container for tools. No configuration changes are required. Closes #1216770.
  • The debug-hooks subcommand is now implemented. Closed issue #1027876
  • The status subcommand now accepts service/unit filters, to limit output; see “juju help status” for more information. Partially addresses issue #1121916
  • The juju cli command now returns a better error message when run against an environment that has not been bootstrapped. Closes issue #1028053.

Resolved issues

  • The destroy-environment subcommand now requests confirmation before running, with an option to quietly proceed (-y/--yes). Closed issue #1121914.
  • The cost calculation for the openstack provider was updated to bias towards the lowest number of CPU cores when choosing a machine type matching the provided instance constraints. This closes issue #1210086.
  • The unit agent now removes stale sockets when restarting. Closes #1212916
  • A bug relating to the formatting of tool versions has been resolved. This formatting error was only visible when using the -v or --debug flag. Closes #1216285.
  • lp #1206628 a bug related to the name of an upstart job created for a service unit has been corrected.
  • lp # 1207230 a bug related specifying the type of local provider container has been corrected.

Known issues

  • The JUJU_ENV environment variable is not being passed down to plugins. LP #1216254.

Testing on Canonistack and HP Cloud

A publicly readable bucket has been created for holding the Juju tools on Canonistack. To use it, put this in your ~/.juju/environments.yaml (all on one line):

public-bucket-url: https://swift.canonistack.canonical.com/v1/AUTH_526ad877f3e3464589dc1145dfeaac60

For HP Cloud the public bucket is available at:

public-bucket-url: https://region-a.geo-1.objects.hpcloudsvc.com/v1/60502529753910
juju-core 1.14.0

A new stable release of Juju, juju-core 1.14.0, is now available.

Getting Juju

juju-core 1.14.0 is available in Saucy and backported to earlier series in the following PPA

https://launchpad.net/~juju/+archive/stable

New and Notable

  • The Microsoft Azure provider is now considered ready for use.
  • lp # 1184786 juju cli now gives a clearer error message when the deprecated default-image-id key is found.
  • When debugging hooks, juju will set the $PS1 value to $JUJU_UNIT_NAME:$JUJU_HOOK_NAME %. This results in a shorter and more useful prompt. Closes ##1212903
  • Juju plugins no longer required the caller to escape all arguments passed to the plugin.
  • The new command “juju unset …” allows a configuration setting for a service to be removed. Removing the configuration setting will revert it to the charm default. This is the first step in solving #1194945.
  • The Windows Azure provider now defines a public storage container for tools. No configuration changes are required. Closes #1216770.
  • The debug-hooks subcommand is now implemented. Closed issue #1027876
  • The status subcommand now accepts service/unit filters, to limit output; see “juju help status” for more information. Partially addresses issue #1121916
  • The juju cli command now returns a better error message when run against an environment that has not been bootstrapped. Closes issue #1028053.

Resolved issues

  • The destroy-environment subcommand now requests confirmation before running, with an option to quietly proceed (-y/--yes). Closed issue #1121914.
  • The cost calculation for the openstack provider was updated to bias towards the lowest number of CPU cores when choosing a machine type matching the provided instance constraints. This closes issue #1210086.
  • The unit agent now removes stale sockets when restarting. Closes #1212916
  • A bug relating to the formatting of tool versions has been resolved. This formatting error was only visible when using the -v or --debug flag. Closes #1216285.
  • lp #1206628 a bug related to the name of an upstart job created for a service unit has been corrected.
  • lp # 1207230 a bug related specifying the type of local provider container has been corrected.

Known issues

  • The JUJU_ENV environment variable is not being passed down to plugins. LP #1216254.

Testing on Canonistack and HP Cloud

A publicly readable bucket has been created for holding the Juju tools on Canonistack. To use it, put this in your ~/.juju/environments.yaml (all on one line):

public-bucket-url: https://swift.canonistack.canonical.com/v1/AUTH_526ad877f3e3464589dc1145dfeaac60

For HP Cloud the public bucket is available at:

public-bucket-url: https://region-a.geo-1.objects.hpcloudsvc.com/v1/60502529753910