Agent Base Configuration

Set System Defaults

On Unix, the entrypoint for the agent commands is a shell script <OSVCBIN>/om that supports defaults injection.

In most situations, this configuration file does not need modification.

Defaults file location:

System

Location

Debian-like

/etc/default/opensvc

Red Hat-like

/etc/sysconfig/opensvc

HP-UX

/etc/rc.config.d/opensvc

AIX

/etc/default/opensvc

SunOS

/etc/default/opensvc

Tru64

/etc/default/opensvc

FreeBSD

/etc/defaults/opensvc

Darwin

/etc/defaults/opensvc

In this sourced file, you can export systems environment variables like LD_PRELOAD or LD_LIBRARY_PATH, and set the following OpenSVC-specific variables.

Variable

Default

Role

OSVC_BOOT_OPTS

Additional parameters passed to the ‘svcmgr boot’ command upon system startup

OSVC_ROOT_PATH

Developpers can set this to their git repository to use the agent from there. If left unset, the agent file are expected in the Filesystem Hierarchy Standard locations.

OSVC_PYTHON

python

Define which python executable to use. ex: /usr/local/python-2.7.3/bin/python

OSVC_PYTHON_ARGS

Additional parameters passed to the python interpreter. ex: debug parameters

Set Node Environment

om cluster set --kw node.env=PRD

The node.env setting is used to enforce the following policies:

  • Only PRD services are allowed to start on a PRD node

  • Only PRD nodes are allowed to push data to a PRD node

Supported node.env values:

Env

Behaves As

Description

PRD

PRD

Production

PPRD

PRD

Pre Production

REC

not PRD

Prod-like testing

INT

not PRD

Integration

DEV

not PRD

Development

TST

not PRD

Testing (Default)

TMP

not PRD

Temporary

DRP

not PRD

Disaster recovery

FOR

not PRD

Training

PRA

not PRD

Disaster recovery

PRJ

not PRD

Project

STG

not PRD

Staging

The setting is stored in <OSVCETC>/cluster.conf.

Set Schedules

The agent executes periodic tasks. All tasks have a default schedule, which you may want to change.

A schedule configuration can be applied using

om node set --kw "brocade.schedule=02:00-04:00@120 sat,sun"

Node schedules are defined in <OSVCETC>/node.conf, where the above command would produce this section:

[brocade]
schedule = 02:00-04:00@120 sat,sun

The live scheduler configuration and states can be extracted with

$ om node print schedule
Action                  Last Run             Config Parameter           Schedule Definition                               
|- auto_reboot          2017-09-30 16:59:19  reboot.schedule            16:00-17:00@1 sat:last,tue-mon:last * %2+1,feb-apr
|- auto_rotate_root_pw  -                    rotate_root_pw.schedule    -
|- checks               2017-10-01 17:43:29  checks.schedule            ["16:00-21:00@30 *:last", "! * wed", "*@1"]
|- collect_stats        2017-10-01 17:42:29  stats_collection.schedule  @10
|- compliance_auto      2017-10-01 00:01:22  compliance.schedule        00:00-01:00@61
|- dequeue_actions      2017-01-30 10:02:01  dequeue_actions.schedule   -
|- pushasset            2017-10-01 00:06:22  asset.schedule             00:00-06:00@361 mon-sun
|- pushbrocade          -                    brocade.schedule           -
|- pushcentera          -                    centera.schedule           -
|- pushdcs              -                    dcs.schedule               -
|- pushdisks            2017-10-01 00:03:22  disks.schedule             00:00-06:00@361 mon-sun
|- pushemcvnx           -                    emcvnx.schedule            -
|- pusheva              -                    eva.schedule               -
|- pushfreenas          -                    freenas.schedule           -
|- pushgcedisks         -                    gcedisks.schedule          -
|- pushhds              -                    hds.schedule               -
|- pushhp3par           -                    hp3par.schedule            -
|- pushibmds            -                    ibmds.schedule             -
|- pushibmsvc           -                    ibmsvc.schedule            -
|- pushnecism           -                    necism.schedule            -
|- pushnetapp           -                    netapp.schedule            -
|- pushnsr              -                    nsr.schedule               -
|- pushpatch            2017-10-01 00:16:01  patches.schedule           00:00-06:00@361 mon-sun
|- pushpkg              2017-10-01 00:12:01  packages.schedule          00:00-06:00@361 mon-sun
|- pushstats            2017-10-01 17:41:29  stats.schedule             ["00:00-23:59@10"]
|- pushsym              -                    sym.schedule               -
|- pushvioserver        -                    vioserver.schedule         -
|- pushxtremio          -                    xtremio.schedule           -
`- sysreport            2017-10-01 00:25:02  sysreport.schedule         00:00-06:00@361 mon-sun

See also

Scheduler

Register on a Collector

Set a Collector Url

By default, the agent does not communicate with a collector.

To enable communications with a collector, the node.dbopensvc node configuration parameter must be set. The simplest expression is:

om cluster set --kw node.dbopensvc=collector.opensvc.com

Here the protocol and path are omitted. In this case, the https protocol is selected, and the path set to a value matching the standard collector integration.

Advanced Url Formats

The following expressions are also supported:

om cluster set --kw node.dbopensvc=https://collector.opensvc.com
om cluster set --kw node.dbopensvc=https://collector.opensvc.com/feed/default/call/xmlrpc

The compliance framework uses a separate xmlrpc entrypoint. The node.dbcompliance can be set to override the default, which is deduced from the node.dbopensvc value.

om cluster set --kw node.dbcompliance=https://collector.opensvc.com/init/compliance/call/xmlrpc

Register the Node

The collector requires the nodes to provide an authentication token (shared secret) with each request. The token is forged by the collector and stored on the node in <OSVCETC>/node.conf. The token initialization is handled by the command:

om node register

Collectors in SaaS mode, like https://collector.opensvc.com, require that you prove your identity. The command is thus:

om node register --user my.self@my.com [--app MYAPP]

If --app is not specified the collector automatically chooses one the user is responsible of.

A successful register is followed by a node discovery, so the collector has detailled information about the node and can serve contextualized compliance rulesets up front. The discovery is also scheduled daily, and can be manually replayed with:

om node pushasset
om node pushpkg
om node pushpatch
om node pushstats
om node checks

To disable collector communications, use:

om cluster unset --kw node.dbopensvc
om cluster unset --kw node.dbcompliance

Or if the settings were added to node.conf

om node unset --kw node.dbopensvc
om node unset --kw node.dbcompliance

Extra System Configurations

HP-UX

The python package provided by HP will output garbage on exec because it won’t find terminfo at the expected places. To fix that, you have to export TERMINFO=/usr/share/lib/terminfo from /etc/profile

The HP-UX base system does not provide tools to handle scsi persistent reservations. You have to install the scu tool if you want to activate this feature.

Linux LVM2

OpenSVC controls volume group activation and desactivation. Most Linux distributions activate all visible volume groups at boot, some even re-activate them upon de-activation events. These mecanisms can be disabled using the following setup. It also provides another protection against unwanted volume group activation from a secondary cluster node.

This setup tells LVM2 commands to activate only the objects tagged with the hostname. Opensvc makes sure the tags are set on start and unset on stop. Opensvc also purges all tags before adding the one it needs to activate a volume group, so opensvc can satisfy a start request on a service uncleanly shut down.

/etc/lvm/lvm.conf

Add the following root-level configuration node:

tags {
    hosttags = 1
    local {}
}

And add the local tag to all local volume groups. For example:

sudo vgchange --addtag local rootvg

Finally you need to rebuild the initrd/initramfs to prevent shared vg activation at boot.

/etc/lvm/lvm_$HOSTNAME.conf

echo activation { volume_list = [\"@local\", \"@$HOSTNAME\"] } >/etc/lvm/lvm_$HOSTNAME.conf

Windows

Dependencies

The OpenSVC agent on Windows depends on:

  • Python 2.6+

  • Python win32 library

  • Microsoft fcinfo for Fibre Channel SAN reporting (optional)

The provided OpenSVC executable installer brings everything except fcinfo tool.

Silent Install

It’s possible to trigger a silent install by using the /S (uppercase) command line switch:

OpenSVC.X.Y.exe /S

There’s also a command line option to specify the target installation folder (no quotes in folder name even with spaces inside):

OpenSVC.X.Y.exe /S  /D=C:\My Path with spaces

GUI Install

Double click on OpenSVC.X.Y.exe and follow install wizard

Upgrade

Upgrading the OpenSVC package manually is the same as an installation from scratch:

OpenSVC.X.Z.exe /S

The installer deals with installation directory detection, and upgrade software in the accurate folder. It’s still a best practice to have a system/data backup before upgrading OpenSVC software.

Mac OS X

Install

curl -o /tmp/opensvc.latest.pkg https://repo.opensvc.com/macos-pkg/current
installer -pkg /tmp/opensvc.latest.pkg  -target /

Uninstall

As MacOS X does not provide a clean way to remove packages, we do it by ourselves

Warning

Backup any configuration file in <OSVCETC> before removing them from the hard disk drive

rm -f /Library/LaunchDaemons/com.opensvc.svcmgr.plist
pkgutil --forget com.opensvc.agent