Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Excerpt

Released on March 2, 2016.

New Features

  • Trend detection:

    • A first version of trend detection is available via the squirro client.

  • Dashboard:
    • Table widget allows numerical aggregations now

Improvements

  • Filter refactoring:
    • Filters, Smart Filters and Saved Searches are now unified in a new side menu for easier access. The top level tab 'Smart Filters' has been removed.

  • Pipelet descriptions support markdown notation.

  • Object title is no longer an optional attribute.

  • Numerical facet values are casted to integer or floats in case this facet is specified as a numerical typed facet.

Bug Fixes

  • Frontend assets caching is fixed.

  • Bulkindexing keywords containing spaces is fixed.

 

Upgrade Instructions

Note

These are the upgrade instructions if you upgrade from version 2.2.6. If you are upgrading from an older version e.g. 2.1.5 please follow the upgrade instructions on Squirro 2.2.0 - Release Notes and Squirro 2.2.6 - Release Notes.

Warning

This is not the latest version of Squirro. To upgrade to this version, please ensure that you point the squirro yum repository to version '2.3.0' and not to 'latest'.

Note

If you are using Squirro in a Box, then there are additional steps involved. In this case, please contact support.

 

1. Upgrade Storage Nodes

If your storage node runs in the same Virtual Machine or Operating System as your cluster node, skip this step. Otherwise upgrade all storage nodes one at a time by running:

Code Block
languagebash
[squirro@storagenode01 ~] sudo yum update

2. Upgrade Cluster Nodes

Code Block
languagebash
[squirro@clusternode01 ~] sudo yum update squirro-cluster-node-users
[squirro@clusternode01 ~] sudo yum update

As with this release a new service is added, some additional steps are required.

On a single-cluster-node setup you need to do the following steps on the cluster node:

Code Block
[squirro@clusternode01 ~] sudo service monit restart
[squirro@clusternode01 ~] sudo service nginx restart

If you run Squirro in a multi-cluster-node environment you need to do the following additional steps on each cluster node:

Code Block
[squirro@clusternode01 ~] sudo sed -i -e 's/db_endpoint_discovery = false/db_endpoint_discovery = true/' /etc/squirro/trends.ini
[squirro@clusternode01 ~] sudo mkdir /mnt/gv0/trends_data
[squirro@clusternode01 ~] sudo chown sqtrends:squirro /mnt/gv0/trends_data
[squirro@clusternode01 ~] sudo sed '$ a [offline_processing]\ndata_directory = /mnt/gv0/trends_data' -i /etc/squirro/trends.ini
[squirro@clusternode01 ~] sudo service monit restart
[squirro@clusternode01 ~] sudo service nginx restart

 

3. Resolve configurations (if required)

The Squirro packages attempt to upgrade *.ini and *.conf configuration files automatically. However if you have made local modifications, the upgrade results in *.rpmnew files that you would need to merge manually. We recommend

  1. backing up the previous *.ini and *.conf files to *.ini.orig and *.conf.orig,
  2. renaming the *.ini.rpmnew and *.conf.rpmnew files to *.ini and *.conf respectively
  3. inspecting all the *.orig files individually and porting any settings manually

Use the following scripts to look for and resolve unresolved configuration files on each cluster node:

 

Code Block
title1. First remember all *.rpmnew because we are about to rename them, but we want to remember the list for step 3 below
[squirro@clusternode01 ~] sudo su
[root@clusternode01 ~] FILES_TO_RESOLVE=`ls /etc/squirro/*.ini.rpmnew /etc/nginx/conf.d/*.conf.rpmnew /etc/monit.d/*.rpmnew 2> /dev/null | sed -e "s/\.rpmnew//"`

 

 

Code Block
title2. Back up original config files and make the *.rpmnew files the real configuration files (\cp and \rm are used to avoid interactive alias of the commands and to ensure that the config file retains the ownership of the original config file)
[root@clusternode01 ~] for CONFIG_FILE in ${FILES_TO_RESOLVE}; do cp ${CONFIG_FILE} ${CONFIG_FILE}.orig; cp ${CONFIG_FILE}.rpmnew ${CONFIG_FILE}; rm ${CONFIG_FILE}.rpmnew; done

 

 

Code Block
title3. Finally merge the commands. Exit vim via :q<enter>:q<enter> to accept a new config files verbatim
[root@clusternode01 ~] for CONFIG_FILE in ${FILES_TO_RESOLVE}; do vim -O ${CONFIG_FILE}.orig ${CONFIG_FILE}; done 

 

 

 

If you have previously edited configuration files manually then consult the Configuration Conflicts section for how to identify and resolve any possible conflicts.