aboutsummaryrefslogtreecommitdiffstats
path: root/sysutils/logstash
diff options
context:
space:
mode:
authormadpilot <madpilot@FreeBSD.org>2013-07-18 07:20:21 +0800
committermadpilot <madpilot@FreeBSD.org>2013-07-18 07:20:21 +0800
commit2fe8644d9a2f5e6484513dbc855404d593e1f1cb (patch)
tree9949d0a8d553dd752525ff90d320eb8e0db46051 /sysutils/logstash
parent19ab19440e217fd63ddc8cdc64ea26cada3c497c (diff)
downloadfreebsd-ports-gnome-2fe8644d9a2f5e6484513dbc855404d593e1f1cb.tar.gz
freebsd-ports-gnome-2fe8644d9a2f5e6484513dbc855404d593e1f1cb.tar.zst
freebsd-ports-gnome-2fe8644d9a2f5e6484513dbc855404d593e1f1cb.zip
Logstash is a tool for managing events and logs. You can use it to
collect logs, parse them, and store them for later use (like, for searching). Speaking of searching, logstash comes with a web interface for searching and drilling into all of your logs. WWW: http://logstash.net/ PR: ports/168266 Submitted by: Daniel Solsona <daniel@ish.com.au>, Regis A. Despres <regis.despres@gmail.com>
Diffstat (limited to 'sysutils/logstash')
-rw-r--r--sysutils/logstash/Makefile50
-rw-r--r--sysutils/logstash/distinfo2
-rw-r--r--sysutils/logstash/files/elasticsearch.yml.sample337
-rw-r--r--sysutils/logstash/files/logstash.conf.sample38
-rw-r--r--sysutils/logstash/files/logstash.in81
-rw-r--r--sysutils/logstash/pkg-descr6
-rw-r--r--sysutils/logstash/pkg-plist13
7 files changed, 527 insertions, 0 deletions
diff --git a/sysutils/logstash/Makefile b/sysutils/logstash/Makefile
new file mode 100644
index 000000000000..57054d0b0f56
--- /dev/null
+++ b/sysutils/logstash/Makefile
@@ -0,0 +1,50 @@
+# Created by: Daniel Solsona <daniel@ish.com.au>, Guido Falsi <madpilot@FreeBSD.org>
+# $FreeBSD$
+
+PORTNAME= logstash
+PORTVERSION= 1.1.13
+CATEGORIES= sysutils java
+MASTER_SITES= https://logstash.objects.dreamhost.com/release/ \
+ http://semicomplete.com/files/logstash/
+DISTNAME= ${PORTNAME}-${PORTVERSION}-flatjar
+EXTRACT_SUFX= .jar
+EXTRACT_ONLY=
+
+MAINTAINER= regis.despres@gmail.com
+COMMENT= Tool for managing events and logs
+
+USE_JAVA= yes
+JAVA_VERSION= 1.5+
+
+NO_BUILD= yes
+
+USE_RC_SUBR= logstash
+
+LOGSTASH_HOME?= ${PREFIX}/${PORTNAME}
+LOGSTASH_HOME_REL?= ${LOGSTASH_HOME:S,^${PREFIX}/,,}
+LOGSTASH_JAR?= ${DISTNAME}${EXTRACT_SUFX}
+LOGSTASH_RUN?= /var/run/${PORTNAME}
+LOGSTASH_DATA_DIR?= /var/db/${PORTNAME}
+
+SUB_LIST= LOGSTASH_DATA_DIR=${LOGSTASH_DATA_DIR} JAVA_HOME=${JAVA_HOME} \
+ LOGSTASH_HOME=${LOGSTASH_HOME} LOGSTASH_JAR=${LOGSTASH_JAR}
+PLIST_SUB+= LOGSTASH_HOME=${LOGSTASH_HOME_REL} LOGSTASH_JAR=${LOGSTASH_JAR} \
+ LOGSTASH_RUN=${LOGSTASH_RUN} \
+ LOGSTASH_DATA_DIR=${LOGSTASH_DATA_DIR}
+
+do-install:
+ ${MKDIR} ${LOGSTASH_RUN}
+ ${MKDIR} ${ETCDIR}
+ ${MKDIR} ${LOGSTASH_HOME}
+ ${MKDIR} ${LOGSTASH_DATA_DIR}
+ ${INSTALL_DATA} ${DISTDIR}/${DIST_SUBDIR}/${LOGSTASH_JAR} ${LOGSTASH_HOME}
+ ${INSTALL_DATA} ${FILESDIR}/logstash.conf.sample ${ETCDIR}
+ @if [ ! -f ${ETCDIR}/logstash.conf ]; then \
+ ${CP} -p ${ETCDIR}/logstash.conf.sample ${ETCDIR}/logstash.conf ; \
+ fi
+ ${INSTALL_DATA} ${FILESDIR}/elasticsearch.yml.sample ${ETCDIR}
+ @if [ ! -f ${ETCDIR}/elasticsearch.yml ]; then \
+ ${CP} -p ${ETCDIR}/elasticsearch.yml.sample ${ETCDIR}/elasticsearch.yml ; \
+ fi
+
+.include <bsd.port.mk>
diff --git a/sysutils/logstash/distinfo b/sysutils/logstash/distinfo
new file mode 100644
index 000000000000..f3fdd78ce0d7
--- /dev/null
+++ b/sysutils/logstash/distinfo
@@ -0,0 +1,2 @@
+SHA256 (logstash-1.1.13-flatjar.jar) = 5ba0639ff4da064c2a4f6a04bd7006b1997a6573859d3691e210b6855e1e47f1
+SIZE (logstash-1.1.13-flatjar.jar) = 69485313
diff --git a/sysutils/logstash/files/elasticsearch.yml.sample b/sysutils/logstash/files/elasticsearch.yml.sample
new file mode 100644
index 000000000000..9bcb096ad02b
--- /dev/null
+++ b/sysutils/logstash/files/elasticsearch.yml.sample
@@ -0,0 +1,337 @@
+##################### ElasticSearch Configuration Example #####################
+
+# This file contains an overview of various configuration settings,
+# targeted at operations staff. Application developers should
+# consult the guide at <http://elasticsearch.org/guide>.
+#
+# The installation procedure is covered at
+# <http://elasticsearch.org/guide/reference/setup/installation.html>.
+#
+# ElasticSearch comes with reasonable defaults for most settings,
+# so you can try it out without bothering with configuration.
+#
+# Most of the time, these defaults are just fine for running a production
+# cluster. If you're fine-tuning your cluster, or wondering about the
+# effect of certain configuration option, please _do ask_ on the
+# mailing list or IRC channel [http://elasticsearch.org/community].
+
+# Any element in the configuration can be replaced with environment variables
+# by placing them in ${...} notation. For example:
+#
+# node.rack: ${RACK_ENV_VAR}
+
+# See <http://elasticsearch.org/guide/reference/setup/configuration.html>
+# for information on supported formats and syntax for the configuration file.
+
+
+################################### Cluster ###################################
+
+# Cluster name identifies your cluster for auto-discovery. If you're running
+# multiple clusters on the same network, make sure you're using unique names.
+#
+# cluster.name: elasticsearch
+
+
+#################################### Node #####################################
+
+# Node names are generated dynamically on startup, so you're relieved
+# from configuring them manually. You can tie this node to a specific name:
+#
+# node.name: "Franz Kafka"
+
+# Every node can be configured to allow or deny being eligible as the master,
+# and to allow or deny to store the data.
+#
+# Allow this node to be eligible as a master node (enabled by default):
+#
+# node.master: true
+#
+# Allow this node to store data (enabled by default):
+#
+# node.data: true
+
+# You can exploit these settings to design advanced cluster topologies.
+#
+# 1. You want this node to never become a master node, only to hold data.
+# This will be the "workhorse" of your cluster.
+#
+# node.master: false
+# node.data: true
+#
+# 2. You want this node to only serve as a master: to not store any data and
+# to have free resources. This will be the "coordinator" of your cluster.
+#
+# node.master: true
+# node.data: false
+#
+# 3. You want this node to be neither master nor data node, but
+# to act as a "search load balancer" (fetching data from nodes,
+# aggregating results, etc.)
+#
+# node.master: false
+# node.data: false
+
+# Use the Cluster Health API [http://localhost:9200/_cluster/health], the
+# Node Info API [http://localhost:9200/_cluster/nodes] or GUI tools
+# such as <http://github.com/lukas-vlcek/bigdesk> and
+# <http://mobz.github.com/elasticsearch-head> to inspect the cluster state.
+
+# A node can have generic attributes associated with it, which can later be used
+# for customized shard allocation filtering, or allocation awareness. An attribute
+# is a simple key value pair, similar to node.key: value, here is an example:
+#
+# node.rack: rack314
+
+
+#################################### Index ####################################
+
+# You can set a number of options (such as shard/replica options, mapping
+# or analyzer definitions, translog settings, ...) for indices globally,
+# in this file.
+#
+# Note, that it makes more sense to configure index settings specifically for
+# a certain index, either when creating it or by using the index templates API.
+#
+# See <http://elasticsearch.org/guide/reference/index-modules/> and
+# <http://elasticsearch.org/guide/reference/api/admin-indices-create-index.html>
+# for more information.
+
+# Set the number of shards (splits) of an index (5 by default):
+#
+# index.number_of_shards: 5
+
+# Set the number of replicas (additional copies) of an index (1 by default):
+#
+# index.number_of_replicas: 1
+
+# Note, that for development on a local machine, with small indices, it usually
+# makes sense to "disable" the distributed features:
+#
+# index.number_of_shards: 1
+# index.number_of_replicas: 0
+
+# These settings directly affect the performance of index and search operations
+# in your cluster. Assuming you have enough machines to hold shards and
+# replicas, the rule of thumb is:
+#
+# 1. Having more *shards* enhances the _indexing_ performance and allows to
+# _distribute_ a big index across machines.
+# 2. Having more *replicas* enhances the _search_ performance and improves the
+# cluster _availability_.
+#
+# The "number_of_shards" is a one-time setting for an index.
+#
+# The "number_of_replicas" can be increased or decreased anytime,
+# by using the Index Update Settings API.
+#
+# ElasticSearch takes care about load balancing, relocating, gathering the
+# results from nodes, etc. Experiment with different settings to fine-tune
+# your setup.
+
+# Use the Index Status API (<http://localhost:9200/A/_status>) to inspect
+# the index status.
+
+
+#################################### Paths ####################################
+
+# Path to directory containing configuration (this file and logging.yml):
+#
+# path.conf: /path/to/conf
+
+# Path to directory where to store index data allocated for this node.
+#
+# path.data: /path/to/data
+#
+# Can optionally include more than one location, causing data to be striped across
+# the locations on a file level, favouring locations with most free
+# space on creation. For example:
+#
+# path.data: /path/to/data1,/path/to/data2
+
+# Path to temporary files:
+#
+# path.work: /path/to/work
+
+# Path to log files:
+#
+# path.logs: /path/to/logs
+
+# Path to where plugins are installed:
+#
+# path.plugins: /path/to/plugins
+
+
+################################### Memory ####################################
+
+# ElasticSearch performs poorly when JVM starts swapping: you should ensure that
+# it _never_ swaps.
+#
+# Set this property to true to lock the memory:
+#
+# bootstrap.mlockall: true
+
+# Make sure that the ES_MIN_MEM and ES_MAX_MEM environment variables are set
+# to the same value, and that the machine has enough memory to allocate
+# for ElasticSearch, leaving enough memory for the operating system itself.
+#
+# You should also make sure that the ElasticSearch process is allowed to lock
+# the memory, eg. by using `ulimit -l unlimited`.
+
+
+############################## Network And HTTP ###############################
+
+# ElasticSearch, by default, binds itself to the 0.0.0.0 address, and listens
+# on port [9200-9300] for HTTP traffic and on port [9300-9400] for node-to-node
+# communication. (the range means that if the port is busy, it will automatically
+# try the next port).
+
+# Set the bind address specifically (IPv4 or IPv6):
+#
+# network.bind_host: 192.168.0.1
+
+# Set the address other nodes will use to communicate with this node. If not
+# set, it is automatically derived. It must point to an actual IP address.
+#
+# network.publish_host: 192.168.0.1
+
+# Set both 'bind_host' and 'publish_host':
+#
+# network.host: 192.168.0.1
+
+# Set a custom port for the node to node communication (9300 by default):
+#
+# transport.port: 9300
+
+# Enable compression for all communication between nodes (disabled by default):
+#
+# transport.tcp.compress: true
+
+# Set a custom port to listen for HTTP traffic:
+#
+# http.port: 9200
+
+# Set a custom allowed content length:
+#
+# http.max_content_length: 100mb
+
+# Disable HTTP completely:
+#
+# http.enabled: false
+
+
+################################### Gateway ###################################
+
+# The gateway allows for persisting the cluster state between full cluster
+# restarts. Every change to the state (such as adding an index) will be stored
+# in the gateway, and when the cluster starts up for the first time,
+# it will read its state from the gateway.
+
+# There are several types of gateway implementations. For more information,
+# see <http://elasticsearch.org/guide/reference/modules/gateway>.
+
+# The default gateway type is the "local" gateway (recommended):
+#
+# gateway.type: local
+
+# Settings below control how and when to start the initial recovery process on
+# a full cluster restart (to reuse as much local data as possible).
+
+# Allow recovery process after N nodes in a cluster are up:
+#
+# gateway.recover_after_nodes: 1
+
+# Set the timeout to initiate the recovery process, once the N nodes
+# from previous setting are up (accepts time value):
+#
+# gateway.recover_after_time: 5m
+
+# Set how many nodes are expected in this cluster. Once these N nodes
+# are up, begin recovery process immediately:
+#
+# gateway.expected_nodes: 2
+
+
+############################# Recovery Throttling #############################
+
+# These settings allow to control the process of shards allocation between
+# nodes during initial recovery, replica allocation, rebalancing,
+# or when adding and removing nodes.
+
+# Set the number of concurrent recoveries happening on a node:
+#
+# 1. During the initial recovery
+#
+# cluster.routing.allocation.node_initial_primaries_recoveries: 4
+#
+# 2. During adding/removing nodes, rebalancing, etc
+#
+# cluster.routing.allocation.node_concurrent_recoveries: 2
+
+# Set to throttle throughput when recovering (eg. 100mb, by default unlimited):
+#
+# indices.recovery.max_size_per_sec: 0
+
+# Set to limit the number of open concurrent streams when
+# recovering a shard from a peer:
+#
+# indices.recovery.concurrent_streams: 5
+
+
+################################## Discovery ##################################
+
+# Discovery infrastructure ensures nodes can be found within a cluster
+# and master node is elected. Multicast discovery is the default.
+
+# Set to ensure a node sees N other master eligible nodes to be considered
+# operational within the cluster. Set this option to a higher value (2-4)
+# for large clusters:
+#
+# discovery.zen.minimum_master_nodes: 1
+
+# Set the time to wait for ping responses from other nodes when discovering.
+# Set this option to a higher value on a slow or congested network
+# to minimize discovery failures:
+#
+# discovery.zen.ping.timeout: 3s
+
+# See <http://elasticsearch.org/guide/reference/modules/discovery/zen.html>
+# for more information.
+
+# Unicast discovery allows to explicitly control which nodes will be used
+# to discover the cluster. It can be used when multicast is not present,
+# or to restrict the cluster communication-wise.
+#
+# 1. Disable multicast discovery (enabled by default):
+#
+# discovery.zen.ping.multicast.enabled: false
+#
+# 2. Configure an initial list of master nodes in the cluster
+# to perform discovery when new nodes (master or data) are started:
+#
+# discovery.zen.ping.unicast.hosts: ["host1", "host2:port", "host3[portX-portY]"]
+
+# EC2 discovery allows to use AWS EC2 API in order to perform discovery.
+#
+# You have to install the cloud-aws plugin for enabling the EC2 discovery.
+#
+# See <http://elasticsearch.org/guide/reference/modules/discovery/ec2.html>
+# for more information.
+#
+# See <http://elasticsearch.org/tutorials/2011/08/22/elasticsearch-on-ec2.html>
+# for a step-by-step tutorial.
+
+
+################################## Slow Log ##################################
+
+# Shard level query and fetch threshold logging.
+
+#index.search.slowlog.level: TRACE
+#index.search.slowlog.threshold.query.warn: 10s
+#index.search.slowlog.threshold.query.info: 5s
+#index.search.slowlog.threshold.query.debug: 2s
+#index.search.slowlog.threshold.query.trace: 500ms
+
+#index.search.slowlog.threshold.fetch.warn: 1s
+#index.search.slowlog.threshold.fetch.info: 800ms
+#index.search.slowlog.threshold.fetch.debug: 500ms
+#index.search.slowlog.threshold.fetch.trace: 200ms
diff --git a/sysutils/logstash/files/logstash.conf.sample b/sysutils/logstash/files/logstash.conf.sample
new file mode 100644
index 000000000000..cd6ab9a6fcad
--- /dev/null
+++ b/sysutils/logstash/files/logstash.conf.sample
@@ -0,0 +1,38 @@
+input {
+ file {
+ type => "system logs"
+
+ # # Wildcards work, here :)
+ # path => [ "/var/log/*.log", "/var/log/messages", "/var/log/syslog" ]
+ path => [ "/var/log/messages" ]
+ }
+
+ #file {
+ # type => "Hudson-access"
+ # path => "/var/log/www/hudson.ish.com.au-access_log"
+ #}
+
+ #file {
+ # type => "Syslog"
+ # path => "/var/log/messages"
+ #}
+}
+
+output {
+ # Emit events to stdout for easy debugging of what is going through
+ # logstash.
+ #stdout { }
+
+ # This will use elasticsearch to store your logs.
+ # The 'embedded' option will cause logstash to run the elasticsearch
+ # server in the same process, so you don't have to worry about
+ # how to download, configure, or run elasticsearch!
+ elasticsearch {
+ embedded => true
+ #embedded_http_port => 9200
+ #cluster => elasticsearch
+ #host => host
+ #port => port
+
+ }
+}
diff --git a/sysutils/logstash/files/logstash.in b/sysutils/logstash/files/logstash.in
new file mode 100644
index 000000000000..fedec80b52f0
--- /dev/null
+++ b/sysutils/logstash/files/logstash.in
@@ -0,0 +1,81 @@
+#!/bin/sh
+
+# $FreeBSD$
+#
+# PROVIDE: logstash
+# REQUIRE: LOGIN
+# KEYWORD: shutdown
+#
+#
+# Configuration settings for logstash in /etc/rc.conf:
+#
+# logstash_enable (bool):
+# Set to "NO" by default.
+# Set it to "YES" to enable logstash
+#
+# logstash_mode :
+# Set to "standalone" by default.
+# Valid options:
+# "standalone": agent, web & elasticsearch
+# "web": Starts logstash as a web ui
+# "agent": Justs works as a log shipper
+#
+# logstash_logging (bool):
+# Set to "NO" by default.
+# Set it to "YES" to enable logstash logging to file
+# Default output to /var/log/logstash.log
+#
+
+. /etc/rc.subr
+
+name=logstash
+rcvar=logstash_enable
+
+load_rc_config ${name}
+
+: ${logstash_enable="NO"}
+: ${logstash_home="%%LOGSTASH_HOME%%"}
+: ${logstash_config="%%PREFIX%%/etc/${name}/${name}.conf"}
+: ${logstash_jar="%%LOGSTASH_HOME%%/%%LOGSTASH_JAR%%"}
+: ${logstash_java_home="%%JAVA_HOME%%"}
+: ${logstash_log="NO"}
+: ${logstash_mode="standalone"}
+: ${logstash_port="9292"}
+: ${logstash_elastic_backend=""}
+: ${logstash_log_file="${logdir}/${name}.log"}
+: ${logstash_elastic_datadir="%%LOGSTASH_DATA_DIR%%"}
+
+piddir=/var/run/${name}
+pidfile=${piddir}/${name}.pid
+
+if [ -d $piddir ]; then
+ mkdir -p $piddir
+fi
+
+logdir="/var/log"
+command="/usr/sbin/daemon"
+
+java_cmd="${logstash_java_home}/bin/java"
+procname="${java_cmd}"
+
+logstash_chdir=${logstash_home}
+logstash_log_options=""
+logstash_elastic_options=""
+
+if checkyesno logstash_log; then
+ logstash_log_options=" --log ${logstash_log_file}"
+fi
+
+if [ ${logstash_mode} = "standalone" ]; then
+ logstash_args="agent -f ${logstash_config} -- web --port ${logstash_port} --backend elasticsearch:///?local ${logstash_log_options}"
+ logstash_elastic_options="-Des.path.data=${logstash_elastic_datadir}"
+elif [ ${logstash_mode} = "agent" ]; then
+ logstash_args="agent -f ${logstash_config} ${logstash_log_options}"
+elif [ ${logstash_mode} = "web" ]; then
+ logstash_args="web --port ${logstash_port} --backend elasticsearch://${logstash_elastic_backend}/ ${logstash_log_options}"
+fi
+
+command_args="-f -p ${pidfile} ${java_cmd} ${logstash_elastic_options} -jar ${logstash_jar} ${logstash_args}"
+required_files="${java_cmd} ${logstash_config}"
+
+run_rc_command "$1"
diff --git a/sysutils/logstash/pkg-descr b/sysutils/logstash/pkg-descr
new file mode 100644
index 000000000000..4c3a547cb228
--- /dev/null
+++ b/sysutils/logstash/pkg-descr
@@ -0,0 +1,6 @@
+Logstash is a tool for managing events and logs. You can use it to
+collect logs, parse them, and store them for later use (like, for
+searching). Speaking of searching, logstash comes with a web interface
+for searching and drilling into all of your logs.
+
+WWW: http://logstash.net/
diff --git a/sysutils/logstash/pkg-plist b/sysutils/logstash/pkg-plist
new file mode 100644
index 000000000000..04d23a140975
--- /dev/null
+++ b/sysutils/logstash/pkg-plist
@@ -0,0 +1,13 @@
+%%LOGSTASH_HOME%%/%%LOGSTASH_JAR%%
+@exec mkdir -p %%LOGSTASH_RUN%%
+@exec mkdir -p %%LOGSTASH_DATA_DIR%%
+@unexec if cmp -s %D/%%ETCDIR%%/logstash.conf.sample %D/%%ETCDIR%%/logstash.conf; then rm -f %D/%%ETCDIR%%/logstash.conf; fi
+%%ETCDIR%%/logstash.conf.sample
+@exec if [ ! -f %D/%%ETCDIR%%/logstash.conf ] ; then cp -p %D/%F %B/logstash.conf; fi
+@unexec if cmp -s %D/%%ETCDIR%%/elasticsearch.yml.sample %D/%%ETCDIR%%/elasticsearch.yml; then rm -f %D/%%ETCDIR%%/elasticsearch.yml; fi
+%%ETCDIR%%/elasticsearch.yml.sample
+@exec if [ ! -f %D/%%ETCDIR%%/elasticsearch.yml ] ; then cp -p %D/%F %B/elasticsearch.yml; fi
+@dirrmtry %%LOGSTASH_DATA_DIR%%
+@dirrmtry %%LOGSTASH_HOME%%
+@dirrmtry %%ETCDIR%%
+@dirrmtry %%LOGSTASH_RUN%%