90 lines
		
	
	
		
			5.0 KiB
		
	
	
	
		
			Bash
		
	
	
		
			Executable File
		
	
	
	
	
			
		
		
	
	
			90 lines
		
	
	
		
			5.0 KiB
		
	
	
	
		
			Bash
		
	
	
		
			Executable File
		
	
	
	
	
| #!/usr/bin/env bash
 | |
| 
 | |
| # Copyright (c) 2021 Battelle Energy Alliance, LLC.  All rights reserved.
 | |
| 
 | |
| set -e
 | |
| 
 | |
| # if any pipelines are volume-mounted inside this docker container, they should belong to subdirectories under this path
 | |
| HOST_PIPELINES_DIR="/usr/share/logstash/malcolm-pipelines.available"
 | |
| 
 | |
| # runtime pipelines parent directory
 | |
| export PIPELINES_DIR="/usr/share/logstash/malcolm-pipelines"
 | |
| 
 | |
| # runtime pipeliens configuration file
 | |
| export PIPELINES_CFG="/usr/share/logstash/config/pipelines.yml"
 | |
| 
 | |
| # for each pipeline in /usr/share/logstash/malcolm-pipelines, append the contents of this file to the dynamically-generated
 | |
| # pipeline section in pipelines.yml (then delete 00_config.conf before starting)
 | |
| export PIPELINE_EXTRA_CONF_FILE="00_config.conf"
 | |
| 
 | |
| # files defining IP->host and MAC->host mapping
 | |
| INPUT_CIDR_MAP="/usr/share/logstash/config/cidr-map.txt"
 | |
| INPUT_HOST_MAP="/usr/share/logstash/config/host-map.txt"
 | |
| INPUT_MIXED_MAP="/usr/share/logstash/config/net-map.json"
 | |
| 
 | |
| # the name of the enrichment pipeline subdirectory under $PIPELINES_DIR
 | |
| ENRICHMENT_PIPELINE=${LOGSTASH_ENRICHMENT_PIPELINE:-"enrichment"}
 | |
| 
 | |
| # the name of the pipeline(s) to which input will send logs for parsing (comma-separated list, no quotes)
 | |
| PARSE_PIPELINE_ADDRESSES=${LOGSTASH_PARSE_PIPELINE_ADDRESSES:-"zeek-parse"}
 | |
| 
 | |
| # pipeline addresses for forwarding from Logstash to Elasticsearch (both "internal" and "external" pipelines)
 | |
| export ELASTICSEARCH_PIPELINE_ADDRESS_INTERNAL=${LOGSTASH_ELASTICSEARCH_PIPELINE_ADDRESS_INTERNAL:-"internal-es"}
 | |
| export ELASTICSEARCH_PIPELINE_ADDRESS_EXTERNAL=${LOGSTASH_ELASTICSEARCH_PIPELINE_ADDRESS_EXTERNAL:-"external-es"}
 | |
| ELASTICSEARCH_OUTPUT_PIPELINE_ADDRESSES=${LOGSTASH_ELASTICSEARCH_OUTPUT_PIPELINE_ADDRESSES:-"$ELASTICSEARCH_PIPELINE_ADDRESS_INTERNAL,$ELASTICSEARCH_PIPELINE_ADDRESS_EXTERNAL"}
 | |
| 
 | |
| # ip-to-segment-logstash.py translate $INPUT_CIDR_MAP, $INPUT_HOST_MAP, $INPUT_MIXED_MAP into this logstash filter file
 | |
| NETWORK_MAP_OUTPUT_FILTER="$PIPELINES_DIR"/"$ENRICHMENT_PIPELINE"/16_host_segment_filters.conf
 | |
| 
 | |
| ####################################################################################################################
 | |
| 
 | |
| # copy over pipeline filters from host-mapped volumes (if any) into their final resting places
 | |
| find "$HOST_PIPELINES_DIR" -mindepth 1 -maxdepth 1 -type d -print0 2>/dev/null | sort -z | \
 | |
|   xargs -0 -n 1 -I '{}' bash -c '
 | |
|   PIPELINE_NAME="$(basename "{}")"
 | |
|   PIPELINES_DEST_DIR="$PIPELINES_DIR"/"$PIPELINE_NAME"
 | |
|   mkdir -p "$PIPELINES_DEST_DIR"
 | |
|   cp -f "{}"/* "$PIPELINES_DEST_DIR"/
 | |
| '
 | |
| 
 | |
| # dynamically generate final pipelines.yml configuration file from all of the pipeline directories
 | |
| > "$PIPELINES_CFG"
 | |
| find "$PIPELINES_DIR" -mindepth 1 -maxdepth 1 -type d -print0 2>/dev/null | sort -z | \
 | |
|   xargs -0 -n 1 -I '{}' bash -c '
 | |
|   PIPELINE_NAME="$(basename "{}")"
 | |
|   PIPELINE_ADDRESS_NAME="$(cat "{}"/*.conf | sed -e "s/:[\}]*.*\(}\)/\1/" | envsubst | grep -P "\baddress\s*=>" | awk "{print \$3}" | sed "s/[\"'']//g" | head -n 1)"
 | |
|   if [[ -n "$ES_EXTERNAL_HOSTS" ]] || [[ "$PIPELINE_ADDRESS_NAME" != "$ELASTICSEARCH_PIPELINE_ADDRESS_EXTERNAL" ]]; then
 | |
|     echo "- pipeline.id: malcolm-$PIPELINE_NAME"       >> "$PIPELINES_CFG"
 | |
|     echo "  path.config: "{}""                         >> "$PIPELINES_CFG"
 | |
|     cat "{}"/"$PIPELINE_EXTRA_CONF_FILE" 2>/dev/null   >> "$PIPELINES_CFG"
 | |
|     rm -f "{}"/"$PIPELINE_EXTRA_CONF_FILE"
 | |
|     echo                                               >> "$PIPELINES_CFG"
 | |
|     echo                                               >> "$PIPELINES_CFG"
 | |
|   fi
 | |
| '
 | |
| 
 | |
| # create filters for network segment and host mapping in the enrichment directory
 | |
| rm -f "$NETWORK_MAP_OUTPUT_FILTER"
 | |
| /usr/local/bin/ip-to-segment-logstash.py --mixed "$INPUT_MIXED_MAP" --segment "$INPUT_CIDR_MAP" --host "$INPUT_HOST_MAP" -o "$NETWORK_MAP_OUTPUT_FILTER"
 | |
| 
 | |
| if [[ -z "$ES_EXTERNAL_HOSTS" ]]; then
 | |
|   # external ES host destination is not specified, remove external destination from enrichment pipeline output
 | |
|   ELASTICSEARCH_OUTPUT_PIPELINE_ADDRESSES="$(echo "$ELASTICSEARCH_OUTPUT_PIPELINE_ADDRESSES" | sed "s/,[[:blank:]]*$ELASTICSEARCH_PIPELINE_ADDRESS_EXTERNAL//")"
 | |
| fi
 | |
| 
 | |
| # insert quotes around the elasticsearch parsing and output pipeline list
 | |
| MALCOLM_PARSE_PIPELINE_ADDRESSES=$(printf '"%s"\n' "${PARSE_PIPELINE_ADDRESSES//,/\",\"}")
 | |
| MALCOLM_ELASTICSEARCH_OUTPUT_PIPELINES=$(printf '"%s"\n' "${ELASTICSEARCH_OUTPUT_PIPELINE_ADDRESSES//,/\",\"}")
 | |
| 
 | |
| # do a manual global replace on these particular values in the config files, as Logstash doesn't like the environment variables with quotes in them
 | |
| find "$PIPELINES_DIR" -type f -name "*.conf" -exec sed -i "s/_MALCOLM_ELASTICSEARCH_OUTPUT_PIPELINES_/${MALCOLM_ELASTICSEARCH_OUTPUT_PIPELINES}/g" "{}" \; 2>/dev/null
 | |
| find "$PIPELINES_DIR" -type f -name "*.conf" -exec sed -i "s/_MALCOLM_PARSE_PIPELINE_ADDRESSES_/${MALCOLM_PARSE_PIPELINE_ADDRESSES}/g" "{}" \; 2>/dev/null
 | |
| 
 | |
| # import trusted CA certificates if necessary
 | |
| /usr/local/bin/jdk-cacerts-auto-import.sh || true
 | |
| 
 | |
| # start logstash (adapted from docker-entrypoint)
 | |
| env2yaml /usr/share/logstash/config/logstash.yml
 | |
| export LS_JAVA_OPTS="-Dls.cgroup.cpuacct.path.override=/ -Dls.cgroup.cpu.path.override=/ $LS_JAVA_OPTS"
 | |
| exec logstash
 |