log-aggregator

command module
v0.0.0-...-56ddd06 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Jan 13, 2020 License: MIT Imports: 18 Imported by: 0

README

Log Aggregator (naming is hard) GoDoc CircleCI Go Report Card

Reads logs from Journald, annotates/transforms, and forwards to AWS Kinesis Firehose.

What is this for?

This is a lightweight process for reading all logs from the system, and forwarding them to a central location for searching and long-term archival.

It also supports performing minimal transformation and annotation operations on each log entry.

FAQs

Is this a package or a binary?

It's both!

While the binary is designed to fit our use-case, it is entirely interface driven, and can be used as a package to power your own processing pipeline.

See the Usage section for more details.

Is this production ready?

We think so. It has been running in our production environments since April 2017 without any major issues.

Why a custom solution vs. something like Fluentd?

We actually did run Fluentd in the past. Prior to switching our Docker log driver to Journald, we found that Fluentd was using unusually high CPU to watch the Docker logs directory.

After switching to Journald, we also found that (at the time) the Fluentd/Ruby Journald integration was dropping fields.

We decided to go with a home-built solution once we needed to add JSON parsing and the lazy Kubernetes initialization.

Because our use-case was so straight forward, it was actually simpler to accomplish what we needed in a standalone program, rather than trying to write plugins to fit into the Fluentd architecture.

It also had the bonus of lowering our resource usage.

Design Considerations

There are a few environmental factors that drove us to this solution:

  1. Because we use Kubernetes to run all of our applications, our low level machine configurations are all homogeneous
  2. Logging in all programming languages was standardized (JSON, consistent field names)
  3. Can't rely on Kubernetes to be running to collect logs (want logs in case something with K8s catches on fire)
Built In Features
  • Input: Journald
  • Output: AWS Kinesis Firehose
  • Transformations
    • AWS: adds aws.instance_id, aws.local_hostname, aws.local_ipv4
    • Journal: Rename MESSAGE field to log
    • K8s: Add Pod metadata if the log comes from a Kubernetes Pod
    • Kibana: insert @timestamp field in the format Kibana expects
    • JSON: attempt to parse the log line as JSON, and if successful set the ts field as the log entry time
Kubernetes Support

While the log-aggregator does need access to Kubernetes APIs in order to annotate logs from Pods, it is also an invaluable tool in debugging instance startup issues.

Because of this, the log-aggregator (installed as a systemd unit) is started as soon as the instance network is online. This immediately provides logs that can be reviewed if the instance is having startup issues, without having to SSH to a running instance.

On machines that will run Kubernetes, a configuration file will eventually be written (by the Kubernetes bootstrap process) that tells the kubelet how to talk to the Kubernetes API. The log-aggregator can watch for this file, and as soon as it is detected, it enables the Kubernetes annotation transformer.

AWS Instance Info

Due to some legacy reasons, instead of reading the instance info from the metadata service, it relies on reading those values from environment variables that are set by another systemd unit that is already pre-installed on our AMIs.

Usage

The log-aggregator can be used as a package (to construct your own pipelines), or as a binary if your use-case fits ours.

As a Package

The main unit of abstraction is a Pipeline, which is configured with the following components:

  • Source: produces log records
  • Destination: saves log records, and reports progress
  • Cursor: provides a starting point for the Source, and persists the Destination progress so that processing can be resumed on restarts from a last-known checkpoint
  • Transformers: transform log records prior to sending to the destination.

Here is a simple pipeline that uses a mock source and destination, and applies the JSON transformer.

package main

import (
	"time"

	"github.com/wearefair/log-aggregator/pkg/cursor"
	"github.com/wearefair/log-aggregator/pkg/destinations/stdout"
	"github.com/wearefair/log-aggregator/pkg/pipeline"
	"github.com/wearefair/log-aggregator/pkg/sources/mock"
	"github.com/wearefair/log-aggregator/pkg/transform"
	"github.com/wearefair/log-aggregator/pkg/transform/json"
)

func main() {
	cursor, _ := cursor.New("/var/log/log-aggregator.cursor")

	// produce a log every 2 seconds
	source = mock.New(time.Second * 2)
	destination = stdout.New()

	logPipeline, _ := pipeline.New(pipeline.Config{
		MaxBuffer:    200,
		Cursor:       cursor,
		Input:        source,
		Destination:  destination,
		Transformers: []transform.Transformer{json.Transform},
	})

	logPipeline.Start()
}
As a Binary

You can run the aggregator any way you like, we happen to use systemd to launch it as part of our base image.

[Unit]
Description=log-aggregator
After=network-online.target
Requires=network-online.target

[Service]
Environment="FAIR_LOG_CURSOR_PATH=/var/log/log-aggregator.cursor"
Environment="FAIR_LOG_FIREHOSE_STREAM=<your-firehose-stream>"
Environment="FAIR_LOG_K8_CONFIG_PATH=<file-path-for-kubelet.conf>"

EnvironmentFile=/etc/fair/log-aggregator-custom.env
ExecStart=/var/fair/bin/log-aggregator

Restart=always
RestartSec=5

[Install]
WantedBy=multi-user.target
Configuring

There are no cli-flags, all configuration is done via environment variables.

Required Environment Variables
  • FAIR_LOG_CURSOR_PATH: The path to save the cursor position to
  • FAIR_LOG_FIREHOSE_STREAM: The Firehose stream name to export to
Optional Environment Variables
  • FAIR_LOG_FIREHOSE_CREDENTIALS_ENDPOINT: Override the metadata service endpoint to use for credentials
  • FAIR_LOG_K8_CONFIG_PATH: The path to watch for the Kubernetes config file
  • FAIR_LOG_K8_CONTAINER_NAME_REGEX: Override the built-in regex for extracting the Pod name
  • FAIR_LOG_MOCK_SOURCE: Enable a mock source instead of journald (for testing)
  • FAIR_LOG_MOCK_DESTINATION: Enable a mock destination (stdout) instead of Kinesis Firehose (for testing)
  • EC2_METADATA_INSTANCE_ID: For the AWS transformer
  • EC2_METADATA_LOCAL_IPV4: For the AWS transformer
  • EC2_METADATA_LOCAL_HOSTNAME: Used by the AWS and the K8s transformer (for the Node name)
  • ENV=production: Turns on JSON logging for the aggregator's own logs

Building/Developing

It requires dep for dependencies.

For osx you can run go build

For linux, run make build-linux, as it requires the systemd development headers.

Documentation

The Go Gopher

There is no documentation for this package.

Directories

Path Synopsis
pkg
destinations/stdout
Package stdout provides a simple destination that prints all records to STDOUT.
Package stdout provides a simple destination that prints all records to STDOUT.
sources/journal
Package journal for linux These methods require being compiled on a system with the systemd headers
Package journal for linux These methods require being compiled on a system with the systemd headers
transform/aws
Package aws reads ec2 metadata instance environment variables and sets them on the logs.
Package aws reads ec2 metadata instance environment variables and sets them on the logs.

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL