Logs are an essential debbuging tool when developing. For simple applications doing a tail -f or docker logs can be enough to find what we are looking for.

But for more complex applications or applications which logs are very verbose (hello Java!), finding the exact piece we want can be cumberstone.

In production systems, it´s common to have some sort of centralized logging platform where we can use powerfull queries to search the logs. Two popular tools are the Elastic Stack with Kibana and Grafana Loki.

So why not having something similar on our local envrionment? Kibana uses ElasticSearch as storage, which can be a little heavy. Grafana Loki is a much more lightweight alternative. I also prefer Grafana as visualization tool.

Unlike Elastic Stack, Loki doens´t index our logs. Instead logs are stored in plaintext and tagged with a set of label names and values, where only the label pairs are indexed. This makes it cheaper to operate than a full index.

Logs in Loki are queried using LogQL, a powerfull query language, inspired by PromQL, the Prometheus Query language.

A more detailed comparison between Elastic stack and Loki can be found here.

The demo project

To demonstrate the use of Loki, we will use Docker Compose to start a sample application and all the required components for Loki to work. You can follow along with the complete source on GitHub.

The “Loki stack” consists of the following components:

We will also use echo-server as example application that will generate some logs.

You can check the full docker-compose file. Next, I will explain in detail the more relevant parts.

The Loki container

    image: grafana/loki:2.2.1
      - "3100"
      - ./config/loki/loki-config.yaml:/mnt/config/loki-config.yaml
    command: -config.file=/mnt/config/loki-config.yaml
      - loki

Here we start our Loki instance, using a config file. The config file allow us to configure some properties of Loki, like the retention time for our logs. You can check all available options here, But at least for development, you should not need to change much of the default config.

Loki has an Api that will listen on the configured server.http_listen_port (3100 by default), Api that will be used by the clients to push logs to Loki.

The Grafana container

    image: grafana/grafana:latest
      - "3000:3000"
      - grafana-data:/var/lib/grafana
      - ./config/grafana/provisioning/:/etc/grafana/provisioning/
      - loki

This container runs a standard Grafana install.

We are taking advantage of the Provisioning feature, which allows to automatically configure any Datasources or Dashboard on Grafana startup. Otherwsie, we would need to manually create the Loki datasource on Grafana interface.

Shipping the container logs to Loki

Now that we have Grafana and Loki configured, we need a way to effectively ship the logs to Loki.

There are two ways to do this: Using Promtail or the Log driver plugin.

I don´t have a formed opinion about which way is the best. Promtail is more flexible and it´s not limited to Docker logs. You could use the same Promtail instance to ingest log files for example, so if your application writes logs to file system instead of stdout you have to use Promtail.

Another difference is that if you want to have a global configuration to ship all your container logs to Loki, you would need to install Loki directly on the host machine or expose the 3100 in the Loki container, as the Docker Daemon would need to be able to reach the Loki API. Not a big deal, but not needed with Promtail, if it´s deployed on the same Docker network as loki, the communication will be internal.

With Promtail, the source logs are read directly from /var/lib/docker/containers in the host filesystem, so you have to mount this volume in the Promtail container, where with the Docker Log driver, you don´t need to.

In this demo I am using Promtail, but using the Log driver on my local PC. It´s up to you to see which one fits better with your requirements.

If you want to know more about the Loki log driver, you can read more about the available configurations in Configuring the Docker Driver.

For Promtail, we have the following definition in our Docker compose file.

    image: grafana/promtail:2.2.1
      - ./config/loki/promtail-config.yaml:/mnt/config/promtail-config.yaml
      - /var/lib/docker/containers:/host/containers
    command: -config.file /mnt/config/promtail-config.yaml
      - loki

Note that we mount the /var/lib/docker/containers, since it´s the place where the Docker logs are stored in the host filesystem and where Promtail will read from.

Similar to Loki, Promtail configuration is done with a config file promtail-config.yaml.

  - url: http://loki:3100/loki/api/v1/push

  - job_name: docker
      - targets:
          - localhost
          job: dockerlogs
          __path__: /host/containers/*/*log
      - json:
            output: log
            stream: stream
      - json:
          source: attrs
      - regex:
          expression: (?P<image_name>(?:[^|]*[^|])).(?P<container_name>(?:[^|]*[^|])).(?P<image_id>(?:[^|]*[^|])).(?P<container_id>(?:[^|]*[^|]))
          source: tag
      - timestamp:
          format: RFC3339Nano
          source: time
      - labels:
      - output:
          source: output

The clients section contains the endpoint information where the logs will be shipped. In our example, we configure it the to the respective Loki push url.

The scrape_configs section is where we define the configuration for the jobs that will scrape the logs.

We defined a job named docker whose __path__ property indicates where the logs are located on the file system. Since we are mounting /var/lib/docker/containers into /host/containers, we configure that path here.

 job_name: docker
      - targets:
          - localhost
          job: dockerlogs
          __path__: /host/containers/*/*log

The “pipeline_stages” field allow to configure a log processing pipeline, that will be executed for each log line and can be used for example to modify the contents of each log before being ingested or to add labels. Labels that will be indexed by Loki, and will be very useful for searching.

You can read more about Pipeline stages here.

In this particular example, we configure the pipeline to parse the log as JSON, extract the “tag” field, parse the contents of that field using a regex and expose the result as labels.

But where that field comes from? The tag field is a standard option of the docker logging configuration, than is used to idenfity the container where the logs come from.

By default, the system uses the first 12 characters of the container ID, but we can override it to a more meaningfull value. That can be done in the log driver configuration. In our docker-compose we do it like this:

  driver: json-file
    tag: "{{.ImageName}}|{{.Name}}|{{.ImageFullID}}|{{.FullID}}"

Docker supports some special template markup that we can use - see Customize log driver output.

So our tag will consist of the Image Name, Container Name, Image Id and Container ID. This info will be populated by docker in each log line, and Promtail will parse it and create the appropriate labels.

We should now have everything we need to start our demo.

Show me the logs!

show me the logs meme

Run docker-compose up to start all the containers.

If everything went fine you can then open grafana in your browser at localhost:3000. You can use admin/testloki credentials to login.

You should now see the Grafana homepage. Head over to “Configuration -> DataSources” on the left menu, and you should see Loki datasource automatically configured:

Grafana Loki

You can click on the “Test” button to check that Grafana and Loki can communicate.

Next we can go to the Loki interface, by clicking on the “Explorer” menu on the left sidebar. We should already be able to see some logs from Grafana, Loki and Promtail containers.

If You open the “Log Labels” menu you should see all the labels that we configured in Promtail and that were extracted from the logs.

Loki logs view

You can select one of them to display the associated logs, or with a LogQL query directly.

For example, to see all the logs of Loki container itself, you can use the following query:


You can also use a regular expression on the query. For example, to get the logs for all the applications in our project, we can do:


We can do now some requests to our Demo app by opening http://localhost:3001. Logs should appear shortly on Grafana.

We should be able use the following LogQL query, to get the application logs:


Filtering by labels is nice, but to use our logs in full, we have to be able to search in the logs content.

We can use a filter expression in LogQL to do just that.

For example, to get all the logs containg the “name” text in our application, we can do:

{container_name="logs-with-loki_demo-app_1"} |= "name"

Log search

To know more about the syntax of LogQL and all the possible queries, this cheat sheet is a good place start, followed by the official LogQL Documentation.

And that´s it. We know have a centralized way to visualize the logs for all our applications in a nice interface and with powerfull search capabilities.

Send all your container logs to Loki.

Right now, only logs of the containers defined in the docker compose file will be ingested by Loki, because we configured the log driver in locally to the compose file.

I prefer to have Loki globally in my machine, so all the container logs, doens´t matter if they are originated from Compose project, or standalone docker container, are indexed and stored in the same place. You can do that by configuring the log driver globally in the Docker daemon, instead of in the Compose file.

For that, you should open /etc/docker/daemon.json and add the following contents: (create the file if not exists):

    "log-driver": "json-file",
    "log-opts": {
        "tag": "{{.ImageName}}|{{.Name}}|{{.ImageFullID}}|{{.FullID}}"

Restart your docker daemon to apply the new configuration.

Note that, this configuration will only work for newly created containers.

You have to make sure that Loki and Promtail are always running. I would recommend having a docker-compose file with Loki, Grafana and Promtail and add restart: always option, or running the containers manually using docker run with the --restart always flag.

If you want to use the Log plugin instead of Promtail, check this Guide.


Hope you enjoyed this article.

Grafana Loki is a great tool that can help a lot searching in your application logs. You should give it a try.