Grafana Monitoring on a Raspberry Pi

Published | 4 min read

As you might have seen from my last few posts I have quite a lot running on my Raspberry Pi.

I am currently using a Raspberry Pi 2 B which is a great device but only has 1GB of RAM and 900 MHz CPU. So I am a little worried sometimes that I am going to overload it with all the docker services I am running on it.

I use Grafana a lot at work and love it, so I thought it would be good to use it to monitor my Raspberry Pi.

Monitoring Aim

With any monitoring, it is important to know what you want to keep an eye on.

In my case I am interested in the following:

  • CPU - If the CPU ends up running at 100% a lot of the time, I might need to scale down the services running on it
  • Memory - With only 1GB of memory I need to keep an eye on how much I run on it.
  • Hard Disk Space - I have a 32GB SD card in my Pi but I have had one fill up before which makes the whole thing unresponsive.
  • Container data - I want to know which containers are causing high CPU and Memory usage.

What we need

There are probably quite a few services that work with Grafana for monitoring. However, I am using the following:

  • Grafana - obviously!
  • Prometheus - for gathering the data in a time series.
  • cAdvisor - A container monitor from Google to monitor the resources used by containers.
  • Node Exporter - Prometheus exporter for hardware and OS metrics
  • Docker - obviously
  • Traefik - I use this as my reverse proxy if you don’t have a reverse proxy set up you can follow my previous post, Traefik vs Nginx for Reverse Proxy with Docker on a Raspberry Pi.

Docker Compose Set up

I will go straight to the Docker Compose file you need and will explain what you need to change for your setup:

version: '3.4'

    image: prom/prometheus:latest
    container_name: monitoring_prometheus
    restart: unless-stopped
      - ./data/prometheus/config:/etc/prometheus/
      - ./data/prometheus/data:/prometheus
      - '--config.file=/etc/prometheus/prometheus.yml'
      - '--storage.tsdb.path=/prometheus'
      - 9090
      - cadvisor:cadvisor
      - node-exporter:node-exporter
      - pi

    image: prom/node-exporter:latest
    container_name: monitoring_node_exporter
    restart: unless-stopped
      - 9100
      - pi

    image: braingamer/cadvisor-arm:latest
    container_name: monitoring_cadvisor
    privileged: true
    restart: unless-stopped
      - /:/rootfs:ro
      - /var/run:/var/run:rw
      - /sys:/sys:ro
      - /var/lib/docker/:/var/lib/docker:ro
      - /dev/kmsg
      - 8080
      - pi

    image: grafana/grafana:latest
    container_name: monitoring_grafana
    restart: unless-stopped
      - prometheus:prometheus
      - ./data/grafana:/var/lib/grafana
      - GF_USERS_ALLOW_SIGN_UP=false
      - 'traefik.enable=true'
      - 'traefik.http.routers.grafana.rule=PathPrefix(`/grafana{regex:$$|/.*}`)'
      - ''
      - 'traefik.frontend.headers.customRequestHeaders=Authorization:-'
      - pi

    external: true

You only need to change the following environment variables to match the domain for your Raspberry Pi. If you are running locally you could use localhost and http://localhost/grafana.


You will notice I am using braingamer/cadvisor-arm:latest for cAdvisor. This is because the official Google image doesn’t support ARM and is marked as deprecated. Of course, if you wanted to you could build your own docker image.

This is the Prometheus config file I am using, prometheus.yml:

# my global config
  scrape_interval: 120s # By default, scrape targets every 15 seconds.
  evaluation_interval: 120s # By default, scrape targets every 15 seconds.
  # scrape_timeout is set to the global default (10s).

  # Attach these labels to any time series or alerts when communicating with
  # external systems (federation, remote storage, Alertmanager).
    monitor: 'my-project'

# Load and evaluate rules in this file every 'evaluation_interval' seconds.
  # - "alert.rules"
  # - "first.rules"
  # - "second.rules"

# A scrape configuration containing exactly one endpoint to scrape:
# Here it's Prometheus itself.
  # The job name is added as a label `job=<job_name>` to any timeseries scraped from this config.
  - job_name: 'prometheus'

    # Override the global default and scrape targets from this job every 5 seconds.
    scrape_interval: 120s

    # metrics_path defaults to '/metrics'
    # scheme defaults to 'http'.

      - targets: ['localhost:9090', 'cadvisor:8080', 'node-exporter:9100']

Dashboard Set Up

This is what my Grafana dashboard looks like. If you want something similar then you can copy my dashboard json.

Grafana Dashboard

There are quite a few other metrics available from node-exporter and cAdvisor I could have used. Let me know in the comments if find any useful ones.

Was this post useful?
If you found this post useful and would like to support me, you can do so by buying me a coffee. Donations help keep this blog ad free.


Adding Inline Javascript to a React App

Adding Inline Javascript to a React App

  • 11 February 2022
Let me start this with a disclaimer, I am not a React expert and I am sure there are probably other ways to do this. If you know of a better way then please let me know in the comments! I have recently added Giscus to my blog which is a great way to ...
Privacy Focused Blog Services

Privacy Focused Blog Services

  • 28 January 2022
Towards the end of 2021, I started getting into web3 a lot more. Not only did I join the Developer DAO but I also launched my own NFT collection and wrote a few blog posts on the topic. One of the key components of web3 is the ownership of data. In w...
What is DeFi? Understanding Decentralized Finance

What is DeFi? Understanding Decentralized Finance

  • 19 November 2021
If you are new to the crypto space the concepts of decentralised finance can seem a bit overwhelming. There are so many new terms to try and understand, it can be challenging to see how everything fits together. I have been working in the traditional...
Using Hashnode as a Source for Gatsby

Using Hashnode as a Source for Gatsby

  • 05 November 2021
I have done quite a few website migrations over the years. My blog like many first started out on WordPress. After several attempts at optimisation, I ended up generating a static version of my WordPress website. On my static site generation journey...
What is Web3? The Decentralised Internet Explained

What is Web3? The Decentralised Internet Explained

  • 15 October 2021
If you have stumbled upon this article you have probably seen the term web3 being mentioned around the internet. Especially on Twitter at the moment it seems to be the latest buzzword. With all the hype around NFTs and crypto at the moment, a lot of ...
Ditch the Ads: The Future of Website Monetization

Ditch the Ads: The Future of Website Monetization

  • 08 October 2021
Do you remember that scene in Ready Player One, when IOI are showing what the future of the Oasis will look like once they take over? “We have determined that we’ll be able to fill 80% of the user’s display with advertising before inducing seizures”...
Dealing with Imposter Syndrome as a Software Developer

Dealing with Imposter Syndrome as a Software Developer

  • 28 May 2021
I have been a professional software developer for over a decade and I have been writing code for over 25 years. However, sometimes I still feel like a fraud. It turns out, I am not the only one that feels this way and they have a name for it, it is c...
Mocking API calls using WireMock

Mocking API calls using WireMock

  • 14 May 2021
It is rare in software development that you are building something in complete isolation from everything else. Generally, you are going to be making calls to other systems or components. If you are lucky you are building against an existing API that ...

Alex Hyett

Alex Hyett

Software Developer, Entrepreneur, Father, and Husband. Engineering Lead at

Want to get in touch? You can find me here: