Using GoAccess with Nginx to replace Google Analytics

post, Jan 25, 2021 on Mitja Felicijan's blog

Introduction

I know! You cannot simply replace Google Analytics with parsing access logs and +displaying a couple of charts. But to be honest, I actually never used Google +Analytics to the fullest extent and was usually interested in seeing page hits +and which pages were visited most often.

I recently moved my blog from Firebase to a VPS and also decided to remove +Google Analytics tracking code from the site since its quite malicious and +tracks users across other pages also and is creating a profile of a user, and +I've had it. But I also need some insight of what is happening on a server and +which content is being read the most etc.

I have looked at many existing solutions like:

But the more I looked at them the more I noticed that I am replacing one evil +with another one. Don't get me wrong. Some of these solutions are absolutely +fantastic but would require installation of databases and something like PHP or +Node. And I was not ready to put those things on my fresh server. Also having +Docker installed is out of the question.

Opting for log parsing

So, I defaulted to parsing already existing logs and generating HTML reports +from this data.

I found this amazing software GoAccess which provides +all the functionalities I need, and it's a single binary. Written in Go.

GoAccess can be used in two different modes.

GoAccess Terminal
Running in a terminal
GoAccess HTML
Running in a browser

I, however, need this to run in a browser. So, the second option is the way to +go. The Idea is to periodically run cronjob and export this report into a folder +that gets then server by Nginx behind a Basic authentication.

Getting Nginx ready

I choose Ubuntu on DigitalOcean. First I +installed Nginx, and +Letsencrypt certbot and all the +necessary dependencies.

# log in as root user
+sudo su -
+
+# first let's update the system
+apt update && apt upgrade -y
+
+# let's install
+apt install nginx certbot python3-certbot-nginx apache2-utils
+

After all this is installed we can create a new configuration for a statistics. +Stats will be available at stats.domain.com.

# creates directory where html will be hosted
+mkdir -p /var/www/html/stats.domain.com
+
+cp /etc/nginx/sites-available/default /etc/nginx/sites-available/stats.domain.com
+nano /etc/nginx/sites-available/stats.domain.com
+
server {
+  root /var/www/html/stats.domain.com;
+  server_name stats.domain.com;
+
+  index index.html;
+  location / {
+    try_files $uri $uri/ =404;
+  }
+}
+

Now we check if the configuration is ok. We can do this with nginx -t. If all +is ok, we can restart Nginx with service nginx restart.

After all that you should add A record for this domain that points to IP of a +droplet.

Before enabling SSL you should test if DNS records have propagated with curl stats.domain.com.

Now, it's time to provision TLS certificate. To achieve this, you execute +command certbot --nginx. Follow the wizard and when you are asked about +redirection always choose 2 (always redirect to HTTPS).

When this is done you can visit https://stats.domain.com and you should get 404 +not found error which is correct.

Getting GoAccess ready

If you are using Debian like system GoAccess should be available in repository. +Otherwise refer to the official website.

apt install goaccess
+

To enable Geo location we also need one additiona thing.

cd /var/www/html/stats.stats.com
+wget https://github.com/P3TERX/GeoLite.mmdb/raw/download/GeoLite2-City.mmdb
+

Now we create a shell script that will be executed every 10 minutes.

nano /var/www/html/stats.domain.com/generate-stats.sh
+

Contents of this file should look like this.

#!/bin/sh
+
+zcat -f /var/log/nginx/access.log* > /var/log/nginx/access-all.log
+
+goaccess \
+  --log-file=/var/log/nginx/access-all.log \
+  --log-format=COMBINED \
+  --exclude-ip=0.0.0.0 \
+  --geoip-database=/var/www/html/stats.domain.com/GeoLite2-City.mmdb \
+  --ignore-crawlers \
+  --real-os \
+  --output=/var/www/html/stats.domain.com/index.html
+
+rm /var/log/nginx/access-all.log
+

Because after a while nginx creates multiple files with access logs we use +zcat to extract Gziped contents and create +a file that has all the access logs. After this file is used we delete it.

If you want to exclude your home IP's result look at the --exclude-ip option +in script and instead of 0.0.0.0 add your own home IP address. You can find +your home IP by executing curl ifconfig.me from your local machine and NOT +from the droplet.

Test the script by executing sh /var/www/html/stats.domain.com/generate-stats.sh and then checking +https://stats.domain.com. If you can see stats instead of 404 than you are +set.

It's time to add this script to cron with cron -e.

*/10 * * * * sh /var/www/html/stats.domain.com/generate-stats.sh
+

Securing with Basic authentication

You probably don't want stats to be publicly available, so we should create a +user and a password for Basic authentication.

First we create a password for a user stats with htpasswd -c /etc/nginx/.htpasswd stats.

Now we update config file with nano /etc/nginx/sites-available/stats.domain.com. You probably noticed that the +file looks a bit different from before. This is because certbot added +additional rules for SSL.

Your location portion the config file should now look like. You should add +auth_basic and auth_basic_user_file lines to the file.

location / {
+  try_files $uri $uri/ =404;
+  auth_basic "Private Property";
+  auth_basic_user_file /etc/nginx/.htpasswd;
+}
+

Test if config is still ok with nginx -t and if it is you can restart Nginx +with service nginx restart.

If you now visit https://stats.domain.com you should be prompted for username +and password. If not, try reopening your browser.

That is all. You now have analytics for your server that gets refreshed every 10 +minutes.