Introduction
YCCaster is a multi platform NTRIP caster designed to work as part of modern web applications. In this tutorial, we’ll discuss how to install YCCaster on your Ubuntu 20.04 server, adjust the firewall, configure access rights for bases and rovers, and set up monitoring.
Initial server setup
First of all, we need a virtual server on Ubuntu 20.04. Since our caster is quite lightweight, a minimal Linux server from Linode or DigitalOcean will work. I recommend these hosting providers for their newbie friendliness and reasonable pricing.
For the purposes of this tutorial, I will be using a virtual server from Linode. If you do not have an account, you need to register. Then log into your account and create a new server.

Installation.
To install the caster, we need to establish an ssh connection to our server. The easiest option is to use the Linode LISH console, the terminal will open in a new browser window. If you are using Linux or MacOS, open your terminal and establish an ssh connection. To log in, you will need the password that was set when creating the server.
ssh root@178.79.191.219
For security reasons, it is better to run caster with non root user permissions. For this we will create a separate user.
useradd caster
Then download caster from YCCaster website to a new folder and make it executable.
mkdir -p /var/caster && wget https://yccaster.s3.eu-central-1.amazonaws.com/bin/1.0.1/linux-amd64/yccaster -O /var/caster/yccaster && chmod +x /var/caster/yccaster
Generate initial config file.
cd /var/caster && ./yccaster init
Now we are ready to launch the caster but to work in the background it needs a linux service that will launch the caster if the server restarted or if the caster crashed. To create the service, execute this command.
cat << EOF > /etc/systemd/system/caster.service
[Unit]
Description=YCCaster
Requires=network-online.target
After=network-online.target
[Service]
Type=simple
User=caster
Group=caster
Restart=always
RestartSec=10
WorkingDirectory=/var/caster
ExecStart=/var/caster/yccaster
StandardOutput=append:/var/log/caster.log
StandardError=append:/var/log/caster.log
[Install]
WantedBy=multi-user.target
EOF
Then enable caster service.
systemctl enable caster.service
And start caster
systemctl start caster.service
Now the caster is up and running. To check it, you can use this command.
systemctl status caster.service
It should display active status
root@localhost:/var/caster# systemctl status caster.service
● caster.service - YCCaster
Loaded: loaded (/etc/systemd/system/caster.service; enabled; vendor preset>
Active: active (running) since Thu 2021-09-02 05:41:54 UTC; 8s ago
Main PID: 16298 (yccaster)
Tasks: 5 (limit: 1040)
Memory: 1.5M
CGroup: /system.slice/caster.service
└─16298 /var/caster/yccaster
Sep 02 05:41:54 localhost systemd[1]: Started YCCaster.
lines 1-10/10 (END)
Testing
Now when the caster is running we can test it and establish a connection between base and rover. To emulate base and rover I will use YCServer. It is an NTRIP server and client application for Android. You can download it from the play market.
I will use two instances of YCServer (probably, you will need two phones to run them simultaneously).
First instance will generate fake data and act as a base station. Second instance will act as a rover, receive data from the caster and save it to file.


Base and rover Authorization
The caster that we launched does not check passwords for connecting bases and rovers. Everyone can connect to it and transmit and receive data. If we need to restrict access and allow only authorized clients to exchange data, we need to change the configuration file and add the following section to it:
configuration:
auth:
- type: file
options:
mount-points: mountpoints.yml
clients: clients.yml
In this section, we stated that authorization will be based on files and that the
mountpoints.yml file will be used to authorize base stations and clients.yml for rovers
Create a new mountpoints.yml file in the same directory as the configuration file.
- mount-point: NICOSIARTKBASE
password: 12345
description:
identifier: Nicosia
format: RTCM 3.2
format-details: 1006(15),1008(15),1013(60),1019,1020,1033(15),1075(1)
carrier: 2
nav-system: GPS+GLO+GAL
network: EUREF
country: CYP
latitude: 35.15
longitude: 33.37
nmea: 0
solution: 0
generator: u-blox zed-f9p
compr-encryp: none
authentication: B
fee: N
bitrate: 6200
misc: Nicosia district base
- mount-point: PAPHOSRTKBASE
password: 678910
description:
identifier: Paphos
format: RTCM 3.2
format-details: 1006(15),1008(15),1013(60),1019,1020,1033(15),1075(1)
carrier: 2
nav-system: GPS+GLO+GAL
network: EUREF
country: CYP
latitude: 34.77
longitude: 32.41
nmea: 0
solution: 0
generator: u-blox zed-f9p
compr-encryp: none
authentication: B
fee: N
bitrate: 6200
misc: Paphos district base
In this file, we have specified 2 mount points. In addition to the name and password, you can specify a description of the stream, which will be used to generate the NTRIP SOURCETABLE.
Next, let's create the clients.yml file in the same directory.
- username: firstrover
password: 12345
- username: secondrover
password: 12345
- username: thirdrover
password: 12345
In it, I indicated the names of 3 rovers and their passwords.
In order for the changes to take effect, you need to restart the caster.
systemctl restart caster.service
Now only the base stations and rovers listed in the corresponding files will have access to the caster.