2020-07-30 22:53:20 +00:00
|
|
|
import time
|
|
|
|
import json
|
|
|
|
import subprocess
|
|
|
|
import os
|
|
|
|
from influxdb import InfluxDBClient
|
|
|
|
|
|
|
|
# InfluxDB Settings
|
|
|
|
DB_ADDRESS = os.environ.get('INFLUX_DB_ADDRESS')
|
|
|
|
DB_PORT = int(os.environ.get('INFLUX_DB_PORT'))
|
|
|
|
DB_USER = os.environ.get('INFLUX_DB_USER')
|
|
|
|
DB_PASSWORD = os.environ.get('INFLUX_DB_PASSWORD')
|
|
|
|
DB_DATABASE = os.environ.get('INFLUX_DB_DATABASE')
|
2020-07-31 03:45:23 +00:00
|
|
|
DB_TAGS = os.environ.get('INFLUX_DB_TAGS')
|
2020-07-30 22:53:20 +00:00
|
|
|
|
|
|
|
# Speedtest Settings
|
2020-07-31 05:10:51 +00:00
|
|
|
# Time between tests (in minutes, converts to seconds).
|
|
|
|
TEST_INTERVAL = int(os.environ.get('SPEEDTEST_INTERVAL')) * 60
|
|
|
|
# Time before retrying a failed Speedtest (in minutes, converts to seconds).
|
|
|
|
TEST_FAIL_INTERVAL = int(os.environ.get('SPEEDTEST_FAIL_INTERVAL')) * 60
|
2020-07-30 22:53:20 +00:00
|
|
|
|
|
|
|
influxdb_client = InfluxDBClient(
|
|
|
|
DB_ADDRESS, DB_PORT, DB_USER, DB_PASSWORD, None)
|
|
|
|
|
|
|
|
|
|
|
|
def init_db():
|
|
|
|
databases = influxdb_client.get_list_database()
|
|
|
|
|
|
|
|
if len(list(filter(lambda x: x['name'] == DB_DATABASE, databases))) == 0:
|
|
|
|
influxdb_client.create_database(
|
|
|
|
DB_DATABASE) # Create if does not exist.
|
|
|
|
else:
|
2020-07-31 03:45:23 +00:00
|
|
|
# Switch to if does exist.
|
|
|
|
influxdb_client.switch_database(DB_DATABASE)
|
|
|
|
|
|
|
|
|
2020-07-30 22:53:20 +00:00
|
|
|
def pkt_loss(data):
|
|
|
|
if 'packetLoss' in data.keys():
|
2020-07-31 18:53:44 +00:00
|
|
|
return int(data['packetLoss'])
|
2020-07-31 03:45:23 +00:00
|
|
|
else:
|
2020-07-30 22:53:20 +00:00
|
|
|
return 0
|
|
|
|
|
2020-07-31 03:45:23 +00:00
|
|
|
|
|
|
|
def tag_selection(data):
|
|
|
|
tags = DB_TAGS
|
2020-07-31 04:13:14 +00:00
|
|
|
if tags is None:
|
|
|
|
return None
|
|
|
|
# tag_switch takes in _data and attaches CLIoutput to more readable ids
|
2020-07-31 03:45:23 +00:00
|
|
|
tag_switch = {
|
|
|
|
'isp': data['isp'],
|
|
|
|
'interface': data['interface']['name'],
|
|
|
|
'internal_ip': data['interface']['internalIp'],
|
|
|
|
'interface_mac': data['interface']['macAddr'],
|
|
|
|
'vpn_enabled': (False if data['interface']['isVpn'] == 'false' else True),
|
|
|
|
'external_ip': data['interface']['externalIp'],
|
|
|
|
'server_id': data['server']['id'],
|
|
|
|
'server_name': data['server']['name'],
|
|
|
|
'server_location': data['server']['location'],
|
|
|
|
'server_country': data['server']['country'],
|
|
|
|
'server_host': data['server']['host'],
|
|
|
|
'server_port': data['server']['port'],
|
|
|
|
'server_ip': data['server']['ip'],
|
|
|
|
'speedtest_id': data['result']['id'],
|
|
|
|
'speedtest_url': data['result']['url']
|
|
|
|
}
|
2020-07-31 04:13:14 +00:00
|
|
|
|
2020-07-31 03:45:23 +00:00
|
|
|
options = {}
|
2020-07-31 04:13:14 +00:00
|
|
|
tags = tags.split(',')
|
|
|
|
for tag in tags:
|
|
|
|
# split the tag string, strip and add selected tags to {options} with corresponding tag_switch data
|
|
|
|
tag = tag.strip()
|
|
|
|
options[tag] = tag_switch[tag]
|
|
|
|
return options
|
2020-07-31 03:45:23 +00:00
|
|
|
|
|
|
|
|
2020-07-30 22:53:20 +00:00
|
|
|
def format_for_influx(cliout):
|
|
|
|
data = json.loads(cliout)
|
|
|
|
# There is additional data in the speedtest-cli output but it is likely not necessary to store.
|
|
|
|
influx_data = [
|
|
|
|
{
|
|
|
|
'measurement': 'ping',
|
|
|
|
'time': data['timestamp'],
|
|
|
|
'fields': {
|
|
|
|
'jitter': data['ping']['jitter'],
|
|
|
|
'latency': data['ping']['latency']
|
|
|
|
}
|
|
|
|
},
|
|
|
|
{
|
|
|
|
'measurement': 'download',
|
|
|
|
'time': data['timestamp'],
|
|
|
|
'fields': {
|
|
|
|
# Byte to Megabit
|
|
|
|
'bandwidth': data['download']['bandwidth'] / 125000,
|
|
|
|
'bytes': data['download']['bytes'],
|
|
|
|
'elapsed': data['download']['elapsed']
|
|
|
|
}
|
|
|
|
},
|
|
|
|
{
|
|
|
|
'measurement': 'upload',
|
|
|
|
'time': data['timestamp'],
|
|
|
|
'fields': {
|
|
|
|
# Byte to Megabit
|
|
|
|
'bandwidth': data['upload']['bandwidth'] / 125000,
|
|
|
|
'bytes': data['upload']['bytes'],
|
|
|
|
'elapsed': data['upload']['elapsed']
|
|
|
|
}
|
|
|
|
},
|
|
|
|
{
|
|
|
|
'measurement': 'packetLoss',
|
|
|
|
'time': data['timestamp'],
|
|
|
|
'fields': {
|
|
|
|
'packetLoss': pkt_loss(data)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
]
|
2020-07-31 04:13:14 +00:00
|
|
|
tags = tag_selection(data)
|
|
|
|
if tags is None:
|
|
|
|
return influx_data
|
|
|
|
else:
|
|
|
|
for measurement in influx_data:
|
|
|
|
measurement['tags'] = tags
|
|
|
|
return influx_data
|
2020-07-30 22:53:20 +00:00
|
|
|
|
|
|
|
|
|
|
|
def main():
|
|
|
|
init_db() # Setup the database if it does not already exist.
|
|
|
|
|
|
|
|
while (1): # Run a Speedtest and send the results to influxDB indefinitely.
|
|
|
|
speedtest = subprocess.run(
|
|
|
|
["speedtest", "--accept-license", "--accept-gdpr", "-f", "json"], capture_output=True)
|
|
|
|
|
|
|
|
if speedtest.returncode == 0: # Speedtest was successful.
|
|
|
|
data = format_for_influx(speedtest.stdout)
|
|
|
|
print("Speedtest Successful:")
|
|
|
|
if influxdb_client.write_points(data) == True:
|
|
|
|
print("Data written to DB successfully")
|
|
|
|
time.sleep(TEST_INTERVAL)
|
|
|
|
else: # Speedtest failed.
|
|
|
|
print("Speedtest Failed:")
|
|
|
|
print(speedtest.stderr)
|
|
|
|
print(speedtest.stdout)
|
|
|
|
time.sleep(TEST_FAIL_INTERVAL)
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
print('Speedtest CLI Data Logger to InfluxDB')
|
2020-07-31 03:45:23 +00:00
|
|
|
main()
|