mirror of
https://github.com/pommi/python-itho-wpu.git
synced 2024-12-03 15:45:09 +01:00
style: apply flake8 and black formatting
To check for errors: $ pre-commit run --all-files To install as pre-commit hook: $ pre-commit install
This commit is contained in:
parent
8c7d0ed947
commit
751182b70d
15
.pre-commit-config.yaml
Normal file
15
.pre-commit-config.yaml
Normal file
@ -0,0 +1,15 @@
|
||||
repos:
|
||||
- repo: local
|
||||
hooks:
|
||||
- id: flake8
|
||||
name: flake8
|
||||
language: system
|
||||
entry: flake8
|
||||
types: [python]
|
||||
stages: [commit]
|
||||
- id: black
|
||||
name: black
|
||||
language: system
|
||||
entry: black
|
||||
types: [python]
|
||||
stages: [commit]
|
@ -14,27 +14,31 @@ import sys
|
||||
|
||||
|
||||
def parse_args():
|
||||
parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter,
|
||||
description='Convert Itho Servicetool database to SQLite')
|
||||
parser.add_argument('--itho-db', nargs='?', required=True, help="Itho Database file")
|
||||
parser.add_argument('--sqlite-db', nargs='?', default='heatpump.sqlite', help="Itho Database file")
|
||||
parser.add_argument('--force', action='store_true', help="Force overwrite SQLite database")
|
||||
parser = argparse.ArgumentParser(
|
||||
formatter_class=argparse.ArgumentDefaultsHelpFormatter,
|
||||
description="Convert Itho Servicetool database to SQLite",
|
||||
)
|
||||
parser.add_argument("--itho-db", nargs="?", required=True, help="Itho Database file")
|
||||
parser.add_argument(
|
||||
"--sqlite-db", nargs="?", default="heatpump.sqlite", help="Itho Database file"
|
||||
)
|
||||
parser.add_argument("--force", action="store_true", help="Force overwrite SQLite database")
|
||||
args = parser.parse_args()
|
||||
return args
|
||||
|
||||
|
||||
def convert(par_file, sqlite_db):
|
||||
par_file = par_file.replace("$", "\\$")
|
||||
par_conn = pyodbc.connect(f'DRIVER={{MDBTools}};DBQ={par_file};')
|
||||
par_conn.setencoding('UTF-8')
|
||||
par_conn.setdecoding(pyodbc.SQL_CHAR, encoding='UTF-8')
|
||||
par_conn = pyodbc.connect(f"DRIVER={{MDBTools}};DBQ={par_file};")
|
||||
par_conn.setencoding("UTF-8")
|
||||
par_conn.setdecoding(pyodbc.SQL_CHAR, encoding="UTF-8")
|
||||
par_cur = par_conn.cursor()
|
||||
|
||||
sqlite_db = db.sqlite(sqlite_db)
|
||||
|
||||
tables = []
|
||||
for table_info in par_cur.tables(tableType='TABLE'):
|
||||
if re.match('^(VersieBeheer|Data[Ll]abel|Parameterlijst)', table_info.table_name):
|
||||
for table_info in par_cur.tables(tableType="TABLE"):
|
||||
if re.match("^(VersieBeheer|Data[Ll]abel|Parameterlijst)", table_info.table_name):
|
||||
tables.append(table_info.table_name)
|
||||
|
||||
for t in sorted(tables):
|
||||
@ -46,10 +50,25 @@ def convert(par_file, sqlite_db):
|
||||
for r in sorted(rows):
|
||||
data.append((r.Index, r.Naam, r.Tekst_NL, r.Tooltip_NL, r.Eenheid_NL))
|
||||
if re.match("^Parameterlijst", t):
|
||||
par_cur.execute(f"select Index, Naam, Naam_fabriek, Min, Max, Default, Tekst_NL, Omschrijving_NL, Eenheid_NL from {t}")
|
||||
par_cur.execute(
|
||||
"select Index, Naam, Naam_fabriek, Min, Max, Default, "
|
||||
f"Tekst_NL, Omschrijving_NL, Eenheid_NL from {t}"
|
||||
)
|
||||
rows = par_cur.fetchall()
|
||||
for r in sorted(rows):
|
||||
data.append((r.Index, r.Naam, r.Naam_fabriek, r.Min, r.Max, r.Default, r.Tekst_NL, r.Omschrijving_NL, r.Eenheid_NL))
|
||||
data.append(
|
||||
(
|
||||
r.Index,
|
||||
r.Naam,
|
||||
r.Naam_fabriek,
|
||||
r.Min,
|
||||
r.Max,
|
||||
r.Default,
|
||||
r.Tekst_NL,
|
||||
r.Omschrijving_NL,
|
||||
r.Eenheid_NL,
|
||||
)
|
||||
)
|
||||
if re.match("^VersieBeheer", t):
|
||||
par_cur.execute(f"select VersieNummer, DataLabel, ParameterLijst from {t}")
|
||||
rows = par_cur.fetchall()
|
||||
|
38
db.py
38
db.py
@ -2,7 +2,7 @@ import sqlite3
|
||||
from sqlite3 import Error
|
||||
|
||||
|
||||
class sqlite():
|
||||
class sqlite:
|
||||
def __init__(self, db_file):
|
||||
self.conn = self.connect(db_file)
|
||||
|
||||
@ -34,15 +34,17 @@ class sqlite():
|
||||
print("Error:", e)
|
||||
|
||||
def create_table(self, t):
|
||||
if t.startswith('datalabel'):
|
||||
if t.startswith("datalabel"):
|
||||
query = """CREATE TABLE {} (
|
||||
id real,
|
||||
name text,
|
||||
title text,
|
||||
tooltip text,
|
||||
unit text
|
||||
);""".format(t)
|
||||
elif t.startswith('parameterlijst'):
|
||||
);""".format(
|
||||
t
|
||||
)
|
||||
elif t.startswith("parameterlijst"):
|
||||
query = """
|
||||
CREATE TABLE {} (
|
||||
id real,
|
||||
@ -54,32 +56,42 @@ class sqlite():
|
||||
title text,
|
||||
description text,
|
||||
unit text
|
||||
);""".format(t)
|
||||
elif t.startswith('versiebeheer'):
|
||||
);""".format(
|
||||
t
|
||||
)
|
||||
elif t.startswith("versiebeheer"):
|
||||
query = """
|
||||
CREATE TABLE {} (
|
||||
version integer primary key,
|
||||
datalabel integer,
|
||||
parameterlist integer
|
||||
);""".format(t)
|
||||
);""".format(
|
||||
t
|
||||
)
|
||||
self.execute(query)
|
||||
self.conn.commit()
|
||||
|
||||
def insert(self, t, data):
|
||||
if t.startswith('datalabel'):
|
||||
if t.startswith("datalabel"):
|
||||
query = """
|
||||
INSERT INTO {} (id, name, title, tooltip, unit)
|
||||
VALUES (?, ?, ?, ?, ?);
|
||||
""".format(t)
|
||||
elif t.startswith('parameterlijst'):
|
||||
""".format(
|
||||
t
|
||||
)
|
||||
elif t.startswith("parameterlijst"):
|
||||
query = """
|
||||
INSERT INTO {} (id, name, name_factory, min, max, def, title, description, unit)
|
||||
VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?);
|
||||
""".format(t)
|
||||
elif t.startswith('versiebeheer'):
|
||||
""".format(
|
||||
t
|
||||
)
|
||||
elif t.startswith("versiebeheer"):
|
||||
query = """
|
||||
INSERT INTO {} (version, datalabel, parameterlist)
|
||||
VALUES (?, ?, ?);
|
||||
""".format(t)
|
||||
""".format(
|
||||
t
|
||||
)
|
||||
self.executemany(query, data)
|
||||
self.conn.commit()
|
||||
|
137
itho-wpu.py
137
itho-wpu.py
@ -11,7 +11,7 @@ import db
|
||||
from collections import namedtuple
|
||||
from itho_i2c import I2CMaster, I2CSlave
|
||||
|
||||
logger = logging.getLogger('stdout')
|
||||
logger = logging.getLogger("stdout")
|
||||
logger.setLevel(logging.INFO)
|
||||
stdout_log_handler = logging.StreamHandler(sys.stdout)
|
||||
stdout_log_handler.setFormatter(logging.Formatter("%(message)s"))
|
||||
@ -27,28 +27,43 @@ actions = {
|
||||
|
||||
|
||||
def parse_args():
|
||||
parser = argparse.ArgumentParser(description='Itho WPU i2c master')
|
||||
parser = argparse.ArgumentParser(description="Itho WPU i2c master")
|
||||
|
||||
parser.add_argument('--action', nargs='?', required=True,
|
||||
choices=actions.keys(), help="Execute an action")
|
||||
parser.add_argument('--loglevel', nargs='?',
|
||||
choices=["debug", "info", "warning", "error", "critical"],
|
||||
help="Loglevel")
|
||||
parser.add_argument('--timestamp', action='store_true', help="Show timestamp in output")
|
||||
parser.add_argument('--master-only', action='store_true', help="Only run I2C master")
|
||||
parser.add_argument('--slave-only', action='store_true', help="Only run I2C slave")
|
||||
parser.add_argument('--slave-timeout', nargs='?', type=int, default=60,
|
||||
help="Slave timeout in seconds when --slave-only")
|
||||
parser.add_argument('--no-cache', action='store_true',
|
||||
help="Don't use local cache")
|
||||
parser.add_argument('--export-to-influxdb', action='store_true',
|
||||
help="Export results to InfluxDB")
|
||||
parser.add_argument(
|
||||
"--action",
|
||||
nargs="?",
|
||||
required=True,
|
||||
choices=actions.keys(),
|
||||
help="Execute an action",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--loglevel",
|
||||
nargs="?",
|
||||
choices=["debug", "info", "warning", "error", "critical"],
|
||||
help="Loglevel",
|
||||
)
|
||||
parser.add_argument("--timestamp", action="store_true", help="Show timestamp in output")
|
||||
parser.add_argument("--master-only", action="store_true", help="Only run I2C master")
|
||||
parser.add_argument("--slave-only", action="store_true", help="Only run I2C slave")
|
||||
parser.add_argument(
|
||||
"--slave-timeout",
|
||||
nargs="?",
|
||||
type=int,
|
||||
default=60,
|
||||
help="Slave timeout in seconds when --slave-only",
|
||||
)
|
||||
parser.add_argument("--no-cache", action="store_true", help="Don't use local cache")
|
||||
parser.add_argument(
|
||||
"--export-to-influxdb",
|
||||
action="store_true",
|
||||
help="Export results to InfluxDB",
|
||||
)
|
||||
|
||||
args = parser.parse_args()
|
||||
return args
|
||||
|
||||
|
||||
class IthoWPU():
|
||||
class IthoWPU:
|
||||
def __init__(self, master_only, slave_only, slave_timeout, no_cache):
|
||||
self.master_only = master_only
|
||||
self.slave_only = slave_only
|
||||
@ -56,13 +71,13 @@ class IthoWPU():
|
||||
self._q = queue.Queue()
|
||||
self.no_cache = no_cache
|
||||
self.cache = IthoWPUCache()
|
||||
self.nodeid = self.get('getnodeid')
|
||||
self.datatype = self.get('getdatatype')
|
||||
self.heatpump_db = db.sqlite('heatpump.sqlite')
|
||||
self.nodeid = self.get("getnodeid")
|
||||
self.datatype = self.get("getdatatype")
|
||||
self.heatpump_db = db.sqlite("heatpump.sqlite")
|
||||
|
||||
def get(self, action):
|
||||
if not self.no_cache:
|
||||
response = self.cache.get(action.replace('get', ''))
|
||||
response = self.cache.get(action.replace("get", ""))
|
||||
if response is not None:
|
||||
logger.debug(f"Response (from cache): {response}")
|
||||
return response
|
||||
@ -85,7 +100,7 @@ class IthoWPU():
|
||||
if not self.master_only:
|
||||
slave.close()
|
||||
|
||||
self.cache.set(action.replace('get', ''), response)
|
||||
self.cache.set(action.replace("get", ""), response)
|
||||
|
||||
return response
|
||||
|
||||
@ -97,32 +112,37 @@ class IthoWPU():
|
||||
def get_datalog_structure(self):
|
||||
listversion = self.get_listversion_from_nodeid()
|
||||
datalabel_version = self.heatpump_db.execute(
|
||||
f"SELECT datalabel FROM versiebeheer WHERE version = {listversion}")[0]['datalabel']
|
||||
f"SELECT datalabel FROM versiebeheer WHERE version = {listversion}"
|
||||
)[0]["datalabel"]
|
||||
if datalabel_version is None or not type(datalabel_version) == int:
|
||||
logger.error(f"Datalabel not found in database for version {listversion}")
|
||||
return None
|
||||
datalabel = self.heatpump_db.execute(
|
||||
f"SELECT name, title, tooltip, unit FROM datalabel_v{datalabel_version} order by id")
|
||||
f"SELECT name, title, tooltip, unit FROM datalabel_v{datalabel_version} order by id"
|
||||
)
|
||||
|
||||
if len(self.datatype[5:-1]) != len(datalabel):
|
||||
logger.warning(f"Number of datatype items ({len(self.datatype[5:-1])}) is not equal to the number of datalabels ({len(datalabel)}) in the database.")
|
||||
logger.warning(
|
||||
f"Number of datatype items ({len(self.datatype[5:-1])}) is not equal to "
|
||||
f"the number of datalabels ({len(datalabel)}) in the database."
|
||||
)
|
||||
|
||||
Field = namedtuple('Field', 'index type label description')
|
||||
Field = namedtuple("Field", "index type label description")
|
||||
|
||||
datalog = []
|
||||
index = 0
|
||||
for dl, dt in zip(datalabel, self.datatype[5:-1]):
|
||||
description = dl['title'].title()
|
||||
if dl['unit'] is not None:
|
||||
description = dl["title"].title()
|
||||
if dl["unit"] is not None:
|
||||
description = f"{description} ({dl['unit']})"
|
||||
description = f"{description} ({dl['name'].lower()})"
|
||||
datalog.append(Field(index, int(dt, 0), dl['name'].lower(), description))
|
||||
datalog.append(Field(index, int(dt, 0), dl["name"].lower(), description))
|
||||
|
||||
if dt in ['0x0', '0xc']:
|
||||
if dt in ["0x0", "0xc"]:
|
||||
index = index + 1
|
||||
elif dt in ['0x10', '0x12', '0x92']:
|
||||
elif dt in ["0x10", "0x12", "0x92"]:
|
||||
index = index + 2
|
||||
elif dt in ['0x20']:
|
||||
elif dt in ["0x20"]:
|
||||
index = index + 4
|
||||
else:
|
||||
logger.error(f"Unknown data type for label {dl['name']}: {dt}")
|
||||
@ -134,10 +154,10 @@ class IthoWPUCache:
|
||||
def __init__(self):
|
||||
self._cache_file = "itho-wpu-cache.json"
|
||||
self._cache_data = {
|
||||
'nodeid': None,
|
||||
'serial': None,
|
||||
'datatype': None,
|
||||
'schema_version': '1',
|
||||
"nodeid": None,
|
||||
"serial": None,
|
||||
"datatype": None,
|
||||
"schema_version": "1",
|
||||
}
|
||||
self._read_cache()
|
||||
|
||||
@ -148,17 +168,17 @@ class IthoWPUCache:
|
||||
with open(self._cache_file) as cache_file:
|
||||
cache_data = json.load(cache_file)
|
||||
logger.debug(f"Loading local cache: {json.dumps(cache_data)}")
|
||||
for key in ['nodeid', 'serial', 'datatype']:
|
||||
for key in ["nodeid", "serial", "datatype"]:
|
||||
if key in cache_data:
|
||||
self._cache_data[key] = cache_data[key]
|
||||
|
||||
def _write_cache(self):
|
||||
with open(self._cache_file, 'w') as cache_file:
|
||||
with open(self._cache_file, "w") as cache_file:
|
||||
logger.debug(f"Writing to local cache: {json.dumps(self._cache_data)}")
|
||||
json.dump(self._cache_data, cache_file)
|
||||
|
||||
def get(self, action):
|
||||
if action not in ['nodeid', 'serial', 'datatype']:
|
||||
if action not in ["nodeid", "serial", "datatype"]:
|
||||
logger.debug(f"Cache for '{action}' is not supported")
|
||||
return None
|
||||
logger.debug(f"Reading '{action}' from local cache")
|
||||
@ -167,7 +187,7 @@ class IthoWPUCache:
|
||||
return self._cache_data[action]
|
||||
|
||||
def set(self, action, value):
|
||||
if action not in ['nodeid', 'serial', 'datatype']:
|
||||
if action not in ["nodeid", "serial", "datatype"]:
|
||||
logger.debug(f"Cache for '{action}' is not supported")
|
||||
return None
|
||||
logger.debug(f"Writing '{action}' to local cache: {value}")
|
||||
@ -177,8 +197,10 @@ class IthoWPUCache:
|
||||
|
||||
def is_messageclass_valid(action, response):
|
||||
if int(response[1], 0) != actions[action][0] and int(response[2], 0) != actions[action][1]:
|
||||
logger.error(f"Response MessageClass != {actions[action][0]} {actions[action][1]} "
|
||||
f"({action}), but {response[1]} {response[2]}")
|
||||
logger.error(
|
||||
f"Response MessageClass != {actions[action][0]} {actions[action][1]} "
|
||||
f"({action}), but {response[1]} {response[2]}"
|
||||
)
|
||||
return False
|
||||
return True
|
||||
|
||||
@ -194,6 +216,7 @@ def process_response(action, response, args, wpu):
|
||||
measurements = process_datalog(response, wpu)
|
||||
if args.export_to_influxdb:
|
||||
from itho_export import export_to_influxdb
|
||||
|
||||
export_to_influxdb(action, measurements)
|
||||
elif action == "getnodeid":
|
||||
process_nodeid(response)
|
||||
@ -208,18 +231,20 @@ def process_nodeid(response):
|
||||
"type": {
|
||||
13: "WPU",
|
||||
15: "Autotemp",
|
||||
}
|
||||
},
|
||||
}
|
||||
}
|
||||
manufacturergroup = ((int(response[5], 0) << 8) + int(response[6], 0))
|
||||
manufacturergroup = (int(response[5], 0) << 8) + int(response[6], 0)
|
||||
manufacturer = hardware_info[int(response[7], 0)]["name"]
|
||||
hardwaretype = hardware_info[int(response[7], 0)]["type"][int(response[8], 0)]
|
||||
productversion = int(response[9], 0)
|
||||
listversion = int(response[10], 0)
|
||||
|
||||
logger.info(f"ManufacturerGroup: {manufacturergroup}, Manufacturer: {manufacturer}, "
|
||||
f"HardwareType: {hardwaretype}, ProductVersion: {productversion}, "
|
||||
f"ListVersion: {listversion}")
|
||||
logger.info(
|
||||
f"ManufacturerGroup: {manufacturergroup}, Manufacturer: {manufacturer}, "
|
||||
f"HardwareType: {hardwaretype}, ProductVersion: {productversion}, "
|
||||
f"ListVersion: {listversion}"
|
||||
)
|
||||
|
||||
|
||||
def process_serial(response):
|
||||
@ -232,24 +257,24 @@ def process_datalog(response, wpu):
|
||||
message = response[5:]
|
||||
measurements = {}
|
||||
for d in datalog:
|
||||
if d.type == 0x0 or d.type == 0xc:
|
||||
m = message[d.index:d.index+1]
|
||||
if d.type == 0x0 or d.type == 0xC:
|
||||
m = message[d.index : d.index + 1] # noqa: E203
|
||||
num = int(m[0], 0)
|
||||
elif d.type == 0x10:
|
||||
m = message[d.index:d.index+2]
|
||||
num = ((int(m[0], 0) << 8) + int(m[1], 0))
|
||||
m = message[d.index : d.index + 2] # noqa: E203
|
||||
num = (int(m[0], 0) << 8) + int(m[1], 0)
|
||||
elif d.type == 0x12:
|
||||
m = message[d.index:d.index+2]
|
||||
m = message[d.index : d.index + 2] # noqa: E203
|
||||
num = round((int(m[0], 0) << 8) + int(m[1], 0) / 100, 2)
|
||||
elif d.type == 0x92:
|
||||
m = message[d.index:d.index+2]
|
||||
num = ((int(m[0], 0) << 8) + int(m[1], 0))
|
||||
m = message[d.index : d.index + 2] # noqa: E203
|
||||
num = (int(m[0], 0) << 8) + int(m[1], 0)
|
||||
if num >= 32768:
|
||||
num -= 65536
|
||||
num = round(num / 100, 2)
|
||||
elif d.type == 0x20:
|
||||
m = message[d.index:d.index+4]
|
||||
num = ((int(m[0], 0) << 24) + (int(m[1], 0) << 16) + (int(m[2], 0) << 8) + int(m[3], 0))
|
||||
m = message[d.index : d.index + 4] # noqa: E203
|
||||
num = (int(m[0], 0) << 24) + (int(m[1], 0) << 16) + (int(m[2], 0) << 8) + int(m[3], 0)
|
||||
else:
|
||||
logger.error(f"Unknown message type for datalog {d.name}: {d.type}")
|
||||
logger.info(f"{d.description}: {num}")
|
||||
|
@ -6,11 +6,11 @@ def export_to_influxdb(action, measurements):
|
||||
from influxdb import InfluxDBClient
|
||||
|
||||
influx_client = InfluxDBClient(
|
||||
host=os.getenv('INFLUXDB_HOST', 'localhost'),
|
||||
port=os.getenv('INFLUXDB_PORT', 8086),
|
||||
username=os.getenv('INFLUXDB_USERNAME', 'root'),
|
||||
password=os.getenv('INFLUXDB_PASSWORD', 'root'),
|
||||
database=os.getenv('INFLUXDB_DATABASE')
|
||||
host=os.getenv("INFLUXDB_HOST", "localhost"),
|
||||
port=os.getenv("INFLUXDB_PORT", 8086),
|
||||
username=os.getenv("INFLUXDB_USERNAME", "root"),
|
||||
password=os.getenv("INFLUXDB_PASSWORD", "root"),
|
||||
database=os.getenv("INFLUXDB_DATABASE"),
|
||||
)
|
||||
json_body = [
|
||||
{
|
||||
@ -22,4 +22,4 @@ def export_to_influxdb(action, measurements):
|
||||
try:
|
||||
influx_client.write_points(json_body)
|
||||
except Exception as e:
|
||||
print('Failed to write to influxdb: ', e)
|
||||
print("Failed to write to influxdb: ", e)
|
||||
|
@ -32,7 +32,7 @@ class I2CRaw:
|
||||
|
||||
def read_i2c_block_data(self, n_bytes):
|
||||
data_raw = self.fr.read(n_bytes)
|
||||
unpack_format = 'B'*n_bytes
|
||||
unpack_format = "B" * n_bytes
|
||||
return list(struct.unpack(unpack_format, data_raw))
|
||||
|
||||
def close(self):
|
||||
@ -80,7 +80,7 @@ class I2CMaster:
|
||||
self.i.close()
|
||||
|
||||
|
||||
class I2CSlave():
|
||||
class I2CSlave:
|
||||
def __init__(self, address, queue):
|
||||
self.address = address
|
||||
self.queue = queue
|
||||
|
9
pyproject.toml
Normal file
9
pyproject.toml
Normal file
@ -0,0 +1,9 @@
|
||||
[tool.black]
|
||||
line-length = 99
|
||||
include = '\.pyi?$'
|
||||
exclude = '''
|
||||
/(
|
||||
\.git
|
||||
| \.venv
|
||||
)/
|
||||
'''
|
Loading…
Reference in New Issue
Block a user