Finish solar server prototype

This commit is contained in:
Tanner Collin 2021-07-10 04:02:44 +00:00
parent 40eba650da
commit c52268be2c
10 changed files with 236 additions and 622 deletions

View File

@ -1,21 +1,14 @@
#!/usr/bin/env python3
# Stolen from https://github.com/ksheumaker/homeassistant-apsystems_ecur
# cheers to ksheumaker and HAEdwin
import asyncio
import socket
import binascii
import datetime
import json
import logging
_LOGGER = logging.getLogger(__name__)
from pprint import pprint
class APSystemsInvalidData(Exception):
pass
class APSystemsECUR:
def __init__(self, ipaddr, port=8899, raw_ecu=None, raw_inverter=None):
self.ipaddr = ipaddr
self.port = port
@ -30,18 +23,18 @@ class APSystemsECUR:
self.cmd_attempts = 3
# how big of a buffer to read at a time from the socket
self.recv_size = 4096
self.recv_size = 4096*16
self.qs1_ids = [ "802", "801" ]
self.yc600_ids = [ "406", "407", "408", "409" ]
self.yc1000_ids = [ "501", "502", "503", "504" ]
self.qs1_ids = [ '802', '801' ]
self.yc600_ids = [ '406', '407', '408', '409' ]
self.yc1000_ids = [ '501', '502', '503', '504' ]
self.cmd_suffix = "END\n"
self.ecu_query = "APS1100160001" + self.cmd_suffix
self.inverter_query_prefix = "APS1100280002"
self.cmd_suffix = 'END\n'
self.ecu_query = 'APS1100160001' + self.cmd_suffix
self.inverter_query_prefix = 'APS1100280002'
self.inverter_query_suffix = self.cmd_suffix
self.inverter_signal_prefix = "APS1100280030"
self.inverter_signal_prefix = 'APS1100280030'
self.inverter_signal_suffix = self.cmd_suffix
self.inverter_byte_start = 26
@ -65,94 +58,54 @@ class APSystemsECUR:
self.reader = None
self.writer = None
def dump(self):
print(f"ECU : {self.ecu_id}")
print(f"Firmware : {self.firmware}")
print(f"TZ : {self.timezone}")
print(f"Qty of inverters : {self.qty_of_inverters}")
async def async_read_from_socket(self):
self.read_buffer = b''
end_data = None
while end_data != self.recv_suffix:
self.read_buffer += await self.reader.read(self.recv_size)
size = len(self.read_buffer)
end_data = self.read_buffer[size-4:]
return self.read_buffer
async def async_send_read_from_socket(self, cmd):
current_attempt = 0
while current_attempt < self.cmd_attempts:
current_attempt += 1
self.writer.write(cmd.encode('utf-8'))
await self.writer.drain()
try:
return await asyncio.wait_for(self.async_read_from_socket(),
timeout=self.timeout)
except Exception as err:
pass
self.writer.close()
raise APSystemsInvalidData(f"Incomplete data from ECU after {current_attempt} attempts, cmd='{cmd.rstrip()}' data={self.read_buffer}")
async def async_query_ecu(self):
self.reader, self.writer = await asyncio.open_connection(self.ipaddr, self.port)
_LOGGER.info(f"Connected to {self.ipaddr} {self.port}")
cmd = self.ecu_query
self.ecu_raw_data = await self.async_send_read_from_socket(cmd)
self.process_ecu_data()
cmd = self.inverter_query_prefix + self.ecu_id + self.inverter_query_suffix
self.inverter_raw_data = await self.async_send_read_from_socket(cmd)
cmd = self.inverter_signal_prefix + self.ecu_id + self.inverter_signal_suffix
self.inverter_raw_signal = await self.async_send_read_from_socket(cmd)
self.writer.close()
data = self.process_inverter_data()
data["ecu_id"] = self.ecu_id
data["today_energy"] = self.today_energy
data["lifetime_energy"] = self.lifetime_energy
data["current_power"] = self.current_power
return(data)
def query_ecu(self):
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
sock.connect((self.ipaddr,self.port))
logging.debug('ecu query: %s', self.ecu_query)
sock.sendall(self.ecu_query.encode('utf-8'))
self.ecu_raw_data = sock.recv(self.recv_size)
logging.debug('raw ecu data: %s', self.ecu_raw_data)
logging.debug('ecu id: %s', self.ecu_raw_data[13:25])
power = self.ecu_raw_data[31:35]
logging.debug('current power: %s', int.from_bytes(power, byteorder='big'))
self.process_ecu_data()
sock.shutdown(socket.SHUT_RDWR)
sock.close()
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
sock.connect((self.ipaddr,self.port))
cmd = self.inverter_query_prefix + self.ecu_id + self.inverter_query_suffix
logging.debug('inverter data cmd: %s', cmd)
sock.sendall(cmd.encode('utf-8'))
self.inverter_raw_data = sock.recv(self.recv_size)
logging.debug('raw inverter data: %s', self.inverter_raw_data)
sock.shutdown(socket.SHUT_RDWR)
sock.close()
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
sock.connect((self.ipaddr,self.port))
cmd = self.inverter_signal_prefix + self.ecu_id + self.inverter_signal_suffix
logging.debug('inverter signal cmd: %s', cmd)
sock.sendall(cmd.encode('utf-8'))
self.inverter_raw_signal = sock.recv(self.recv_size)
logging.debug('raw signal data: %s', self.inverter_raw_signal)
sock.shutdown(socket.SHUT_RDWR)
sock.close()
data = self.process_inverter_data()
data["ecu_id"] = self.ecu_id
data["today_energy"] = self.today_energy
data["lifetime_energy"] = self.lifetime_energy
data["current_power"] = self.current_power
data['ecu_id'] = self.ecu_id
data['today_energy'] = self.today_energy
data['lifetime_energy'] = self.lifetime_energy
data['current_power'] = self.current_power
return(data)
@ -161,21 +114,21 @@ class APSystemsECUR:
return int(binascii.b2a_hex(codec[(start):(start+2)]), 16)
except ValueError as err:
debugdata = binascii.b2a_hex(codec)
raise APSystemsInvalidData(f"Unable to convert binary to int location={start} data={debugdata}")
raise APSystemsInvalidData(f'Unable to convert binary to int location={start} data={debugdata}')
def aps_short(self, codec, start):
try:
return int(binascii.b2a_hex(codec[(start):(start+1)]), 8)
except ValueError as err:
debugdata = binascii.b2a_hex(codec)
raise APSystemsInvalidData(f"Unable to convert binary to short int location={start} data={debugdata}")
raise APSystemsInvalidData(f'Unable to convert binary to short int location={start} data={debugdata}')
def aps_double(self, codec, start):
try:
return int (binascii.b2a_hex(codec[(start):(start+4)]), 16)
except ValueError as err:
debugdata = binascii.b2a_hex(codec)
raise APSystemsInvalidData(f"Unable to convert binary to double location={start} data={debugdata}")
raise APSystemsInvalidData(f'Unable to convert binary to double location={start} data={debugdata}')
def aps_bool(self, codec, start):
return bool(binascii.b2a_hex(codec[(start):(start+2)]))
@ -188,38 +141,41 @@ class APSystemsECUR:
def aps_timestamp(self, codec, start, amount):
timestr=str(binascii.b2a_hex(codec[start:(start+amount)]))[2:(amount+2)]
return timestr[0:4]+"-"+timestr[4:6]+"-"+timestr[6:8]+" "+timestr[8:10]+":"+timestr[10:12]+":"+timestr[12:14]
return timestr[0:4]+'-'+timestr[4:6]+'-'+timestr[6:8]+' '+timestr[8:10]+':'+timestr[10:12]+':'+timestr[12:14]
def check_ecu_checksum(self, data, cmd):
datalen = len(data) - 1
logging.debug('datalen: %s', datalen)
try:
checksum = int(data[5:9])
logging.debug('checksum: %s', checksum)
except ValueError as err:
debugdata = binascii.b2a_hex(data)
raise APSystemsInvalidData(f"Error getting checksum int from '{cmd}' data={debugdata}")
raise APSystemsInvalidData(f'Error getting checksum int from {cmd} data={debugdata}')
if datalen != checksum:
debugdata = binascii.b2a_hex(data)
raise APSystemsInvalidData(f"Checksum on '{cmd}' failed checksum={checksum} datalen={datalen} data={debugdata}")
raise APSystemsInvalidData(f'Checksum on {cmd} failed checksum={checksum} datalen={datalen} data={debugdata}')
start_str = self.aps_str(data, 0, 3)
end_str = self.aps_str(data, len(data) - 4, 3)
if start_str != 'APS':
debugdata = binascii.b2a_hex(data)
raise APSystemsInvalidData(f"Result on '{cmd}' incorrect start signature '{start_str}' != APS data={debugdata}")
raise APSystemsInvalidData(f'Result on {cmd} incorrect start signature {start_str} != APS data={debugdata}')
if end_str != 'END':
debugdata = binascii.b2a_hex(data)
raise APSystemsInvalidData(f"Result on '{cmd}' incorrect end signature '{end_str}' != END data={debugdata}")
raise APSystemsInvalidData(f'Result on {cmd} incorrect end signature {end_str} != END data={debugdata}')
return True
def process_ecu_data(self, data=None):
logging.debug('Processing ECU data...')
if not data:
data = self.ecu_raw_data
self.check_ecu_checksum(data, "ECU Query")
self.check_ecu_checksum(data, 'ECU Query')
self.ecu_id = self.aps_str(data, 13, 12)
self.qty_of_inverters = self.aps_int(data, 46)
@ -231,12 +187,13 @@ class APSystemsECUR:
def process_signal_data(self, data=None):
logging.debug('Processing signal data...')
signal_data = {}
if not data:
data = self.inverter_raw_signal
self.check_ecu_checksum(data, "Signal Query")
self.check_ecu_checksum(data, 'Signal Query')
if not self.qty_of_inverters:
return signal_data
@ -255,10 +212,11 @@ class APSystemsECUR:
return signal_data
def process_inverter_data(self, data=None):
logging.debug('Processing inverter data...')
if not data:
data = self.inverter_raw_data
self.check_ecu_checksum(data, "Inverter data")
self.check_ecu_checksum(data, 'Inverter data')
output = {}
@ -266,9 +224,9 @@ class APSystemsECUR:
inverter_qty = self.aps_int(data, 17)
self.last_update = timestamp
output["timestamp"] = timestamp
output["inverter_qty"] = inverter_qty
output["inverters"] = {}
output['timestamp'] = timestamp
output['inverter_qty'] = inverter_qty
output['inverters'] = {}
# this is the start of the loop of inverters
location = self.inverter_byte_start
@ -281,22 +239,22 @@ class APSystemsECUR:
inv={}
inverter_uid = self.aps_uid(data, location)
inv["uid"] = inverter_uid
inv['uid'] = inverter_uid
location += 6
inv["online"] = self.aps_bool(data, location)
inv['online'] = self.aps_bool(data, location)
location += 1
inv["unknown"] = self.aps_str(data, location, 2)
inv['unknown'] = self.aps_str(data, location, 2)
location += 2
inv["frequency"] = self.aps_int(data, location) / 10
inv['frequency'] = self.aps_int(data, location) / 10
location += 2
inv["temperature"] = self.aps_int(data, location) - 100
inv['temperature'] = self.aps_int(data, location) - 100
location += 2
inv["signal"] = signal.get(inverter_uid, 0)
inv['signal'] = signal.get(inverter_uid, 0)
# the first 3 digits determine the type of inverter
inverter_type = inverter_uid[0:3]
@ -313,11 +271,11 @@ class APSystemsECUR:
inv.update(channel_data)
else:
raise APSystemsInvalidData(f"Unsupported inverter type {inverter_type}")
raise APSystemsInvalidData(f'Unsupported inverter type {inverter_type}')
inverters[inverter_uid] = inv
output["inverters"] = inverters
output['inverters'] = inverters
return (output)
def process_yc1000(self, data, location):
@ -349,10 +307,10 @@ class APSystemsECUR:
voltages.append(voltage)
output = {
"model" : "YC1000",
"channel_qty" : 4,
"power" : power,
"voltage" : voltages
'model' : 'YC1000',
'channel_qty' : 4,
'power' : power,
'voltage' : voltages
}
return (output, location)
@ -381,10 +339,10 @@ class APSystemsECUR:
voltages.append(voltage)
output = {
"model" : "QS1",
"channel_qty" : 4,
"power" : power,
"voltage" : voltages
'model' : 'QS1',
'channel_qty' : 4,
'power' : power,
'voltage' : voltages
}
return (output, location)
@ -402,12 +360,10 @@ class APSystemsECUR:
location += 2
output = {
"model" : "YC600",
"channel_qty" : 2,
"power" : power,
"voltage" : voltages,
'model' : 'YC600',
'channel_qty' : 2,
'power' : power,
'voltage' : voltages,
}
return (output, location)

View File

@ -1,154 +0,0 @@
import logging
import voluptuous as vol
import traceback
from datetime import timedelta
from .APSystemsECUR import APSystemsECUR, APSystemsInvalidData
from homeassistant.helpers.discovery import load_platform
import homeassistant.helpers.config_validation as cv
from homeassistant.const import CONF_HOST
from homeassistant.helpers.entity import Entity
from homeassistant.util import Throttle
from homeassistant.helpers.update_coordinator import (
CoordinatorEntity,
DataUpdateCoordinator,
UpdateFailed,
)
_LOGGER = logging.getLogger(__name__)
from .const import DOMAIN
CONF_INTERVAL = "interval"
CONFIG_SCHEMA = vol.Schema({
DOMAIN : vol.Schema({
vol.Required(CONF_HOST): cv.string,
vol.Optional(CONF_INTERVAL) : cv.time_period_seconds
})
}, extra=vol.ALLOW_EXTRA)
PLATFORMS = [ "sensor", "binary_sensor" ]
## handle all the communications with the ECUR class and deal with our need for caching, etc
class ECUR():
def __init__(self, ipaddr):
self.ecu = APSystemsECUR(ipaddr)
self.cache_count = 0
self.cache_max = 5
self.data_from_cache = False
self.querying = True
self.error_messge = ""
self.cached_data = {}
async def stop_query(self):
self.querying = False
async def start_query(self):
self.querying = True
async def update(self):
data = {}
# if we aren't actively quering data, pull data form the cache
# this is so we can stop querying after sunset
if not self.querying:
_LOGGER.debug("Not querying ECU due to stopped")
data = self.cached_data
self.data_from_cache = True
data["data_from_cache"] = self.data_from_cache
data["querying"] = self.querying
return self.cached_data
_LOGGER.debug("Querying ECU")
try:
data = await self.ecu.async_query_ecu()
_LOGGER.debug("Got data from ECU")
# we got good results, so we store it and set flags about our
# cache state
self.cached_data = data
self.cache_count = 0
self.data_from_cache = False
self.error_message = ""
except APSystemsInvalidData as err:
msg = f"Using cached data from last successful communication from ECU. Error: {err}"
_LOGGER.warning(msg)
# we got invalid data, so we need to pull from cache
self.error_msg = msg
self.cache_count += 1
self.data_from_cache = True
data = self.cached_data
if self.cache_count > self.cache_max:
raise Exception(f"Error using cached data for more than {self.cache_max} times.")
except Exception as err:
msg = f"Using cached data from last successful communication from ECU. Error: {err}"
_LOGGER.warning(msg)
# we got invalid data, so we need to pull from cache
self.error_msg = msg
self.cache_count += 1
self.data_from_cache = True
data = self.cached_data
if self.cache_count > self.cache_max:
raise Exception(f"Error using cached data for more than {self.cache_max} times.")
data["data_from_cache"] = self.data_from_cache
data["querying"] = self.querying
_LOGGER.debug(f"Returning {data}")
return data
async def async_setup(hass, config):
""" Setup the APsystems platform """
hass.data.setdefault(DOMAIN, {})
host = config[DOMAIN].get(CONF_HOST)
interval = config[DOMAIN].get(CONF_INTERVAL)
if not interval:
interval = timedelta(seconds=60)
ecu = ECUR(host)
coordinator = DataUpdateCoordinator(
hass,
_LOGGER,
name=DOMAIN,
update_method=ecu.update,
update_interval=interval,
)
await coordinator.async_refresh()
hass.data[DOMAIN] = {
"ecu" : ecu,
"coordinator" : coordinator
}
async def handle_stop_query(call):
await ecu.stop_query()
coordinator.async_refresh()
async def handle_start_query(call):
await ecu.start_query()
coordinator.async_refresh()
hass.services.async_register(DOMAIN, "start_query", handle_start_query)
hass.services.async_register(DOMAIN, "stop_query", handle_stop_query)
for component in PLATFORMS:
load_platform(hass, component, DOMAIN, {}, config)
return True

View File

@ -1,86 +0,0 @@
"""Example integration using DataUpdateCoordinator."""
from datetime import timedelta
import logging
import async_timeout
from homeassistant.components.binary_sensor import (
BinarySensorEntity,
)
from homeassistant.helpers.update_coordinator import (
CoordinatorEntity,
DataUpdateCoordinator,
UpdateFailed,
)
from .const import (
DOMAIN,
RELOAD_ICON,
CACHE_ICON
)
_LOGGER = logging.getLogger(__name__)
async def async_setup_platform(hass, config, add_entities, discovery_info=None):
ecu = hass.data[DOMAIN].get("ecu")
coordinator = hass.data[DOMAIN].get("coordinator")
sensors = [
APSystemsECUBinarySensor(coordinator, ecu, "data_from_cache",
label="Using Cached Data", icon=CACHE_ICON),
APSystemsECUBinarySensor(coordinator, ecu, "querying",
label="Querying Enabled", icon=RELOAD_ICON),
]
add_entities(sensors)
class APSystemsECUBinarySensor(CoordinatorEntity, BinarySensorEntity):
def __init__(self, coordinator, ecu, field, label=None, devclass=None, icon=None):
super().__init__(coordinator)
self.coordinator = coordinator
self._ecu = ecu
self._field = field
self._label = label
if not label:
self._label = field
self._icon = icon
self._name = f"ECU {self._label}"
self._state = None
@property
def unique_id(self):
return f"{self._ecu.ecu.ecu_id}_{self._field}"
@property
def name(self):
return self._name
@property
def is_on(self):
return self.coordinator.data.get(self._field)
@property
def icon(self):
return self._icon
@property
def device_state_attributes(self):
attrs = {
"ecu_id" : self._ecu.ecu.ecu_id,
"firmware" : self._ecu.ecu.firmware,
"last_update" : self._ecu.ecu.last_update
}
return attrs

View File

@ -1,7 +0,0 @@
DOMAIN = 'apsystems_ecur'
SOLAR_ICON = "mdi:solar-power"
FREQ_ICON = "mdi:sine-wave"
SIGNAL_ICON = "mdi:signal"
RELOAD_ICON = "mdi:reload"
CACHE_ICON = "mdi:cached"

View File

@ -1,10 +0,0 @@
{
"codeowners": ["@ksheumaker"],
"config_flow": false,
"dependencies": [],
"documentation": "https://github.com/ksheumaker/homeassistant-apsystems_ecur",
"domain": "apsystems_ecur",
"name": "APSystems PV solar ECU-R",
"version": "1.0.1",
"requirements": []
}

View File

@ -1,208 +0,0 @@
"""Example integration using DataUpdateCoordinator."""
from datetime import timedelta
import logging
import async_timeout
from homeassistant.helpers.entity import Entity
from homeassistant.helpers.update_coordinator import (
CoordinatorEntity,
DataUpdateCoordinator,
UpdateFailed,
)
from .const import (
DOMAIN,
SOLAR_ICON,
FREQ_ICON,
SIGNAL_ICON
)
from homeassistant.const import (
DEVICE_CLASS_TEMPERATURE,
DEVICE_CLASS_POWER,
DEVICE_CLASS_ENERGY,
DEVICE_CLASS_VOLTAGE,
DEVICE_CLASS_ENERGY,
ENERGY_KILO_WATT_HOUR,
POWER_WATT,
VOLT,
TEMP_CELSIUS,
PERCENTAGE,
FREQUENCY_HERTZ
)
_LOGGER = logging.getLogger(__name__)
async def async_setup_platform(hass, config, add_entities, discovery_info=None):
ecu = hass.data[DOMAIN].get("ecu")
coordinator = hass.data[DOMAIN].get("coordinator")
sensors = [
APSystemsECUSensor(coordinator, ecu, "current_power",
label="Current Power", unit=POWER_WATT,
devclass=DEVICE_CLASS_POWER, icon=SOLAR_ICON),
APSystemsECUSensor(coordinator, ecu, "today_energy",
label="Today Energy", unit=ENERGY_KILO_WATT_HOUR,
devclass=DEVICE_CLASS_ENERGY, icon=SOLAR_ICON),
APSystemsECUSensor(coordinator, ecu, "lifetime_energy",
label="Lifetime Energy", unit=ENERGY_KILO_WATT_HOUR,
devclass=DEVICE_CLASS_ENERGY, icon=SOLAR_ICON),
]
inverters = coordinator.data.get("inverters", {})
for uid,inv_data in inverters.items():
#_LOGGER.warning(f"Inverter {uid} {inv_data.get('channel_qty')}")
sensors.extend([
APSystemsECUInverterSensor(coordinator, ecu, uid,
"temperature", label="Temperature",
devclass=DEVICE_CLASS_TEMPERATURE, unit=TEMP_CELSIUS),
APSystemsECUInverterSensor(coordinator, ecu, uid,
"frequency", label="Frequency", unit=FREQUENCY_HERTZ,
devclass=None, icon=FREQ_ICON),
APSystemsECUInverterSensor(coordinator, ecu, uid,
"voltage", label="Voltage", unit=VOLT,
devclass=DEVICE_CLASS_VOLTAGE),
APSystemsECUInverterSensor(coordinator, ecu, uid,
"signal", label="Signal", unit=PERCENTAGE,
icon=SIGNAL_ICON)
])
for i in range(0, inv_data.get("channel_qty", 0)):
sensors.append(
APSystemsECUInverterSensor(coordinator, ecu, uid, f"power",
index=i, label=f"Power Ch {i+1}", unit=POWER_WATT,
devclass=DEVICE_CLASS_POWER, icon=SOLAR_ICON)
)
add_entities(sensors)
class APSystemsECUInverterSensor(CoordinatorEntity, Entity):
def __init__(self, coordinator, ecu, uid, field, index=0, label=None, icon=None, unit=None, devclass=None):
super().__init__(coordinator)
self.coordinator = coordinator
self._index = index
self._uid = uid
self._ecu = ecu
self._field = field
self._devclass = devclass
self._label = label
if not label:
self._label = field
self._icon = icon
self._unit = unit
self._name = f"Inverter {self._uid} {self._label}"
self._state = None
@property
def unique_id(self):
field = self._field
if self._index != None:
field = f"{field}_{self._index}"
return f"{self._ecu.ecu.ecu_id}_{self._uid}_{field}"
@property
def device_class(self):
return self._devclass
@property
def name(self):
return self._name
@property
def state(self):
#_LOGGER.warning(f"State called for {self._field}")
if self._field == "voltage":
return self.coordinator.data.get("inverters", {}).get(self._uid, {}).get("voltage", [])[0]
elif self._field == "power":
#_LOGGER.warning(f"POWER {self._uid} {self._index}")
return self.coordinator.data.get("inverters", {}).get(self._uid, {}).get("power", [])[self._index]
else:
return self.coordinator.data.get("inverters", {}).get(self._uid, {}).get(self._field)
@property
def icon(self):
return self._icon
@property
def unit_of_measurement(self):
return self._unit
@property
def device_state_attributes(self):
attrs = {
"ecu_id" : self._ecu.ecu.ecu_id,
"last_update" : self._ecu.ecu.last_update,
}
return attrs
class APSystemsECUSensor(CoordinatorEntity, Entity):
def __init__(self, coordinator, ecu, field, label=None, icon=None, unit=None, devclass=None):
super().__init__(coordinator)
self.coordinator = coordinator
self._ecu = ecu
self._field = field
self._label = label
if not label:
self._label = field
self._icon = icon
self._unit = unit
self._devclass = devclass
self._name = f"ECU {self._label}"
self._state = None
@property
def unique_id(self):
return f"{self._ecu.ecu.ecu_id}_{self._field}"
@property
def name(self):
return self._name
@property
def device_class(self):
return self._devclass
@property
def state(self):
#_LOGGER.warning(f"State called for {self._field}")
return self.coordinator.data.get(self._field)
@property
def icon(self):
return self._icon
@property
def unit_of_measurement(self):
return self._unit
@property
def device_state_attributes(self):
attrs = {
"ecu_id" : self._ecu.ecu.ecu_id,
"firmware" : self._ecu.ecu.firmware,
"last_update" : self._ecu.ecu.last_update
}
return attrs

View File

@ -1,5 +0,0 @@
start_query:
description: Start querying the ECU for new data (i.e. when sunrise)
stop_query:
description: Stop querying the ECU for new data (i.e. when sunset)

View File

@ -1,25 +0,0 @@
#!/usr/bin/env python3
from APSystemsECUR import APSystemsECUR
import time
import asyncio
from pprint import pprint
ecu_ip = "192.168.0.251"
sleep = 60
loop = asyncio.get_event_loop()
ecu = APSystemsECUR(ecu_ip)
while True:
try:
data = loop.run_until_complete(ecu.async_query_ecu())
print(f"[OK] Timestamp: {data.get('timestamp')} Current Power: {data.get('current_power')}")
pprint(data)
except Exception as err:
print(f"[ERROR] {err}")
print(f"Sleeping for {sleep} sec")
time.sleep(sleep)

View File

@ -1,30 +1,165 @@
import os
import logging
logging.basicConfig(
format='[%(asctime)s] %(levelname)s %(module)s/%(funcName)s: - %(message)s',
level=logging.DEBUG if os.environ.get('DEBUG') else logging.INFO)
logging.getLogger('aiohttp').setLevel(logging.DEBUG if os.environ.get('DEBUG') else logging.WARNING)
import asyncio
from aiohttp import web
from datetime import datetime, timedelta
import pytz
from APSystemsECUR import APSystemsECUR
ECU_IP = '192.168.69.103'
LISTEN_IP = '192.168.69.100'
ecu = APSystemsECUR(ECU_IP)
app = web.Application()
prev_ecu_timestamp = None
solar_data = {}
from influxdb import InfluxDBClient
client = InfluxDBClient('localhost', 8086, database='solar2')
async def proxy(reader, writer):
message = await reader.read(1024)
addr = writer.get_extra_info('peername')
try:
print('Recvd from {}: {}'.format(addr[0], message))
logging.debug('Recvd from {}: {}'.format(addr[0], message))
offset = 32
date_time_obj = datetime.strptime(str(message)[offset:offset+14], '%Y%m%d%H%M%S') + timedelta(minutes=-5)
send_str = '101' + datetime.strftime(date_time_obj, '%Y%m%d%H%M%S')
send_data = send_str.encode()
print('Sending to {}: {}'.format(addr[0], send_data))
logging.debug('Sending to {}: {}'.format(addr[0], send_data))
writer.write(send_data)
await writer.drain()
except ValueError:
print('Ignored unnecessary data')
logging.debug('Ignored unnecessary data')
writer.close()
async def main():
async def run_proxies():
for port in [8995, 8996, 8997]:
server = await asyncio.start_server(proxy, '192.168.69.69', port)
server = await asyncio.start_server(proxy, LISTEN_IP, port)
logging.info('Started TCP listener server on %s:%s', LISTEN_IP, port)
task = asyncio.create_task(server.serve_forever())
# block here for now
await task
async def get_data():
global prev_ecu_timestamp
global solar_data
asyncio.run(main())
await asyncio.sleep(1)
while True:
try:
logging.debug('Grabbing ECU data...')
data = ecu.query_ecu()
logging.debug('Good read, timestamp: %s Mountain Time, current power: %s', data['timestamp'], data['current_power'])
if data['timestamp'] != prev_ecu_timestamp:
total = 0
timestamp = datetime.utcnow()
for i in data['inverters'].values():
total += i['power'][0]
total += i['power'][1]
data['actual_total'] = total
solar_data = data
logging.info('Solar data updated, ecu time: %s Mountain, ecu total: %s, actual total: %s', data['timestamp'], data['current_power'], total)
points = []
for i in data['inverters'].values():
points.append({
'time': timestamp,
'measurement': 'inverter',
'tags': {'ecu': data['ecu_id'], 'inverter': i['uid']},
'fields': {
'ecu_time': data['timestamp'],
'online': i['online'],
'frequency': i['frequency'],
'temperature': i['temperature'],
}
})
points.append({
'time': timestamp,
'measurement': 'panel',
'tags': {'ecu': data['ecu_id'], 'inverter': i['uid'], 'channel': '0'},
'fields': {
'ecu_time': data['timestamp'],
'power': i['power'][0],
'voltage': i['voltage'][0]
}
})
points.append({
'time': timestamp,
'measurement': 'panel',
'tags': {'ecu': data['ecu_id'], 'inverter': i['uid'], 'channel': '1'},
'fields': {
'ecu_time': data['timestamp'],
'power': i['power'][1],
'voltage': i['voltage'][1]
}
})
points.append({
'time': timestamp,
'measurement': 'ecu',
'tags': {'ecu': data['ecu_id']},
'fields': {
'ecu_total': data['current_power'],
'ecu_time': data['timestamp'],
'actual_total': data['actual_total'],
'today_energy': data['today_energy'],
'lifetime_energy': data['lifetime_energy'],
}
})
client.write_points(points)
logging.info('Wrote %s points to InfluxDB', len(points))
prev_ecu_timestamp = data['timestamp']
except Exception as err:
logging.error('Error: ' + str(err))
await asyncio.sleep(120)
async def index(request):
return web.Response(text='hello world', content_type='text/html')
async def data(request):
return web.json_response(solar_data)
async def history(request):
try:
date = datetime.strptime(request.match_info['date'], '%Y-%m-%d')
tz = pytz.timezone('America/Edmonton')
date = tz.localize(date)
except ValueError:
raise web.HTTPNotFound
start = int(date.timestamp())
end = date + timedelta(days=1)
end = int(end.timestamp())
q = 'select * from ecu where time >= {}s and time < {}s'.format(start, end)
history = list(client.query(q).get_points())
return web.json_response(history)
if __name__ == '__main__':
app.router.add_get('/', index)
app.router.add_get('/data', data)
app.router.add_get('/history/{date}', history)
loop = asyncio.get_event_loop()
loop.create_task(run_proxies())
loop.create_task(get_data())
web.run_app(app, port=6901)

18
server/requirements.txt Normal file
View File

@ -0,0 +1,18 @@
aiohttp==3.7.4.post0
async-timeout==3.0.1
attrs==21.2.0
certifi==2021.5.30
chardet==4.0.0
ciso8601==2.1.3
idna==2.10
influxdb==5.3.1
msgpack==1.0.2
multidict==5.1.0
pkg-resources==0.0.0
python-dateutil==2.8.1
pytz==2021.1
requests==2.25.1
six==1.16.0
typing-extensions==3.10.0.0
urllib3==1.26.5
yarl==1.6.3