-
Notifications
You must be signed in to change notification settings - Fork 15
/
energy_csv.py
executable file
·138 lines (126 loc) · 5.92 KB
/
energy_csv.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
#!/usr/bin/env python
"""Example exec module to use the Anker API for export of daily Solarbank Energy Data.
This method will prompt for the Anker account details if not pre-set in the
header. Then you can specify a start day and the number of days for data
extraction from the Anker Cloud.
Note: The Solar production and Solarbank discharge can be queried across the
full range. The solarbank charge however can be queried only as total for an
interval (e.g. day). Therefore when solarbank charge data is also selected for
export, an additional API query per day is required. The received daily values
will be exported into a csv file.
"""
import asyncio
import csv
from datetime import datetime
import json
import logging
from pathlib import Path
from aiohttp import ClientSession
from api import api # pylint: disable=no-name-in-module
import common
# use Console logger from common module
CONSOLE: logging.Logger = common.CONSOLE
# enable debug mode for the console handler
# CONSOLE.handlers[0].setLevel(logging.DEBUG)
async def main() -> bool:
"""Run main to export energy history from cloud."""
CONSOLE.info("Exporting daily Energy data for Anker Solarbank:")
try:
async with ClientSession() as websession:
CONSOLE.info("\nTrying authentication...")
myapi = api.AnkerSolixApi(
common.user(), common.password(), common.country(), websession, CONSOLE
)
if await myapi.async_authenticate():
CONSOLE.info("OK")
else:
CONSOLE.info(
"CACHED"
) # Login validation will be done during first API call
# Refresh the site and device info of the API
CONSOLE.info("\nUpdating Site Info...")
if (await myapi.update_sites()) == {}:
CONSOLE.info("NO INFO")
return False
CONSOLE.info("OK")
CONSOLE.info("\nSites: %s", len(myapi.sites))
CONSOLE.debug(json.dumps(myapi.sites, indent=2))
for site_id, site in myapi.sites.items():
site_name = (site.get("site_info") or {}).get("site_name") or ""
CONSOLE.info("Found site %s ID %s", site_name, site_id)
try:
daystr = input(
"\nEnter start day for daily energy data (yyyy-mm-dd) or enter to skip site: "
)
if daystr == "":
CONSOLE.info(
"Skipped site %s, checking for next site...", site_name
)
continue
startday = datetime.fromisoformat(daystr)
numdays = int(input("How many days to query (1-366): "))
daytotals = input(
"Do you want to include daily total data (e.g. solarbank charge, grid import/export) which may require several API queries per day? (Y/N): "
)
daytotals = daytotals.upper() in ["Y", "YES", "TRUE", 1]
prefix = input(
f"CSV filename prefix for export (daily_energy_{daystr}): "
)
if prefix == "":
prefix = f"daily_energy_{daystr}"
filename = f"{prefix}_{site_name}.csv"
except ValueError:
return False
# delay requests, limit appears to be around 25 per minute
if numdays > 10:
myapi.requestDelay(2.5)
else:
myapi.requestDelay(0.3)
CONSOLE.info(
"Queries may take up to %s seconds with %.1f seconds delay...please wait...",
round((2 * numdays * daytotals + 5) * myapi.requestDelay()),
myapi.requestDelay(),
)
data = await myapi.energy_daily(
siteId=site_id,
deviceSn="", # mandatory parameter but can be empty since not distinguished for site energy stats
startDay=startday,
numDays=numdays,
dayTotals=daytotals,
# TODO(#SMARTPLUG): Add Smartplug type once supported
devTypes={
api.SolixDeviceType.INVERTER.value,
api.SolixDeviceType.SOLARBANK.value,
api.SolixDeviceType.SMARTMETER.value,
}, # include all possible energy stats per site
)
CONSOLE.debug(json.dumps(data, indent=2))
# Write csv file
if len(data) > 0:
with Path.open( # noqa: ASYNC230
Path(filename), "w", newline="", encoding="utf-8"
) as csvfile:
fieldnames = (next(iter(data.values()))).keys()
writer = csv.DictWriter(csvfile, fieldnames=fieldnames)
writer.writeheader()
writer.writerows(data.values())
CONSOLE.info(
"\nCompleted: Successfully exported data to %s",
Path.resolve(Path(filename)),
)
else:
CONSOLE.info(
"No data received for site %s ID %s", site_name, site_id
)
return False
return True
except Exception as err: # pylint: disable=broad-exception-caught # noqa: BLE001
CONSOLE.error("%s: %s", type(err), err)
return False
# run async main
if __name__ == "__main__":
try:
if not asyncio.run(main(), debug=False):
CONSOLE.warning("Aborted!")
except Exception as exception: # pylint: disable=broad-exception-caught # noqa: BLE001
CONSOLE.exception("%s: %s", type(exception), exception)