Move tomorrowio coordinator to separate module (#117537)
* Move tomorrowio coordinator to separate module * Adjust imports
This commit is contained in:
parent
b015dbfccb
commit
7ceaf2d3f0
4 changed files with 283 additions and 273 deletions
|
@ -2,129 +2,24 @@
|
|||
|
||||
from __future__ import annotations
|
||||
|
||||
import asyncio
|
||||
from datetime import timedelta
|
||||
from math import ceil
|
||||
from typing import Any
|
||||
|
||||
from pytomorrowio import TomorrowioV4
|
||||
from pytomorrowio.const import CURRENT, FORECASTS
|
||||
from pytomorrowio.exceptions import (
|
||||
CantConnectException,
|
||||
InvalidAPIKeyException,
|
||||
RateLimitedException,
|
||||
UnknownException,
|
||||
)
|
||||
from pytomorrowio.const import CURRENT
|
||||
|
||||
from homeassistant.components.sensor import DOMAIN as SENSOR_DOMAIN
|
||||
from homeassistant.components.weather import DOMAIN as WEATHER_DOMAIN
|
||||
from homeassistant.config_entries import ConfigEntry
|
||||
from homeassistant.const import (
|
||||
CONF_API_KEY,
|
||||
CONF_LATITUDE,
|
||||
CONF_LOCATION,
|
||||
CONF_LONGITUDE,
|
||||
)
|
||||
from homeassistant.core import HomeAssistant, callback
|
||||
from homeassistant.const import CONF_API_KEY
|
||||
from homeassistant.core import HomeAssistant
|
||||
from homeassistant.helpers.aiohttp_client import async_get_clientsession
|
||||
from homeassistant.helpers.device_registry import DeviceEntryType, DeviceInfo
|
||||
from homeassistant.helpers.update_coordinator import (
|
||||
CoordinatorEntity,
|
||||
DataUpdateCoordinator,
|
||||
UpdateFailed,
|
||||
)
|
||||
from homeassistant.helpers.update_coordinator import CoordinatorEntity
|
||||
|
||||
from .const import (
|
||||
ATTRIBUTION,
|
||||
CONF_TIMESTEP,
|
||||
DOMAIN,
|
||||
INTEGRATION_NAME,
|
||||
LOGGER,
|
||||
TMRW_ATTR_CARBON_MONOXIDE,
|
||||
TMRW_ATTR_CHINA_AQI,
|
||||
TMRW_ATTR_CHINA_HEALTH_CONCERN,
|
||||
TMRW_ATTR_CHINA_PRIMARY_POLLUTANT,
|
||||
TMRW_ATTR_CLOUD_BASE,
|
||||
TMRW_ATTR_CLOUD_CEILING,
|
||||
TMRW_ATTR_CLOUD_COVER,
|
||||
TMRW_ATTR_CONDITION,
|
||||
TMRW_ATTR_DEW_POINT,
|
||||
TMRW_ATTR_EPA_AQI,
|
||||
TMRW_ATTR_EPA_HEALTH_CONCERN,
|
||||
TMRW_ATTR_EPA_PRIMARY_POLLUTANT,
|
||||
TMRW_ATTR_FEELS_LIKE,
|
||||
TMRW_ATTR_FIRE_INDEX,
|
||||
TMRW_ATTR_HUMIDITY,
|
||||
TMRW_ATTR_NITROGEN_DIOXIDE,
|
||||
TMRW_ATTR_OZONE,
|
||||
TMRW_ATTR_PARTICULATE_MATTER_10,
|
||||
TMRW_ATTR_PARTICULATE_MATTER_25,
|
||||
TMRW_ATTR_POLLEN_GRASS,
|
||||
TMRW_ATTR_POLLEN_TREE,
|
||||
TMRW_ATTR_POLLEN_WEED,
|
||||
TMRW_ATTR_PRECIPITATION,
|
||||
TMRW_ATTR_PRECIPITATION_PROBABILITY,
|
||||
TMRW_ATTR_PRECIPITATION_TYPE,
|
||||
TMRW_ATTR_PRESSURE,
|
||||
TMRW_ATTR_PRESSURE_SURFACE_LEVEL,
|
||||
TMRW_ATTR_SOLAR_GHI,
|
||||
TMRW_ATTR_SULPHUR_DIOXIDE,
|
||||
TMRW_ATTR_TEMPERATURE,
|
||||
TMRW_ATTR_TEMPERATURE_HIGH,
|
||||
TMRW_ATTR_TEMPERATURE_LOW,
|
||||
TMRW_ATTR_UV_HEALTH_CONCERN,
|
||||
TMRW_ATTR_UV_INDEX,
|
||||
TMRW_ATTR_VISIBILITY,
|
||||
TMRW_ATTR_WIND_DIRECTION,
|
||||
TMRW_ATTR_WIND_GUST,
|
||||
TMRW_ATTR_WIND_SPEED,
|
||||
)
|
||||
from .const import ATTRIBUTION, DOMAIN, INTEGRATION_NAME
|
||||
from .coordinator import TomorrowioDataUpdateCoordinator
|
||||
|
||||
PLATFORMS = [SENSOR_DOMAIN, WEATHER_DOMAIN]
|
||||
|
||||
|
||||
@callback
|
||||
def async_get_entries_by_api_key(
|
||||
hass: HomeAssistant, api_key: str, exclude_entry: ConfigEntry | None = None
|
||||
) -> list[ConfigEntry]:
|
||||
"""Get all entries for a given API key."""
|
||||
return [
|
||||
entry
|
||||
for entry in hass.config_entries.async_entries(DOMAIN)
|
||||
if entry.data[CONF_API_KEY] == api_key
|
||||
and (exclude_entry is None or exclude_entry != entry)
|
||||
]
|
||||
|
||||
|
||||
@callback
|
||||
def async_set_update_interval(
|
||||
hass: HomeAssistant, api: TomorrowioV4, exclude_entry: ConfigEntry | None = None
|
||||
) -> timedelta:
|
||||
"""Calculate update_interval."""
|
||||
# We check how many Tomorrow.io configured instances are using the same API key and
|
||||
# calculate interval to not exceed allowed numbers of requests. Divide 90% of
|
||||
# max_requests by the number of API calls because we want a buffer in the
|
||||
# number of API calls left at the end of the day.
|
||||
entries = async_get_entries_by_api_key(hass, api.api_key, exclude_entry)
|
||||
minutes = ceil(
|
||||
(24 * 60 * len(entries) * api.num_api_requests)
|
||||
/ (api.max_requests_per_day * 0.9)
|
||||
)
|
||||
LOGGER.debug(
|
||||
(
|
||||
"Number of config entries: %s\n"
|
||||
"Number of API Requests per call: %s\n"
|
||||
"Max requests per day: %s\n"
|
||||
"Update interval: %s minutes"
|
||||
),
|
||||
len(entries),
|
||||
api.num_api_requests,
|
||||
api.max_requests_per_day,
|
||||
minutes,
|
||||
)
|
||||
return timedelta(minutes=minutes)
|
||||
|
||||
|
||||
async def async_setup_entry(hass: HomeAssistant, entry: ConfigEntry) -> bool:
|
||||
"""Set up Tomorrow.io API from a config entry."""
|
||||
hass.data.setdefault(DOMAIN, {})
|
||||
|
@ -164,166 +59,6 @@ async def async_unload_entry(hass: HomeAssistant, config_entry: ConfigEntry) ->
|
|||
return unload_ok
|
||||
|
||||
|
||||
class TomorrowioDataUpdateCoordinator(DataUpdateCoordinator[dict[str, Any]]): # pylint: disable=hass-enforce-coordinator-module
|
||||
"""Define an object to hold Tomorrow.io data."""
|
||||
|
||||
def __init__(self, hass: HomeAssistant, api: TomorrowioV4) -> None:
|
||||
"""Initialize."""
|
||||
self._api = api
|
||||
self.data = {CURRENT: {}, FORECASTS: {}}
|
||||
self.entry_id_to_location_dict: dict[str, str] = {}
|
||||
self._coordinator_ready: asyncio.Event | None = None
|
||||
|
||||
super().__init__(hass, LOGGER, name=f"{DOMAIN}_{self._api.api_key_masked}")
|
||||
|
||||
def add_entry_to_location_dict(self, entry: ConfigEntry) -> None:
|
||||
"""Add an entry to the location dict."""
|
||||
latitude = entry.data[CONF_LOCATION][CONF_LATITUDE]
|
||||
longitude = entry.data[CONF_LOCATION][CONF_LONGITUDE]
|
||||
self.entry_id_to_location_dict[entry.entry_id] = f"{latitude},{longitude}"
|
||||
|
||||
async def async_setup_entry(self, entry: ConfigEntry) -> None:
|
||||
"""Load config entry into coordinator."""
|
||||
# If we haven't loaded any data yet, register all entries with this API key and
|
||||
# get the initial data for all of them. We do this because another config entry
|
||||
# may start setup before we finish setting the initial data and we don't want
|
||||
# to do multiple refreshes on startup.
|
||||
if self._coordinator_ready is None:
|
||||
LOGGER.debug(
|
||||
"Setting up coordinator for API key %s, loading data for all entries",
|
||||
self._api.api_key_masked,
|
||||
)
|
||||
self._coordinator_ready = asyncio.Event()
|
||||
for entry_ in async_get_entries_by_api_key(self.hass, self._api.api_key):
|
||||
self.add_entry_to_location_dict(entry_)
|
||||
LOGGER.debug(
|
||||
"Loaded %s entries, initiating first refresh",
|
||||
len(self.entry_id_to_location_dict),
|
||||
)
|
||||
await self.async_config_entry_first_refresh()
|
||||
self._coordinator_ready.set()
|
||||
else:
|
||||
# If we have an event, we need to wait for it to be set before we proceed
|
||||
await self._coordinator_ready.wait()
|
||||
# If we're not getting new data because we already know this entry, we
|
||||
# don't need to schedule a refresh
|
||||
if entry.entry_id in self.entry_id_to_location_dict:
|
||||
return
|
||||
LOGGER.debug(
|
||||
(
|
||||
"Adding new entry to existing coordinator for API key %s, doing a "
|
||||
"partial refresh"
|
||||
),
|
||||
self._api.api_key_masked,
|
||||
)
|
||||
# We need a refresh, but it's going to be a partial refresh so we can
|
||||
# minimize repeat API calls
|
||||
self.add_entry_to_location_dict(entry)
|
||||
await self.async_refresh()
|
||||
|
||||
self.update_interval = async_set_update_interval(self.hass, self._api)
|
||||
self._async_unsub_refresh()
|
||||
if self._listeners:
|
||||
self._schedule_refresh()
|
||||
|
||||
async def async_unload_entry(self, entry: ConfigEntry) -> bool | None:
|
||||
"""Unload a config entry from coordinator.
|
||||
|
||||
Returns whether coordinator can be removed as well because there are no
|
||||
config entries tied to it anymore.
|
||||
"""
|
||||
self.entry_id_to_location_dict.pop(entry.entry_id)
|
||||
self.update_interval = async_set_update_interval(self.hass, self._api, entry)
|
||||
return not self.entry_id_to_location_dict
|
||||
|
||||
async def _async_update_data(self) -> dict[str, Any]:
|
||||
"""Update data via library."""
|
||||
data: dict[str, Any] = {}
|
||||
# If we are refreshing because of a new config entry that's not already in our
|
||||
# data, we do a partial refresh to avoid wasted API calls.
|
||||
if self.data and any(
|
||||
entry_id not in self.data for entry_id in self.entry_id_to_location_dict
|
||||
):
|
||||
data = self.data
|
||||
|
||||
LOGGER.debug(
|
||||
"Fetching data for %s entries",
|
||||
len(set(self.entry_id_to_location_dict) - set(data)),
|
||||
)
|
||||
for entry_id, location in self.entry_id_to_location_dict.items():
|
||||
if entry_id in data:
|
||||
continue
|
||||
entry = self.hass.config_entries.async_get_entry(entry_id)
|
||||
assert entry
|
||||
try:
|
||||
data[entry_id] = await self._api.realtime_and_all_forecasts(
|
||||
[
|
||||
# Weather
|
||||
TMRW_ATTR_TEMPERATURE,
|
||||
TMRW_ATTR_HUMIDITY,
|
||||
TMRW_ATTR_PRESSURE,
|
||||
TMRW_ATTR_WIND_SPEED,
|
||||
TMRW_ATTR_WIND_DIRECTION,
|
||||
TMRW_ATTR_CONDITION,
|
||||
TMRW_ATTR_VISIBILITY,
|
||||
TMRW_ATTR_OZONE,
|
||||
TMRW_ATTR_WIND_GUST,
|
||||
TMRW_ATTR_CLOUD_COVER,
|
||||
TMRW_ATTR_PRECIPITATION_TYPE,
|
||||
# Sensors
|
||||
TMRW_ATTR_CARBON_MONOXIDE,
|
||||
TMRW_ATTR_CHINA_AQI,
|
||||
TMRW_ATTR_CHINA_HEALTH_CONCERN,
|
||||
TMRW_ATTR_CHINA_PRIMARY_POLLUTANT,
|
||||
TMRW_ATTR_CLOUD_BASE,
|
||||
TMRW_ATTR_CLOUD_CEILING,
|
||||
TMRW_ATTR_CLOUD_COVER,
|
||||
TMRW_ATTR_DEW_POINT,
|
||||
TMRW_ATTR_EPA_AQI,
|
||||
TMRW_ATTR_EPA_HEALTH_CONCERN,
|
||||
TMRW_ATTR_EPA_PRIMARY_POLLUTANT,
|
||||
TMRW_ATTR_FEELS_LIKE,
|
||||
TMRW_ATTR_FIRE_INDEX,
|
||||
TMRW_ATTR_NITROGEN_DIOXIDE,
|
||||
TMRW_ATTR_OZONE,
|
||||
TMRW_ATTR_PARTICULATE_MATTER_10,
|
||||
TMRW_ATTR_PARTICULATE_MATTER_25,
|
||||
TMRW_ATTR_POLLEN_GRASS,
|
||||
TMRW_ATTR_POLLEN_TREE,
|
||||
TMRW_ATTR_POLLEN_WEED,
|
||||
TMRW_ATTR_PRECIPITATION_TYPE,
|
||||
TMRW_ATTR_PRESSURE_SURFACE_LEVEL,
|
||||
TMRW_ATTR_SOLAR_GHI,
|
||||
TMRW_ATTR_SULPHUR_DIOXIDE,
|
||||
TMRW_ATTR_UV_INDEX,
|
||||
TMRW_ATTR_UV_HEALTH_CONCERN,
|
||||
TMRW_ATTR_WIND_GUST,
|
||||
],
|
||||
[
|
||||
TMRW_ATTR_TEMPERATURE_LOW,
|
||||
TMRW_ATTR_TEMPERATURE_HIGH,
|
||||
TMRW_ATTR_DEW_POINT,
|
||||
TMRW_ATTR_HUMIDITY,
|
||||
TMRW_ATTR_WIND_SPEED,
|
||||
TMRW_ATTR_WIND_DIRECTION,
|
||||
TMRW_ATTR_CONDITION,
|
||||
TMRW_ATTR_PRECIPITATION,
|
||||
TMRW_ATTR_PRECIPITATION_PROBABILITY,
|
||||
],
|
||||
nowcast_timestep=entry.options[CONF_TIMESTEP],
|
||||
location=location,
|
||||
)
|
||||
except (
|
||||
CantConnectException,
|
||||
InvalidAPIKeyException,
|
||||
RateLimitedException,
|
||||
UnknownException,
|
||||
) as error:
|
||||
raise UpdateFailed from error
|
||||
|
||||
return data
|
||||
|
||||
|
||||
class TomorrowioEntity(CoordinatorEntity[TomorrowioDataUpdateCoordinator]):
|
||||
"""Base Tomorrow.io Entity."""
|
||||
|
||||
|
|
273
homeassistant/components/tomorrowio/coordinator.py
Normal file
273
homeassistant/components/tomorrowio/coordinator.py
Normal file
|
@ -0,0 +1,273 @@
|
|||
"""The Tomorrow.io integration."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import asyncio
|
||||
from datetime import timedelta
|
||||
from math import ceil
|
||||
from typing import Any
|
||||
|
||||
from pytomorrowio import TomorrowioV4
|
||||
from pytomorrowio.const import CURRENT, FORECASTS
|
||||
from pytomorrowio.exceptions import (
|
||||
CantConnectException,
|
||||
InvalidAPIKeyException,
|
||||
RateLimitedException,
|
||||
UnknownException,
|
||||
)
|
||||
|
||||
from homeassistant.config_entries import ConfigEntry
|
||||
from homeassistant.const import (
|
||||
CONF_API_KEY,
|
||||
CONF_LATITUDE,
|
||||
CONF_LOCATION,
|
||||
CONF_LONGITUDE,
|
||||
)
|
||||
from homeassistant.core import HomeAssistant, callback
|
||||
from homeassistant.helpers.update_coordinator import DataUpdateCoordinator, UpdateFailed
|
||||
|
||||
from .const import (
|
||||
CONF_TIMESTEP,
|
||||
DOMAIN,
|
||||
LOGGER,
|
||||
TMRW_ATTR_CARBON_MONOXIDE,
|
||||
TMRW_ATTR_CHINA_AQI,
|
||||
TMRW_ATTR_CHINA_HEALTH_CONCERN,
|
||||
TMRW_ATTR_CHINA_PRIMARY_POLLUTANT,
|
||||
TMRW_ATTR_CLOUD_BASE,
|
||||
TMRW_ATTR_CLOUD_CEILING,
|
||||
TMRW_ATTR_CLOUD_COVER,
|
||||
TMRW_ATTR_CONDITION,
|
||||
TMRW_ATTR_DEW_POINT,
|
||||
TMRW_ATTR_EPA_AQI,
|
||||
TMRW_ATTR_EPA_HEALTH_CONCERN,
|
||||
TMRW_ATTR_EPA_PRIMARY_POLLUTANT,
|
||||
TMRW_ATTR_FEELS_LIKE,
|
||||
TMRW_ATTR_FIRE_INDEX,
|
||||
TMRW_ATTR_HUMIDITY,
|
||||
TMRW_ATTR_NITROGEN_DIOXIDE,
|
||||
TMRW_ATTR_OZONE,
|
||||
TMRW_ATTR_PARTICULATE_MATTER_10,
|
||||
TMRW_ATTR_PARTICULATE_MATTER_25,
|
||||
TMRW_ATTR_POLLEN_GRASS,
|
||||
TMRW_ATTR_POLLEN_TREE,
|
||||
TMRW_ATTR_POLLEN_WEED,
|
||||
TMRW_ATTR_PRECIPITATION,
|
||||
TMRW_ATTR_PRECIPITATION_PROBABILITY,
|
||||
TMRW_ATTR_PRECIPITATION_TYPE,
|
||||
TMRW_ATTR_PRESSURE,
|
||||
TMRW_ATTR_PRESSURE_SURFACE_LEVEL,
|
||||
TMRW_ATTR_SOLAR_GHI,
|
||||
TMRW_ATTR_SULPHUR_DIOXIDE,
|
||||
TMRW_ATTR_TEMPERATURE,
|
||||
TMRW_ATTR_TEMPERATURE_HIGH,
|
||||
TMRW_ATTR_TEMPERATURE_LOW,
|
||||
TMRW_ATTR_UV_HEALTH_CONCERN,
|
||||
TMRW_ATTR_UV_INDEX,
|
||||
TMRW_ATTR_VISIBILITY,
|
||||
TMRW_ATTR_WIND_DIRECTION,
|
||||
TMRW_ATTR_WIND_GUST,
|
||||
TMRW_ATTR_WIND_SPEED,
|
||||
)
|
||||
|
||||
|
||||
@callback
|
||||
def async_get_entries_by_api_key(
|
||||
hass: HomeAssistant, api_key: str, exclude_entry: ConfigEntry | None = None
|
||||
) -> list[ConfigEntry]:
|
||||
"""Get all entries for a given API key."""
|
||||
return [
|
||||
entry
|
||||
for entry in hass.config_entries.async_entries(DOMAIN)
|
||||
if entry.data[CONF_API_KEY] == api_key
|
||||
and (exclude_entry is None or exclude_entry != entry)
|
||||
]
|
||||
|
||||
|
||||
@callback
|
||||
def async_set_update_interval(
|
||||
hass: HomeAssistant, api: TomorrowioV4, exclude_entry: ConfigEntry | None = None
|
||||
) -> timedelta:
|
||||
"""Calculate update_interval."""
|
||||
# We check how many Tomorrow.io configured instances are using the same API key and
|
||||
# calculate interval to not exceed allowed numbers of requests. Divide 90% of
|
||||
# max_requests by the number of API calls because we want a buffer in the
|
||||
# number of API calls left at the end of the day.
|
||||
entries = async_get_entries_by_api_key(hass, api.api_key, exclude_entry)
|
||||
minutes = ceil(
|
||||
(24 * 60 * len(entries) * api.num_api_requests)
|
||||
/ (api.max_requests_per_day * 0.9)
|
||||
)
|
||||
LOGGER.debug(
|
||||
(
|
||||
"Number of config entries: %s\n"
|
||||
"Number of API Requests per call: %s\n"
|
||||
"Max requests per day: %s\n"
|
||||
"Update interval: %s minutes"
|
||||
),
|
||||
len(entries),
|
||||
api.num_api_requests,
|
||||
api.max_requests_per_day,
|
||||
minutes,
|
||||
)
|
||||
return timedelta(minutes=minutes)
|
||||
|
||||
|
||||
class TomorrowioDataUpdateCoordinator(DataUpdateCoordinator[dict[str, Any]]):
|
||||
"""Define an object to hold Tomorrow.io data."""
|
||||
|
||||
def __init__(self, hass: HomeAssistant, api: TomorrowioV4) -> None:
|
||||
"""Initialize."""
|
||||
self._api = api
|
||||
self.data = {CURRENT: {}, FORECASTS: {}}
|
||||
self.entry_id_to_location_dict: dict[str, str] = {}
|
||||
self._coordinator_ready: asyncio.Event | None = None
|
||||
|
||||
super().__init__(hass, LOGGER, name=f"{DOMAIN}_{self._api.api_key_masked}")
|
||||
|
||||
def add_entry_to_location_dict(self, entry: ConfigEntry) -> None:
|
||||
"""Add an entry to the location dict."""
|
||||
latitude = entry.data[CONF_LOCATION][CONF_LATITUDE]
|
||||
longitude = entry.data[CONF_LOCATION][CONF_LONGITUDE]
|
||||
self.entry_id_to_location_dict[entry.entry_id] = f"{latitude},{longitude}"
|
||||
|
||||
async def async_setup_entry(self, entry: ConfigEntry) -> None:
|
||||
"""Load config entry into coordinator."""
|
||||
# If we haven't loaded any data yet, register all entries with this API key and
|
||||
# get the initial data for all of them. We do this because another config entry
|
||||
# may start setup before we finish setting the initial data and we don't want
|
||||
# to do multiple refreshes on startup.
|
||||
if self._coordinator_ready is None:
|
||||
LOGGER.debug(
|
||||
"Setting up coordinator for API key %s, loading data for all entries",
|
||||
self._api.api_key_masked,
|
||||
)
|
||||
self._coordinator_ready = asyncio.Event()
|
||||
for entry_ in async_get_entries_by_api_key(self.hass, self._api.api_key):
|
||||
self.add_entry_to_location_dict(entry_)
|
||||
LOGGER.debug(
|
||||
"Loaded %s entries, initiating first refresh",
|
||||
len(self.entry_id_to_location_dict),
|
||||
)
|
||||
await self.async_config_entry_first_refresh()
|
||||
self._coordinator_ready.set()
|
||||
else:
|
||||
# If we have an event, we need to wait for it to be set before we proceed
|
||||
await self._coordinator_ready.wait()
|
||||
# If we're not getting new data because we already know this entry, we
|
||||
# don't need to schedule a refresh
|
||||
if entry.entry_id in self.entry_id_to_location_dict:
|
||||
return
|
||||
LOGGER.debug(
|
||||
(
|
||||
"Adding new entry to existing coordinator for API key %s, doing a "
|
||||
"partial refresh"
|
||||
),
|
||||
self._api.api_key_masked,
|
||||
)
|
||||
# We need a refresh, but it's going to be a partial refresh so we can
|
||||
# minimize repeat API calls
|
||||
self.add_entry_to_location_dict(entry)
|
||||
await self.async_refresh()
|
||||
|
||||
self.update_interval = async_set_update_interval(self.hass, self._api)
|
||||
self._async_unsub_refresh()
|
||||
if self._listeners:
|
||||
self._schedule_refresh()
|
||||
|
||||
async def async_unload_entry(self, entry: ConfigEntry) -> bool | None:
|
||||
"""Unload a config entry from coordinator.
|
||||
|
||||
Returns whether coordinator can be removed as well because there are no
|
||||
config entries tied to it anymore.
|
||||
"""
|
||||
self.entry_id_to_location_dict.pop(entry.entry_id)
|
||||
self.update_interval = async_set_update_interval(self.hass, self._api, entry)
|
||||
return not self.entry_id_to_location_dict
|
||||
|
||||
async def _async_update_data(self) -> dict[str, Any]:
|
||||
"""Update data via library."""
|
||||
data: dict[str, Any] = {}
|
||||
# If we are refreshing because of a new config entry that's not already in our
|
||||
# data, we do a partial refresh to avoid wasted API calls.
|
||||
if self.data and any(
|
||||
entry_id not in self.data for entry_id in self.entry_id_to_location_dict
|
||||
):
|
||||
data = self.data
|
||||
|
||||
LOGGER.debug(
|
||||
"Fetching data for %s entries",
|
||||
len(set(self.entry_id_to_location_dict) - set(data)),
|
||||
)
|
||||
for entry_id, location in self.entry_id_to_location_dict.items():
|
||||
if entry_id in data:
|
||||
continue
|
||||
entry = self.hass.config_entries.async_get_entry(entry_id)
|
||||
assert entry
|
||||
try:
|
||||
data[entry_id] = await self._api.realtime_and_all_forecasts(
|
||||
[
|
||||
# Weather
|
||||
TMRW_ATTR_TEMPERATURE,
|
||||
TMRW_ATTR_HUMIDITY,
|
||||
TMRW_ATTR_PRESSURE,
|
||||
TMRW_ATTR_WIND_SPEED,
|
||||
TMRW_ATTR_WIND_DIRECTION,
|
||||
TMRW_ATTR_CONDITION,
|
||||
TMRW_ATTR_VISIBILITY,
|
||||
TMRW_ATTR_OZONE,
|
||||
TMRW_ATTR_WIND_GUST,
|
||||
TMRW_ATTR_CLOUD_COVER,
|
||||
TMRW_ATTR_PRECIPITATION_TYPE,
|
||||
# Sensors
|
||||
TMRW_ATTR_CARBON_MONOXIDE,
|
||||
TMRW_ATTR_CHINA_AQI,
|
||||
TMRW_ATTR_CHINA_HEALTH_CONCERN,
|
||||
TMRW_ATTR_CHINA_PRIMARY_POLLUTANT,
|
||||
TMRW_ATTR_CLOUD_BASE,
|
||||
TMRW_ATTR_CLOUD_CEILING,
|
||||
TMRW_ATTR_CLOUD_COVER,
|
||||
TMRW_ATTR_DEW_POINT,
|
||||
TMRW_ATTR_EPA_AQI,
|
||||
TMRW_ATTR_EPA_HEALTH_CONCERN,
|
||||
TMRW_ATTR_EPA_PRIMARY_POLLUTANT,
|
||||
TMRW_ATTR_FEELS_LIKE,
|
||||
TMRW_ATTR_FIRE_INDEX,
|
||||
TMRW_ATTR_NITROGEN_DIOXIDE,
|
||||
TMRW_ATTR_OZONE,
|
||||
TMRW_ATTR_PARTICULATE_MATTER_10,
|
||||
TMRW_ATTR_PARTICULATE_MATTER_25,
|
||||
TMRW_ATTR_POLLEN_GRASS,
|
||||
TMRW_ATTR_POLLEN_TREE,
|
||||
TMRW_ATTR_POLLEN_WEED,
|
||||
TMRW_ATTR_PRECIPITATION_TYPE,
|
||||
TMRW_ATTR_PRESSURE_SURFACE_LEVEL,
|
||||
TMRW_ATTR_SOLAR_GHI,
|
||||
TMRW_ATTR_SULPHUR_DIOXIDE,
|
||||
TMRW_ATTR_UV_INDEX,
|
||||
TMRW_ATTR_UV_HEALTH_CONCERN,
|
||||
TMRW_ATTR_WIND_GUST,
|
||||
],
|
||||
[
|
||||
TMRW_ATTR_TEMPERATURE_LOW,
|
||||
TMRW_ATTR_TEMPERATURE_HIGH,
|
||||
TMRW_ATTR_DEW_POINT,
|
||||
TMRW_ATTR_HUMIDITY,
|
||||
TMRW_ATTR_WIND_SPEED,
|
||||
TMRW_ATTR_WIND_DIRECTION,
|
||||
TMRW_ATTR_CONDITION,
|
||||
TMRW_ATTR_PRECIPITATION,
|
||||
TMRW_ATTR_PRECIPITATION_PROBABILITY,
|
||||
],
|
||||
nowcast_timestep=entry.options[CONF_TIMESTEP],
|
||||
location=location,
|
||||
)
|
||||
except (
|
||||
CantConnectException,
|
||||
InvalidAPIKeyException,
|
||||
RateLimitedException,
|
||||
UnknownException,
|
||||
) as error:
|
||||
raise UpdateFailed from error
|
||||
|
||||
return data
|
|
@ -38,7 +38,7 @@ from homeassistant.helpers.entity_platform import AddEntitiesCallback
|
|||
from homeassistant.util.unit_conversion import DistanceConverter, SpeedConverter
|
||||
from homeassistant.util.unit_system import US_CUSTOMARY_SYSTEM
|
||||
|
||||
from . import TomorrowioDataUpdateCoordinator, TomorrowioEntity
|
||||
from . import TomorrowioEntity
|
||||
from .const import (
|
||||
DOMAIN,
|
||||
TMRW_ATTR_CARBON_MONOXIDE,
|
||||
|
@ -69,6 +69,7 @@ from .const import (
|
|||
TMRW_ATTR_UV_INDEX,
|
||||
TMRW_ATTR_WIND_GUST,
|
||||
)
|
||||
from .coordinator import TomorrowioDataUpdateCoordinator
|
||||
|
||||
|
||||
@dataclass(frozen=True)
|
||||
|
|
|
@ -37,7 +37,7 @@ from homeassistant.helpers.entity_platform import AddEntitiesCallback
|
|||
from homeassistant.helpers.sun import is_up
|
||||
from homeassistant.util import dt as dt_util
|
||||
|
||||
from . import TomorrowioDataUpdateCoordinator, TomorrowioEntity
|
||||
from . import TomorrowioEntity
|
||||
from .const import (
|
||||
CLEAR_CONDITIONS,
|
||||
CONDITIONS,
|
||||
|
@ -60,6 +60,7 @@ from .const import (
|
|||
TMRW_ATTR_WIND_DIRECTION,
|
||||
TMRW_ATTR_WIND_SPEED,
|
||||
)
|
||||
from .coordinator import TomorrowioDataUpdateCoordinator
|
||||
|
||||
|
||||
async def async_setup_entry(
|
||||
|
|
Loading…
Add table
Reference in a new issue