Support for overriding dish IP and port
Probably not terribly useful unless someone needs to tunnel through a different network to get to their dish, but it makes testing the dish unreachable case a lot easier. This was complicated a bit by the fact that a channel (and therefor the dish IP and port) is needed to get the list of alert types via reflection due to prior changes. This exposed some issues with the error message for dish unreachable, so fixed those.
This commit is contained in:
parent
1659133168
commit
a4bf2d1625
6 changed files with 122 additions and 86 deletions
|
@ -43,6 +43,10 @@ def create_arg_parser(output_description, bulk_history=True):
|
||||||
parser.bulk_history = bulk_history
|
parser.bulk_history = bulk_history
|
||||||
|
|
||||||
group = parser.add_argument_group(title="General options")
|
group = parser.add_argument_group(title="General options")
|
||||||
|
group.add_argument("-g",
|
||||||
|
"--target",
|
||||||
|
help="host:port of dish to query, default is the standard IP address "
|
||||||
|
"and port (192.168.100.1:9200)")
|
||||||
group.add_argument("-h", "--help", action="help", help="Be helpful")
|
group.add_argument("-h", "--help", action="help", help="Be helpful")
|
||||||
group.add_argument("-t",
|
group.add_argument("-t",
|
||||||
"--loop-interval",
|
"--loop-interval",
|
||||||
|
@ -129,14 +133,14 @@ def conn_error(opts, msg, *args):
|
||||||
|
|
||||||
class GlobalState:
|
class GlobalState:
|
||||||
"""A class for keeping state across loop iterations."""
|
"""A class for keeping state across loop iterations."""
|
||||||
def __init__(self):
|
def __init__(self, target=None):
|
||||||
# counter for bulk_history:
|
# counter for bulk_history:
|
||||||
self.counter = None
|
self.counter = None
|
||||||
# counter for history stats:
|
# counter for history stats:
|
||||||
self.counter_stats = None
|
self.counter_stats = None
|
||||||
self.timestamp = None
|
self.timestamp = None
|
||||||
self.dish_id = None
|
self.dish_id = None
|
||||||
self.context = starlink_grpc.ChannelContext()
|
self.context = starlink_grpc.ChannelContext(target=target)
|
||||||
|
|
||||||
def shutdown(self):
|
def shutdown(self):
|
||||||
self.context.close()
|
self.context.close()
|
||||||
|
@ -183,12 +187,12 @@ def get_data(opts, gstate, add_item, add_sequence, add_bulk=None):
|
||||||
if "status" in opts.mode:
|
if "status" in opts.mode:
|
||||||
if opts.need_id and gstate.dish_id is None:
|
if opts.need_id and gstate.dish_id is None:
|
||||||
conn_error(opts, "Dish unreachable and ID unknown, so not recording state")
|
conn_error(opts, "Dish unreachable and ID unknown, so not recording state")
|
||||||
else:
|
return 1
|
||||||
if opts.verbose:
|
if opts.verbose:
|
||||||
print("Dish unreachable")
|
print("Dish unreachable")
|
||||||
if "status" in opts.mode:
|
|
||||||
add_item("state", "DISH_UNREACHABLE", "status")
|
add_item("state", "DISH_UNREACHABLE", "status")
|
||||||
return 0
|
return 0
|
||||||
|
conn_error(opts, "Failure getting status: %s", str(e))
|
||||||
return 1
|
return 1
|
||||||
if opts.need_id:
|
if opts.need_id:
|
||||||
gstate.dish_id = status_data["id"]
|
gstate.dish_id = status_data["id"]
|
||||||
|
|
|
@ -278,7 +278,7 @@ def main():
|
||||||
|
|
||||||
logging.basicConfig(format="%(levelname)s: %(message)s")
|
logging.basicConfig(format="%(levelname)s: %(message)s")
|
||||||
|
|
||||||
gstate = dish_common.GlobalState()
|
gstate = dish_common.GlobalState(target=opts.target)
|
||||||
gstate.points = []
|
gstate.points = []
|
||||||
gstate.deferred_points = []
|
gstate.deferred_points = []
|
||||||
gstate.timebase_synced = opts.skip_query
|
gstate.timebase_synced = opts.skip_query
|
||||||
|
|
|
@ -111,7 +111,7 @@ def main():
|
||||||
|
|
||||||
logging.basicConfig(format="%(levelname)s: %(message)s")
|
logging.basicConfig(format="%(levelname)s: %(message)s")
|
||||||
|
|
||||||
gstate = dish_common.GlobalState()
|
gstate = dish_common.GlobalState(target=opts.target)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
next_loop = time.monotonic()
|
next_loop = time.monotonic()
|
||||||
|
|
|
@ -148,34 +148,39 @@ def loop_body(opts, gstate):
|
||||||
return rc
|
return rc
|
||||||
|
|
||||||
|
|
||||||
def ensure_schema(opts, conn):
|
def ensure_schema(opts, conn, context):
|
||||||
cur = conn.cursor()
|
cur = conn.cursor()
|
||||||
cur.execute("PRAGMA user_version")
|
cur.execute("PRAGMA user_version")
|
||||||
version = cur.fetchone()
|
version = cur.fetchone()
|
||||||
if version and version[0] == SCHEMA_VERSION:
|
if version and version[0] == SCHEMA_VERSION:
|
||||||
cur.close()
|
cur.close()
|
||||||
return
|
return 0
|
||||||
|
|
||||||
|
try:
|
||||||
if not version or not version[0]:
|
if not version or not version[0]:
|
||||||
if opts.verbose:
|
if opts.verbose:
|
||||||
print("Initializing new database")
|
print("Initializing new database")
|
||||||
create_tables(conn, "")
|
create_tables(conn, context, "")
|
||||||
elif version[0] > SCHEMA_VERSION and not opts.force:
|
elif version[0] > SCHEMA_VERSION and not opts.force:
|
||||||
logging.error("Cowardly refusing to downgrade from schema version %s", version[0])
|
logging.error("Cowardly refusing to downgrade from schema version %s", version[0])
|
||||||
raise Terminated
|
return 1
|
||||||
else:
|
else:
|
||||||
print("Converting from schema version:", version[0])
|
print("Converting from schema version:", version[0])
|
||||||
convert_tables(conn)
|
convert_tables(conn, context)
|
||||||
|
|
||||||
cur.execute("PRAGMA user_version={0}".format(SCHEMA_VERSION))
|
cur.execute("PRAGMA user_version={0}".format(SCHEMA_VERSION))
|
||||||
cur.close()
|
|
||||||
conn.commit()
|
conn.commit()
|
||||||
|
return 0
|
||||||
|
except starlink_grpc.GrpcError as e:
|
||||||
|
dish_common.conn_error(opts, "Failure reflecting status fields: %s", str(e))
|
||||||
|
return 1
|
||||||
|
finally:
|
||||||
|
cur.close()
|
||||||
|
|
||||||
|
|
||||||
def create_tables(conn, suffix):
|
def create_tables(conn, context, suffix):
|
||||||
tables = {}
|
tables = {}
|
||||||
name_groups = starlink_grpc.status_field_names()
|
name_groups = starlink_grpc.status_field_names(context=context)
|
||||||
type_groups = starlink_grpc.status_field_types()
|
type_groups = starlink_grpc.status_field_types(context=context)
|
||||||
tables["status"] = zip(name_groups, type_groups)
|
tables["status"] = zip(name_groups, type_groups)
|
||||||
|
|
||||||
name_groups = starlink_grpc.history_stats_field_names()
|
name_groups = starlink_grpc.history_stats_field_names()
|
||||||
|
@ -220,8 +225,8 @@ def create_tables(conn, suffix):
|
||||||
return column_info
|
return column_info
|
||||||
|
|
||||||
|
|
||||||
def convert_tables(conn):
|
def convert_tables(conn, context):
|
||||||
new_column_info = create_tables(conn, "_new")
|
new_column_info = create_tables(conn, context, "_new")
|
||||||
conn.row_factory = sqlite3.Row
|
conn.row_factory = sqlite3.Row
|
||||||
old_cur = conn.cursor()
|
old_cur = conn.cursor()
|
||||||
new_cur = conn.cursor()
|
new_cur = conn.cursor()
|
||||||
|
@ -245,7 +250,7 @@ def main():
|
||||||
|
|
||||||
logging.basicConfig(format="%(levelname)s: %(message)s")
|
logging.basicConfig(format="%(levelname)s: %(message)s")
|
||||||
|
|
||||||
gstate = dish_common.GlobalState()
|
gstate = dish_common.GlobalState(target=opts.target)
|
||||||
gstate.points = []
|
gstate.points = []
|
||||||
gstate.deferred_points = []
|
gstate.deferred_points = []
|
||||||
|
|
||||||
|
@ -254,7 +259,9 @@ def main():
|
||||||
|
|
||||||
rc = 0
|
rc = 0
|
||||||
try:
|
try:
|
||||||
ensure_schema(opts, gstate.sql_conn)
|
rc = ensure_schema(opts, gstate.sql_conn, gstate.context)
|
||||||
|
if rc:
|
||||||
|
sys.exit(rc)
|
||||||
next_loop = time.monotonic()
|
next_loop = time.monotonic()
|
||||||
while True:
|
while True:
|
||||||
rc = loop_body(opts, gstate)
|
rc = loop_body(opts, gstate)
|
||||||
|
|
|
@ -85,13 +85,18 @@ def print_header(opts):
|
||||||
header.append(name)
|
header.append(name)
|
||||||
|
|
||||||
if opts.satus_mode:
|
if opts.satus_mode:
|
||||||
status_names, obstruct_names, alert_names = starlink_grpc.status_field_names()
|
context = starlink_grpc.ChannelContext(target=opts.target)
|
||||||
|
try:
|
||||||
|
name_groups = starlink_grpc.status_field_names(context=context)
|
||||||
|
except starlink_grpc.GrpcError as e:
|
||||||
|
dish_common.conn_error(opts, "Failure reflecting status field names: %s", str(e))
|
||||||
|
return 1
|
||||||
if "status" in opts.mode:
|
if "status" in opts.mode:
|
||||||
header_add(status_names)
|
header_add(name_groups[0])
|
||||||
if "obstruction_detail" in opts.mode:
|
if "obstruction_detail" in opts.mode:
|
||||||
header_add(obstruct_names)
|
header_add(name_groups[1])
|
||||||
if "alert_detail" in opts.mode:
|
if "alert_detail" in opts.mode:
|
||||||
header_add(alert_names)
|
header_add(name_groups[2])
|
||||||
|
|
||||||
if opts.bulk_mode:
|
if opts.bulk_mode:
|
||||||
general, bulk = starlink_grpc.history_bulk_field_names()
|
general, bulk = starlink_grpc.history_bulk_field_names()
|
||||||
|
@ -114,6 +119,7 @@ def print_header(opts):
|
||||||
header_add(usage)
|
header_add(usage)
|
||||||
|
|
||||||
print(",".join(header))
|
print(",".join(header))
|
||||||
|
return 0
|
||||||
|
|
||||||
|
|
||||||
def loop_body(opts, gstate):
|
def loop_body(opts, gstate):
|
||||||
|
@ -180,10 +186,10 @@ def main():
|
||||||
logging.basicConfig(format="%(levelname)s: %(message)s")
|
logging.basicConfig(format="%(levelname)s: %(message)s")
|
||||||
|
|
||||||
if opts.print_header:
|
if opts.print_header:
|
||||||
print_header(opts)
|
rc = print_header(opts)
|
||||||
sys.exit(0)
|
sys.exit(rc)
|
||||||
|
|
||||||
gstate = dish_common.GlobalState()
|
gstate = dish_common.GlobalState(target=opts.target)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
next_loop = time.monotonic()
|
next_loop = time.monotonic()
|
||||||
|
|
|
@ -345,9 +345,9 @@ class ChannelContext:
|
||||||
`close()` should be called on the object when it is no longer
|
`close()` should be called on the object when it is no longer
|
||||||
in use.
|
in use.
|
||||||
"""
|
"""
|
||||||
def __init__(self, target="192.168.100.1:9200"):
|
def __init__(self, target=None):
|
||||||
self.channel = None
|
self.channel = None
|
||||||
self.target = target
|
self.target = "192.168.100.1:9200" if target is None else target
|
||||||
|
|
||||||
def get_channel(self):
|
def get_channel(self):
|
||||||
reused = True
|
reused = True
|
||||||
|
@ -362,19 +362,53 @@ class ChannelContext:
|
||||||
self.channel = None
|
self.channel = None
|
||||||
|
|
||||||
|
|
||||||
def status_field_names():
|
def call_with_channel(function, *args, context=None, **kwargs):
|
||||||
|
"""Call a function with a channel object.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
function: Function to call with channel as first arg.
|
||||||
|
args: Additional args to pass to function
|
||||||
|
context (ChannelContext): Optionally provide a channel for (re)use.
|
||||||
|
If not set, a new default channel will be used and then closed.
|
||||||
|
kwargs: Additional keyword args to pass to function.
|
||||||
|
"""
|
||||||
|
if context is None:
|
||||||
|
with grpc.insecure_channel("192.168.100.1:9200") as channel:
|
||||||
|
return function(channel, *args, **kwargs)
|
||||||
|
|
||||||
|
while True:
|
||||||
|
channel, reused = context.get_channel()
|
||||||
|
try:
|
||||||
|
return function(channel, *args, **kwargs)
|
||||||
|
except grpc.RpcError:
|
||||||
|
context.close()
|
||||||
|
if not reused:
|
||||||
|
raise
|
||||||
|
|
||||||
|
|
||||||
|
def status_field_names(context=None):
|
||||||
"""Return the field names of the status data.
|
"""Return the field names of the status data.
|
||||||
|
|
||||||
Note:
|
Note:
|
||||||
See module level docs regarding brackets in field names.
|
See module level docs regarding brackets in field names.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
context (ChannelContext): Optionally provide a channel for (re)use
|
||||||
|
with reflection service.
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
A tuple with 3 lists, with status data field names, alert detail field
|
A tuple with 3 lists, with status data field names, alert detail field
|
||||||
names, and obstruction detail field names, in that order.
|
names, and obstruction detail field names, in that order.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
GrpcError: No user terminal is currently available to resolve imports
|
||||||
|
via reflection.
|
||||||
"""
|
"""
|
||||||
if imports_pending:
|
if imports_pending:
|
||||||
with grpc.insecure_channel("192.168.100.1:9200") as channel:
|
try:
|
||||||
resolve_imports(channel)
|
call_with_channel(resolve_imports, context=context)
|
||||||
|
except grpc.RpcError as e:
|
||||||
|
raise GrpcError(e)
|
||||||
alert_names = []
|
alert_names = []
|
||||||
for field in dish_pb2.DishAlerts.DESCRIPTOR.fields:
|
for field in dish_pb2.DishAlerts.DESCRIPTOR.fields:
|
||||||
alert_names.append("alert_" + field.name)
|
alert_names.append("alert_" + field.name)
|
||||||
|
@ -402,19 +436,29 @@ def status_field_names():
|
||||||
], alert_names
|
], alert_names
|
||||||
|
|
||||||
|
|
||||||
def status_field_types():
|
def status_field_types(context=None):
|
||||||
"""Return the field types of the status data.
|
"""Return the field types of the status data.
|
||||||
|
|
||||||
Return the type classes for each field. For sequence types, the type of
|
Return the type classes for each field. For sequence types, the type of
|
||||||
element in the sequence is returned, not the type of the sequence.
|
element in the sequence is returned, not the type of the sequence.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
context (ChannelContext): Optionally provide a channel for (re)use
|
||||||
|
with reflection service.
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
A tuple with 3 lists, with status data field types, alert detail field
|
A tuple with 3 lists, with status data field types, alert detail field
|
||||||
types, and obstruction detail field types, in that order.
|
types, and obstruction detail field types, in that order.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
GrpcError: No user terminal is currently available to resolve imports
|
||||||
|
via reflection.
|
||||||
"""
|
"""
|
||||||
if imports_pending:
|
if imports_pending:
|
||||||
with grpc.insecure_channel("192.168.100.1:9200") as channel:
|
try:
|
||||||
resolve_imports(channel)
|
call_with_channel(resolve_imports, context=context)
|
||||||
|
except grpc.RpcError as e:
|
||||||
|
raise GrpcError(e)
|
||||||
return [
|
return [
|
||||||
str, # id
|
str, # id
|
||||||
str, # hardware_version
|
str, # hardware_version
|
||||||
|
@ -450,26 +494,14 @@ def get_status(context=None):
|
||||||
Raises:
|
Raises:
|
||||||
grpc.RpcError: Communication or service error.
|
grpc.RpcError: Communication or service error.
|
||||||
"""
|
"""
|
||||||
if context is None:
|
def grpc_call(channel):
|
||||||
with grpc.insecure_channel("192.168.100.1:9200") as channel:
|
|
||||||
if imports_pending:
|
if imports_pending:
|
||||||
resolve_imports(channel)
|
resolve_imports(channel)
|
||||||
stub = device_pb2_grpc.DeviceStub(channel)
|
stub = device_pb2_grpc.DeviceStub(channel)
|
||||||
response = stub.Handle(device_pb2.Request(get_status={}))
|
response = stub.Handle(device_pb2.Request(get_status={}))
|
||||||
return response.dish_get_status
|
return response.dish_get_status
|
||||||
|
|
||||||
while True:
|
return call_with_channel(grpc_call, context=context)
|
||||||
channel, reused = context.get_channel()
|
|
||||||
try:
|
|
||||||
if imports_pending:
|
|
||||||
resolve_imports(channel)
|
|
||||||
stub = device_pb2_grpc.DeviceStub(channel)
|
|
||||||
response = stub.Handle(device_pb2.Request(get_status={}))
|
|
||||||
return response.dish_get_status
|
|
||||||
except grpc.RpcError:
|
|
||||||
context.close()
|
|
||||||
if not reused:
|
|
||||||
raise
|
|
||||||
|
|
||||||
|
|
||||||
def get_id(context=None):
|
def get_id(context=None):
|
||||||
|
@ -506,8 +538,7 @@ def status_data(context=None):
|
||||||
values, in that order.
|
values, in that order.
|
||||||
|
|
||||||
Raises:
|
Raises:
|
||||||
GrpcError: Failed getting history info from the Starlink user
|
GrpcError: Failed getting status info from the Starlink user terminal.
|
||||||
terminal.
|
|
||||||
"""
|
"""
|
||||||
try:
|
try:
|
||||||
status = get_status(context)
|
status = get_status(context)
|
||||||
|
@ -712,26 +743,14 @@ def get_history(context=None):
|
||||||
Raises:
|
Raises:
|
||||||
grpc.RpcError: Communication or service error.
|
grpc.RpcError: Communication or service error.
|
||||||
"""
|
"""
|
||||||
if context is None:
|
def grpc_call(channel):
|
||||||
with grpc.insecure_channel("192.168.100.1:9200") as channel:
|
|
||||||
if imports_pending:
|
if imports_pending:
|
||||||
resolve_imports(channel)
|
resolve_imports(channel)
|
||||||
stub = device_pb2_grpc.DeviceStub(channel)
|
stub = device_pb2_grpc.DeviceStub(channel)
|
||||||
response = stub.Handle(device_pb2.Request(get_history={}))
|
response = stub.Handle(device_pb2.Request(get_history={}))
|
||||||
return response.dish_get_history
|
return response.dish_get_history
|
||||||
|
|
||||||
while True:
|
return call_with_channel(grpc_call, context=context)
|
||||||
channel, reused = context.get_channel()
|
|
||||||
try:
|
|
||||||
if imports_pending:
|
|
||||||
resolve_imports(channel)
|
|
||||||
stub = device_pb2_grpc.DeviceStub(channel)
|
|
||||||
response = stub.Handle(device_pb2.Request(get_history={}))
|
|
||||||
return response.dish_get_history
|
|
||||||
except grpc.RpcError:
|
|
||||||
context.close()
|
|
||||||
if not reused:
|
|
||||||
raise
|
|
||||||
|
|
||||||
|
|
||||||
def _compute_sample_range(history, parse_samples, start=None, verbose=False):
|
def _compute_sample_range(history, parse_samples, start=None, verbose=False):
|
||||||
|
|
Loading…
Reference in a new issue