Revert "Make dish_obstruction_map work in Docker image"

This reverts commit 1545b0fcc9.
This commit is contained in:
sparky8512 2022-12-01 17:51:24 -08:00
parent 1545b0fcc9
commit f23cf2d756
5 changed files with 21 additions and 304 deletions

View file

@ -11,7 +11,6 @@ pip3 install \
urllib3==1.26.12 \
influxdb-client==1.34.0 reactivex==4.0.4 \
paho-mqtt==1.6.1 \
pypng==0.20220715.0 \
typing_extensions==4.4.0 \
yagrc==1.1.1 grpcio-reflection==1.50.0 protobuf==4.21.9

View file

@ -2,28 +2,17 @@
"""Prometheus exporter for Starlink user terminal data info.
This script pulls the current status info and/or metrics computed from the
history data and makes it available via HTTP in the format Prometheus expects.
history data and makes it available via HTTP in the format Prometeus expects.
"""
import logging
import sys
from http import HTTPStatus
from http.server import BaseHTTPRequestHandler, ThreadingHTTPServer
import logging
import signal
import sys
import threading
import dish_common
class Terminated(Exception):
pass
def handle_sigterm(signum, frame):
# Turn SIGTERM into an exception so main loop can clean up
raise Terminated
class MetricInfo:
unit = ""
kind = "gauge"
@ -136,7 +125,13 @@ class MetricValue:
return f"{label_str} {self.value}"
opts = None
gstate = None
def parse_args():
global opts
parser = dish_common.create_arg_parser(
output_description="Prometheus exporter", bulk_history=False
)
@ -145,10 +140,12 @@ def parse_args():
group.add_argument("--address", default="0.0.0.0", help="IP address to listen on")
group.add_argument("--port", default=8080, type=int, help="Port to listen on")
return dish_common.run_arg_parser(parser)
opts = dish_common.run_arg_parser(parser)
def prometheus_export(opts, gstate):
def prometheus_export():
global opts, gstate
raw_data = {}
def data_add_item(name, value, category):
@ -246,16 +243,7 @@ def prometheus_export(opts, gstate):
class MetricsRequestHandler(BaseHTTPRequestHandler):
def do_GET(self):
path = self.path.partition("?")[0]
if path != "/":
self.send_error(HTTPStatus.NOT_FOUND)
return
opts = self.server.opts
gstate = self.server.gstate
with gstate.lock:
content = prometheus_export(opts, gstate)
content = prometheus_export()
self.send_response(HTTPStatus.OK)
self.send_header("Content-type", "text/plain")
self.send_header("Content-Length", len(content))
@ -264,28 +252,16 @@ class MetricsRequestHandler(BaseHTTPRequestHandler):
def main():
opts = parse_args()
global opts, gstate
parse_args()
logging.basicConfig(format="%(levelname)s: %(message)s", stream=sys.stderr)
gstate = dish_common.GlobalState(target=opts.target)
gstate.lock = threading.Lock()
httpd = ThreadingHTTPServer((opts.address, opts.port), MetricsRequestHandler)
httpd.daemon_threads = False
httpd.opts = opts
httpd.gstate = gstate
signal.signal(signal.SIGTERM, handle_sigterm)
print("HTTP listening on port", opts.port)
try:
httpd.serve_forever()
except (KeyboardInterrupt, Terminated):
pass
finally:
httpd.server_close()
httpd.gstate.shutdown()
httpd.serve_forever()
sys.exit()

View file

@ -13,25 +13,11 @@ except ModuleNotFoundError:
file=sys.stderr)
sys.exit(1)
def wrap_ser(self, **kwargs):
ret = device_pb2.Request.SerializeToString(self, **kwargs)
print("request:", ret)
return ret
def wrap_des(s):
print("response:", s)
#return "Hello"
return device_pb2.Response.FromString(s)
# Note that if you remove the 'with' clause here, you need to separately
# call channel.close() when you're done with the gRPC connection.
with grpc.insecure_channel("192.168.100.1:9200") as channel:
method = channel.unary_unary('/SpaceX.API.Device.Device/Handle',
request_serializer=wrap_ser,
response_deserializer=None
#response_deserializer=wrap_des
)
response = method(device_pb2.Request(get_status={}), timeout=10)
stub = device_pb2_grpc.DeviceStub(channel)
response = stub.Handle(device_pb2.Request(get_status={}), timeout=10)
# Dump everything
print(response)

View file

@ -1,239 +0,0 @@
#!/usr/bin/python3
"""Dump gRPC service information from a file or a reflection server
This script will query a gRPC reflection server for descriptor information of
all services supported by the server, excluding the reflection service itself,
and dump it in a textual format, or read previous saved descriptor information
and dump that.
Although the default target option is the local IP and port number used by the
gRPC service on a Starlink user terminal, this script is otherwise not
specific to Starlink and should work for any gRPC server that does not require
SSL and that has the reflection service enabled.
"""
import argparse
import logging
import sys
import time
import grpc
from google.protobuf import descriptor_pb2
from yagrc import dump
from yagrc import reflector
TARGET_DEFAULT = "192.168.100.1:9200"
RETRY_DELAY_DEFAULT = 0
def parse_args():
parser = argparse.ArgumentParser(
description="Textually dump a serialized FileDescriptorSet (protoset) from "
"either a file or a gRPC reflection server")
parser.add_argument("in_file",
nargs="?",
metavar="IN_FILE",
help="File from which to read protoset data instead of getting "
"it via reflection")
parser.add_argument("-g",
"--target",
default=TARGET_DEFAULT,
help="host:port of device to query, default: " + TARGET_DEFAULT)
parser.add_argument("-r",
"--retry-delay",
type=float,
default=float(RETRY_DELAY_DEFAULT),
help="Time in seconds to wait before retrying after network "
"error or 0 for no retry, default: " + str(RETRY_DELAY_DEFAULT))
opts = parser.parse_args()
return opts
def defix(name, prefix):
if prefix:
packages = prefix.split(".")[1:]
if packages:
packages[0] = "." + packages[0]
for package in packages:
if name.startswith(package + "."):
name = name.removeprefix(package + ".")
else:
break
return name
def dump_service(indent, service, prefix):
name = [indent, "service", service.name]
if service.options.deprecated:
name.append("deprecated")
print(*name)
indent = indent + " "
for method in service.method:
items = [indent, " rpc ", method.name, "("]
if method.client_streaming:
items.append("stream ")
items.extend([defix(method.input_type, prefix), ") returns ("])
if method.server_streaming:
items.append("stream ")
items.extend([defix(method.output_type, prefix), ")"])
if method.options.deprecated:
items.append("deprecated")
print(*items, sep="")
def dump_enum(indent, enum):
name = [indent, "enum", enum.name]
if enum.options.deprecated:
name.append("deprecated")
print(*name)
indent = indent + " "
for value in enum.value:
items = [indent, value.name, "=", value.number]
if value.options.deprecated:
items.append("deprecated")
print(*items)
def field_type(field, prefix):
return defix(
field.type_name if field.type_name else descriptor_pb2.FieldDescriptorProto.Type.Name(
field.type).removeprefix("TYPE_").lower(), prefix)
def dump_field(indent, field, prefix, maps):
items = [indent]
if field.type_name and not field.type_name.startswith("."):
full_type = prefix + "." + field.type_name
else:
full_type = field.type_name
if (field.label == descriptor_pb2.FieldDescriptorProto.Label.LABEL_REPEATED
and full_type in maps):
key_type, value_type = maps[full_type]
items.append("".join(
["map<", defix(key_type, prefix), ", ",
defix(value_type, prefix), ">"]))
else:
if field.label != descriptor_pb2.FieldDescriptorProto.Label.LABEL_OPTIONAL:
items.append(
descriptor_pb2.FieldDescriptorProto.Label.Name(
field.label).removeprefix("LABEL_").lower())
items.append(field_type(field, prefix))
items.extend([field.name, "=", field.number])
if field.options.deprecated:
items.append("deprecated")
print(*items)
def dump_message(indent, message, prefix, maps):
full_name = prefix + "." + message.name
if message.options.map_entry:
if len(message.field
) >= 2 and message.field[0].name == "key" and message.field[1].name == "value":
maps[full_name] = (field_type(message.field[0],
None), field_type(message.field[1], None))
# else just assume it is malformed and ignore it
return
name = [indent, "message", message.name]
if message.options.deprecated:
name.append("deprecated")
print(*name)
indent = indent + " "
for nested_message in message.nested_type:
dump_message(indent, nested_message, full_name, maps)
for enum in message.enum_type:
dump_enum(indent, enum)
oneof_emitted = set()
for field in message.field:
if field.HasField("oneof_index") and field.oneof_index < len(message.oneof_decl):
# Nothing in descriptor proto mandates that all the oneof fields
# for a given index be grouped together, so make a separate pass
# for each unique index and keep track of which were already seen.
if field.oneof_index not in oneof_emitted:
print(indent, "oneof", message.oneof_decl[field.oneof_index].name)
indent = indent + " "
for oneof_field in message.field:
if oneof_field.HasField(
"oneof_index") and oneof_field.oneof_index == field.oneof_index:
dump_field(indent, oneof_field, full_name, {})
oneof_emitted.add(field.oneof_index)
else:
dump_field(indent, field, full_name, maps)
# Any leftover oneof decls will be for an empty oneof
for index, decl in enumerate(message.oneof_decl):
if index not in oneof_emitted:
print(indent, "oneof", decl.name)
def dump_protoset(protoset):
file_desc_set = descriptor_pb2.FileDescriptorSet.FromString(protoset)
for file in file_desc_set.file:
name = ["file", file.name]
if file.options.deprecated:
name.append("deprecated")
print(*name)
if file.package:
print(" package", file.package)
prefix = "." + file.package
else:
prefix = ""
for depend in file.dependency:
print(" import", depend)
for service in file.service:
dump_service(" ", service, prefix)
maps = {}
for message in file.message_type:
dump_message(" ", message, prefix, maps)
for enum in file.enum_type:
dump_enum(" ", enum)
print()
def reflect_protoset(opts):
while True:
try:
with grpc.insecure_channel(opts.target) as channel:
return dump.dump_protocols(channel)
break
except reflector.ServiceError as e:
logging.error("Problem with reflection service: %s", str(e))
# Only retry on network-related errors, not service errors
return None
except grpc.RpcError as e:
# grpc.RpcError error message is not very useful, but grpc.Call has
# something slightly better
if isinstance(e, grpc.Call):
msg = e.details()
else:
msg = "Unknown communication or service error"
print("Problem communicating with reflection service:", msg)
if opts.retry_delay > 0.0:
time.sleep(opts.retry_delay)
else:
return None
def main():
opts = parse_args()
logging.basicConfig(format="%(levelname)s: %(message)s")
if opts.in_file is not None:
try:
with open(opts.in_file, mode="rb") as infile:
protoset = infile.read()
except OSError as e:
logging.error("Failed to read file %s: %s", opts.in_file, str(e))
protoset = None
else:
protoset = reflect_protoset(opts)
if protoset:
dump_protoset(protoset)
else:
sys.exit(1)
if __name__ == "__main__":
main()

View file

@ -719,12 +719,7 @@ def status_data(
# Special case translate this to equivalent old name
state = "SEARCHING"
else:
try:
state = dish_pb2.DishOutage.Cause.Name(status.outage.cause)
except ValueError:
# Unlikely, but possible if dish is running newer firmware
# than protocol data pulled via reflection
state = str(status.outage.cause)
state = dish_pb2.DishOutage.Cause.Name(status.outage.cause)
else:
state = "CONNECTED"