Remove need to generate grpc modules via protoc
If the spacex grpc modules are not available in the import path, will now fall back to using reflection to get them dynamically. I'm not real happy with the mess this made of the import lines, though (and neither is pylint...), so I may hack on that a little further when I get the time. Add a requirements.txt file to enable installation of all prerequisites so users don't have to follow the individual instructions for each dependency package. At some point, I'll really need to add proper Python packaging so the whole thing can just be installed via pip. Rearrange the README a bit, since some of the sections have increased or decreased in relevance over time.
This commit is contained in:
parent
381cfbee00
commit
491227ddb4
3 changed files with 106 additions and 57 deletions
123
README.md
123
README.md
|
@ -7,11 +7,20 @@ For more information on what Starlink is, see [starlink.com](https://www.starlin
|
||||||
|
|
||||||
Most of the scripts here are [Python](https://www.python.org/) scripts. To use them, you will either need Python installed on your system or you can use the Docker image. If you use the Docker image, you can skip the rest of the prerequisites other than making sure the dish IP is reachable and Docker itself. For Linux systems, the python package from your distribution should be fine, as long as it is Python 3. The JSON script should actually work with Python 2.7, but the grpc scripts all require Python 3 (and Python 2.7 is past end-of-life, so is not recommended anyway).
|
Most of the scripts here are [Python](https://www.python.org/) scripts. To use them, you will either need Python installed on your system or you can use the Docker image. If you use the Docker image, you can skip the rest of the prerequisites other than making sure the dish IP is reachable and Docker itself. For Linux systems, the python package from your distribution should be fine, as long as it is Python 3. The JSON script should actually work with Python 2.7, but the grpc scripts all require Python 3 (and Python 2.7 is past end-of-life, so is not recommended anyway).
|
||||||
|
|
||||||
`parseJsonHistory.py` operates on a JSON format data representation of the protocol buffer messages, such as that output by [gRPCurl](https://github.com/fullstorydev/grpcurl). The command lines below assume `grpcurl` is installed in the runtime PATH. If that's not the case, just substitute in the full path to the command.
|
|
||||||
|
|
||||||
All the tools that pull data from the dish expect to be able to reach it at the dish's fixed IP address of 192.168.100.1, as do the Starlink [Android app](https://play.google.com/store/apps/details?id=com.starlink.mobile), [iOS app](https://apps.apple.com/us/app/starlink/id1537177988), and the browser app you can run directly from http://192.168.100.1. When using a router other than the one included with the Starlink installation kit, this usually requires some additional router configuration to make it work. That configuration is beyond the scope of this document, but if the Starlink app doesn't work on your home network, then neither will these scripts. That being said, you do not need the Starlink app installed to make use of these scripts.
|
All the tools that pull data from the dish expect to be able to reach it at the dish's fixed IP address of 192.168.100.1, as do the Starlink [Android app](https://play.google.com/store/apps/details?id=com.starlink.mobile), [iOS app](https://apps.apple.com/us/app/starlink/id1537177988), and the browser app you can run directly from http://192.168.100.1. When using a router other than the one included with the Starlink installation kit, this usually requires some additional router configuration to make it work. That configuration is beyond the scope of this document, but if the Starlink app doesn't work on your home network, then neither will these scripts. That being said, you do not need the Starlink app installed to make use of these scripts.
|
||||||
|
|
||||||
The scripts that don't use `grpcurl` to pull data require the `grpcio` Python package at runtime and generating the necessary gRPC protocol code requires the `grpcio-tools` package. Information about how to install both can be found at https://grpc.io/docs/languages/python/quickstart/
|
Running the scripts within a [Docker](https://www.docker.com/) container requires Docker to be installed. Information about how to install that can be found at https://docs.docker.com/engine/install/
|
||||||
|
|
||||||
|
`parseJsonHistory.py` operates on a JSON format data representation of the protocol buffer messages, such as that output by [gRPCurl](https://github.com/fullstorydev/grpcurl). The command lines below assume `grpcurl` is installed in the runtime PATH. If that's not the case, just substitute in the full path to the command.
|
||||||
|
|
||||||
|
### Required Python modules
|
||||||
|
|
||||||
|
If you don't care about the details or minimizing your package requirements, you can skip the rest of this section and just do this to install latest versions of a superset of required modules:
|
||||||
|
```shell script
|
||||||
|
pip install --upgrade -r requirements.txt
|
||||||
|
```
|
||||||
|
|
||||||
|
The scripts that don't use `grpcurl` to pull data require the `grpcio` Python package at runtime and the optional step of generating the gRPC protocol module code requires the `grpcio-tools` package. Information about how to install both can be found at https://grpc.io/docs/languages/python/quickstart/. If you skip generation of the gRPC protocol modules, the scripts will instead require the `yagrc` Python package. Information about how to install that is at https://github.com/sparky8512/yagrc.
|
||||||
|
|
||||||
The scripts that use [MQTT](https://mqtt.org/) for output require the `paho-mqtt` Python package. Information about how to install that can be found at https://www.eclipse.org/paho/index.php?page=clients/python/index.php
|
The scripts that use [MQTT](https://mqtt.org/) for output require the `paho-mqtt` Python package. Information about how to install that can be found at https://www.eclipse.org/paho/index.php?page=clients/python/index.php
|
||||||
|
|
||||||
|
@ -19,31 +28,12 @@ The scripts that use [InfluxDB](https://www.influxdata.com/products/influxdb/) f
|
||||||
|
|
||||||
Note that the Python package versions available from various Linux distributions (ie: installed via `apt-get` or similar) tend to run a bit behind those available to install via `pip`. While the distro packages should work OK as long as they aren't extremely old, they may not work as well as the later versions.
|
Note that the Python package versions available from various Linux distributions (ie: installed via `apt-get` or similar) tend to run a bit behind those available to install via `pip`. While the distro packages should work OK as long as they aren't extremely old, they may not work as well as the later versions.
|
||||||
|
|
||||||
Running the scripts within a [Docker](https://www.docker.com/) container requires Docker to be installed. Information about how to install that can be found at https://docs.docker.com/engine/install/
|
### Generating the gRPC protocol modules
|
||||||
|
|
||||||
## Usage
|
This step is no longer required, as the grpc scripts can now get the protocol modules classes at run time via reflection, but generating the protocol modules will improve script startup time, and it would be a good idea to at least stash away the protoset file emitted by `grpcurl` in case SpaceX ever turns off server reflection in the dish software.
|
||||||
|
|
||||||
Of the 3 groups below, the grpc scripts are really the only ones being actively developed. The others are mostly by way of example of what could be done with the underlying data.
|
The grpc scripts require some generated code to support the specific gRPC protocol messages used. These would normally be generated from .proto files that specify those messages, but to date (2020-Dec), SpaceX has not publicly released such files. The gRPC service running on the dish appears to have [server reflection](https://github.com/grpc/grpc/blob/master/doc/server-reflection.md) enabled, though. `grpcurl` can use that to extract a protoset file, and the `protoc` compiler can use that to make the necessary generated code:
|
||||||
|
```shell script
|
||||||
### The JSON parser script
|
|
||||||
|
|
||||||
`parseJsonHistory.py` takes input from a file and writes its output to standard output. The easiest way to use it is to pipe the `grpcurl` command directly into it. For example:
|
|
||||||
```
|
|
||||||
grpcurl -plaintext -d {\"get_history\":{}} 192.168.100.1:9200 SpaceX.API.Device.Device/Handle | python parseJsonHistory.py
|
|
||||||
```
|
|
||||||
For more usage options, run:
|
|
||||||
```
|
|
||||||
python parseJsonHistory.py -h
|
|
||||||
```
|
|
||||||
|
|
||||||
When used as-is, `parseJsonHistory.py` will summarize packet loss information from the data the dish records. There's other bits of data in there, though, so that script (or more likely the parsing logic it uses, which now resides in `starlink_json.py`) could be used as a starting point or example of how to iterate through it. Most of the data displayed in the Statistics page of the Starlink app appears to come from this same `get_history` gRPC response. See the file `get_history_notes.txt` for some ramblings on how to interpret it.
|
|
||||||
|
|
||||||
The one bit of functionality this script has over the grpc scripts is that it supports capturing the grpcurl output to a file and reading from that, which may be useful if you're collecting data in one place but analyzing it in another. Otherwise, it's probably better to use `dish_grpc_text.py`, described below.
|
|
||||||
|
|
||||||
### The grpc scripts
|
|
||||||
|
|
||||||
This set of scripts can do the gRPC communication directly, but they require some generated code to support the specific gRPC protocol messages used. These would normally be generated from .proto files that specify those messages, but to date (2020-Dec), SpaceX has not publicly released such files. The gRPC service running on the dish appears to have [server reflection](https://github.com/grpc/grpc/blob/master/doc/server-reflection.md) enabled, though. `grpcurl` can use that to extract a protoset file, and the `protoc` compiler can use that to make the necessary generated code:
|
|
||||||
```
|
|
||||||
grpcurl -plaintext -protoset-out dish.protoset 192.168.100.1:9200 describe SpaceX.API.Device.Device
|
grpcurl -plaintext -protoset-out dish.protoset 192.168.100.1:9200 describe SpaceX.API.Device.Device
|
||||||
mkdir src
|
mkdir src
|
||||||
cd src
|
cd src
|
||||||
|
@ -58,24 +48,30 @@ python3 -m grpc_tools.protoc --descriptor_set_in=../dish.protoset --python_out=.
|
||||||
```
|
```
|
||||||
Then move the resulting files to where the Python scripts can find them in the import path, such as in the same directory as the scripts themselves.
|
Then move the resulting files to where the Python scripts can find them in the import path, such as in the same directory as the scripts themselves.
|
||||||
|
|
||||||
Once those are available, the `dish_grpc_text.py` script can be used in place of the `grpcurl | parseJsonHistory.py` pipeline; however, the command line interface is slightly different because `dish_grpc_text.py` supports additional functionality. The equivalent command to `grpcurl | parseJsonHistory.py` would be:
|
## Usage
|
||||||
```
|
|
||||||
python3 dish_grpc_text.py ping_drop
|
Of the 3 groups below, the grpc scripts are really the only ones being actively developed. The others are mostly by way of example of what could be done with the underlying data.
|
||||||
|
|
||||||
|
### The grpc scripts
|
||||||
|
|
||||||
|
This set of scripts includes `dish_grpc_text.py`, `dish_grpc_influx.py`, `dish_grpc_sqlite.py`, and `dish_grpc_mqtt.py`. They mostly support the same functionality, but write their output in different ways. `dish_grpc_text.py` writes data to standard output, `dish_grpc_influx.py` sends it to an InfluxDB server, `dish_grpc_sqlite.py` writes it a sqlite database, and `dish_grpc_mqtt.py` sends it to a MQTT broker.
|
||||||
|
|
||||||
|
All 4 scripts support processing status data and/or history data in various modes. The status data is mostly what appears related to the dish in the Debug Data section of the Starlink app, whereas most of the data displayed in the Statistics page of the Starlink app comes from the history data. Specific status or history data groups can be selected by including their mode names on the command line. Run the scripts with `-h` command line option to get a list of available modes. See the documentation at the top of `starlink_grpc.py` for detail on what each of the fields means within each mode group.
|
||||||
|
|
||||||
|
For example, all the currently available status groups can be output by doing:
|
||||||
|
```shell script
|
||||||
|
python3 dish_grpc_text.py status obstruction_detail alert_detail
|
||||||
```
|
```
|
||||||
|
|
||||||
By default, `dish_grpc_text.py` (and `parseJsonHistory.py`) will output in CSV format. You can use the `-v` option to instead output in a (slightly) more human-readable format.
|
By default, `dish_grpc_text.py` (and `parseJsonHistory.py`, described below) will output in CSV format. You can use the `-v` option to instead output in a (slightly) more human-readable format.
|
||||||
|
|
||||||
To collect and record packet loss summary stats at the top of every hour, you could put something like the following in your user crontab (assuming you have moved the scripts to ~/bin and made them executable):
|
To collect and record packet loss summary stats at the top of every hour, you could put something like the following in your user crontab (assuming you have moved the scripts to ~/bin and made them executable):
|
||||||
```
|
```
|
||||||
00 * * * * [ -e ~/dishStats.csv ] || ~/bin/dish_grpc_text.py -H >~/dishStats.csv; ~/bin/dish_grpc_text.py ping_drop >>~/dishStats.csv
|
00 * * * * [ -e ~/dishStats.csv ] || ~/bin/dish_grpc_text.py -H >~/dishStats.csv; ~/bin/dish_grpc_text.py ping_drop >>~/dishStats.csv
|
||||||
```
|
```
|
||||||
|
|
||||||
`dish_grpc_influx.py`, `dish_grpc_sqlite.py`, and `dish_grpc_mqtt.py` are similar, but they send their output to an InfluxDB server, a sqlite database, and a MQTT broker, respectively. Run them with `-h` command line option for details on how to specify server and/or database options.
|
|
||||||
|
|
||||||
All 4 scripts support processing status data in addition to the history data. The status data is mostly what appears related to the dish in the Debug Data section of the Starlink app. Specific status or history data groups can be selected by including their mode names on the command line. Run the scripts with `-h` command line option to get a list of available modes. See the documentation at the top of `starlink_grpc.py` for detail on what each of the fields means within each mode group.
|
|
||||||
|
|
||||||
By default, all of these scripts will pull data once, send it off to the specified data backend, and then exit. They can instead be made to run in a periodic loop by passing a `-t` option to specify loop interval, in seconds. For example, to capture status information to a InfluxDB server every 30 seconds, you could do something like this:
|
By default, all of these scripts will pull data once, send it off to the specified data backend, and then exit. They can instead be made to run in a periodic loop by passing a `-t` option to specify loop interval, in seconds. For example, to capture status information to a InfluxDB server every 30 seconds, you could do something like this:
|
||||||
```
|
```shell script
|
||||||
python3 dish_grpc_influx.py -t 30 [... probably other args to specify server options ...] status
|
python3 dish_grpc_influx.py -t 30 [... probably other args to specify server options ...] status
|
||||||
```
|
```
|
||||||
|
|
||||||
|
@ -85,41 +81,40 @@ Some of the scripts (currently only the InfluxDB one) also support specifying op
|
||||||
|
|
||||||
`dish_grpc_influx.py`, `dish_grpc_sqlite.py`, and `dish_grpc_text.py` also support a bulk history mode that collects and writes the full second-by-second data instead of summary stats. To select bulk mode, use `bulk_history` for the mode argument. You'll probably also want to use the `-t` option to have it run in a loop.
|
`dish_grpc_influx.py`, `dish_grpc_sqlite.py`, and `dish_grpc_text.py` also support a bulk history mode that collects and writes the full second-by-second data instead of summary stats. To select bulk mode, use `bulk_history` for the mode argument. You'll probably also want to use the `-t` option to have it run in a loop.
|
||||||
|
|
||||||
|
### The JSON parser script
|
||||||
|
|
||||||
|
`parseJsonHistory.py` takes input from a file and writes its output to standard output. The easiest way to use it is to pipe the `grpcurl` command directly into it. For example:
|
||||||
|
```shell script
|
||||||
|
grpcurl -plaintext -d {\"get_history\":{}} 192.168.100.1:9200 SpaceX.API.Device.Device/Handle | python parseJsonHistory.py
|
||||||
|
```
|
||||||
|
For more usage options, run:
|
||||||
|
```shell script
|
||||||
|
python parseJsonHistory.py -h
|
||||||
|
```
|
||||||
|
|
||||||
|
When used as-is, `parseJsonHistory.py` will summarize packet loss information from the data the dish records. There's other bits of data in there, though, so that script (or more likely the parsing logic it uses, which now resides in `starlink_json.py`) could be used as a starting point or example of how to iterate through it.
|
||||||
|
|
||||||
|
The one bit of functionality this script has over the grpc scripts is that it supports capturing the grpcurl output to a file and reading from that, which may be useful if you're collecting data in one place but analyzing it in another. Otherwise, it's probably better to use `dish_grpc_text.py`, described above.
|
||||||
|
|
||||||
### Other scripts
|
### Other scripts
|
||||||
|
|
||||||
`dump_dish_status.py` is a simple example of how to use the grpc modules (the ones generated by protoc, not `starlink_grpc`) directly. Just run it as:
|
`dump_dish_status.py` is a simple example of how to use the grpc modules (the ones generated by protoc, not `starlink_grpc`) directly. Just run it as:
|
||||||
```
|
```shell script
|
||||||
python3 dump_dish_status.py
|
python3 dump_dish_status.py
|
||||||
```
|
```
|
||||||
and revel in copious amounts of dish status information. OK, maybe it's not as impressive as all that. This one is really just meant to be a starting point for real functionality to be added to it.
|
and revel in copious amounts of dish status information. OK, maybe it's not as impressive as all that. This one is really just meant to be a starting point for real functionality to be added to it.
|
||||||
|
|
||||||
`poll_history.py` is another silly example, but this one illustrates how to periodically poll the status and/or bulk history data using the `starlink_grpc` module's API. It's not really useful by itself, but if you really want to, you can run it as:
|
`poll_history.py` is another silly example, but this one illustrates how to periodically poll the status and/or bulk history data using the `starlink_grpc` module's API. It's not really useful by itself, but if you really want to, you can run it as:
|
||||||
```
|
```shell script
|
||||||
python3 poll_history.py
|
python3 poll_history.py
|
||||||
```
|
```
|
||||||
Possibly more simple examples to come, as the other scripts have started getting a bit complicated.
|
Possibly more simple examples to come, as the other scripts have started getting a bit complicated.
|
||||||
|
|
||||||
## To Be Done (Maybe)
|
|
||||||
|
|
||||||
Maybe more data backend options. If there's one you'd like to see supported, please open a feature request issue.
|
|
||||||
|
|
||||||
There are `reboot` and `dish_stow` requests in the Device protocol, too, so it should be trivial to write a command that initiates dish reboot and stow operations. These are easy enough to do with `grpcurl`, though, as there is no need to parse through the response data. For that matter, they're easy enough to do with the Starlink app.
|
|
||||||
|
|
||||||
Proper Python packaging, since some of the scripts are no longer self-contained.
|
|
||||||
|
|
||||||
The requirement to run `grpcurl` and `protoc` could be eliminated by adding support for use of gRPC server reflection directly in the grpc scripts. This would sidestep any packaging questions about whether or not the protoc-generated files could be redistributed.
|
|
||||||
|
|
||||||
## Other Tidbits
|
|
||||||
|
|
||||||
The Starlink Android app actually uses port 9201 instead of 9200. Both appear to expose the same gRPC service, but the one on port 9201 uses [gRPC-Web](https://github.com/grpc/grpc/blob/master/doc/PROTOCOL-WEB.md), which can use HTTP/1.1, whereas the one on port 9200 uses HTTP/2, which is what most gRPC tools expect.
|
|
||||||
|
|
||||||
The Starlink router also exposes a gRPC service, on ports 9000 (HTTP/2.0) and 9001 (HTTP/1.1).
|
|
||||||
|
|
||||||
## Docker for InfluxDB ( & MQTT under development )
|
## Docker for InfluxDB ( & MQTT under development )
|
||||||
|
|
||||||
Initialization of the container can be performed with the following command:
|
Initialization of the container can be performed with the following command:
|
||||||
|
|
||||||
```
|
```shell script
|
||||||
docker run -d -t --name='starlink-grpc-tools' -e INFLUXDB_HOST={InfluxDB Hostname} \
|
docker run -d -t --name='starlink-grpc-tools' -e INFLUXDB_HOST={InfluxDB Hostname} \
|
||||||
-e INFLUXDB_PORT={Port, 8086 usually} \
|
-e INFLUXDB_PORT={Port, 8086 usually} \
|
||||||
-e INFLUXDB_USER={Optional, InfluxDB Username} \
|
-e INFLUXDB_USER={Optional, InfluxDB Username} \
|
||||||
|
@ -136,6 +131,26 @@ The `dish_grpc_influx.py -v status alert_detail` is optional and omitting it wil
|
||||||
|
|
||||||
You'll probably want to run with the `-t` option to `dish_grpc_influx.py` to collect status information periodically for this to be meaningful.
|
You'll probably want to run with the `-t` option to `dish_grpc_influx.py` to collect status information periodically for this to be meaningful.
|
||||||
|
|
||||||
|
## To Be Done (Maybe)
|
||||||
|
|
||||||
|
Maybe more data backend options. If there's one you'd like to see supported, please open a feature request issue.
|
||||||
|
|
||||||
|
There are `reboot` and `dish_stow` requests in the Device protocol, too, so it should be trivial to write a command that initiates dish reboot and stow operations. These are easy enough to do with `grpcurl`, though, as there is no need to parse through the response data. For that matter, they're easy enough to do with the Starlink app.
|
||||||
|
|
||||||
|
Proper Python packaging, since the dependency list keeps growing....
|
||||||
|
|
||||||
|
Some of the functionality implemented in the `starlink-grpc` module could be ported into `starlink-json` easily enough, but this won't be a priority unless someone asks for it.
|
||||||
|
|
||||||
|
## Other Tidbits
|
||||||
|
|
||||||
|
The Starlink Android app actually uses port 9201 instead of 9200. Both appear to expose the same gRPC service, but the one on port 9201 uses [gRPC-Web](https://github.com/grpc/grpc/blob/master/doc/PROTOCOL-WEB.md), which can use HTTP/1.1, whereas the one on port 9200 uses HTTP/2, which is what most gRPC tools expect.
|
||||||
|
|
||||||
|
The Starlink router also exposes a gRPC service, on ports 9000 (HTTP/2.0) and 9001 (HTTP/1.1).
|
||||||
|
|
||||||
|
The file `get_history_notes.txt` has my original ramblings on how to interpret the history buffer data (with the JSON format naming). It may be of interest if you're interested in pulling the `get_history` grpc data directly and don't want to dig through the convoluted logic in the `starlink-grpc` module.
|
||||||
|
|
||||||
## Related Projects
|
## Related Projects
|
||||||
|
|
||||||
[ChuckTSI's Better Than Nothing Web Interface](https://github.com/ChuckTSI/BetterThanNothingWebInterface) uses grpcurl and PHP to provide a spiffy web UI for some of the same data this project works on.
|
[ChuckTSI's Better Than Nothing Web Interface](https://github.com/ChuckTSI/BetterThanNothingWebInterface) uses grpcurl and PHP to provide a spiffy web UI for some of the same data this project works on.
|
||||||
|
|
||||||
|
[starlink-cli](https://github.com/starlink-community/starlink-cli) is another command line tool for interacting with the Starlink gRPC services, including the one on the Starlink router, in case Go is more your thing.
|
||||||
|
|
6
requirements.txt
Normal file
6
requirements.txt
Normal file
|
@ -0,0 +1,6 @@
|
||||||
|
grpcio>=1.12.0
|
||||||
|
grpcio-tools>=1.20.0
|
||||||
|
protobuf>=3.6.0
|
||||||
|
yagrc>=1.0.1
|
||||||
|
paho-mqtt>=1.5.1
|
||||||
|
influxdb>=5.3.1
|
|
@ -302,9 +302,29 @@ import statistics
|
||||||
|
|
||||||
import grpc
|
import grpc
|
||||||
|
|
||||||
from spacex.api.device import device_pb2
|
try:
|
||||||
from spacex.api.device import device_pb2_grpc
|
from spacex.api.device import device_pb2
|
||||||
from spacex.api.device import dish_pb2
|
from spacex.api.device import device_pb2_grpc
|
||||||
|
from spacex.api.device import dish_pb2
|
||||||
|
import_ok = True
|
||||||
|
except ImportError:
|
||||||
|
from yagrc import importer
|
||||||
|
import_ok = False
|
||||||
|
|
||||||
|
|
||||||
|
def import_protocols(channel):
|
||||||
|
grpc_importer = importer.GrpcImporter()
|
||||||
|
grpc_importer.configure(
|
||||||
|
channel, filenames=["spacex/api/device/device.proto", "spacex/api/device/dish.proto"])
|
||||||
|
|
||||||
|
global device_pb2
|
||||||
|
global device_pb2_grpc
|
||||||
|
global dish_pb2
|
||||||
|
from spacex.api.device import device_pb2
|
||||||
|
from spacex.api.device import device_pb2_grpc
|
||||||
|
from spacex.api.device import dish_pb2
|
||||||
|
global import_ok
|
||||||
|
import_ok = True
|
||||||
|
|
||||||
|
|
||||||
class GrpcError(Exception):
|
class GrpcError(Exception):
|
||||||
|
@ -426,6 +446,8 @@ def get_status(context=None):
|
||||||
"""
|
"""
|
||||||
if context is None:
|
if context is None:
|
||||||
with grpc.insecure_channel("192.168.100.1:9200") as channel:
|
with grpc.insecure_channel("192.168.100.1:9200") as channel:
|
||||||
|
if not import_ok:
|
||||||
|
import_protocols(channel)
|
||||||
stub = device_pb2_grpc.DeviceStub(channel)
|
stub = device_pb2_grpc.DeviceStub(channel)
|
||||||
response = stub.Handle(device_pb2.Request(get_status={}))
|
response = stub.Handle(device_pb2.Request(get_status={}))
|
||||||
return response.dish_get_status
|
return response.dish_get_status
|
||||||
|
@ -433,6 +455,8 @@ def get_status(context=None):
|
||||||
while True:
|
while True:
|
||||||
channel, reused = context.get_channel()
|
channel, reused = context.get_channel()
|
||||||
try:
|
try:
|
||||||
|
if not import_ok:
|
||||||
|
import_protocols(channel)
|
||||||
stub = device_pb2_grpc.DeviceStub(channel)
|
stub = device_pb2_grpc.DeviceStub(channel)
|
||||||
response = stub.Handle(device_pb2.Request(get_status={}))
|
response = stub.Handle(device_pb2.Request(get_status={}))
|
||||||
return response.dish_get_status
|
return response.dish_get_status
|
||||||
|
@ -683,6 +707,8 @@ def get_history(context=None):
|
||||||
"""
|
"""
|
||||||
if context is None:
|
if context is None:
|
||||||
with grpc.insecure_channel("192.168.100.1:9200") as channel:
|
with grpc.insecure_channel("192.168.100.1:9200") as channel:
|
||||||
|
if not import_ok:
|
||||||
|
import_protocols(channel)
|
||||||
stub = device_pb2_grpc.DeviceStub(channel)
|
stub = device_pb2_grpc.DeviceStub(channel)
|
||||||
response = stub.Handle(device_pb2.Request(get_history={}))
|
response = stub.Handle(device_pb2.Request(get_history={}))
|
||||||
return response.dish_get_history
|
return response.dish_get_history
|
||||||
|
@ -690,6 +716,8 @@ def get_history(context=None):
|
||||||
while True:
|
while True:
|
||||||
channel, reused = context.get_channel()
|
channel, reused = context.get_channel()
|
||||||
try:
|
try:
|
||||||
|
if not import_ok:
|
||||||
|
import_protocols(channel)
|
||||||
stub = device_pb2_grpc.DeviceStub(channel)
|
stub = device_pb2_grpc.DeviceStub(channel)
|
||||||
response = stub.Handle(device_pb2.Request(get_history={}))
|
response = stub.Handle(device_pb2.Request(get_history={}))
|
||||||
return response.dish_get_history
|
return response.dish_get_history
|
||||||
|
|
Loading…
Reference in a new issue