commit | 88c027bac3b6ac40d1ea65f563bf1f9de8a9cc19 | [log] [tgz] |
---|---|---|
author | Chengyuan Zhang <[email protected]> | Wed Feb 19 18:07:44 2020 -0800 |
committer | GitHub <[email protected]> | Wed Feb 19 18:07:44 2020 -0800 |
tree | b6c1681a5300c25cd20601c01ac978d4e69b28f2 | |
parent | 19de229b7f5fa5cf2b5259c04b7d7e896270dea3 [diff] |
xds: report load stats for all clusters over a single LRS client (#6706) Current implementation of client side load reporting is incorrect. Mainly each gRPC channel should have at most one LRS stream based on the current design of using a single management server. In this change: - Each LoadStatsStore instance is associated with clusterName:clusterServiceName. clusterName and clusterServiceName (nullable) is required to construct an LoadStatsStore instance. - The semantics is that an LoadStatsStore is responsible for recording loads sent to that cluster service of the cluster. The queried load report (via LoadStatsStore#generateLoadReport()) will have cluster_name and cluster_service_name (if not null) set. - A LoadReportClient is responsible for reporting loads for all clusters. Add LoadStatsStore to LoadReportClient via LoadReportClient#addLoadStatsStore(clusterName, clusterServiceName, loadStatsStore). This should be done before LoadReportClient#startLoadReporting() is called due to the above open question. - An XdsClient contains a single LoadReportClient instance. Its APIs XdsClient#reportClientStats(clusterName, clusterServiceName, loadStatsStore) calls LoadReportClient#addLoadStatsStore(clusterName, clusterServiceName, loadStatsStore) and then starts it. XdsClient#cancelClientStatsReport(clusterName, clusterServiceName) calls LoadReportClient#removeLoadStatsStore(clusterName, clusterServiceName) and stops it. LoadReportClient#addLoadStatsStore(clusterName, clusterServiceName, loadStatsStore) cannot be called repeatedly as once the load reporting started, we cannot change the cluster to report loads for. However, we are able to do report then cancel then report then cancel and so on. - Refactored EdsLoadBalancer a bit, to accommodate the new APIs of enabling/disabling load reporting. The ClusterEndpointsLoadBalancer instance carries its own LoadStatsStore and controls start/cancel of load reporting. - The interface for LoadReportClient is eliminated. LoadReportClient will completely be a subcomponent of XdsClient. Note: Currently we assume no cluster/eds service switch, which means we will report load for a single cluster/eds service. So we make the restriction to LoadReportClient#addLoadStatsStore() API that it cannot be called after load reporting has already started. This restriction will be removed after the above open question is resolved.
gRPC-Java works with JDK 7. gRPC-Java clients are supported on Android API levels 14 and up (Ice Cream Sandwich and later). Deploying gRPC servers on an Android device is not supported.
TLS usage typically requires using Java 8, or Play Services Dynamic Security Provider on Android. Please see the Security Readme.
For a guided tour, take a look at the quick start guide or the more explanatory gRPC basics.
The examples and the Android example are standalone projects that showcase the usage of gRPC.
Download the JARs. Or for Maven with non-Android, add to your pom.xml
:
<dependency> <groupId>io.grpc</groupId> <artifactId>grpc-netty-shaded</artifactId> <version>1.27.1</version> </dependency> <dependency> <groupId>io.grpc</groupId> <artifactId>grpc-protobuf</artifactId> <version>1.27.1</version> </dependency> <dependency> <groupId>io.grpc</groupId> <artifactId>grpc-stub</artifactId> <version>1.27.1</version> </dependency>
Or for Gradle with non-Android, add to your dependencies:
implementation 'io.grpc:grpc-netty-shaded:1.27.1' implementation 'io.grpc:grpc-protobuf:1.27.1' implementation 'io.grpc:grpc-stub:1.27.1'
For Android client, use grpc-okhttp
instead of grpc-netty-shaded
and grpc-protobuf-lite
instead of grpc-protobuf
:
implementation 'io.grpc:grpc-okhttp:1.27.1' implementation 'io.grpc:grpc-protobuf-lite:1.27.1' implementation 'io.grpc:grpc-stub:1.27.1'
Development snapshots are available in Sonatypes's snapshot repository.
For protobuf-based codegen, you can put your proto files in the src/main/proto
and src/test/proto
directories along with an appropriate plugin.
For protobuf-based codegen integrated with the Maven build system, you can use protobuf-maven-plugin (Eclipse and NetBeans users should also look at os-maven-plugin
's IDE documentation):
<build> <extensions> <extension> <groupId>kr.motd.maven</groupId> <artifactId>os-maven-plugin</artifactId> <version>1.6.2</version> </extension> </extensions> <plugins> <plugin> <groupId>org.xolstice.maven.plugins</groupId> <artifactId>protobuf-maven-plugin</artifactId> <version>0.6.1</version> <configuration> <protocArtifact>com.google.protobuf:protoc:3.11.0:exe:${os.detected.classifier}</protocArtifact> <pluginId>grpc-java</pluginId> <pluginArtifact>io.grpc:protoc-gen-grpc-java:1.27.1:exe:${os.detected.classifier}</pluginArtifact> </configuration> <executions> <execution> <goals> <goal>compile</goal> <goal>compile-custom</goal> </goals> </execution> </executions> </plugin> </plugins> </build>
For protobuf-based codegen integrated with the Gradle build system, you can use protobuf-gradle-plugin:
plugins { id 'com.google.protobuf' version '0.8.8' } protobuf { protoc { artifact = "com.google.protobuf:protoc:3.11.0" } plugins { grpc { artifact = 'io.grpc:protoc-gen-grpc-java:1.27.1' } } generateProtoTasks { all()*.plugins { grpc {} } } }
The prebuilt protoc-gen-grpc-java binary uses glibc on Linux. If you are compiling on Alpine Linux, you may want to use the Alpine grpc-java package which uses musl instead.
APIs annotated with @Internal
are for internal use by the gRPC library and should not be used by gRPC users. APIs annotated with @ExperimentalApi
are subject to change in future releases, and library code that other projects may depend on should not use these APIs.
We recommend using the grpc-java-api-checker (an Error Prone plugin) to check for usages of @ExperimentalApi
and @Internal
in any library code that depends on gRPC. It may also be used to check for @Internal
usage or unintended @ExperimentalApi
consumption in non-library code.
If you are making changes to gRPC-Java, see the compiling instructions.
At a high level there are three distinct layers to the library: Stub, Channel, and Transport.
The Stub layer is what is exposed to most developers and provides type-safe bindings to whatever datamodel/IDL/interface you are adapting. gRPC comes with a plugin to the protocol-buffers compiler that generates Stub interfaces out of .proto
files, but bindings to other datamodel/IDL are easy and encouraged.
The Channel layer is an abstraction over Transport handling that is suitable for interception/decoration and exposes more behavior to the application than the Stub layer. It is intended to be easy for application frameworks to use this layer to address cross-cutting concerns such as logging, monitoring, auth, etc.
The Transport layer does the heavy lifting of putting and taking bytes off the wire. The interfaces to it are abstract just enough to allow plugging in of different implementations. Note the transport layer API is considered internal to gRPC and has weaker API guarantees than the core API under package io.grpc
.
gRPC comes with three Transport implementations: