Status: This is an early experimental alpha version designed to gauge interest and showcase the possibilities. Please reach out if you are interested in the approach, its direct application, or its development. Minimal tests are running, as the project goals are still evolving. Feedback and assistance are welcome.
Help distinguish between normal and malicious use of GraphQL APIs of any size and complexity.
The aim is not to accurately predict or even estimate performance, time taken, or resource utilization for GraphQL queries, as these should not have an impact on what valid API uses are. It is understood that this project could be used for some related request analysis but, at this time, this is not the primary goal.
API security teams got used to various combinations of rate and endpoint constraints with REST and other http-based APIs. The model of the world these approaches use matches the reality of use cases only as much as underlying APIs do. When singular real-world use cases translate to many requests and/or endpoints the mapping becomes broken. Distinguishing logic begins to operate on secondary, derivative information, without the use case context. This yields compromises being made that affect the recognition quality.
Despite this many attempt to apply the same thinking to GraphQL APIs, presumably because that is what they are familiar with and/or have the tools for. That is a failure in the following ways:
- GraphQL requests are dynamic so a single request to a single endpoint can be an issue.
- Not leveraging better alignment of GraphQL to use cases (closer to one request per case).
This lead to the spread of blind GraphQL request depth restrictions, another failure:
- Even shallow GraphQL requests can be "dangerous" using batching, especially with fragments.
- Even deep GraphQL requests can be perfectly safe, e.g. navigating "to one" relations.
More recently, query/request complexity accounting started popping up, in a variety of ways. These are better but, at least in some cases, appear to be rudimentary and/or force an oversimplification that replaces one kind of secondary/derivative information (e.g. REST API request rate) with another, only slightly better single number, without accounting for resource type access constraints.
Paniql analyses a given GraphQL request and, given the GraphQL schema annotated with Paniql directives, estimates the access and a number of different standard work type quantities for each type and field, while supporting servers able to optimize queries by "looking ahead" (see 1, 2, 3, 4, 5, ... for commentaries, variety of programming languages), These quantities are standardized and not rolled-up. Instead, they are all independently available.
Have a look at some report examples and corresponding sample schema and requests for illustration. Paniql is meant to allow us to analyse queries such as:
{
folder(id: 123) {
name
files {
id
name
}
}
}
and deduce something like t:
-
At most 1
Folder
instance will be accessed -
At most 1
Folder.name
will be accessed. -
Between 0 and (say) 1000
File
instances will be accessed, with some probability distribution. -
Equal number of
File.id
andFile.name
fields will be accessed. -
(Possibly)
File.id
andFile.name
may be very efficiently fetched in bulk, together with the containingFolder
.
Reasoning behind avoiding concrete cost units or custom work types is that in multi-service environments it becomes almost impossible to synchronize different standards and/or perform what would be the equivalent of "(cost) currency conversion". This is in addition to focusing on primary, first-order information mapped from the use cases and away from the non-goal - it isn't about the performance or resource utilization but the validity of the usage.
Prerequisite: Java 11 or greater and the paniql-commandline
jar.
java -jar <path-to>paniql-commandline-version.jar <path-to-schema> <path-to-request>
Get relevant dependencies. Unless you want console printing and/or command-line support,
you'll just need the paniql-core
with its transitive dependencies (graphql-java
).
Gradle ( |
Maven ( |
---|---|
dependencies {
implementation 'net.susnjar.paniql:core:0.1.24.21'
} |
<dependency>
<groupId>net.susnjar.paniql</groupId>
<artifactId>core</artifactId>
<version>0.1.24.21</version>
</dependency> |
Import relevant classes:
import net.susnjar.paniql.Environment;
import net.susnjar.paniql.pricing.Invoice;
Instantiate the Environment object and keep it handy:
final Environment environment = new Environment(Path.of("/some/dir/api-schema.graphqls"));
Obtain an "Invoice":
final Invoice invoice1 = environment.invoice("{ folder(id: 123) { id } }");
final Invoice invoice2 = environment.invoice(someRequestString);
final Invoice invoice3 = environment.invoice(parsedRequestDocument);
Inspect it as desired. Example to follow also relies on the following imports:
import java.util.Map;
import net.susnjar.paniql.pricing.Price;
import net.susnjar.paniql.pricing.WorkType;
Example Invoice inspection:
final Map<OutputTypeModel, Price> resourceCosts = invoice.getResourceCosts();
for (final Map.Entry<OutputTypeModel, Price> entry: resourceCosts.entrySet()) {
final String resourceTypeName = entry.getKey().getFullyQualifiedName();
final Price price = entry.getValue();
System.out.println(resourceTypeName + ":");
for (final WorkType workType: WorkType.values()) {
System.out.println(" - " + workType.getHeading() + ": " + price.get(workType).getMaximum());
}
}
Note that the invoice divides the items into three categories:
- Accessing instances of types declared as resources using the
@paniqlResource
directive, as shown in the above example, accessed usinginvoice.getResourceCosts()
. - Accessing instances of other types not marked as resources, using
invoice.getPartCosts()
, also yieldingMap.Entry<OutputTypeModel, Price>
. - Accessing fields, using
invoice.getFieldCosts()
, yielding a similarMap<FieldDefModel, Price>
.
Also note that Invoice, Price and Bounds support basic arithmetic (addition, multiplication) so that they can be easily aggregated. See their methods for further reference.
Paniql is not a complete solution, only a part of it. It only analyses the requests and does not make any decisions. Actual decision-making and related constraints are (presently) not handled by Paniql.
Recommended approach is to use Paniql together with a sliding time window quota system that can additionally have distinct quotas for different effective users (authenticated and/or not), client types (if known, e.g. which application/agent is making requests), client network locations (if known) and paths that requests took to arrive (e.g. which network interface and/or gateway). Since Paniql exposes many distinct metrics, the quotas would express different constrains for each metric of interest.
For example, any combination of the following could be envisioned as quotas:
-
Unauthenticated users have "0 per request" quotas on fields or types requiring authentication.
-
Unauthenticated users have certain quotas on authentication/login operations.
-
Different applications (client types) have "0 per request" quotas on operations they are not intended or built for. Those would imply "counterfeit" clients.
-
Different applications and users can have appropriate and distinct quotas that would meaningfully let them to their jobs and not more simply because they themselves are not expected to be able to handle more.
-
Multiple separate time window lengths can be specified, e.g. per second, per minute, per 15 minutes, per hour, per 8 hours, per day, such that quotas for longer windows are less than corresponding multiples of quotas of smaller windows.
Such quota/user tracking logic could also be present in or added to a variety of existing API gateways. Additionally, as a Java project, Paniql could be used within Java GraphQL projects as an embedded first level of defence against malicious requests. Note that, while current implementation is in Java, this is not meant to be tied to Java at all.
Though small, the project is divided into three modules:
- core - the code expected to be needed in production.
Main classes of interest to start with are:
- Environment - reusable representation the annotated GraphQL API schema model.
- Invoice - the result of the analysis.
- print - the code that produces formatted ASCII text reports of Invoice(s).
- commandline - a command-line utility that utilizes the other two projects to allow "playing" with Paniql from command line.
Documentation is pretty limited at the moment. Please see the test classes for code examples:
- Get+apply feedback + help
- More documentation
- Fix bugs
- Figure out what else to do, if there's interest