diff --git a/api/LICENSE b/api/LICENSE index affd2b9..93114d5 100644 --- a/api/LICENSE +++ b/api/LICENSE @@ -1,4 +1,4 @@ -This licence applies to the file: event.schema.json +This licence applies to the files: event.schema.json (and its derivatives) The source of this file is: https://raw.githubusercontent.com/getsentry/sentry-data-schemas/main/LICENSE Copyright (c) 2020 Sentry (https://sentry.io) and individual contributors. diff --git a/api/README.md b/api/README.md index 49f92c3..12dc586 100644 --- a/api/README.md +++ b/api/README.md @@ -40,12 +40,3 @@ In short, the more reasons to just use the "upstream" API. Said in another way: we act more as the "relay" than as "getsentry/sentry", because we do ingest straight in the main process. So we should adhere to the relay's spec. - -### Notes on use: - -Bugsink, as it stands, doesn't use event.schema.json much. - -* We have `--valid-only` as a param on `send_json`, but I appear to have used that only sporadically (back in nov 2023) -* We _could_ at some point in the future [offer the option to] throw events through a validator before proceeding with - digesting. At that point we'll re-vendor event.schema.json (from the sentry-data-schemas repo) -* Reading this file is useful, but we can do that straight from the source. diff --git a/api/event.schema.altered.json b/api/event.schema.altered.json new file mode 100644 index 0000000..c7e44eb --- /dev/null +++ b/api/event.schema.altered.json @@ -0,0 +1,4372 @@ +{ + "$schema": "http://json-schema.org/draft-07/schema#", + "title": "Event", + "description": " The sentry v7 event structure.", + "anyOf": [ + { + "type": "object", + "properties": { + "breadcrumbs": { + "anyOf": [ + { + "description": " List of breadcrumbs recorded before this event.", + "default": null, + "type": [ + "object", + "null" + ], + "required": [ + "values" + ], + "properties": { + "values": { + "type": "array", + "items": { + "anyOf": [ + { + "$ref": "#/definitions/Breadcrumb" + }, + { + "type": "null" + } + ] + } + } + } + }, + { + "type": "array", + "items": { + "anyOf": [ + { + "$ref": "#/definitions/Breadcrumb" + }, + { + "type": "null" + } + ] + } + } + ] + }, + "contexts": { + "description": " Contexts describing the environment (e.g. device, os or browser).", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Contexts" + }, + { + "type": "null" + } + ] + }, + "culprit": { + "description": " Custom culprit of the event.\n\n This field is deprecated and shall not be set by client SDKs.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "debug_meta": { + "description": " Meta data for event processing and debugging.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/DebugMeta" + }, + { + "type": "null" + } + ] + }, + "dist": { + "description": " Program's distribution identifier.\n\n The distribution of the application.\n\n Distributions are used to disambiguate build or deployment variants of the same release of\n an application. For example, the dist can be the build number of an XCode build or the\n version code of an Android build.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "environment": { + "description": " The environment name, such as `production` or `staging`.\n\n ```json\n { \"environment\": \"production\" }\n ```", + "default": null, + "type": [ + "string", + "null" + ] + }, + "errors": { + "description": " Errors encountered during processing. Intended to be phased out in favor of\n annotation/metadata system.", + "default": null, + "type": [ + "array", + "null" + ], + "items": { + "anyOf": [ + { + "$ref": "#/definitions/EventProcessingError" + }, + { + "type": "null" + } + ] + } + }, + "event_id": { + "description": " Unique identifier of this event.\n\n Hexadecimal string representing a uuid4 value. The length is exactly 32 characters. Dashes\n are not allowed. Has to be lowercase.\n\n Even though this field is backfilled on the server with a new uuid4, it is strongly\n recommended to generate that uuid4 clientside. There are some features like user feedback\n which are easier to implement that way, and debugging in case events get lost in your\n Sentry installation is also easier.\n\n Example:\n\n ```json\n {\n \"event_id\": \"fc6d8c0c43fc4630ad850ee518f1b9d0\"\n }\n ```", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/EventId" + }, + { + "type": "null" + } + ] + }, + "exception": { + "anyOf": [ + { + "description": " One or multiple chained (nested) exceptions.", + "default": null, + "type": [ + "object", + "null" + ], + "required": [ + "values" + ], + "properties": { + "values": { + "type": "array", + "items": { + "anyOf": [ + { + "$ref": "#/definitions/Exception" + }, + { + "type": "null" + } + ] + } + } + } + }, + { + "type": "array", + "items": { + "anyOf": [ + { + "$ref": "#/definitions/Exception" + }, + { + "type": "null" + } + ] + } + } + ] + }, + "extra": { + "description": " Arbitrary extra information set by the user.\n\n ```json\n {\n \"extra\": {\n \"my_key\": 1,\n \"some_other_value\": \"foo bar\"\n }\n }```", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": true + }, + "fingerprint": { + "description": " Manual fingerprint override.\n\n A list of strings used to dictate how this event is supposed to be grouped with other\n events into issues. For more information about overriding grouping see [Customize Grouping\n with Fingerprints](https://docs.sentry.io/data-management/event-grouping/).\n\n ```json\n {\n \"fingerprint\": [\"myrpc\", \"POST\", \"/foo.bar\"]\n }", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Fingerprint" + }, + { + "type": "null" + } + ] + }, + "level": { + "description": " Severity level of the event. Defaults to `error`.\n\n Example:\n\n ```json\n {\"level\": \"warning\"}\n ```", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Level" + }, + { + "type": "null" + } + ] + }, + "logentry": { + "description": " Custom parameterized message for this event.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/LogEntry" + }, + { + "type": "null" + } + ] + }, + "logger": { + "description": " Logger that created the event.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "modules": { + "description": " Name and versions of all installed modules/packages/dependencies in the current\n environment/application.\n\n ```json\n { \"django\": \"3.0.0\", \"celery\": \"4.2.1\" }\n ```\n\n In Python this is a list of installed packages as reported by `pkg_resources` together with\n their reported version string.\n\n This is primarily used for suggesting to enable certain SDK integrations from within the UI\n and for making informed decisions on which frameworks to support in future development\n efforts.", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": { + "type": [ + "string", + "null" + ] + } + }, + "platform": { + "description": " Platform identifier of this event (defaults to \"other\").\n\n A string representing the platform the SDK is submitting from. This will be used by the\n Sentry interface to customize various components in the interface, but also to enter or\n skip stacktrace processing.\n\n Acceptable values are: `as3`, `c`, `cfml`, `cocoa`, `csharp`, `elixir`, `haskell`, `go`,\n `groovy`, `java`, `javascript`, `native`, `node`, `objc`, `other`, `perl`, `php`, `python`,\n `ruby`", + "default": null, + "type": [ + "string", + "null" + ] + }, + "received": { + "description": " Timestamp when the event has been received by Sentry.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Timestamp" + }, + { + "type": "null" + } + ] + }, + "release": { + "description": " The release version of the application.\n\n **Release versions must be unique across all projects in your organization.** This value\n can be the git SHA for the given project, or a product identifier with a semantic version.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "request": { + "description": " Information about a web request that occurred during the event.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Request" + }, + { + "type": "null" + } + ] + }, + "sdk": { + "description": " Information about the Sentry SDK that generated this event.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/ClientSdkInfo" + }, + { + "type": "null" + } + ] + }, + "server_name": { + "description": " Server or device name the event was generated on.\n\n This is supposed to be a hostname.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "stacktrace": { + "description": " Event stacktrace.\n\n DEPRECATED: Prefer `threads` or `exception` depending on which is more appropriate.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Stacktrace" + }, + { + "type": "null" + } + ] + }, + "tags": { + "description": " Custom tags for this event.\n\n A map or list of tags for this event. Each tag must be less than 200 characters.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Tags" + }, + { + "type": "null" + } + ] + }, + "threads": { + "anyOf": [ + { + "description": " Threads that were active when the event occurred.", + "default": null, + "type": [ + "object", + "null" + ], + "required": [ + "values" + ], + "properties": { + "values": { + "type": "array", + "items": { + "anyOf": [ + { + "$ref": "#/definitions/Thread" + }, + { + "type": "null" + } + ] + } + } + } + }, + { + "type": "array", + "items": { + "anyOf": [ + { + "$ref": "#/definitions/Thread" + }, + { + "type": "null" + } + ] + } + } + ] + }, + "time_spent": { + "description": " Time since the start of the transaction until the error occurred.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "timestamp": { + "description": " Timestamp when the event was created.\n\n Indicates when the event was created in the Sentry SDK. The format is either a string as\n defined in [RFC 3339](https://tools.ietf.org/html/rfc3339) or a numeric (integer or float)\n value representing the number of seconds that have elapsed since the [Unix\n epoch](https://en.wikipedia.org/wiki/Unix_time).\n\n Timezone is assumed to be UTC if missing.\n\n Sub-microsecond precision is not preserved with numeric values due to precision\n limitations with floats (at least in our systems). With that caveat in mind, just send\n whatever is easiest to produce.\n\n All timestamps in the event protocol are formatted this way.\n\n # Example\n\n All of these are the same date:\n\n ```json\n { \"timestamp\": \"2011-05-02T17:41:36Z\" }\n { \"timestamp\": \"2011-05-02T17:41:36\" }\n { \"timestamp\": \"2011-05-02T17:41:36.000\" }\n { \"timestamp\": 1304358096.0 }\n ```", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Timestamp" + }, + { + "type": "null" + } + ] + }, + "transaction": { + "description": " Transaction name of the event.\n\n For example, in a web app, this might be the route name (`\"/users//\"` or\n `UserView`), in a task queue it might be the function + module name.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "transaction_info": { + "description": " Additional information about the name of the transaction.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/TransactionInfo" + }, + { + "type": "null" + } + ] + }, + "type": { + "description": " Type of the event. Defaults to `default`.\n\n The event type determines how Sentry handles the event and has an impact on processing, rate\n limiting, and quotas. There are three fundamental classes of event types:\n\n - **Error monitoring events**: Processed and grouped into unique issues based on their\n exception stack traces and error messages.\n - **Security events**: Derived from Browser security violation reports and grouped into\n unique issues based on the endpoint and violation. SDKs do not send such events.\n - **Transaction events** (`transaction`): Contain operation spans and collected into traces\n for performance monitoring.\n\n Transactions must explicitly specify the `\"transaction\"` event type. In all other cases,\n Sentry infers the appropriate event type from the payload and overrides the stated type.\n SDKs should not send an event type other than for transactions.\n\n Example:\n\n ```json\n {\n \"type\": \"transaction\",\n \"spans\": []\n }\n ```", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/EventType" + }, + { + "type": "null" + } + ] + }, + "user": { + "description": " Information about the user who triggered this event.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/User" + }, + { + "type": "null" + } + ] + }, + "version": { + "description": " Version", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ], + "definitions": { + "Addr": { + "type": "string" + }, + "AppContext": { + "description": " Application information.\n\n App context describes the application. As opposed to the runtime, this is the actual\n application that was running and carries metadata about the current session.", + "anyOf": [ + { + "type": "object", + "properties": { + "app_build": { + "description": " Internal build ID as it appears on the platform.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "app_identifier": { + "description": " Version-independent application identifier, often a dotted bundle ID.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "app_memory": { + "description": " Amount of memory used by the application in bytes.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "app_name": { + "description": " Application name as it appears on the platform.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "app_start_time": { + "description": " Start time of the app.\n\n Formatted UTC timestamp when the user started the application.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "app_version": { + "description": " Application version as it appears on the platform.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "build_type": { + "description": " String identifying the kind of build. For example, `testflight`.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "device_app_hash": { + "description": " Application-specific device identifier.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "in_foreground": { + "description": " A flag indicating whether the app is in foreground or not. An app is in foreground when it's visible to the user.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "view_names": { + "description": " The names of the currently visible views.", + "default": null, + "type": [ + "array", + "null" + ], + "items": { + "type": [ + "string", + "null" + ] + } + } + }, + "additionalProperties": false + } + ] + }, + "AppleDebugImage": { + "description": " Legacy apple debug images (MachO).\n\n This was also used for non-apple platforms with similar debug setups.", + "anyOf": [ + { + "type": "object", + "required": [ + "image_addr", + "image_size", + "name", + "uuid" + ], + "properties": { + "arch": { + "description": " CPU architecture target.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "cpu_subtype": { + "description": " MachO CPU subtype identifier.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "cpu_type": { + "description": " MachO CPU type identifier.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "image_addr": { + "description": " Starting memory address of the image (required).", + "anyOf": [ + { + "$ref": "#/definitions/Addr" + }, + { + "type": "null" + } + ] + }, + "image_size": { + "description": " Size of the image in bytes (required).", + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "image_vmaddr": { + "description": " Loading address in virtual memory.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Addr" + }, + { + "type": "null" + } + ] + }, + "name": { + "description": " Path and name of the debug image (required).", + "type": [ + "string", + "null" + ] + }, + "uuid": { + "description": " The unique UUID of the image.", + "type": [ + "string", + "null" + ], + "format": "uuid" + } + }, + "additionalProperties": false + } + ] + }, + "Breadcrumb": { + "description": " The Breadcrumbs Interface specifies a series of application events, or \"breadcrumbs\", that\n occurred before an event.\n\n An event may contain one or more breadcrumbs in an attribute named `breadcrumbs`. The entries\n are ordered from oldest to newest. Consequently, the last entry in the list should be the last\n entry before the event occurred.\n\n While breadcrumb attributes are not strictly validated in Sentry, a breadcrumb is most useful\n when it includes at least a `timestamp` and `type`, `category` or `message`. The rendering of\n breadcrumbs in Sentry depends on what is provided.\n\n The following example illustrates the breadcrumbs part of the event payload and omits other\n attributes for simplicity.\n\n ```json\n {\n \"breadcrumbs\": {\n \"values\": [\n {\n \"timestamp\": \"2016-04-20T20:55:53.845Z\",\n \"message\": \"Something happened\",\n \"category\": \"log\",\n \"data\": {\n \"foo\": \"bar\",\n \"blub\": \"blah\"\n }\n },\n {\n \"timestamp\": \"2016-04-20T20:55:53.847Z\",\n \"type\": \"navigation\",\n \"data\": {\n \"from\": \"/login\",\n \"to\": \"/dashboard\"\n }\n }\n ]\n }\n }\n ```", + "anyOf": [ + { + "type": "object", + "properties": { + "category": { + "description": " A dotted string indicating what the crumb is or from where it comes. _Optional._\n\n Typically it is a module name or a descriptive string. For instance, _ui.click_ could be\n used to indicate that a click happened in the UI or _flask_ could be used to indicate that\n the event originated in the Flask framework.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "data": { + "description": " Arbitrary data associated with this breadcrumb.\n\n Contains a dictionary whose contents depend on the breadcrumb `type`. Additional parameters\n that are unsupported by the type are rendered as a key/value table.", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": true + }, + "event_id": { + "description": " Identifier of the event this breadcrumb belongs to.\n\n Sentry events can appear as breadcrumbs in other events as long as they have occurred in the\n same organization. This identifier links to the original event.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/EventId" + }, + { + "type": "null" + } + ] + }, + "level": { + "description": " Severity level of the breadcrumb. _Optional._\n\n Allowed values are, from highest to lowest: `fatal`, `error`, `warning`, `info`, and\n `debug`. Levels are used in the UI to emphasize and deemphasize the crumb. Defaults to\n `info`.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Level" + }, + { + "type": "null" + } + ] + }, + "message": { + "description": " Human readable message for the breadcrumb.\n\n If a message is provided, it is rendered as text with all whitespace preserved. Very long\n text might be truncated in the UI.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "timestamp": { + "description": " The timestamp of the breadcrumb. Recommended.\n\n A timestamp representing when the breadcrumb occurred. The format is either a string as\n defined in [RFC 3339](https://tools.ietf.org/html/rfc3339) or a numeric (integer or float)\n value representing the number of seconds that have elapsed since the [Unix\n epoch](https://en.wikipedia.org/wiki/Unix_time).\n\n Breadcrumbs are most useful when they include a timestamp, as it creates a timeline leading\n up to an event.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Timestamp" + }, + { + "type": "null" + } + ] + }, + "type": { + "description": " The type of the breadcrumb. _Optional_, defaults to `default`.\n\n - `default`: Describes a generic breadcrumb. This is typically a log message or\n user-generated breadcrumb. The `data` field is entirely undefined and as such, completely\n rendered as a key/value table.\n\n - `navigation`: Describes a navigation breadcrumb. A navigation event can be a URL change\n in a web application, or a UI transition in a mobile or desktop application, etc.\n\n Such a breadcrumb's `data` object has the required fields `from` and `to`, which\n represent an application route/url each.\n\n - `http`: Describes an HTTP request breadcrumb. This represents an HTTP request transmitted\n from your application. This could be an AJAX request from a web application, or a\n server-to-server HTTP request to an API service provider, etc.\n\n Such a breadcrumb's `data` property has the fields `url`, `method`, `status_code`\n (integer) and `reason` (string).", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "BrowserContext": { + "description": " Web browser information.", + "anyOf": [ + { + "type": "object", + "properties": { + "name": { + "description": " Display name of the browser application.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "version": { + "description": " Version string of the browser.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "CError": { + "description": " POSIX signal with optional extended data.\n\n Error codes set by Linux system calls and some library functions as specified in ISO C99,\n POSIX.1-2001, and POSIX.1-2008. See\n [`errno(3)`](https://man7.org/linux/man-pages/man3/errno.3.html) for more information.", + "anyOf": [ + { + "type": "object", + "properties": { + "name": { + "description": " Optional name of the errno constant.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "number": { + "description": " The error code as specified by ISO C99, POSIX.1-2001 or POSIX.1-2008.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "int64" + } + }, + "additionalProperties": false + } + ] + }, + "ClientSdkInfo": { + "description": " The SDK Interface describes the Sentry SDK and its configuration used to capture and transmit an event.", + "anyOf": [ + { + "type": "object", + "required": [ + "name", + "version" + ], + "properties": { + "features": { + "description": " List of features that are enabled in the SDK. _Optional._\n\n A list of feature names identifying enabled SDK features. This list\n should contain all enabled SDK features. On some SDKs, enabling a feature in the\n options also adds an integration. We encourage tracking such features with either\n integrations or features but not both to reduce the payload size.", + "default": null, + "type": [ + "array", + "null" + ], + "items": { + "type": [ + "string", + "null" + ] + } + }, + "integrations": { + "description": " List of integrations that are enabled in the SDK. _Optional._\n\n The list should have all enabled integrations, including default integrations. Default\n integrations are included because different SDK releases may contain different default\n integrations.", + "default": null, + "type": [ + "array", + "null" + ], + "items": { + "type": [ + "string", + "null" + ] + } + }, + "name": { + "description": " Unique SDK name. _Required._\n\n The name of the SDK. The format is `entity.ecosystem[.flavor]` where entity identifies the\n developer of the SDK, ecosystem refers to the programming language or platform where the\n SDK is to be used and the optional flavor is used to identify standalone SDKs that are part\n of a major ecosystem.\n\n Official Sentry SDKs use the entity `sentry`, as in `sentry.python` or\n `sentry.javascript.react-native`. Please use a different entity for your own SDKs.", + "type": [ + "string", + "null" + ] + }, + "packages": { + "description": " List of installed and loaded SDK packages. _Optional._\n\n A list of packages that were installed as part of this SDK or the activated integrations.\n Each package consists of a name in the format `source:identifier` and `version`. If the\n source is a Git repository, the `source` should be `git`, the identifier should be a\n checkout link and the version should be a Git reference (branch, tag or SHA).", + "default": null, + "type": [ + "array", + "null" + ], + "items": { + "anyOf": [ + { + "$ref": "#/definitions/ClientSdkPackage" + }, + { + "type": "null" + } + ] + } + }, + "version": { + "description": " The version of the SDK. _Required._\n\n It should have the [Semantic Versioning](https://semver.org/) format `MAJOR.MINOR.PATCH`,\n without any prefix (no `v` or anything else in front of the major version number).\n\n Examples: `0.1.0`, `1.0.0`, `4.3.12`", + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "ClientSdkPackage": { + "description": " An installed and loaded package as part of the Sentry SDK.", + "anyOf": [ + { + "type": "object", + "properties": { + "name": { + "description": " Name of the package.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "version": { + "description": " Version of the package.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "CloudResourceContext": { + "description": " Cloud Resource Context.\n\n This context describes the cloud resource the event originated from.\n\n Example:\n\n ```json\n \"cloud_resource\": {\n \"cloud.account.id\": \"499517922981\",\n \"cloud.provider\": \"aws\",\n \"cloud.platform\": \"aws_ec2\",\n \"cloud.region\": \"us-east-1\",\n \"cloud.vavailability_zone\": \"us-east-1e\",\n \"host.id\": \"i-07d3301208fe0a55a\",\n \"host.type\": \"t2.large\"\n }\n ```", + "anyOf": [ + { + "type": "object", + "properties": { + "cloud.account.id": { + "description": " The cloud account ID the resource is assigned to.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "cloud.availability_zone": { + "description": " The zone where the resource is running.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "cloud.platform": { + "description": " The cloud platform in use.\n The prefix of the service SHOULD match the one specified in cloud_provider.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "cloud.provider": { + "description": " Name of the cloud provider.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "cloud.region": { + "description": " The geographical region the resource is running.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "host.id": { + "description": " Unique host ID.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "host.type": { + "description": " Machine type of the host.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "CodeId": { + "type": "string" + }, + "Context": { + "description": " A context describes environment info (e.g. device, os or browser).", + "anyOf": [ + { + "$ref": "#/definitions/DeviceContext" + }, + { + "$ref": "#/definitions/OsContext" + }, + { + "$ref": "#/definitions/RuntimeContext" + }, + { + "$ref": "#/definitions/AppContext" + }, + { + "$ref": "#/definitions/BrowserContext" + }, + { + "$ref": "#/definitions/GpuContext" + }, + { + "$ref": "#/definitions/TraceContext" + }, + { + "$ref": "#/definitions/ProfileContext" + }, + { + "$ref": "#/definitions/ReplayContext" + }, + { + "$ref": "#/definitions/UserReportV2Context" + }, + { + "$ref": "#/definitions/MonitorContext" + }, + { + "$ref": "#/definitions/ResponseContext" + }, + { + "$ref": "#/definitions/OtelContext" + }, + { + "$ref": "#/definitions/CloudResourceContext" + }, + { + "$ref": "#/definitions/NelContext" + }, + { + "$ref": "#/definitions/PerformanceScoreContext" + }, + { + "type": "object", + "additionalProperties": true + } + ] + }, + "ContextInner": { + "anyOf": [ + { + "$ref": "#/definitions/Context" + } + ] + }, + "Contexts": { + "description": " The Contexts interface provides additional context data. Typically, this is data related to the\n current user and the environment. For example, the device or application version. Its canonical\n name is `contexts`.\n\n The `contexts` type can be used to define arbitrary contextual data on the event. It accepts an\n object of key/value pairs. The key is the \u201calias\u201d of the context and can be freely chosen.\n However, as per policy, it should match the type of the context unless there are two values for\n a type. You can omit `type` if the key name is the type.\n\n Unknown data for the contexts is rendered as a key/value list.\n\n For more details about sending additional data with your event, see the [full documentation on\n Additional Data](https://docs.sentry.io/enriching-error-data/additional-data/).", + "anyOf": [ + { + "type": "object", + "additionalProperties": { + "anyOf": [ + { + "$ref": "#/definitions/ContextInner" + }, + { + "type": "null" + } + ] + } + } + ] + }, + "Cookies": { + "description": " A map holding cookies.", + "anyOf": [ + { + "anyOf": [ + { + "type": "object", + "additionalProperties": { + "type": [ + "string", + "null" + ] + } + }, + { + "type": "array", + "items": { + "type": [ + "array", + "null" + ], + "items": [ + { + "type": [ + "string", + "null" + ] + }, + { + "type": [ + "string", + "null" + ] + } + ], + "maxItems": 2, + "minItems": 2 + } + } + ] + } + ] + }, + "DebugId": { + "type": "string" + }, + "DebugImage": { + "description": " A debug information file (debug image).", + "anyOf": [ + { + "$ref": "#/definitions/AppleDebugImage" + }, + { + "$ref": "#/definitions/NativeDebugImage" + }, + { + "$ref": "#/definitions/NativeDebugImage" + }, + { + "$ref": "#/definitions/NativeDebugImage" + }, + { + "$ref": "#/definitions/NativeDebugImage" + }, + { + "$ref": "#/definitions/NativeDebugImage" + }, + { + "$ref": "#/definitions/ProguardDebugImage" + }, + { + "$ref": "#/definitions/NativeDebugImage" + }, + { + "$ref": "#/definitions/SourceMapDebugImage" + }, + { + "$ref": "#/definitions/JvmDebugImage" + }, + { + "type": "object", + "additionalProperties": true + } + ] + }, + "DebugMeta": { + "description": " Debugging and processing meta information.\n\n The debug meta interface carries debug information for processing errors and crash reports.\n Sentry amends the information in this interface.\n\n Example (look at field types to see more detail):\n\n ```json\n {\n \"debug_meta\": {\n \"images\": [],\n \"sdk_info\": {\n \"sdk_name\": \"iOS\",\n \"version_major\": 10,\n \"version_minor\": 3,\n \"version_patchlevel\": 0\n }\n }\n }\n ```", + "anyOf": [ + { + "type": "object", + "properties": { + "images": { + "description": " List of debug information files (debug images).", + "default": null, + "type": [ + "array", + "null" + ], + "items": { + "anyOf": [ + { + "$ref": "#/definitions/DebugImage" + }, + { + "type": "null" + } + ] + } + }, + "sdk_info": { + "description": " Information about the system SDK (e.g. iOS SDK).", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/SystemSdkInfo" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "DeviceContext": { + "description": " Device information.\n\n Device context describes the device that caused the event. This is most appropriate for mobile\n applications.", + "anyOf": [ + { + "type": "object", + "properties": { + "arch": { + "description": " Native cpu architecture of the device.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "battery_level": { + "description": " Current battery level in %.\n\n If the device has a battery, this can be a floating point value defining the battery level\n (in the range 0-100).", + "default": null, + "type": [ + "number", + "null" + ], + "format": "double" + }, + "battery_status": { + "description": " Status of the device's battery.\n\n For example, `Unknown`, `Charging`, `Discharging`, `NotCharging`, `Full`.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "boot_time": { + "description": " Indicator when the device was booted.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "brand": { + "description": " Brand of the device.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "charging": { + "description": " Whether the device was charging or not.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "cpu_description": { + "description": " CPU description.\n\n For example, Intel(R) Core(TM)2 Quad CPU Q6600 @ 2.40GHz.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "device_type": { + "description": " Kind of device the application is running on.\n\n For example, `Unknown`, `Handheld`, `Console`, `Desktop`.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "device_unique_identifier": { + "description": " Unique device identifier.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "external_free_storage": { + "description": " Free size of the attached external storage in bytes (eg: android SDK card).", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "external_storage_size": { + "description": " Total size of the attached external storage in bytes (eg: android SDK card).", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "family": { + "description": " Family of the device model.\n\n This is usually the common part of model names across generations. For instance, `iPhone`\n would be a reasonable family, so would be `Samsung Galaxy`.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "free_memory": { + "description": " How much memory is still available in bytes.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "free_storage": { + "description": " How much storage is free in bytes.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "locale": { + "description": " ISO 639-1 code of the locale the device is set to.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "low_memory": { + "description": " Whether the device was low on memory.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "manufacturer": { + "description": " Manufacturer of the device.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "memory_size": { + "description": " Total memory available in bytes.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "model": { + "description": " Device model.\n\n This, for example, can be `Samsung Galaxy S3`.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "model_id": { + "description": " Device model (internal identifier).\n\n An internal hardware revision to identify the device exactly.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "name": { + "description": " Name of the device.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "online": { + "description": " Whether the device was online or not.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "orientation": { + "description": " Current screen orientation.\n\n This can be a string `portrait` or `landscape` to define the orientation of a device.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "processor_count": { + "description": " Number of \"logical processors\".\n\n For example, 8.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "processor_frequency": { + "description": " Processor frequency in MHz.\n\n Note that the actual CPU frequency might vary depending on current load and\n power conditions, especially on low-powered devices like phones and laptops.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "screen_density": { + "description": " Device screen density.", + "default": null, + "type": [ + "number", + "null" + ], + "format": "double" + }, + "screen_dpi": { + "description": " Screen density as dots-per-inch.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "screen_height_pixels": { + "description": " Height of the screen in pixels.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "screen_resolution": { + "description": " Device screen resolution.\n\n (e.g.: 800x600, 3040x1444)", + "default": null, + "type": [ + "string", + "null" + ] + }, + "screen_width_pixels": { + "description": " Width of the screen in pixels.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "simulator": { + "description": " Simulator/prod indicator.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "storage_size": { + "description": " Total storage size of the device in bytes.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "supports_accelerometer": { + "description": " Whether the accelerometer is available on the device.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "supports_audio": { + "description": " Whether audio is available on the device.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "supports_gyroscope": { + "description": " Whether the gyroscope is available on the device.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "supports_location_service": { + "description": " Whether location support is available on the device.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "supports_vibration": { + "description": " Whether vibration is available on the device.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "timezone": { + "description": " Timezone of the device.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "usable_memory": { + "description": " How much memory is usable for the app in bytes.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "uuid": { + "description": " UUID of the device.", + "default": null, + "type": [ + "string", + "null" + ], + "format": "uuid" + } + }, + "additionalProperties": false + } + ] + }, + "EventId": { + "description": " Wrapper around a UUID with slightly different formatting.", + "anyOf": [ + { + "type": "string", + "format": "uuid" + } + ] + }, + "EventProcessingError": { + "description": " An event processing error.", + "anyOf": [ + { + "type": "object", + "required": [ + "type" + ], + "properties": { + "name": { + "description": " Affected key or deep path.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "type": { + "description": " The error kind.", + "type": [ + "string", + "null" + ] + }, + "value": { + "description": " The original value causing this error.", + "default": null + } + }, + "additionalProperties": false + } + ] + }, + "EventType": { + "description": "The type of an event.\n\nThe event type determines how Sentry handles the event and has an impact on processing, rate limiting, and quotas. There are three fundamental classes of event types:\n\n- **Error monitoring events** (`default`, `error`): Processed and grouped into unique issues based on their exception stack traces and error messages. - **Security events** (`csp`, `hpkp`, `expectct`, `expectstaple`): Derived from Browser security violation reports and grouped into unique issues based on the endpoint and violation. SDKs do not send such events. - **Transaction events** (`transaction`): Contain operation spans and collected into traces for performance monitoring.", + "type": "string", + "enum": [ + "error", + "csp", + "hpkp", + "expectct", + "expectstaple", + "nel", + "transaction", + "userreportv2", + "default" + ] + }, + "Exception": { + "description": " A single exception.\n\n Multiple values inside of an [event](#typedef-Event) represent chained exceptions and should be sorted oldest to newest. For example, consider this Python code snippet:\n\n ```python\n try:\n raise Exception(\"random boring invariant was not met!\")\n except Exception as e:\n raise ValueError(\"something went wrong, help!\") from e\n ```\n\n `Exception` would be described first in the values list, followed by a description of `ValueError`:\n\n ```json\n {\n \"exception\": {\n \"values\": [\n {\"type\": \"Exception\": \"value\": \"random boring invariant was not met!\"},\n {\"type\": \"ValueError\", \"value\": \"something went wrong, help!\"},\n ]\n }\n }\n ```", + "anyOf": [ + { + "type": "object", + "properties": { + "mechanism": { + "description": " Mechanism by which this exception was generated and handled.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Mechanism" + }, + { + "type": "null" + } + ] + }, + "module": { + "description": " The optional module, or package which the exception type lives in.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "stacktrace": { + "description": " Stack trace containing frames of this exception.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Stacktrace" + }, + { + "type": "null" + } + ] + }, + "thread_id": { + "description": " An optional value that refers to a [thread](#typedef-Thread).", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/ThreadId" + }, + { + "type": "null" + } + ] + }, + "type": { + "description": " Exception type, e.g. `ValueError`.\n\n At least one of `type` or `value` is required, otherwise the exception is discarded.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "value": { + "description": " Human readable display value.\n\n At least one of `type` or `value` is required, otherwise the exception is discarded.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/JsonLenientString" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "Fingerprint": { + "description": " A fingerprint value.", + "anyOf": [ + { + "type": "array", + "items": { + "type": "string" + } + } + ] + }, + "Frame": { + "description": " Holds information about a single stacktrace frame.\n\n Each object should contain **at least** a `filename`, `function` or `instruction_addr`\n attribute. All values are optional, but recommended.", + "anyOf": [ + { + "type": "object", + "properties": { + "abs_path": { + "description": " Absolute path to the source file.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/NativeImagePath" + }, + { + "type": "null" + } + ] + }, + "addr_mode": { + "description": " Defines the addressing mode for addresses.\n\n This can be:\n - `\"abs\"` (the default): `instruction_addr` is absolute.\n - `\"rel:$idx\"`: `instruction_addr` is relative to the `debug_meta.image` identified by its index in the list.\n - `\"rel:$uuid\"`: `instruction_addr` is relative to the `debug_meta.image` identified by its `debug_id`.\n\n If one of the `\"rel:XXX\"` variants is given together with `function_id`, the `instruction_addr` is relative\n to the uniquely identified function in the references `debug_meta.image`.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "colno": { + "description": " Column number within the source file, starting at 1.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "context_line": { + "description": " Source code of the current line (`lineno`).", + "default": null, + "type": [ + "string", + "null" + ] + }, + "filename": { + "description": " The source file name (basename only).", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/NativeImagePath" + }, + { + "type": "null" + } + ] + }, + "function": { + "description": " Name of the frame's function. This might include the name of a class.\n\n This function name may be shortened or demangled. If not, Sentry will demangle and shorten\n it for some platforms. The original function name will be stored in `raw_function`.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "function_id": { + "description": " (.NET) The function id / index that uniquely identifies a function inside a module.\n\n This is the `MetadataToken` of a .NET `MethodBase`.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Addr" + }, + { + "type": "null" + } + ] + }, + "image_addr": { + "description": " (C/C++/Native) Start address of the containing code module (image).", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Addr" + }, + { + "type": "null" + } + ] + }, + "in_app": { + "description": " Override whether this frame should be considered part of application code, or part of\n libraries/frameworks/dependencies.\n\n Setting this attribute to `false` causes the frame to be hidden/collapsed by default and\n mostly ignored during issue grouping.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "instruction_addr": { + "description": " (C/C++/Native) An optional instruction address for symbolication.\n\n This should be a string with a hexadecimal number that includes a 0x prefix.\n If this is set and a known image is defined in the\n [Debug Meta Interface]({%- link _documentation/development/sdk-dev/event-payloads/debugmeta.md -%}),\n then symbolication can take place.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Addr" + }, + { + "type": "null" + } + ] + }, + "lineno": { + "description": " Line number within the source file, starting at 1.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "lock": { + "description": " A possible lock (java monitor object) held by this frame.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/LockReason" + }, + { + "type": "null" + } + ] + }, + "module": { + "description": " Name of the module the frame is contained in.\n\n Note that this might also include a class name if that is something the\n language natively considers to be part of the stack (for instance in Java).", + "default": null, + "type": [ + "string", + "null" + ] + }, + "package": { + "description": " Name of the package that contains the frame.\n\n For instance this can be a dylib for native languages, the name of the jar\n or .NET assembly.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "platform": { + "description": " Which platform this frame is from.\n\n This can override the platform for a single frame. Otherwise, the platform of the event is\n assumed. This can be used for multi-platform stack traces, such as in React Native.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "post_context": { + "description": " Source code of the lines after `lineno`.", + "default": null, + "type": [ + "array", + "null" + ], + "items": { + "type": [ + "string", + "null" + ] + } + }, + "pre_context": { + "description": " Source code leading up to `lineno`.", + "default": null, + "type": [ + "array", + "null" + ], + "items": { + "type": [ + "string", + "null" + ] + } + }, + "raw_function": { + "description": " A raw (but potentially truncated) function value.\n\n The original function name, if the function name is shortened or demangled. Sentry shows the\n raw function when clicking on the shortened one in the UI.\n\n If this has the same value as `function` it's best to be omitted. This exists because on\n many platforms the function itself contains additional information like overload specifies\n or a lot of generics which can make it exceed the maximum limit we provide for the field.\n In those cases then we cannot reliably trim down the function any more at a later point\n because the more valuable information has been removed.\n\n The logic to be applied is that an intelligently trimmed function name should be stored in\n `function` and the value before trimming is stored in this field instead. However also this\n field will be capped at 256 characters at the moment which often means that not the entire\n original value can be stored.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "stack_start": { + "description": " Marks this frame as the bottom of a chained stack trace.\n\n Stack traces from asynchronous code consist of several sub traces that are chained together\n into one large list. This flag indicates the root function of a chained stack trace.\n Depending on the runtime and thread, this is either the `main` function or a thread base\n stub.\n\n This field should only be specified when true.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "symbol": { + "description": " Potentially mangled name of the symbol as it appears in an executable.\n\n This is different from a function name by generally being the mangled\n name that appears natively in the binary. This is relevant for languages\n like Swift, C++ or Rust.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "symbol_addr": { + "description": " (C/C++/Native) Start address of the frame's function.\n\n We use the instruction address for symbolication, but this can be used to calculate\n an instruction offset automatically.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Addr" + }, + { + "type": "null" + } + ] + }, + "vars": { + "description": " Mapping of local variables and expression names that were available in this frame.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/FrameVars" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "FrameVars": { + "description": " Frame local variables.", + "anyOf": [ + { + "type": "object", + "additionalProperties": true + } + ] + }, + "Geo": { + "description": " Geographical location of the end user or device.", + "anyOf": [ + { + "type": "object", + "properties": { + "city": { + "description": " Human readable city name.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "country_code": { + "description": " Two-letter country code (ISO 3166-1 alpha-2).", + "default": null, + "type": [ + "string", + "null" + ] + }, + "region": { + "description": " Human readable region name or code.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "subdivision": { + "description": " Human readable subdivision name.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "GpuContext": { + "description": " GPU information.\n\n Example:\n\n ```json\n \"gpu\": {\n \"name\": \"AMD Radeon Pro 560\",\n \"vendor_name\": \"Apple\",\n \"memory_size\": 4096,\n \"api_type\": \"Metal\",\n \"multi_threaded_rendering\": true,\n \"version\": \"Metal\",\n \"npot_support\": \"Full\"\n }\n ```", + "anyOf": [ + { + "type": "object", + "properties": { + "api_type": { + "description": " The device low-level API type.\n\n Examples: `\"Apple Metal\"` or `\"Direct3D11\"`", + "default": null, + "type": [ + "string", + "null" + ] + }, + "graphics_shader_level": { + "description": " Approximate \"shader capability\" level of the graphics device.\n\n For Example: Shader Model 2.0, OpenGL ES 3.0, Metal / OpenGL ES 3.1, 27 (unknown)", + "default": null, + "type": [ + "string", + "null" + ] + }, + "id": { + "description": " The PCI identifier of the graphics device.", + "default": null + }, + "max_texture_size": { + "description": " Largest size of a texture that is supported by the graphics hardware.\n\n For Example: 16384", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "memory_size": { + "description": " The total GPU memory available in Megabytes.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "multi_threaded_rendering": { + "description": " Whether the GPU has multi-threaded rendering or not.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "name": { + "description": " The name of the graphics device.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "npot_support": { + "description": " The Non-Power-Of-Two support.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "supports_compute_shaders": { + "description": " Whether compute shaders are available on the device.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "supports_draw_call_instancing": { + "description": " Whether GPU draw call instancing is supported.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "supports_geometry_shaders": { + "description": " Whether geometry shaders are available on the device.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "supports_ray_tracing": { + "description": " Whether ray tracing is available on the device.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "vendor_id": { + "description": " The PCI vendor identifier of the graphics device.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "vendor_name": { + "description": " The vendor name as reported by the graphics device.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "version": { + "description": " The Version of the graphics device.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "HeaderName": { + "description": " A \"into-string\" type that normalizes header names.", + "anyOf": [ + { + "type": "string" + } + ] + }, + "HeaderValue": { + "description": " A \"into-string\" type that normalizes header values.", + "anyOf": [ + { + "type": "string" + } + ] + }, + "Headers": { + "description": " A map holding headers.", + "anyOf": [ + { + "anyOf": [ + { + "type": "object", + "additionalProperties": { + "anyOf": [ + { + "$ref": "#/definitions/HeaderValue" + }, + { + "type": "null" + } + ] + } + }, + { + "type": "array", + "items": { + "type": [ + "array", + "null" + ], + "items": [ + { + "anyOf": [ + { + "$ref": "#/definitions/HeaderName" + }, + { + "type": "null" + } + ] + }, + { + "anyOf": [ + { + "$ref": "#/definitions/HeaderValue" + }, + { + "type": "null" + } + ] + } + ], + "maxItems": 2, + "minItems": 2 + } + } + ] + } + ] + }, + "InstructionAddrAdjustment": { + "description": "Controls the mechanism by which the `instruction_addr` of a [`Stacktrace`] [`Frame`] is adjusted.\n\nThe adjustment tries to transform *return addresses* to *call addresses* for symbolication. Typically, this adjustment needs to be done for all frames but the first, as the first frame is usually taken directly from the cpu context of a hardware exception or a suspended thread and the stack trace is created from that.\n\nWhen the stack walking implementation truncates frames from the top, `\"all\"` frames should be adjusted. In case the stack walking implementation already does the adjustment when producing stack frames, `\"none\"` should be used here.", + "type": "string", + "enum": [ + "auto", + "all_but_first", + "all", + "none" + ] + }, + "JsonLenientString": { + "description": " A \"into-string\" type of value. All non-string values are serialized as JSON.", + "anyOf": [ + { + "type": "string" + } + ] + }, + "JvmDebugImage": { + "description": " A debug image consisting of source files for a JVM based language.\n\n Examples:\n\n ```json\n {\n \"type\": \"jvm\",\n \"debug_id\": \"395835f4-03e0-4436-80d3-136f0749a893\"\n }\n ```", + "anyOf": [ + { + "type": "object", + "required": [ + "debug_id" + ], + "properties": { + "debug_id": { + "description": " Unique debug identifier of the bundle.", + "anyOf": [ + { + "$ref": "#/definitions/DebugId" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "Level": { + "description": "Severity level of an event or breadcrumb.", + "type": "string", + "enum": [ + "debug", + "info", + "warning", + "error", + "fatal" + ] + }, + "LinuxDistribution": { + "description": " Metadata for the Linux Distribution.", + "anyOf": [ + { + "type": "object", + "properties": { + "name": { + "description": " An index-able name that is stable for each distribution.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "pretty_name": { + "description": " A full rendering of name + version + release name (not available in all distributions).", + "default": null, + "type": [ + "string", + "null" + ] + }, + "version": { + "description": " The version of the distribution (missing in distributions with solely rolling release).", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "LockReason": { + "description": " Represents an instance of a held lock (java monitor object) in a thread.", + "anyOf": [ + { + "type": "object", + "required": [ + "type" + ], + "properties": { + "address": { + "description": " Address of the java monitor object.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "class_name": { + "description": " Class name of the java monitor object.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "package_name": { + "description": " Package name of the java monitor object.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "thread_id": { + "description": " Thread ID that's holding the lock.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/ThreadId" + }, + { + "type": "null" + } + ] + }, + "type": { + "description": " Type of lock on the thread with available options being blocked, waiting, sleeping and locked.", + "anyOf": [ + { + "$ref": "#/definitions/LockReasonType" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "LockReasonType": { + "description": "Possible lock types responsible for a thread's blocked state", + "type": "string", + "enum": [ + "locked", + "waiting", + "sleeping", + "blocked" + ] + }, + "LogEntry": { + "description": " A log entry message.\n\n A log message is similar to the `message` attribute on the event itself but\n can additionally hold optional parameters.\n\n ```json\n {\n \"logentry\": {\n \"message\": \"My raw message with interpreted strings like %s\",\n \"params\": [\"this\"]\n }\n }\n ```\n\n ```json\n {\n \"logentry\": {\n \"message\": \"My raw message with interpreted strings like {foo}\",\n \"params\": {\"foo\": \"this\"}\n }\n }\n ```", + "anyOf": [ + { + "type": "object", + "properties": { + "formatted": { + "description": " The formatted message. If `message` and `params` are given, Sentry\n will attempt to backfill `formatted` if empty.\n\n It must not exceed 8192 characters. Longer messages will be truncated.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Message" + }, + { + "type": "null" + } + ] + }, + "message": { + "description": " The log message with parameter placeholders.\n\n This attribute is primarily used for grouping related events together into issues.\n Therefore this really should just be a string template, i.e. `Sending %d requests` instead\n of `Sending 9999 requests`. The latter is much better at home in `formatted`.\n\n It must not exceed 8192 characters. Longer messages will be truncated.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Message" + }, + { + "type": "null" + } + ] + }, + "params": { + "description": " Parameters to be interpolated into the log message. This can be an array of positional\n parameters as well as a mapping of named arguments to their values.", + "default": null + } + }, + "additionalProperties": false + } + ] + }, + "MachException": { + "description": " Mach exception information.", + "anyOf": [ + { + "type": "object", + "properties": { + "code": { + "description": " The mach exception code.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "exception": { + "description": " The mach exception type.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "int64" + }, + "name": { + "description": " Optional name of the mach exception.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "subcode": { + "description": " The mach exception subcode.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + } + }, + "additionalProperties": false + } + ] + }, + "Mechanism": { + "description": " The mechanism by which an exception was generated and handled.\n\n The exception mechanism is an optional field residing in the [exception](#typedef-Exception).\n It carries additional information about the way the exception was created on the target system.\n This includes general exception values obtained from the operating system or runtime APIs, as\n well as mechanism-specific values.", + "anyOf": [ + { + "type": "object", + "required": [ + "type" + ], + "properties": { + "data": { + "description": " Arbitrary extra data that might help the user understand the error thrown by this mechanism.", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": true + }, + "description": { + "description": " Optional human-readable description of the error mechanism.\n\n May include a possible hint on how to solve this error.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "exception_id": { + "description": " An optional numeric value providing an ID for the exception relative to this specific event.\n It is referenced by the `parent_id` to reconstruct the logical tree of exceptions in an\n exception group.\n\n This should contain an unsigned integer value starting with `0` for the last exception in\n the exception values list, then `1` for the previous exception, etc.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "handled": { + "description": " Flag indicating whether this exception was handled.\n\n This is a best-effort guess at whether the exception was handled by user code or not. For\n example:\n\n - Exceptions leading to a 500 Internal Server Error or to a hard process crash are\n `handled=false`, as the SDK typically has an integration that automatically captures the\n error.\n\n - Exceptions captured using `capture_exception` (called from user code) are `handled=true`\n as the user explicitly captured the exception (and therefore kind of handled it)", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "help_link": { + "description": " Link to online resources describing this error.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "is_exception_group": { + "description": " An optional boolean value, set `true` when the exception is the platform-specific exception\n group type. Defaults to `false`.\n\n For example, exceptions of type `ExceptionGroup` (Python), `AggregateException` (.NET), and\n `AggregateError` (JavaScript) should have `\"is_exception_group\": true`. Other exceptions\n can omit this field.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "meta": { + "description": " Operating system or runtime meta information.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/MechanismMeta" + }, + { + "type": "null" + } + ] + }, + "parent_id": { + "description": " An optional numeric value pointing at the `exception_id` that is the direct parent of this\n exception, used to reconstruct the logical tree of exceptions in an exception group.\n\n The last exception in the exception values list should omit this field, because it is the\n root exception and thus has no parent.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "source": { + "description": " An optional string value describing the source of the exception.\n\n For chained exceptions, this should contain the platform-specific name of the property or\n attribute (on the parent exception) that this exception was acquired from. In the case of\n an array, it should include the zero-based array index as well.\n\n - Python Examples: `\"__context__\"`, `\"__cause__\"`, `\"exceptions[0]\"`, `\"exceptions[1]\"`\n\n - .NET Examples: `\"InnerException\"`, `\"InnerExceptions[0]\"`, `\"InnerExceptions[1]\"`\n\n - JavaScript Examples: `\"cause\"`, `\"errors[0]\"`, `\"errors[1]\"`", + "default": null, + "type": [ + "string", + "null" + ] + }, + "synthetic": { + "description": " If this is set then the exception is not a real exception but some\n form of synthetic error for instance from a signal handler, a hard\n segfault or similar where type and value are not useful for grouping\n or display purposes.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "type": { + "description": " Mechanism type (required).\n\n Required unique identifier of this mechanism determining rendering and processing of the\n mechanism data.\n\n In the Python SDK this is merely the name of the framework integration that produced the\n exception, while for native it is e.g. `\"minidump\"` or `\"applecrashreport\"`.", + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "MechanismMeta": { + "description": " Operating system or runtime meta information to an exception mechanism.\n\n The mechanism metadata usually carries error codes reported by the runtime or operating system,\n along with a platform-dependent interpretation of these codes. SDKs can safely omit code names\n and descriptions for well-known error codes, as it will be filled out by Sentry. For\n proprietary or vendor-specific error codes, adding these values will give additional\n information to the user.", + "anyOf": [ + { + "type": "object", + "properties": { + "errno": { + "description": " Optional ISO C standard error code.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/CError" + }, + { + "type": "null" + } + ] + }, + "mach_exception": { + "description": " A Mach Exception on Apple systems comprising a code triple and optional descriptions.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/MachException" + }, + { + "type": "null" + } + ] + }, + "ns_error": { + "description": " An NSError on Apple systems comprising code and signal.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/NsError" + }, + { + "type": "null" + } + ] + }, + "signal": { + "description": " Information on the POSIX signal.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/PosixSignal" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "Message": { + "anyOf": [ + { + "type": "string" + } + ] + }, + "MonitorContext": { + "description": " Monitor information.", + "anyOf": [ + { + "type": "object", + "additionalProperties": true + } + ] + }, + "NativeDebugImage": { + "description": " A generic (new-style) native platform debug information file.\n\n The `type` key must be one of:\n\n - `macho`\n - `elf`: ELF images are used on Linux platforms. Their structure is identical to other native images.\n - `pe`\n\n Examples:\n\n ```json\n {\n \"type\": \"elf\",\n \"code_id\": \"68220ae2c65d65c1b6aaa12fa6765a6ec2f5f434\",\n \"code_file\": \"/lib/x86_64-linux-gnu/libgcc_s.so.1\",\n \"debug_id\": \"e20a2268-5dc6-c165-b6aa-a12fa6765a6e\",\n \"image_addr\": \"0x7f5140527000\",\n \"image_size\": 90112,\n \"image_vmaddr\": \"0x40000\",\n \"arch\": \"x86_64\"\n }\n ```\n\n ```json\n {\n \"type\": \"pe\",\n \"code_id\": \"57898e12145000\",\n \"code_file\": \"C:\\\\Windows\\\\System32\\\\dbghelp.dll\",\n \"debug_id\": \"9c2a902b-6fdf-40ad-8308-588a41d572a0-1\",\n \"debug_file\": \"dbghelp.pdb\",\n \"image_addr\": \"0x70850000\",\n \"image_size\": \"1331200\",\n \"image_vmaddr\": \"0x40000\",\n \"arch\": \"x86\"\n }\n ```\n\n ```json\n {\n \"type\": \"macho\",\n \"debug_id\": \"84a04d24-0e60-3810-a8c0-90a65e2df61a\",\n \"debug_file\": \"libDiagnosticMessagesClient.dylib\",\n \"code_file\": \"/usr/lib/libDiagnosticMessagesClient.dylib\",\n \"image_addr\": \"0x7fffe668e000\",\n \"image_size\": 8192,\n \"image_vmaddr\": \"0x40000\",\n \"arch\": \"x86_64\",\n }\n ```", + "anyOf": [ + { + "type": "object", + "required": [ + "code_file", + "debug_id" + ], + "properties": { + "arch": { + "description": " CPU architecture target.\n\n Architecture of the module. If missing, this will be backfilled by Sentry.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "code_file": { + "description": " Path and name of the image file (required).\n\n The absolute path to the dynamic library or executable. This helps to locate the file if it is missing on Sentry.\n\n - `pe`: The code file should be provided to allow server-side stack walking of binary crash reports, such as Minidumps.", + "anyOf": [ + { + "$ref": "#/definitions/NativeImagePath" + }, + { + "type": "null" + } + ] + }, + "code_id": { + "description": " Optional identifier of the code file.\n\n - `elf`: If the program was compiled with a relatively recent compiler, this should be the hex representation of the `NT_GNU_BUILD_ID` program header (type `PT_NOTE`), or the value of the `.note.gnu.build-id` note section (type `SHT_NOTE`). Otherwise, leave this value empty.\n\n Certain symbol servers use the code identifier to locate debug information for ELF images, in which case this field should be included if possible.\n\n - `pe`: Identifier of the executable or DLL. It contains the values of the `time_date_stamp` from the COFF header and `size_of_image` from the optional header formatted together into a hex string using `%08x%X` (note that the second value is not padded):\n\n ```text\n time_date_stamp: 0x5ab38077\n size_of_image: 0x9000\n code_id: 5ab380779000\n ```\n\n The code identifier should be provided to allow server-side stack walking of binary crash reports, such as Minidumps.\n\n\n - `macho`: Identifier of the dynamic library or executable. It is the value of the `LC_UUID` load command in the Mach header, formatted as UUID. Can be empty for Mach images, as it is equivalent to the debug identifier.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/CodeId" + }, + { + "type": "null" + } + ] + }, + "debug_checksum": { + "description": " The optional checksum of the debug companion file.\n\n - `pe_dotnet`: This is the hash algorithm and hex-formatted checksum of the associated PDB file.\n This should have the format `$algorithm:$hash`, for example `SHA256:aabbccddeeff...`.\n\n See: ", + "default": null, + "type": [ + "string", + "null" + ] + }, + "debug_file": { + "description": " Path and name of the debug companion file.\n\n - `elf`: Name or absolute path to the file containing stripped debug information for this image. This value might be _required_ to retrieve debug files from certain symbol servers.\n\n - `pe`: Name of the PDB file containing debug information for this image. This value is often required to retrieve debug files from specific symbol servers.\n\n - `macho`: Name or absolute path to the dSYM file containing debug information for this image. This value might be required to retrieve debug files from certain symbol servers.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/NativeImagePath" + }, + { + "type": "null" + } + ] + }, + "debug_id": { + "description": " Unique debug identifier of the image.\n\n - `elf`: Debug identifier of the dynamic library or executable. If a code identifier is available, the debug identifier is the little-endian UUID representation of the first 16-bytes of that\n identifier. Spaces are inserted for readability, note the byte order of the first fields:\n\n ```text\n code id: f1c3bcc0 2798 65fe 3058 404b2831d9e6 4135386c\n debug id: c0bcc3f1-9827-fe65-3058-404b2831d9e6\n ```\n\n If no code id is available, the debug id should be computed by XORing the first 4096 bytes of the `.text` section in 16-byte chunks, and representing it as a little-endian UUID (again swapping the byte order).\n\n - `pe`: `signature` and `age` of the PDB file. Both values can be read from the CodeView PDB70 debug information header in the PE. The value should be represented as little-endian UUID, with the age appended at the end. Note that the byte order of the UUID fields must be swapped (spaces inserted for readability):\n\n ```text\n signature: f1c3bcc0 2798 65fe 3058 404b2831d9e6\n age: 1\n debug_id: c0bcc3f1-9827-fe65-3058-404b2831d9e6-1\n ```\n\n - `macho`: Identifier of the dynamic library or executable. It is the value of the `LC_UUID` load command in the Mach header, formatted as UUID.", + "anyOf": [ + { + "$ref": "#/definitions/DebugId" + }, + { + "type": "null" + } + ] + }, + "image_addr": { + "description": " Starting memory address of the image (required).\n\n Memory address, at which the image is mounted in the virtual address space of the process. Should be a string in hex representation prefixed with `\"0x\"`.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Addr" + }, + { + "type": "null" + } + ] + }, + "image_size": { + "description": " Size of the image in bytes (required).\n\n The size of the image in virtual memory. If missing, Sentry will assume that the image spans up to the next image, which might lead to invalid stack traces.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "image_vmaddr": { + "description": " Loading address in virtual memory.\n\n Preferred load address of the image in virtual memory, as declared in the headers of the\n image. When loading an image, the operating system may still choose to place it at a\n different address.\n\n Symbols and addresses in the native image are always relative to the start of the image and do not consider the preferred load address. It is merely a hint to the loader.\n\n - `elf`/`macho`: If this value is non-zero, all symbols and addresses declared in the native image start at this address, rather than 0. By contrast, Sentry deals with addresses relative to the start of the image. For example, with `image_vmaddr: 0x40000`, a symbol located at `0x401000` has a relative address of `0x1000`.\n\n Relative addresses used in Apple Crash Reports and `addr2line` are usually in the preferred address space, and not relative address space.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Addr" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "NativeImagePath": { + "description": " A type for strings that are generally paths, might contain system user names, but still cannot\n be stripped liberally because it would break processing for certain platforms.\n\n Those strings get special treatment in our PII processor to avoid stripping the basename.", + "anyOf": [ + { + "type": "string" + } + ] + }, + "NelContext": { + "description": " Contains NEL report information.\n\n Network Error Logging (NEL) is a browser feature that allows reporting of failed network\n requests from the client side. See the following resources for more information:\n\n - [W3C Editor's Draft](https://w3c.github.io/network-error-logging/)\n - [MDN](https://developer.mozilla.org/en-US/docs/Web/HTTP/Network_Error_Logging)", + "anyOf": [ + { + "type": "object", + "properties": { + "elapsed_time": { + "description": " The number of milliseconds between the start of the resource fetch and when it was aborted by the user agent.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "error_type": { + "description": " If request failed, the type of its network error. If request succeeded, \"ok\".", + "default": null, + "type": [ + "string", + "null" + ] + }, + "phase": { + "description": " If request failed, the phase of its network error. If request succeeded, \"application\".", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/NetworkReportPhases" + }, + { + "type": "null" + } + ] + }, + "sampling_fraction": { + "description": " The sampling rate.", + "default": null, + "type": [ + "number", + "null" + ], + "format": "double" + }, + "server_ip": { + "description": " Server IP where the requests was sent to.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/String" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "NetworkReportPhases": { + "description": " Describes which phase the error occurred in.", + "anyOf": [ + { + "type": "object", + "additionalProperties": false + }, + { + "type": "object", + "additionalProperties": false + }, + { + "type": "object", + "additionalProperties": false + }, + { + "type": "string" + } + ] + }, + "NsError": { + "description": " NSError informaiton.", + "anyOf": [ + { + "type": "object", + "properties": { + "code": { + "description": " The error code.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "int64" + }, + "domain": { + "description": " A string containing the error domain.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "OsContext": { + "description": " Operating system information.\n\n OS context describes the operating system on which the event was created. In web contexts, this\n is the operating system of the browser (generally pulled from the User-Agent string).", + "anyOf": [ + { + "type": "object", + "properties": { + "build": { + "description": " Internal build number of the operating system.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "distribution": { + "description": " Meta-data for the Linux Distribution.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/LinuxDistribution" + }, + { + "type": "null" + } + ] + }, + "kernel_version": { + "description": " Current kernel version.\n\n This is typically the entire output of the `uname` syscall.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "name": { + "description": " Name of the operating system.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "raw_description": { + "description": " Unprocessed operating system info.\n\n An unprocessed description string obtained by the operating system. For some well-known\n runtimes, Sentry will attempt to parse `name` and `version` from this string, if they are\n not explicitly given.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "rooted": { + "description": " Indicator if the OS is rooted (mobile mostly).", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "version": { + "description": " Version of the operating system.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "OtelContext": { + "description": " OpenTelemetry Context\n\n If an event has this context, it was generated from an OpenTelemetry signal (trace, metric, log).", + "anyOf": [ + { + "type": "object", + "properties": { + "attributes": { + "description": " Attributes of the OpenTelemetry span that maps to a Sentry event.\n\n ", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": true + }, + "resource": { + "description": " Information about an OpenTelemetry resource.\n\n ", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": true + } + }, + "additionalProperties": false + } + ] + }, + "PerformanceScoreContext": { + "description": " Performance Score context.\n\n The performance score context contains the version of the\n profile used to calculate the performance score.", + "anyOf": [ + { + "type": "object", + "properties": { + "score_profile_version": { + "description": " The performance score profile version.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "PosixSignal": { + "description": " POSIX signal with optional extended data.\n\n On Apple systems, signals also carry a code in addition to the signal number describing the\n signal in more detail. On Linux, this code does not exist.", + "anyOf": [ + { + "type": "object", + "properties": { + "code": { + "description": " An optional signal code present on Apple systems.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "int64" + }, + "code_name": { + "description": " Optional name of the errno constant.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "name": { + "description": " Optional name of the errno constant.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "number": { + "description": " The POSIX signal number.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "int64" + } + }, + "additionalProperties": false + } + ] + }, + "ProfileContext": { + "description": " Profile context", + "anyOf": [ + { + "type": "object", + "properties": { + "profile_id": { + "description": " The profile ID.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/EventId" + }, + { + "type": "null" + } + ] + }, + "profiler_id": { + "description": " The profiler ID.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/EventId" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "ProguardDebugImage": { + "description": " Proguard mapping file.\n\n Proguard images refer to `mapping.txt` files generated when Proguard obfuscates function names. The Java SDK integrations assign this file a unique identifier, which has to be included in the list of images.", + "anyOf": [ + { + "type": "object", + "required": [ + "uuid" + ], + "properties": { + "uuid": { + "description": " UUID computed from the file contents, assigned by the Java SDK.", + "type": [ + "string", + "null" + ], + "format": "uuid" + } + }, + "additionalProperties": false + } + ] + }, + "RawStacktrace": { + "description": " A stack trace of a single thread.\n\n A stack trace contains a list of frames, each with various bits (most optional) describing the\n context of that frame. Frames should be sorted from oldest to newest.\n\n For the given example program written in Python:\n\n ```python\n def foo():\n my_var = 'foo'\n raise ValueError()\n\n def main():\n foo()\n ```\n\n A minimalistic stack trace for the above program in the correct order:\n\n ```json\n {\n \"frames\": [\n {\"function\": \"main\"},\n {\"function\": \"foo\"}\n ]\n }\n ```\n\n The top frame fully symbolicated with five lines of source context:\n\n ```json\n {\n \"frames\": [{\n \"in_app\": true,\n \"function\": \"myfunction\",\n \"abs_path\": \"/real/file/name.py\",\n \"filename\": \"file/name.py\",\n \"lineno\": 3,\n \"vars\": {\n \"my_var\": \"'value'\"\n },\n \"pre_context\": [\n \"def foo():\",\n \" my_var = 'foo'\",\n ],\n \"context_line\": \" raise ValueError()\",\n \"post_context\": [\n \"\",\n \"def main():\"\n ],\n }]\n }\n ```\n\n A minimal native stack trace with register values. Note that the `package` event attribute must\n be \"native\" for these frames to be symbolicated.\n\n ```json\n {\n \"frames\": [\n {\"instruction_addr\": \"0x7fff5bf3456c\"},\n {\"instruction_addr\": \"0x7fff5bf346c0\"},\n ],\n \"registers\": {\n \"rip\": \"0x00007ff6eef54be2\",\n \"rsp\": \"0x0000003b710cd9e0\"\n }\n }\n ```", + "anyOf": [ + { + "type": "object", + "required": [ + "frames" + ], + "properties": { + "frames": { + "description": " Required. A non-empty list of stack frames. The list is ordered from caller to callee, or\n oldest to youngest. The last frame is the one creating the exception.", + "type": [ + "array", + "null" + ], + "items": { + "anyOf": [ + { + "$ref": "#/definitions/Frame" + }, + { + "type": "null" + } + ] + } + }, + "instruction_addr_adjustment": { + "description": " Optional. A flag that indicates if, and how, `instruction_addr` values need to be adjusted\n before they are symbolicated.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/InstructionAddrAdjustment" + }, + { + "type": "null" + } + ] + }, + "lang": { + "description": " The language of the stacktrace.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "registers": { + "description": " Register values of the thread (top frame).\n\n A map of register names and their values. The values should contain the actual register\n values of the thread, thus mapping to the last frame in the list.", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": { + "anyOf": [ + { + "$ref": "#/definitions/RegVal" + }, + { + "type": "null" + } + ] + } + }, + "snapshot": { + "description": " Indicates that this stack trace is a snapshot triggered by an external signal.\n\n If this field is `false`, then the stack trace points to the code that caused this stack\n trace to be created. This can be the location of a raised exception, as well as an exception\n or signal handler.\n\n If this field is `true`, then the stack trace was captured as part of creating an unrelated\n event. For example, a thread other than the crashing thread, or a stack trace computed as a\n result of an external kill signal.", + "default": null, + "type": [ + "boolean", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "RegVal": { + "type": "string" + }, + "ReplayContext": { + "description": " Replay context.\n\n The replay context contains the replay_id of the session replay if the event\n occurred during a replay. The replay_id is added onto the dynamic sampling context\n on the javascript SDK which propagates it through the trace. In relay, we take\n this value from the DSC and create a context which contains only the replay_id\n This context is never set on the client for events, only on relay.", + "anyOf": [ + { + "type": "object", + "properties": { + "replay_id": { + "description": " The replay ID.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/EventId" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "Request": { + "description": " Http request information.\n\n The Request interface contains information on a HTTP request related to the event. In client\n SDKs, this can be an outgoing request, or the request that rendered the current web page. On\n server SDKs, this could be the incoming web request that is being handled.\n\n The data variable should only contain the request body (not the query string). It can either be\n a dictionary (for standard HTTP requests) or a raw request body.\n\n ### Ordered Maps\n\n In the Request interface, several attributes can either be declared as string, object, or list\n of tuples. Sentry attempts to parse structured information from the string representation in\n such cases.\n\n Sometimes, keys can be declared multiple times, or the order of elements matters. In such\n cases, use the tuple representation over a plain object.\n\n Example of request headers as object:\n\n ```json\n {\n \"content-type\": \"application/json\",\n \"accept\": \"application/json, application/xml\"\n }\n ```\n\n Example of the same headers as list of tuples:\n\n ```json\n [\n [\"content-type\", \"application/json\"],\n [\"accept\", \"application/json\"],\n [\"accept\", \"application/xml\"]\n ]\n ```\n\n Example of a fully populated request object:\n\n ```json\n {\n \"request\": {\n \"method\": \"POST\",\n \"url\": \"http://absolute.uri/foo\",\n \"query_string\": \"query=foobar&page=2\",\n \"data\": {\n \"foo\": \"bar\"\n },\n \"cookies\": \"PHPSESSID=298zf09hf012fh2; csrftoken=u32t4o3tb3gg43; _gat=1;\",\n \"headers\": {\n \"content-type\": \"text/html\"\n },\n \"env\": {\n \"REMOTE_ADDR\": \"192.168.0.1\"\n }\n }\n }\n ```", + "anyOf": [ + { + "type": "object", + "properties": { + "api_target": { + "description": " The API target/specification that made the request.\n\n Values can be `graphql`, `rest`, etc.\n\n The data field should contain the request and response bodies based on its target specification.\n\n This information can be used for better data scrubbing and normalization.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "body_size": { + "description": " HTTP request body size.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "cookies": { + "description": " The cookie values.\n\n Can be given unparsed as string, as dictionary, or as a list of tuples.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Cookies" + }, + { + "type": "null" + } + ] + }, + "data": { + "description": " Request data in any format that makes sense.\n\n SDKs should discard large and binary bodies by default. Can be given as a string or\n structural data of any format.", + "default": null + }, + "env": { + "description": " Server environment data, such as CGI/WSGI.\n\n A dictionary containing environment information passed from the server. This is where\n information such as CGI/WSGI/Rack keys go that are not HTTP headers.\n\n Sentry will explicitly look for `REMOTE_ADDR` to extract an IP address.", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": true + }, + "fragment": { + "description": " The fragment of the request URI.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "headers": { + "description": " A dictionary of submitted headers.\n\n If a header appears multiple times it, needs to be merged according to the HTTP standard\n for header merging. Header names are treated case-insensitively by Sentry.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Headers" + }, + { + "type": "null" + } + ] + }, + "inferred_content_type": { + "description": " The inferred content type of the request payload.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "method": { + "description": " HTTP request method.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "protocol": { + "description": " HTTP protocol.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "query_string": { + "description": " The query string component of the URL.\n\n Can be given as unparsed string, dictionary, or list of tuples.\n\n If the query string is not declared and part of the `url`, Sentry moves it to the\n query string.", + "default": null, + "anyOf": [ + { + "anyOf": [ + { + "type": "string" + }, + { + "anyOf": [ + { + "type": "object", + "additionalProperties": { + "type": [ + "string", + "null" + ] + } + }, + { + "type": "array", + "items": { + "type": [ + "array", + "null" + ], + "items": [ + { + "type": [ + "string", + "null" + ] + }, + { + "type": [ + "string", + "null" + ] + } + ], + "maxItems": 2, + "minItems": 2 + } + } + ] + } + ] + }, + { + "type": "null" + } + ] + }, + "url": { + "description": " The URL of the request if available.\n\nThe query string can be declared either as part of the `url`, or separately in `query_string`.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "ResponseContext": { + "description": " Response interface that contains information on a HTTP response related to the event.\n\n The data variable should only contain the response body. It can either be\n a dictionary (for standard HTTP responses) or a raw response body.", + "anyOf": [ + { + "type": "object", + "properties": { + "body_size": { + "description": " HTTP response body size.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "cookies": { + "description": " The cookie values.\n\n Can be given unparsed as string, as dictionary, or as a list of tuples.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Cookies" + }, + { + "type": "null" + } + ] + }, + "data": { + "description": " Response data in any format that makes sense.\n\n SDKs should discard large and binary bodies by default. Can be given as a string or\n structural data of any format.", + "default": null + }, + "headers": { + "description": " A dictionary of submitted headers.\n\n If a header appears multiple times it, needs to be merged according to the HTTP standard\n for header merging. Header names are treated case-insensitively by Sentry.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Headers" + }, + { + "type": "null" + } + ] + }, + "inferred_content_type": { + "description": " The inferred content type of the response payload.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "status_code": { + "description": " HTTP status code.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + } + }, + "additionalProperties": false + } + ] + }, + "Route": { + "description": " The route in the application, set by React Native SDK.", + "anyOf": [ + { + "type": "object", + "properties": { + "name": { + "description": " The name of the route.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "params": { + "description": " Parameters assigned to this route.", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": true + } + }, + "additionalProperties": false + } + ] + }, + "RuntimeContext": { + "description": " Runtime information.\n\n Runtime context describes a runtime in more detail. Typically, this context is present in\n `contexts` multiple times if multiple runtimes are involved (for instance, if you have a\n JavaScript application running on top of JVM).", + "anyOf": [ + { + "type": "object", + "properties": { + "build": { + "description": " Application build string, if it is separate from the version.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "name": { + "description": " Runtime name.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "raw_description": { + "description": " Unprocessed runtime info.\n\n An unprocessed description string obtained by the runtime. For some well-known runtimes,\n Sentry will attempt to parse `name` and `version` from this string, if they are not\n explicitly given.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "version": { + "description": " Runtime version string.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "SourceMapDebugImage": { + "description": " A debug image pointing to a source map.\n\n Examples:\n\n ```json\n {\n \"type\": \"sourcemap\",\n \"code_file\": \"https://example.com/static/js/main.min.js\",\n \"debug_id\": \"395835f4-03e0-4436-80d3-136f0749a893\"\n }\n ```\n\n **Note:** Stack frames and the correlating entries in the debug image here\n for `code_file`/`abs_path` are not PII stripped as they need to line up\n perfectly for source map processing.", + "anyOf": [ + { + "type": "object", + "required": [ + "code_file", + "debug_id" + ], + "properties": { + "code_file": { + "description": " Path and name of the image file as URL. (required).\n\n The absolute path to the minified JavaScript file. This helps to correlate the file to the stack trace.", + "type": [ + "string", + "null" + ] + }, + "debug_file": { + "description": " Path and name of the associated source map.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "debug_id": { + "description": " Unique debug identifier of the source map.", + "anyOf": [ + { + "$ref": "#/definitions/DebugId" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "SpanData": { + "description": " Arbitrary additional data on a span.\n\n Besides arbitrary user data, this type also contains SDK-provided fields used by the\n product (see ).", + "anyOf": [ + { + "type": "object", + "properties": { + "ai.completion_tokens.used": { + "description": " The output tokens used by an LLM call (the ones the LLM actually generated)", + "default": null + }, + "ai.input_messages": { + "description": " The input messages to an AI model call", + "default": null + }, + "ai.model_id": { + "description": " The Model ID of an AI pipeline, e.g., gpt-4", + "default": null + }, + "ai.pipeline.name": { + "description": " The 'name' field of the ancestor span with op ai.pipeline.*", + "default": null + }, + "ai.prompt_tokens.used": { + "description": " The input tokens used by an LLM call (usually cheaper than output tokens)", + "default": null + }, + "ai.responses": { + "description": " The responses to an AI model call", + "default": null + }, + "ai.total_tokens.used": { + "description": " The total tokens that were used by an LLM call", + "default": null + }, + "app_start_type": { + "description": " Mobile app start variant.\n\n Can be either \"cold\" or \"warm\".", + "default": null + }, + "browser.name": { + "description": " The client's browser name.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "cache.hit": { + "description": " Whether cache was hit or miss on a read operation.", + "default": null + }, + "cache.item_size": { + "description": " The size of the cache item.", + "default": null + }, + "cache.key": { + "description": " The name of the cache key.", + "default": null + }, + "client.address": { + "description": " The client's IP address.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "code.filepath": { + "description": " The source code file name that identifies the code unit as uniquely as possible.", + "default": null + }, + "code.function": { + "description": " The method or function name, or equivalent.\n\n Usually rightmost part of the code unit's name.", + "default": null + }, + "code.lineno": { + "description": " The line number in `code.filepath` best representing the operation.", + "default": null + }, + "code.namespace": { + "description": " The \"namespace\" within which `code.function` is defined.\n\n Usually the qualified class or module name, such that\n `code.namespace + some separator + code.function`\n form a unique identifier for the code unit.", + "default": null + }, + "db.operation": { + "description": " The name of the operation being executed.\n\n E.g. the MongoDB command name such as findAndModify, or the SQL keyword.\n Based on [OpenTelemetry's call level db attributes](https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/trace/semantic_conventions/database.md#call-level-attributes).", + "default": null + }, + "db.system": { + "description": " An identifier for the database management system (DBMS) product being used.\n\n See [OpenTelemetry docs for a list of well-known identifiers](https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/trace/semantic_conventions/database.md#notes-and-well-known-identifiers-for-dbsystem).", + "default": null + }, + "frames.delay": { + "default": null + }, + "http.decoded_response_content_length": { + "description": " The decoded body size of the response (in bytes).", + "default": null + }, + "http.request_method": { + "description": " The HTTP method used.", + "default": null + }, + "http.response.status_code": { + "description": " The status HTTP response.", + "default": null + }, + "http.response_content_length": { + "description": " The encoded body size of the response (in bytes).", + "default": null + }, + "http.response_transfer_size": { + "description": " The transfer size of the response (in bytes).", + "default": null + }, + "messaging.destination.name": { + "default": null, + "type": [ + "string", + "null" + ] + }, + "messaging.message.body.size": { + "description": " Message Body Size", + "default": null + }, + "messaging.message.id": { + "description": " Message ID", + "default": null, + "type": [ + "string", + "null" + ] + }, + "messaging.message.receive.latency": { + "description": " Message Receive Latency", + "default": null + }, + "messaging.message.retry.count": { + "description": " Message Retry Count", + "default": null + }, + "previousRoute": { + "description": " The previous route in the application\n\n Set by React Native SDK.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Route" + }, + { + "type": "null" + } + ] + }, + "resource.render_blocking_status": { + "description": " The render blocking status of the resource.", + "default": null + }, + "route": { + "description": " The current route in the application.\n\n Set by React Native SDK.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Route" + }, + { + "type": "null" + } + ] + }, + "sentry.environment": { + "description": " The sentry environment.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "sentry.frames.frozen": { + "description": " Frozen Frames", + "default": null + }, + "sentry.frames.slow": { + "description": " Slow Frames", + "default": null + }, + "sentry.frames.total": { + "description": " Total Frames", + "default": null + }, + "sentry.release": { + "description": " The release version of the project.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "sentry.replay.id": { + "description": " Replay ID", + "default": null + }, + "sentry.sdk.name": { + "description": " The sentry SDK (see [`crate::protocol::ClientSdkInfo`]).", + "default": null, + "type": [ + "string", + "null" + ] + }, + "sentry.sdk.version": { + "description": " The sentry SDK version (see [`crate::protocol::ClientSdkInfo`]).", + "default": null, + "type": [ + "string", + "null" + ] + }, + "sentry.segment.name": { + "description": " Name of the segment that this span belongs to (see `segment_id`).\n\n This corresponds to the transaction name in the transaction-based model.\n\n For INP spans, this is the route name where the interaction occurred.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "server.address": { + "description": " Name of the web server host.", + "default": null + }, + "thread.id": { + "description": " ID of thread from where the span originated.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/ThreadId" + }, + { + "type": "null" + } + ] + }, + "thread.name": { + "description": " Label identifying a thread from where the span originated.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "ui.component_name": { + "description": " Name of the UI component (e.g. React).", + "default": null + }, + "url.full": { + "description": " Absolute URL of a network resource.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "url.scheme": { + "description": " The URL scheme, e.g. `\"https\"`.", + "default": null + }, + "user": { + "description": " User Display", + "default": null + }, + "user_agent.original": { + "description": " Value of the HTTP User-Agent header sent by the client.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "SpanId": { + "description": " A 16-character hex string as described in the W3C trace context spec.", + "anyOf": [ + { + "type": "string" + } + ] + }, + "SpanStatus": { + "description": "Trace status.\n\nValues from Mapping to HTTP from ", + "type": "string", + "enum": [ + "ok", + "cancelled", + "unknown", + "invalid_argument", + "deadline_exceeded", + "not_found", + "already_exists", + "permission_denied", + "resource_exhausted", + "failed_precondition", + "aborted", + "out_of_range", + "unimplemented", + "internal_error", + "unavailable", + "data_loss", + "unauthenticated" + ] + }, + "Stacktrace": { + "anyOf": [ + { + "$ref": "#/definitions/RawStacktrace" + } + ] + }, + "String": { + "type": "string" + }, + "SystemSdkInfo": { + "description": " Holds information about the system SDK.\n\n This is relevant for iOS and other platforms that have a system\n SDK. Not to be confused with the client SDK.", + "anyOf": [ + { + "type": "object", + "properties": { + "sdk_name": { + "description": " The internal name of the SDK.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "version_major": { + "description": " The major version of the SDK as integer or 0.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "version_minor": { + "description": " The minor version of the SDK as integer or 0.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "version_patchlevel": { + "description": " The patch version of the SDK as integer or 0.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + } + }, + "additionalProperties": false + } + ] + }, + "TagEntry": { + "anyOf": [ + { + "type": "array", + "items": [ + { + "type": [ + "string", + "null" + ] + }, + { + "type": [ + "string", + "null" + ] + } + ], + "maxItems": 2, + "minItems": 2 + } + ] + }, + "Tags": { + "description": " Manual key/value tag pairs.", + "anyOf": [ + { + "anyOf": [ + { + "type": "object", + "additionalProperties": { + "type": [ + "string", + "null" + ] + } + }, + { + "type": "array", + "items": { + "anyOf": [ + { + "$ref": "#/definitions/TagEntry" + }, + { + "type": "null" + } + ] + } + } + ] + } + ] + }, + "Thread": { + "description": " A process thread of an event.\n\n The Threads Interface specifies threads that were running at the time an event happened. These threads can also contain stack traces.\n\n An event may contain one or more threads in an attribute named `threads`.\n\n The following example illustrates the threads part of the event payload and omits other attributes for simplicity.\n\n ```json\n {\n \"threads\": {\n \"values\": [\n {\n \"id\": \"0\",\n \"name\": \"main\",\n \"crashed\": true,\n \"stacktrace\": {}\n }\n ]\n }\n }\n ```", + "anyOf": [ + { + "type": "object", + "properties": { + "crashed": { + "description": " A flag indicating whether the thread crashed. Defaults to `false`.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "current": { + "description": " A flag indicating whether the thread was in the foreground. Defaults to `false`.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "held_locks": { + "description": " Represents a collection of locks (java monitor objects) held by a thread.\n\n A map of lock object addresses and their respective lock reason/details.", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": { + "anyOf": [ + { + "$ref": "#/definitions/LockReason" + }, + { + "type": "null" + } + ] + } + }, + "id": { + "description": " The ID of the thread. Typically a number or numeric string.\n\n Needs to be unique among the threads. An exception can set the `thread_id` attribute to cross-reference this thread.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/ThreadId" + }, + { + "type": "null" + } + ] + }, + "main": { + "description": " A flag indicating whether the thread was responsible for rendering the user interface.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "name": { + "description": " Display name of this thread.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "stacktrace": { + "description": " Stack trace containing frames of this exception.\n\n The thread that crashed with an exception should not have a stack trace, but instead, the `thread_id` attribute should be set on the exception and Sentry will connect the two.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Stacktrace" + }, + { + "type": "null" + } + ] + }, + "state": { + "description": " Thread state at the time of the crash.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "ThreadId": { + "description": " Represents a thread id.", + "anyOf": [ + { + "type": "integer", + "format": "uint64", + "minimum": 0.0 + }, + { + "type": "string" + } + ] + }, + "Timestamp": { + "description": "Can be a ISO-8601 formatted string or a unix timestamp in seconds (floating point values allowed).\n\nMust be UTC.", + "anyOf": [ + { + "type": "number", + "format": "double" + }, + { + "type": "string" + } + ] + }, + "TraceContext": { + "description": " Trace context", + "anyOf": [ + { + "type": "object", + "required": [ + "span_id", + "trace_id" + ], + "properties": { + "client_sample_rate": { + "description": " The client-side sample rate as reported in the envelope's `trace.sample_rate` header.\n\n The server takes this field from envelope headers and writes it back into the event. Clients\n should not ever send this value.", + "default": null, + "type": [ + "number", + "null" + ], + "format": "double" + }, + "data": { + "description": " Data of the trace's root span.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/SpanData" + }, + { + "type": "null" + } + ] + }, + "exclusive_time": { + "description": " The amount of time in milliseconds spent in this transaction span,\n excluding its immediate child spans.", + "default": null, + "type": [ + "number", + "null" + ], + "format": "double" + }, + "op": { + "description": " Span type (see `OperationType` docs).", + "default": null, + "type": [ + "string", + "null" + ] + }, + "origin": { + "description": " The origin of the trace indicates what created the trace (see [OriginType] docs).", + "default": null, + "type": [ + "string", + "null" + ] + }, + "parent_span_id": { + "description": " The ID of the span enclosing this span.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/SpanId" + }, + { + "type": "null" + } + ] + }, + "sampled": { + "description": " Track whether the trace connected to this event has been sampled entirely.\n\n This flag only applies to events with [`Error`] type that have an associated dynamic sampling context.", + "default": null, + "type": [ + "boolean", + "null" + ] + }, + "span_id": { + "description": " The ID of the span.", + "anyOf": [ + { + "$ref": "#/definitions/SpanId" + }, + { + "type": "null" + } + ] + }, + "status": { + "description": " Whether the trace failed or succeeded. Currently only used to indicate status of individual\n transactions.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/SpanStatus" + }, + { + "type": "null" + } + ] + }, + "trace_id": { + "description": " The trace ID.", + "anyOf": [ + { + "$ref": "#/definitions/TraceId" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "TraceId": { + "description": " A 32-character hex string as described in the W3C trace context spec.", + "anyOf": [ + { + "type": "string" + } + ] + }, + "TransactionInfo": { + "description": " Additional information about the name of the transaction.", + "anyOf": [ + { + "type": "object", + "properties": { + "changes": { + "description": " A list of changes prior to the final transaction name.\n\n This list must be empty if the transaction name is set at the beginning of the transaction\n and never changed. There is no placeholder entry for the initial transaction name.", + "default": null, + "type": [ + "array", + "null" + ], + "items": { + "anyOf": [ + { + "$ref": "#/definitions/TransactionNameChange" + }, + { + "type": "null" + } + ] + } + }, + "original": { + "description": " The unmodified transaction name as obtained by the source.\n\n This value will only be set if the transaction name was modified during event processing.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "propagations": { + "description": " The total number of propagations during the transaction.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "source": { + "description": " Describes how the name of the transaction was determined.\n\n This will be used by the server to decide whether or not to scrub identifiers from the\n transaction name, or replace the entire name with a placeholder.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/TransactionSource" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "TransactionNameChange": { + "anyOf": [ + { + "type": "object", + "properties": { + "propagations": { + "description": " The number of propagations from the start of the transaction to this change.", + "default": null, + "type": [ + "integer", + "null" + ], + "format": "uint64", + "minimum": 0.0 + }, + "source": { + "description": " Describes how the previous transaction name was determined.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/TransactionSource" + }, + { + "type": "null" + } + ] + }, + "timestamp": { + "description": " Timestamp when the transaction name was changed.\n\n This adheres to the event timestamp specification.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Timestamp" + }, + { + "type": "null" + } + ] + } + }, + "additionalProperties": false + } + ] + }, + "TransactionSource": { + "description": "Describes how the name of the transaction was determined.", + "type": "string", + "enum": [ + "custom", + "url", + "route", + "view", + "component", + "sanitized", + "task", + "unknown" + ] + }, + "User": { + "description": " Information about the user who triggered an event.\n\n ```json\n {\n \"user\": {\n \"id\": \"unique_id\",\n \"username\": \"my_user\",\n \"email\": \"foo@example.com\",\n \"ip_address\": \"127.0.0.1\",\n \"subscription\": \"basic\"\n }\n }\n ```", + "anyOf": [ + { + "type": "object", + "properties": { + "data": { + "description": " Additional arbitrary fields, as stored in the database (and sometimes as sent by clients).\n All data from `self.other` should end up here after store normalization.", + "default": null, + "type": [ + "object", + "null" + ], + "additionalProperties": true + }, + "email": { + "description": " Email address of the user.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "geo": { + "description": " Approximate geographical location of the end user or device.", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/Geo" + }, + { + "type": "null" + } + ] + }, + "id": { + "description": " Unique identifier of the user.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "ip_address": { + "description": " Remote IP address of the user. Defaults to \"{{auto}}\".", + "default": null, + "anyOf": [ + { + "$ref": "#/definitions/String" + }, + { + "type": "null" + } + ] + }, + "name": { + "description": " Human readable name of the user.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "segment": { + "description": " The user segment, for apps that divide users in user segments.", + "default": null, + "type": [ + "string", + "null" + ] + }, + "sentry_user": { + "description": " The user string representation as handled in Sentry.\n\n This field is computed by concatenating the name of specific fields of the `User`\n struct with their value. For example, if `id` is set, `sentry_user` will be equal to\n `\"id:id-of-the-user\".", + "default": null, + "type": [ + "string", + "null" + ] + }, + "username": { + "description": " Username of the user.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + }, + "UserReportV2Context": { + "description": " Feedback context.\n\n This contexts contains user feedback specific attributes.\n We don't PII scrub contact_email as that is provided by the user.\n TODO(jferg): rename to FeedbackContext once old UserReport logic is deprecated.", + "anyOf": [ + { + "type": "object", + "properties": { + "contact_email": { + "description": " an email optionally provided by the user, which can be different from user.email", + "default": null, + "type": [ + "string", + "null" + ] + }, + "message": { + "description": " The feedback message which contains what the user has to say.", + "default": null, + "type": [ + "string", + "null" + ] + } + }, + "additionalProperties": false + } + ] + } + } +} \ No newline at end of file diff --git a/bugsink/event_schema.py b/bugsink/event_schema.py index 1ee5905..d68b86f 100644 --- a/bugsink/event_schema.py +++ b/bugsink/event_schema.py @@ -1,5 +1,5 @@ # This file is generated by fetch_event_schema_json.py -# it is based on api/event.schema.json +# it is based on api/event.schema.altered.json # the following license applies: # Copyright (c) 2020 Sentry (https://sentry.io) and individual contributors. # @@ -31,65 +31,96 @@ def validate(data, custom_formats={}, name_prefix=None): if "breadcrumbs" in data_keys: data_keys.remove("breadcrumbs") data__breadcrumbs = data["breadcrumbs"] - if not isinstance(data__breadcrumbs, (dict, NoneType)): - raise JsonSchemaValueException( - "" - + (name_prefix or "data") - + ".breadcrumbs must be object or null" - ) - data__breadcrumbs_is_dict = isinstance(data__breadcrumbs, dict) - if data__breadcrumbs_is_dict: - data__breadcrumbs__missing_keys = ( - set(["values"]) - data__breadcrumbs.keys() - ) - if data__breadcrumbs__missing_keys: - raise JsonSchemaValueException( - "" - + (name_prefix or "data") - + ".breadcrumbs must contain " - ) - data__breadcrumbs_keys = set(data__breadcrumbs.keys()) - if "values" in data__breadcrumbs_keys: - data__breadcrumbs_keys.remove("values") - data__breadcrumbs__values = data__breadcrumbs["values"] - if not isinstance(data__breadcrumbs__values, (list, tuple)): + data__breadcrumbs_any_of_count2 = 0 + if not data__breadcrumbs_any_of_count2: + try: + if not isinstance(data__breadcrumbs, (dict, NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") - + ".breadcrumbs.values must be array" + + ".breadcrumbs must be object or null" ) - data__breadcrumbs__values_is_list = isinstance( - data__breadcrumbs__values, (list, tuple) + data__breadcrumbs_is_dict = isinstance( + data__breadcrumbs, dict ) - if data__breadcrumbs__values_is_list: - data__breadcrumbs__values_len = len( - data__breadcrumbs__values + if data__breadcrumbs_is_dict: + data__breadcrumbs__missing_keys = ( + set(["values"]) - data__breadcrumbs.keys() ) - for ( - data__breadcrumbs__values_x, - data__breadcrumbs__values_item, - ) in enumerate(data__breadcrumbs__values): - data__breadcrumbs__values_item_any_of_count2 = 0 - if not data__breadcrumbs__values_item_any_of_count2: - try: - validate___definitions_breadcrumb( - data__breadcrumbs__values_item, - custom_formats, - (name_prefix or "data") - + ".breadcrumbs.values[{data__breadcrumbs__values_x}]".format( - **locals() - ), + if data__breadcrumbs__missing_keys: + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".breadcrumbs must contain " + ) + data__breadcrumbs_keys = set(data__breadcrumbs.keys()) + if "values" in data__breadcrumbs_keys: + data__breadcrumbs_keys.remove("values") + data__breadcrumbs__values = data__breadcrumbs[ + "values" + ] + if not isinstance( + data__breadcrumbs__values, (list, tuple) + ): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".breadcrumbs.values must be array" + ) + data__breadcrumbs__values_is_list = isinstance( + data__breadcrumbs__values, (list, tuple) + ) + if data__breadcrumbs__values_is_list: + data__breadcrumbs__values_len = len( + data__breadcrumbs__values + ) + for ( + data__breadcrumbs__values_x, + data__breadcrumbs__values_item, + ) in enumerate(data__breadcrumbs__values): + data__breadcrumbs__values_item_any_of_count3 = ( + 0 ) - data__breadcrumbs__values_item_any_of_count2 += ( - 1 - ) - except JsonSchemaValueException: - pass - if not data__breadcrumbs__values_item_any_of_count2: - try: - if not isinstance( - data__breadcrumbs__values_item, - (NoneType), + if ( + not data__breadcrumbs__values_item_any_of_count3 + ): + try: + validate___definitions_breadcrumb( + data__breadcrumbs__values_item, + custom_formats, + (name_prefix or "data") + + ".breadcrumbs.values[{data__breadcrumbs__values_x}]".format( + **locals() + ), + ) + data__breadcrumbs__values_item_any_of_count3 += ( + 1 + ) + except JsonSchemaValueException: + pass + if ( + not data__breadcrumbs__values_item_any_of_count3 + ): + try: + if not isinstance( + data__breadcrumbs__values_item, + (NoneType), + ): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".breadcrumbs.values[{data__breadcrumbs__values_x}]".format( + **locals() + ) + + " must be null" + ) + data__breadcrumbs__values_item_any_of_count3 += ( + 1 + ) + except JsonSchemaValueException: + pass + if ( + not data__breadcrumbs__values_item_any_of_count3 ): raise JsonSchemaValueException( "" @@ -97,37 +128,91 @@ def validate(data, custom_formats={}, name_prefix=None): + ".breadcrumbs.values[{data__breadcrumbs__values_x}]".format( **locals() ) - + " must be null" + + " cannot be validated by any definition" ) - data__breadcrumbs__values_item_any_of_count2 += ( - 1 + data__breadcrumbs_any_of_count2 += 1 + except JsonSchemaValueException: + pass + if not data__breadcrumbs_any_of_count2: + try: + if not isinstance(data__breadcrumbs, (list, tuple)): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".breadcrumbs must be array" + ) + data__breadcrumbs_is_list = isinstance( + data__breadcrumbs, (list, tuple) + ) + if data__breadcrumbs_is_list: + data__breadcrumbs_len = len(data__breadcrumbs) + for ( + data__breadcrumbs_x, + data__breadcrumbs_item, + ) in enumerate(data__breadcrumbs): + data__breadcrumbs_item_any_of_count4 = 0 + if not data__breadcrumbs_item_any_of_count4: + try: + validate___definitions_breadcrumb( + data__breadcrumbs_item, + custom_formats, + (name_prefix or "data") + + ".breadcrumbs[{data__breadcrumbs_x}]".format( + **locals() + ), ) + data__breadcrumbs_item_any_of_count4 += 1 except JsonSchemaValueException: pass - if not data__breadcrumbs__values_item_any_of_count2: + if not data__breadcrumbs_item_any_of_count4: + try: + if not isinstance( + data__breadcrumbs_item, (NoneType) + ): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".breadcrumbs[{data__breadcrumbs_x}]".format( + **locals() + ) + + " must be null" + ) + data__breadcrumbs_item_any_of_count4 += 1 + except JsonSchemaValueException: + pass + if not data__breadcrumbs_item_any_of_count4: raise JsonSchemaValueException( "" + (name_prefix or "data") - + ".breadcrumbs.values[{data__breadcrumbs__values_x}]".format( + + ".breadcrumbs[{data__breadcrumbs_x}]".format( **locals() ) + " cannot be validated by any definition" ) + data__breadcrumbs_any_of_count2 += 1 + except JsonSchemaValueException: + pass + if not data__breadcrumbs_any_of_count2: + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".breadcrumbs cannot be validated by any definition" + ) if "contexts" in data_keys: data_keys.remove("contexts") data__contexts = data["contexts"] - data__contexts_any_of_count3 = 0 - if not data__contexts_any_of_count3: + data__contexts_any_of_count5 = 0 + if not data__contexts_any_of_count5: try: validate___definitions_contexts( data__contexts, custom_formats, (name_prefix or "data") + ".contexts", ) - data__contexts_any_of_count3 += 1 + data__contexts_any_of_count5 += 1 except JsonSchemaValueException: pass - if not data__contexts_any_of_count3: + if not data__contexts_any_of_count5: try: if not isinstance(data__contexts, (NoneType)): raise JsonSchemaValueException( @@ -135,10 +220,10 @@ def validate(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".contexts must be null" ) - data__contexts_any_of_count3 += 1 + data__contexts_any_of_count5 += 1 except JsonSchemaValueException: pass - if not data__contexts_any_of_count3: + if not data__contexts_any_of_count5: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -156,18 +241,18 @@ def validate(data, custom_formats={}, name_prefix=None): if "debug_meta" in data_keys: data_keys.remove("debug_meta") data__debugmeta = data["debug_meta"] - data__debugmeta_any_of_count4 = 0 - if not data__debugmeta_any_of_count4: + data__debugmeta_any_of_count6 = 0 + if not data__debugmeta_any_of_count6: try: validate___definitions_debugmeta( data__debugmeta, custom_formats, (name_prefix or "data") + ".debug_meta", ) - data__debugmeta_any_of_count4 += 1 + data__debugmeta_any_of_count6 += 1 except JsonSchemaValueException: pass - if not data__debugmeta_any_of_count4: + if not data__debugmeta_any_of_count6: try: if not isinstance(data__debugmeta, (NoneType)): raise JsonSchemaValueException( @@ -175,10 +260,10 @@ def validate(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".debug_meta must be null" ) - data__debugmeta_any_of_count4 += 1 + data__debugmeta_any_of_count6 += 1 except JsonSchemaValueException: pass - if not data__debugmeta_any_of_count4: + if not data__debugmeta_any_of_count6: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -217,8 +302,8 @@ def validate(data, custom_formats={}, name_prefix=None): for data__errors_x, data__errors_item in enumerate( data__errors ): - data__errors_item_any_of_count5 = 0 - if not data__errors_item_any_of_count5: + data__errors_item_any_of_count7 = 0 + if not data__errors_item_any_of_count7: try: validate___definitions_eventprocessingerror( data__errors_item, @@ -228,10 +313,10 @@ def validate(data, custom_formats={}, name_prefix=None): **locals() ), ) - data__errors_item_any_of_count5 += 1 + data__errors_item_any_of_count7 += 1 except JsonSchemaValueException: pass - if not data__errors_item_any_of_count5: + if not data__errors_item_any_of_count7: try: if not isinstance(data__errors_item, (NoneType)): raise JsonSchemaValueException( @@ -242,10 +327,10 @@ def validate(data, custom_formats={}, name_prefix=None): ) + " must be null" ) - data__errors_item_any_of_count5 += 1 + data__errors_item_any_of_count7 += 1 except JsonSchemaValueException: pass - if not data__errors_item_any_of_count5: + if not data__errors_item_any_of_count7: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -255,18 +340,18 @@ def validate(data, custom_formats={}, name_prefix=None): if "event_id" in data_keys: data_keys.remove("event_id") data__eventid = data["event_id"] - data__eventid_any_of_count6 = 0 - if not data__eventid_any_of_count6: + data__eventid_any_of_count8 = 0 + if not data__eventid_any_of_count8: try: validate___definitions_eventid( data__eventid, custom_formats, (name_prefix or "data") + ".event_id", ) - data__eventid_any_of_count6 += 1 + data__eventid_any_of_count8 += 1 except JsonSchemaValueException: pass - if not data__eventid_any_of_count6: + if not data__eventid_any_of_count8: try: if not isinstance(data__eventid, (NoneType)): raise JsonSchemaValueException( @@ -274,10 +359,10 @@ def validate(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".event_id must be null" ) - data__eventid_any_of_count6 += 1 + data__eventid_any_of_count8 += 1 except JsonSchemaValueException: pass - if not data__eventid_any_of_count6: + if not data__eventid_any_of_count8: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -286,64 +371,92 @@ def validate(data, custom_formats={}, name_prefix=None): if "exception" in data_keys: data_keys.remove("exception") data__exception = data["exception"] - if not isinstance(data__exception, (dict, NoneType)): - raise JsonSchemaValueException( - "" - + (name_prefix or "data") - + ".exception must be object or null" - ) - data__exception_is_dict = isinstance(data__exception, dict) - if data__exception_is_dict: - data__exception__missing_keys = ( - set(["values"]) - data__exception.keys() - ) - if data__exception__missing_keys: - raise JsonSchemaValueException( - "" - + (name_prefix or "data") - + ".exception must contain " - ) - data__exception_keys = set(data__exception.keys()) - if "values" in data__exception_keys: - data__exception_keys.remove("values") - data__exception__values = data__exception["values"] - if not isinstance(data__exception__values, (list, tuple)): + data__exception_any_of_count9 = 0 + if not data__exception_any_of_count9: + try: + if not isinstance(data__exception, (dict, NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") - + ".exception.values must be array" + + ".exception must be object or null" ) - data__exception__values_is_list = isinstance( - data__exception__values, (list, tuple) - ) - if data__exception__values_is_list: - data__exception__values_len = len( - data__exception__values + data__exception_is_dict = isinstance(data__exception, dict) + if data__exception_is_dict: + data__exception__missing_keys = ( + set(["values"]) - data__exception.keys() ) - for ( - data__exception__values_x, - data__exception__values_item, - ) in enumerate(data__exception__values): - data__exception__values_item_any_of_count7 = 0 - if not data__exception__values_item_any_of_count7: - try: - validate___definitions_exception( - data__exception__values_item, - custom_formats, - (name_prefix or "data") - + ".exception.values[{data__exception__values_x}]".format( - **locals() - ), + if data__exception__missing_keys: + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".exception must contain " + ) + data__exception_keys = set(data__exception.keys()) + if "values" in data__exception_keys: + data__exception_keys.remove("values") + data__exception__values = data__exception["values"] + if not isinstance( + data__exception__values, (list, tuple) + ): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".exception.values must be array" + ) + data__exception__values_is_list = isinstance( + data__exception__values, (list, tuple) + ) + if data__exception__values_is_list: + data__exception__values_len = len( + data__exception__values + ) + for ( + data__exception__values_x, + data__exception__values_item, + ) in enumerate(data__exception__values): + data__exception__values_item_any_of_count10 = ( + 0 ) - data__exception__values_item_any_of_count7 += ( - 1 - ) - except JsonSchemaValueException: - pass - if not data__exception__values_item_any_of_count7: - try: - if not isinstance( - data__exception__values_item, (NoneType) + if ( + not data__exception__values_item_any_of_count10 + ): + try: + validate___definitions_exception( + data__exception__values_item, + custom_formats, + (name_prefix or "data") + + ".exception.values[{data__exception__values_x}]".format( + **locals() + ), + ) + data__exception__values_item_any_of_count10 += ( + 1 + ) + except JsonSchemaValueException: + pass + if ( + not data__exception__values_item_any_of_count10 + ): + try: + if not isinstance( + data__exception__values_item, + (NoneType), + ): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".exception.values[{data__exception__values_x}]".format( + **locals() + ) + + " must be null" + ) + data__exception__values_item_any_of_count10 += ( + 1 + ) + except JsonSchemaValueException: + pass + if ( + not data__exception__values_item_any_of_count10 ): raise JsonSchemaValueException( "" @@ -351,22 +464,76 @@ def validate(data, custom_formats={}, name_prefix=None): + ".exception.values[{data__exception__values_x}]".format( **locals() ) - + " must be null" + + " cannot be validated by any definition" ) - data__exception__values_item_any_of_count7 += ( - 1 + data__exception_any_of_count9 += 1 + except JsonSchemaValueException: + pass + if not data__exception_any_of_count9: + try: + if not isinstance(data__exception, (list, tuple)): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".exception must be array" + ) + data__exception_is_list = isinstance( + data__exception, (list, tuple) + ) + if data__exception_is_list: + data__exception_len = len(data__exception) + for ( + data__exception_x, + data__exception_item, + ) in enumerate(data__exception): + data__exception_item_any_of_count11 = 0 + if not data__exception_item_any_of_count11: + try: + validate___definitions_exception( + data__exception_item, + custom_formats, + (name_prefix or "data") + + ".exception[{data__exception_x}]".format( + **locals() + ), ) + data__exception_item_any_of_count11 += 1 except JsonSchemaValueException: pass - if not data__exception__values_item_any_of_count7: + if not data__exception_item_any_of_count11: + try: + if not isinstance( + data__exception_item, (NoneType) + ): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".exception[{data__exception_x}]".format( + **locals() + ) + + " must be null" + ) + data__exception_item_any_of_count11 += 1 + except JsonSchemaValueException: + pass + if not data__exception_item_any_of_count11: raise JsonSchemaValueException( "" + (name_prefix or "data") - + ".exception.values[{data__exception__values_x}]".format( + + ".exception[{data__exception_x}]".format( **locals() ) + " cannot be validated by any definition" ) + data__exception_any_of_count9 += 1 + except JsonSchemaValueException: + pass + if not data__exception_any_of_count9: + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".exception cannot be validated by any definition" + ) if "extra" in data_keys: data_keys.remove("extra") data__extra = data["extra"] @@ -382,18 +549,18 @@ def validate(data, custom_formats={}, name_prefix=None): if "fingerprint" in data_keys: data_keys.remove("fingerprint") data__fingerprint = data["fingerprint"] - data__fingerprint_any_of_count8 = 0 - if not data__fingerprint_any_of_count8: + data__fingerprint_any_of_count12 = 0 + if not data__fingerprint_any_of_count12: try: validate___definitions_fingerprint( data__fingerprint, custom_formats, (name_prefix or "data") + ".fingerprint", ) - data__fingerprint_any_of_count8 += 1 + data__fingerprint_any_of_count12 += 1 except JsonSchemaValueException: pass - if not data__fingerprint_any_of_count8: + if not data__fingerprint_any_of_count12: try: if not isinstance(data__fingerprint, (NoneType)): raise JsonSchemaValueException( @@ -401,10 +568,10 @@ def validate(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".fingerprint must be null" ) - data__fingerprint_any_of_count8 += 1 + data__fingerprint_any_of_count12 += 1 except JsonSchemaValueException: pass - if not data__fingerprint_any_of_count8: + if not data__fingerprint_any_of_count12: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -413,27 +580,27 @@ def validate(data, custom_formats={}, name_prefix=None): if "level" in data_keys: data_keys.remove("level") data__level = data["level"] - data__level_any_of_count9 = 0 - if not data__level_any_of_count9: + data__level_any_of_count13 = 0 + if not data__level_any_of_count13: try: validate___definitions_level( data__level, custom_formats, (name_prefix or "data") + ".level", ) - data__level_any_of_count9 += 1 + data__level_any_of_count13 += 1 except JsonSchemaValueException: pass - if not data__level_any_of_count9: + if not data__level_any_of_count13: try: if not isinstance(data__level, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".level must be null" ) - data__level_any_of_count9 += 1 + data__level_any_of_count13 += 1 except JsonSchemaValueException: pass - if not data__level_any_of_count9: + if not data__level_any_of_count13: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -442,18 +609,18 @@ def validate(data, custom_formats={}, name_prefix=None): if "logentry" in data_keys: data_keys.remove("logentry") data__logentry = data["logentry"] - data__logentry_any_of_count10 = 0 - if not data__logentry_any_of_count10: + data__logentry_any_of_count14 = 0 + if not data__logentry_any_of_count14: try: validate___definitions_logentry( data__logentry, custom_formats, (name_prefix or "data") + ".logentry", ) - data__logentry_any_of_count10 += 1 + data__logentry_any_of_count14 += 1 except JsonSchemaValueException: pass - if not data__logentry_any_of_count10: + if not data__logentry_any_of_count14: try: if not isinstance(data__logentry, (NoneType)): raise JsonSchemaValueException( @@ -461,10 +628,10 @@ def validate(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".logentry must be null" ) - data__logentry_any_of_count10 += 1 + data__logentry_any_of_count14 += 1 except JsonSchemaValueException: pass - if not data__logentry_any_of_count10: + if not data__logentry_any_of_count14: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -517,18 +684,18 @@ def validate(data, custom_formats={}, name_prefix=None): if "received" in data_keys: data_keys.remove("received") data__received = data["received"] - data__received_any_of_count11 = 0 - if not data__received_any_of_count11: + data__received_any_of_count15 = 0 + if not data__received_any_of_count15: try: validate___definitions_timestamp( data__received, custom_formats, (name_prefix or "data") + ".received", ) - data__received_any_of_count11 += 1 + data__received_any_of_count15 += 1 except JsonSchemaValueException: pass - if not data__received_any_of_count11: + if not data__received_any_of_count15: try: if not isinstance(data__received, (NoneType)): raise JsonSchemaValueException( @@ -536,10 +703,10 @@ def validate(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".received must be null" ) - data__received_any_of_count11 += 1 + data__received_any_of_count15 += 1 except JsonSchemaValueException: pass - if not data__received_any_of_count11: + if not data__received_any_of_count15: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -557,18 +724,18 @@ def validate(data, custom_formats={}, name_prefix=None): if "request" in data_keys: data_keys.remove("request") data__request = data["request"] - data__request_any_of_count12 = 0 - if not data__request_any_of_count12: + data__request_any_of_count16 = 0 + if not data__request_any_of_count16: try: validate___definitions_request( data__request, custom_formats, (name_prefix or "data") + ".request", ) - data__request_any_of_count12 += 1 + data__request_any_of_count16 += 1 except JsonSchemaValueException: pass - if not data__request_any_of_count12: + if not data__request_any_of_count16: try: if not isinstance(data__request, (NoneType)): raise JsonSchemaValueException( @@ -576,10 +743,10 @@ def validate(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".request must be null" ) - data__request_any_of_count12 += 1 + data__request_any_of_count16 += 1 except JsonSchemaValueException: pass - if not data__request_any_of_count12: + if not data__request_any_of_count16: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -588,27 +755,27 @@ def validate(data, custom_formats={}, name_prefix=None): if "sdk" in data_keys: data_keys.remove("sdk") data__sdk = data["sdk"] - data__sdk_any_of_count13 = 0 - if not data__sdk_any_of_count13: + data__sdk_any_of_count17 = 0 + if not data__sdk_any_of_count17: try: validate___definitions_clientsdkinfo( data__sdk, custom_formats, (name_prefix or "data") + ".sdk", ) - data__sdk_any_of_count13 += 1 + data__sdk_any_of_count17 += 1 except JsonSchemaValueException: pass - if not data__sdk_any_of_count13: + if not data__sdk_any_of_count17: try: if not isinstance(data__sdk, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".sdk must be null" ) - data__sdk_any_of_count13 += 1 + data__sdk_any_of_count17 += 1 except JsonSchemaValueException: pass - if not data__sdk_any_of_count13: + if not data__sdk_any_of_count17: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -626,18 +793,18 @@ def validate(data, custom_formats={}, name_prefix=None): if "stacktrace" in data_keys: data_keys.remove("stacktrace") data__stacktrace = data["stacktrace"] - data__stacktrace_any_of_count14 = 0 - if not data__stacktrace_any_of_count14: + data__stacktrace_any_of_count18 = 0 + if not data__stacktrace_any_of_count18: try: validate___definitions_stacktrace( data__stacktrace, custom_formats, (name_prefix or "data") + ".stacktrace", ) - data__stacktrace_any_of_count14 += 1 + data__stacktrace_any_of_count18 += 1 except JsonSchemaValueException: pass - if not data__stacktrace_any_of_count14: + if not data__stacktrace_any_of_count18: try: if not isinstance(data__stacktrace, (NoneType)): raise JsonSchemaValueException( @@ -645,10 +812,10 @@ def validate(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".stacktrace must be null" ) - data__stacktrace_any_of_count14 += 1 + data__stacktrace_any_of_count18 += 1 except JsonSchemaValueException: pass - if not data__stacktrace_any_of_count14: + if not data__stacktrace_any_of_count18: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -657,27 +824,27 @@ def validate(data, custom_formats={}, name_prefix=None): if "tags" in data_keys: data_keys.remove("tags") data__tags = data["tags"] - data__tags_any_of_count15 = 0 - if not data__tags_any_of_count15: + data__tags_any_of_count19 = 0 + if not data__tags_any_of_count19: try: validate___definitions_tags( data__tags, custom_formats, (name_prefix or "data") + ".tags", ) - data__tags_any_of_count15 += 1 + data__tags_any_of_count19 += 1 except JsonSchemaValueException: pass - if not data__tags_any_of_count15: + if not data__tags_any_of_count19: try: if not isinstance(data__tags, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".tags must be null" ) - data__tags_any_of_count15 += 1 + data__tags_any_of_count19 += 1 except JsonSchemaValueException: pass - if not data__tags_any_of_count15: + if not data__tags_any_of_count19: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -686,60 +853,92 @@ def validate(data, custom_formats={}, name_prefix=None): if "threads" in data_keys: data_keys.remove("threads") data__threads = data["threads"] - if not isinstance(data__threads, (dict, NoneType)): - raise JsonSchemaValueException( - "" - + (name_prefix or "data") - + ".threads must be object or null" - ) - data__threads_is_dict = isinstance(data__threads, dict) - if data__threads_is_dict: - data__threads__missing_keys = ( - set(["values"]) - data__threads.keys() - ) - if data__threads__missing_keys: - raise JsonSchemaValueException( - "" + (name_prefix or "data") + ".threads must contain " - ) - data__threads_keys = set(data__threads.keys()) - if "values" in data__threads_keys: - data__threads_keys.remove("values") - data__threads__values = data__threads["values"] - if not isinstance(data__threads__values, (list, tuple)): + data__threads_any_of_count20 = 0 + if not data__threads_any_of_count20: + try: + if not isinstance(data__threads, (dict, NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") - + ".threads.values must be array" + + ".threads must be object or null" ) - data__threads__values_is_list = isinstance( - data__threads__values, (list, tuple) - ) - if data__threads__values_is_list: - data__threads__values_len = len(data__threads__values) - for ( - data__threads__values_x, - data__threads__values_item, - ) in enumerate(data__threads__values): - data__threads__values_item_any_of_count16 = 0 - if not data__threads__values_item_any_of_count16: - try: - validate___definitions_thread( - data__threads__values_item, - custom_formats, - (name_prefix or "data") - + ".threads.values[{data__threads__values_x}]".format( - **locals() - ), + data__threads_is_dict = isinstance(data__threads, dict) + if data__threads_is_dict: + data__threads__missing_keys = ( + set(["values"]) - data__threads.keys() + ) + if data__threads__missing_keys: + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".threads must contain " + ) + data__threads_keys = set(data__threads.keys()) + if "values" in data__threads_keys: + data__threads_keys.remove("values") + data__threads__values = data__threads["values"] + if not isinstance( + data__threads__values, (list, tuple) + ): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".threads.values must be array" + ) + data__threads__values_is_list = isinstance( + data__threads__values, (list, tuple) + ) + if data__threads__values_is_list: + data__threads__values_len = len( + data__threads__values + ) + for ( + data__threads__values_x, + data__threads__values_item, + ) in enumerate(data__threads__values): + data__threads__values_item_any_of_count21 = ( + 0 ) - data__threads__values_item_any_of_count16 += ( - 1 - ) - except JsonSchemaValueException: - pass - if not data__threads__values_item_any_of_count16: - try: - if not isinstance( - data__threads__values_item, (NoneType) + if ( + not data__threads__values_item_any_of_count21 + ): + try: + validate___definitions_thread( + data__threads__values_item, + custom_formats, + (name_prefix or "data") + + ".threads.values[{data__threads__values_x}]".format( + **locals() + ), + ) + data__threads__values_item_any_of_count21 += ( + 1 + ) + except JsonSchemaValueException: + pass + if ( + not data__threads__values_item_any_of_count21 + ): + try: + if not isinstance( + data__threads__values_item, + (NoneType), + ): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".threads.values[{data__threads__values_x}]".format( + **locals() + ) + + " must be null" + ) + data__threads__values_item_any_of_count21 += ( + 1 + ) + except JsonSchemaValueException: + pass + if ( + not data__threads__values_item_any_of_count21 ): raise JsonSchemaValueException( "" @@ -747,22 +946,75 @@ def validate(data, custom_formats={}, name_prefix=None): + ".threads.values[{data__threads__values_x}]".format( **locals() ) - + " must be null" + + " cannot be validated by any definition" ) - data__threads__values_item_any_of_count16 += ( - 1 + data__threads_any_of_count20 += 1 + except JsonSchemaValueException: + pass + if not data__threads_any_of_count20: + try: + if not isinstance(data__threads, (list, tuple)): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".threads must be array" + ) + data__threads_is_list = isinstance( + data__threads, (list, tuple) + ) + if data__threads_is_list: + data__threads_len = len(data__threads) + for data__threads_x, data__threads_item in enumerate( + data__threads + ): + data__threads_item_any_of_count22 = 0 + if not data__threads_item_any_of_count22: + try: + validate___definitions_thread( + data__threads_item, + custom_formats, + (name_prefix or "data") + + ".threads[{data__threads_x}]".format( + **locals() + ), ) + data__threads_item_any_of_count22 += 1 except JsonSchemaValueException: pass - if not data__threads__values_item_any_of_count16: + if not data__threads_item_any_of_count22: + try: + if not isinstance( + data__threads_item, (NoneType) + ): + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".threads[{data__threads_x}]".format( + **locals() + ) + + " must be null" + ) + data__threads_item_any_of_count22 += 1 + except JsonSchemaValueException: + pass + if not data__threads_item_any_of_count22: raise JsonSchemaValueException( "" + (name_prefix or "data") - + ".threads.values[{data__threads__values_x}]".format( + + ".threads[{data__threads_x}]".format( **locals() ) + " cannot be validated by any definition" ) + data__threads_any_of_count20 += 1 + except JsonSchemaValueException: + pass + if not data__threads_any_of_count20: + raise JsonSchemaValueException( + "" + + (name_prefix or "data") + + ".threads cannot be validated by any definition" + ) if "time_spent" in data_keys: data_keys.remove("time_spent") data__timespent = data["time_spent"] @@ -789,18 +1041,18 @@ def validate(data, custom_formats={}, name_prefix=None): if "timestamp" in data_keys: data_keys.remove("timestamp") data__timestamp = data["timestamp"] - data__timestamp_any_of_count17 = 0 - if not data__timestamp_any_of_count17: + data__timestamp_any_of_count23 = 0 + if not data__timestamp_any_of_count23: try: validate___definitions_timestamp( data__timestamp, custom_formats, (name_prefix or "data") + ".timestamp", ) - data__timestamp_any_of_count17 += 1 + data__timestamp_any_of_count23 += 1 except JsonSchemaValueException: pass - if not data__timestamp_any_of_count17: + if not data__timestamp_any_of_count23: try: if not isinstance(data__timestamp, (NoneType)): raise JsonSchemaValueException( @@ -808,10 +1060,10 @@ def validate(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".timestamp must be null" ) - data__timestamp_any_of_count17 += 1 + data__timestamp_any_of_count23 += 1 except JsonSchemaValueException: pass - if not data__timestamp_any_of_count17: + if not data__timestamp_any_of_count23: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -829,18 +1081,18 @@ def validate(data, custom_formats={}, name_prefix=None): if "transaction_info" in data_keys: data_keys.remove("transaction_info") data__transactioninfo = data["transaction_info"] - data__transactioninfo_any_of_count18 = 0 - if not data__transactioninfo_any_of_count18: + data__transactioninfo_any_of_count24 = 0 + if not data__transactioninfo_any_of_count24: try: validate___definitions_transactioninfo( data__transactioninfo, custom_formats, (name_prefix or "data") + ".transaction_info", ) - data__transactioninfo_any_of_count18 += 1 + data__transactioninfo_any_of_count24 += 1 except JsonSchemaValueException: pass - if not data__transactioninfo_any_of_count18: + if not data__transactioninfo_any_of_count24: try: if not isinstance(data__transactioninfo, (NoneType)): raise JsonSchemaValueException( @@ -848,10 +1100,10 @@ def validate(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".transaction_info must be null" ) - data__transactioninfo_any_of_count18 += 1 + data__transactioninfo_any_of_count24 += 1 except JsonSchemaValueException: pass - if not data__transactioninfo_any_of_count18: + if not data__transactioninfo_any_of_count24: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -860,27 +1112,27 @@ def validate(data, custom_formats={}, name_prefix=None): if "type" in data_keys: data_keys.remove("type") data__type = data["type"] - data__type_any_of_count19 = 0 - if not data__type_any_of_count19: + data__type_any_of_count25 = 0 + if not data__type_any_of_count25: try: validate___definitions_eventtype( data__type, custom_formats, (name_prefix or "data") + ".type", ) - data__type_any_of_count19 += 1 + data__type_any_of_count25 += 1 except JsonSchemaValueException: pass - if not data__type_any_of_count19: + if not data__type_any_of_count25: try: if not isinstance(data__type, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".type must be null" ) - data__type_any_of_count19 += 1 + data__type_any_of_count25 += 1 except JsonSchemaValueException: pass - if not data__type_any_of_count19: + if not data__type_any_of_count25: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -889,27 +1141,27 @@ def validate(data, custom_formats={}, name_prefix=None): if "user" in data_keys: data_keys.remove("user") data__user = data["user"] - data__user_any_of_count20 = 0 - if not data__user_any_of_count20: + data__user_any_of_count26 = 0 + if not data__user_any_of_count26: try: validate___definitions_user( data__user, custom_formats, (name_prefix or "data") + ".user", ) - data__user_any_of_count20 += 1 + data__user_any_of_count26 += 1 except JsonSchemaValueException: pass - if not data__user_any_of_count20: + if not data__user_any_of_count26: try: if not isinstance(data__user, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".user must be null" ) - data__user_any_of_count20 += 1 + data__user_any_of_count26 += 1 except JsonSchemaValueException: pass - if not data__user_any_of_count20: + if not data__user_any_of_count26: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -943,8 +1195,8 @@ def validate(data, custom_formats={}, name_prefix=None): def validate___definitions_user(data, custom_formats={}, name_prefix=None): - data_any_of_count21 = 0 - if not data_any_of_count21: + data_any_of_count27 = 0 + if not data_any_of_count27: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -977,27 +1229,27 @@ def validate___definitions_user(data, custom_formats={}, name_prefix=None): if "geo" in data_keys: data_keys.remove("geo") data__geo = data["geo"] - data__geo_any_of_count22 = 0 - if not data__geo_any_of_count22: + data__geo_any_of_count28 = 0 + if not data__geo_any_of_count28: try: validate___definitions_geo( data__geo, custom_formats, (name_prefix or "data") + ".geo", ) - data__geo_any_of_count22 += 1 + data__geo_any_of_count28 += 1 except JsonSchemaValueException: pass - if not data__geo_any_of_count22: + if not data__geo_any_of_count28: try: if not isinstance(data__geo, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".geo must be null" ) - data__geo_any_of_count22 += 1 + data__geo_any_of_count28 += 1 except JsonSchemaValueException: pass - if not data__geo_any_of_count22: + if not data__geo_any_of_count28: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1013,18 +1265,18 @@ def validate___definitions_user(data, custom_formats={}, name_prefix=None): if "ip_address" in data_keys: data_keys.remove("ip_address") data__ipaddress = data["ip_address"] - data__ipaddress_any_of_count23 = 0 - if not data__ipaddress_any_of_count23: + data__ipaddress_any_of_count29 = 0 + if not data__ipaddress_any_of_count29: try: validate___definitions_string( data__ipaddress, custom_formats, (name_prefix or "data") + ".ip_address", ) - data__ipaddress_any_of_count23 += 1 + data__ipaddress_any_of_count29 += 1 except JsonSchemaValueException: pass - if not data__ipaddress_any_of_count23: + if not data__ipaddress_any_of_count29: try: if not isinstance(data__ipaddress, (NoneType)): raise JsonSchemaValueException( @@ -1032,10 +1284,10 @@ def validate___definitions_user(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".ip_address must be null" ) - data__ipaddress_any_of_count23 += 1 + data__ipaddress_any_of_count29 += 1 except JsonSchemaValueException: pass - if not data__ipaddress_any_of_count23: + if not data__ipaddress_any_of_count29: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1085,10 +1337,10 @@ def validate___definitions_user(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count21 += 1 + data_any_of_count27 += 1 except JsonSchemaValueException: pass - if not data_any_of_count21: + if not data_any_of_count27: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -1102,8 +1354,8 @@ def validate___definitions_string(data, custom_formats={}, name_prefix=None): def validate___definitions_geo(data, custom_formats={}, name_prefix=None): - data_any_of_count24 = 0 - if not data_any_of_count24: + data_any_of_count30 = 0 + if not data_any_of_count30: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -1156,10 +1408,10 @@ def validate___definitions_geo(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count24 += 1 + data_any_of_count30 += 1 except JsonSchemaValueException: pass - if not data_any_of_count24: + if not data_any_of_count30: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -1189,8 +1441,8 @@ def validate___definitions_eventtype(data, custom_formats={}, name_prefix=None): def validate___definitions_transactioninfo(data, custom_formats={}, name_prefix=None): - data_any_of_count25 = 0 - if not data_any_of_count25: + data_any_of_count31 = 0 + if not data_any_of_count31: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -1214,8 +1466,8 @@ def validate___definitions_transactioninfo(data, custom_formats={}, name_prefix= for data__changes_x, data__changes_item in enumerate( data__changes ): - data__changes_item_any_of_count26 = 0 - if not data__changes_item_any_of_count26: + data__changes_item_any_of_count32 = 0 + if not data__changes_item_any_of_count32: try: validate___definitions_transactionnamechange( data__changes_item, @@ -1225,10 +1477,10 @@ def validate___definitions_transactioninfo(data, custom_formats={}, name_prefix= **locals() ), ) - data__changes_item_any_of_count26 += 1 + data__changes_item_any_of_count32 += 1 except JsonSchemaValueException: pass - if not data__changes_item_any_of_count26: + if not data__changes_item_any_of_count32: try: if not isinstance(data__changes_item, (NoneType)): raise JsonSchemaValueException( @@ -1239,10 +1491,10 @@ def validate___definitions_transactioninfo(data, custom_formats={}, name_prefix= ) + " must be null" ) - data__changes_item_any_of_count26 += 1 + data__changes_item_any_of_count32 += 1 except JsonSchemaValueException: pass - if not data__changes_item_any_of_count26: + if not data__changes_item_any_of_count32: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1284,18 +1536,18 @@ def validate___definitions_transactioninfo(data, custom_formats={}, name_prefix= if "source" in data_keys: data_keys.remove("source") data__source = data["source"] - data__source_any_of_count27 = 0 - if not data__source_any_of_count27: + data__source_any_of_count33 = 0 + if not data__source_any_of_count33: try: validate___definitions_transactionsource( data__source, custom_formats, (name_prefix or "data") + ".source", ) - data__source_any_of_count27 += 1 + data__source_any_of_count33 += 1 except JsonSchemaValueException: pass - if not data__source_any_of_count27: + if not data__source_any_of_count33: try: if not isinstance(data__source, (NoneType)): raise JsonSchemaValueException( @@ -1303,10 +1555,10 @@ def validate___definitions_transactioninfo(data, custom_formats={}, name_prefix= + (name_prefix or "data") + ".source must be null" ) - data__source_any_of_count27 += 1 + data__source_any_of_count33 += 1 except JsonSchemaValueException: pass - if not data__source_any_of_count27: + if not data__source_any_of_count33: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1320,10 +1572,10 @@ def validate___definitions_transactioninfo(data, custom_formats={}, name_prefix= + str(data_keys) + " properties" ) - data_any_of_count25 += 1 + data_any_of_count31 += 1 except JsonSchemaValueException: pass - if not data_any_of_count25: + if not data_any_of_count31: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -1354,8 +1606,8 @@ def validate___definitions_transactionsource(data, custom_formats={}, name_prefi def validate___definitions_transactionnamechange( data, custom_formats={}, name_prefix=None ): - data_any_of_count28 = 0 - if not data_any_of_count28: + data_any_of_count34 = 0 + if not data_any_of_count34: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -1390,18 +1642,18 @@ def validate___definitions_transactionnamechange( if "source" in data_keys: data_keys.remove("source") data__source = data["source"] - data__source_any_of_count29 = 0 - if not data__source_any_of_count29: + data__source_any_of_count35 = 0 + if not data__source_any_of_count35: try: validate___definitions_transactionsource( data__source, custom_formats, (name_prefix or "data") + ".source", ) - data__source_any_of_count29 += 1 + data__source_any_of_count35 += 1 except JsonSchemaValueException: pass - if not data__source_any_of_count29: + if not data__source_any_of_count35: try: if not isinstance(data__source, (NoneType)): raise JsonSchemaValueException( @@ -1409,10 +1661,10 @@ def validate___definitions_transactionnamechange( + (name_prefix or "data") + ".source must be null" ) - data__source_any_of_count29 += 1 + data__source_any_of_count35 += 1 except JsonSchemaValueException: pass - if not data__source_any_of_count29: + if not data__source_any_of_count35: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1421,18 +1673,18 @@ def validate___definitions_transactionnamechange( if "timestamp" in data_keys: data_keys.remove("timestamp") data__timestamp = data["timestamp"] - data__timestamp_any_of_count30 = 0 - if not data__timestamp_any_of_count30: + data__timestamp_any_of_count36 = 0 + if not data__timestamp_any_of_count36: try: validate___definitions_timestamp( data__timestamp, custom_formats, (name_prefix or "data") + ".timestamp", ) - data__timestamp_any_of_count30 += 1 + data__timestamp_any_of_count36 += 1 except JsonSchemaValueException: pass - if not data__timestamp_any_of_count30: + if not data__timestamp_any_of_count36: try: if not isinstance(data__timestamp, (NoneType)): raise JsonSchemaValueException( @@ -1440,10 +1692,10 @@ def validate___definitions_transactionnamechange( + (name_prefix or "data") + ".timestamp must be null" ) - data__timestamp_any_of_count30 += 1 + data__timestamp_any_of_count36 += 1 except JsonSchemaValueException: pass - if not data__timestamp_any_of_count30: + if not data__timestamp_any_of_count36: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1457,10 +1709,10 @@ def validate___definitions_transactionnamechange( + str(data_keys) + " properties" ) - data_any_of_count28 += 1 + data_any_of_count34 += 1 except JsonSchemaValueException: pass - if not data_any_of_count28: + if not data_any_of_count34: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -1468,8 +1720,8 @@ def validate___definitions_transactionnamechange( def validate___definitions_thread(data, custom_formats={}, name_prefix=None): - data_any_of_count31 = 0 - if not data_any_of_count31: + data_any_of_count37 = 0 + if not data_any_of_count37: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -1513,8 +1765,8 @@ def validate___definitions_thread(data, custom_formats={}, name_prefix=None): data__heldlocks_value = data__heldlocks.get( data__heldlocks_key ) - data__heldlocks_value_any_of_count32 = 0 - if not data__heldlocks_value_any_of_count32: + data__heldlocks_value_any_of_count38 = 0 + if not data__heldlocks_value_any_of_count38: try: validate___definitions_lockreason( data__heldlocks_value, @@ -1524,10 +1776,10 @@ def validate___definitions_thread(data, custom_formats={}, name_prefix=None): **locals() ), ) - data__heldlocks_value_any_of_count32 += 1 + data__heldlocks_value_any_of_count38 += 1 except JsonSchemaValueException: pass - if not data__heldlocks_value_any_of_count32: + if not data__heldlocks_value_any_of_count38: try: if not isinstance( data__heldlocks_value, (NoneType) @@ -1540,10 +1792,10 @@ def validate___definitions_thread(data, custom_formats={}, name_prefix=None): ) + " must be null" ) - data__heldlocks_value_any_of_count32 += 1 + data__heldlocks_value_any_of_count38 += 1 except JsonSchemaValueException: pass - if not data__heldlocks_value_any_of_count32: + if not data__heldlocks_value_any_of_count38: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1555,27 +1807,27 @@ def validate___definitions_thread(data, custom_formats={}, name_prefix=None): if "id" in data_keys: data_keys.remove("id") data__id = data["id"] - data__id_any_of_count33 = 0 - if not data__id_any_of_count33: + data__id_any_of_count39 = 0 + if not data__id_any_of_count39: try: validate___definitions_threadid( data__id, custom_formats, (name_prefix or "data") + ".id", ) - data__id_any_of_count33 += 1 + data__id_any_of_count39 += 1 except JsonSchemaValueException: pass - if not data__id_any_of_count33: + if not data__id_any_of_count39: try: if not isinstance(data__id, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".id must be null" ) - data__id_any_of_count33 += 1 + data__id_any_of_count39 += 1 except JsonSchemaValueException: pass - if not data__id_any_of_count33: + if not data__id_any_of_count39: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1602,18 +1854,18 @@ def validate___definitions_thread(data, custom_formats={}, name_prefix=None): if "stacktrace" in data_keys: data_keys.remove("stacktrace") data__stacktrace = data["stacktrace"] - data__stacktrace_any_of_count34 = 0 - if not data__stacktrace_any_of_count34: + data__stacktrace_any_of_count40 = 0 + if not data__stacktrace_any_of_count40: try: validate___definitions_stacktrace( data__stacktrace, custom_formats, (name_prefix or "data") + ".stacktrace", ) - data__stacktrace_any_of_count34 += 1 + data__stacktrace_any_of_count40 += 1 except JsonSchemaValueException: pass - if not data__stacktrace_any_of_count34: + if not data__stacktrace_any_of_count40: try: if not isinstance(data__stacktrace, (NoneType)): raise JsonSchemaValueException( @@ -1621,10 +1873,10 @@ def validate___definitions_thread(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".stacktrace must be null" ) - data__stacktrace_any_of_count34 += 1 + data__stacktrace_any_of_count40 += 1 except JsonSchemaValueException: pass - if not data__stacktrace_any_of_count34: + if not data__stacktrace_any_of_count40: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1647,10 +1899,10 @@ def validate___definitions_thread(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count31 += 1 + data_any_of_count37 += 1 except JsonSchemaValueException: pass - if not data_any_of_count31: + if not data_any_of_count37: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -1658,8 +1910,8 @@ def validate___definitions_thread(data, custom_formats={}, name_prefix=None): def validate___definitions_threadid(data, custom_formats={}, name_prefix=None): - data_any_of_count35 = 0 - if not data_any_of_count35: + data_any_of_count41 = 0 + if not data_any_of_count41: try: if ( not isinstance(data, (int)) @@ -1676,19 +1928,19 @@ def validate___definitions_threadid(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + " must be bigger than or equal to 0.0" ) - data_any_of_count35 += 1 + data_any_of_count41 += 1 except JsonSchemaValueException: pass - if not data_any_of_count35: + if not data_any_of_count41: try: if not isinstance(data, (str)): raise JsonSchemaValueException( "" + (name_prefix or "data") + " must be string" ) - data_any_of_count35 += 1 + data_any_of_count41 += 1 except JsonSchemaValueException: pass - if not data_any_of_count35: + if not data_any_of_count41: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -1696,8 +1948,8 @@ def validate___definitions_threadid(data, custom_formats={}, name_prefix=None): def validate___definitions_lockreason(data, custom_formats={}, name_prefix=None): - data_any_of_count36 = 0 - if not data_any_of_count36: + data_any_of_count42 = 0 + if not data_any_of_count42: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -1741,18 +1993,18 @@ def validate___definitions_lockreason(data, custom_formats={}, name_prefix=None) if "thread_id" in data_keys: data_keys.remove("thread_id") data__threadid = data["thread_id"] - data__threadid_any_of_count37 = 0 - if not data__threadid_any_of_count37: + data__threadid_any_of_count43 = 0 + if not data__threadid_any_of_count43: try: validate___definitions_threadid( data__threadid, custom_formats, (name_prefix or "data") + ".thread_id", ) - data__threadid_any_of_count37 += 1 + data__threadid_any_of_count43 += 1 except JsonSchemaValueException: pass - if not data__threadid_any_of_count37: + if not data__threadid_any_of_count43: try: if not isinstance(data__threadid, (NoneType)): raise JsonSchemaValueException( @@ -1760,10 +2012,10 @@ def validate___definitions_lockreason(data, custom_formats={}, name_prefix=None) + (name_prefix or "data") + ".thread_id must be null" ) - data__threadid_any_of_count37 += 1 + data__threadid_any_of_count43 += 1 except JsonSchemaValueException: pass - if not data__threadid_any_of_count37: + if not data__threadid_any_of_count43: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1772,27 +2024,27 @@ def validate___definitions_lockreason(data, custom_formats={}, name_prefix=None) if "type" in data_keys: data_keys.remove("type") data__type = data["type"] - data__type_any_of_count38 = 0 - if not data__type_any_of_count38: + data__type_any_of_count44 = 0 + if not data__type_any_of_count44: try: validate___definitions_lockreasontype( data__type, custom_formats, (name_prefix or "data") + ".type", ) - data__type_any_of_count38 += 1 + data__type_any_of_count44 += 1 except JsonSchemaValueException: pass - if not data__type_any_of_count38: + if not data__type_any_of_count44: try: if not isinstance(data__type, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".type must be null" ) - data__type_any_of_count38 += 1 + data__type_any_of_count44 += 1 except JsonSchemaValueException: pass - if not data__type_any_of_count38: + if not data__type_any_of_count44: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -1806,10 +2058,10 @@ def validate___definitions_lockreason(data, custom_formats={}, name_prefix=None) + str(data_keys) + " properties" ) - data_any_of_count36 += 1 + data_any_of_count42 += 1 except JsonSchemaValueException: pass - if not data_any_of_count36: + if not data_any_of_count42: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -1829,11 +2081,11 @@ def validate___definitions_lockreasontype(data, custom_formats={}, name_prefix=N def validate___definitions_tags(data, custom_formats={}, name_prefix=None): - data_any_of_count39 = 0 - if not data_any_of_count39: + data_any_of_count45 = 0 + if not data_any_of_count45: try: - data_any_of_count40 = 0 - if not data_any_of_count40: + data_any_of_count46 = 0 + if not data_any_of_count46: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -1852,10 +2104,10 @@ def validate___definitions_tags(data, custom_formats={}, name_prefix=None): + ".{data_key}".format(**locals()) + " must be string or null" ) - data_any_of_count40 += 1 + data_any_of_count46 += 1 except JsonSchemaValueException: pass - if not data_any_of_count40: + if not data_any_of_count46: try: if not isinstance(data, (list, tuple)): raise JsonSchemaValueException( @@ -1865,8 +2117,8 @@ def validate___definitions_tags(data, custom_formats={}, name_prefix=None): if data_is_list: data_len = len(data) for data_x, data_item in enumerate(data): - data_item_any_of_count41 = 0 - if not data_item_any_of_count41: + data_item_any_of_count47 = 0 + if not data_item_any_of_count47: try: validate___definitions_tagentry( data_item, @@ -1874,10 +2126,10 @@ def validate___definitions_tags(data, custom_formats={}, name_prefix=None): (name_prefix or "data") + "[{data_x}]".format(**locals()), ) - data_item_any_of_count41 += 1 + data_item_any_of_count47 += 1 except JsonSchemaValueException: pass - if not data_item_any_of_count41: + if not data_item_any_of_count47: try: if not isinstance(data_item, (NoneType)): raise JsonSchemaValueException( @@ -1886,29 +2138,29 @@ def validate___definitions_tags(data, custom_formats={}, name_prefix=None): + "[{data_x}]".format(**locals()) + " must be null" ) - data_item_any_of_count41 += 1 + data_item_any_of_count47 += 1 except JsonSchemaValueException: pass - if not data_item_any_of_count41: + if not data_item_any_of_count47: raise JsonSchemaValueException( "" + (name_prefix or "data") + "[{data_x}]".format(**locals()) + " cannot be validated by any definition" ) - data_any_of_count40 += 1 + data_any_of_count46 += 1 except JsonSchemaValueException: pass - if not data_any_of_count40: + if not data_any_of_count46: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) - data_any_of_count39 += 1 + data_any_of_count45 += 1 except JsonSchemaValueException: pass - if not data_any_of_count39: + if not data_any_of_count45: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -1916,8 +2168,8 @@ def validate___definitions_tags(data, custom_formats={}, name_prefix=None): def validate___definitions_tagentry(data, custom_formats={}, name_prefix=None): - data_any_of_count42 = 0 - if not data_any_of_count42: + data_any_of_count48 = 0 + if not data_any_of_count48: try: if not isinstance(data, (list, tuple)): raise JsonSchemaValueException( @@ -1948,10 +2200,10 @@ def validate___definitions_tagentry(data, custom_formats={}, name_prefix=None): raise JsonSchemaValueException( "" + (name_prefix or "data") + "[1] must be string or null" ) - data_any_of_count42 += 1 + data_any_of_count48 += 1 except JsonSchemaValueException: pass - if not data_any_of_count42: + if not data_any_of_count48: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -1959,16 +2211,16 @@ def validate___definitions_tagentry(data, custom_formats={}, name_prefix=None): def validate___definitions_stacktrace(data, custom_formats={}, name_prefix=None): - data_any_of_count43 = 0 - if not data_any_of_count43: + data_any_of_count49 = 0 + if not data_any_of_count49: try: validate___definitions_rawstacktrace( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count43 += 1 + data_any_of_count49 += 1 except JsonSchemaValueException: pass - if not data_any_of_count43: + if not data_any_of_count49: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -1976,8 +2228,8 @@ def validate___definitions_stacktrace(data, custom_formats={}, name_prefix=None) def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=None): - data_any_of_count44 = 0 - if not data_any_of_count44: + data_any_of_count50 = 0 + if not data_any_of_count50: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -2006,8 +2258,8 @@ def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=No for data__frames_x, data__frames_item in enumerate( data__frames ): - data__frames_item_any_of_count45 = 0 - if not data__frames_item_any_of_count45: + data__frames_item_any_of_count51 = 0 + if not data__frames_item_any_of_count51: try: validate___definitions_frame( data__frames_item, @@ -2017,10 +2269,10 @@ def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=No **locals() ), ) - data__frames_item_any_of_count45 += 1 + data__frames_item_any_of_count51 += 1 except JsonSchemaValueException: pass - if not data__frames_item_any_of_count45: + if not data__frames_item_any_of_count51: try: if not isinstance(data__frames_item, (NoneType)): raise JsonSchemaValueException( @@ -2031,10 +2283,10 @@ def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=No ) + " must be null" ) - data__frames_item_any_of_count45 += 1 + data__frames_item_any_of_count51 += 1 except JsonSchemaValueException: pass - if not data__frames_item_any_of_count45: + if not data__frames_item_any_of_count51: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2046,8 +2298,8 @@ def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=No data__instructionaddradjustment = data[ "instruction_addr_adjustment" ] - data__instructionaddradjustment_any_of_count46 = 0 - if not data__instructionaddradjustment_any_of_count46: + data__instructionaddradjustment_any_of_count52 = 0 + if not data__instructionaddradjustment_any_of_count52: try: validate___definitions_instructionaddradjustment( data__instructionaddradjustment, @@ -2055,10 +2307,10 @@ def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=No (name_prefix or "data") + ".instruction_addr_adjustment", ) - data__instructionaddradjustment_any_of_count46 += 1 + data__instructionaddradjustment_any_of_count52 += 1 except JsonSchemaValueException: pass - if not data__instructionaddradjustment_any_of_count46: + if not data__instructionaddradjustment_any_of_count52: try: if not isinstance( data__instructionaddradjustment, (NoneType) @@ -2068,10 +2320,10 @@ def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=No + (name_prefix or "data") + ".instruction_addr_adjustment must be null" ) - data__instructionaddradjustment_any_of_count46 += 1 + data__instructionaddradjustment_any_of_count52 += 1 except JsonSchemaValueException: pass - if not data__instructionaddradjustment_any_of_count46: + if not data__instructionaddradjustment_any_of_count52: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2103,8 +2355,8 @@ def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=No data__registers_value = data__registers.get( data__registers_key ) - data__registers_value_any_of_count47 = 0 - if not data__registers_value_any_of_count47: + data__registers_value_any_of_count53 = 0 + if not data__registers_value_any_of_count53: try: validate___definitions_regval( data__registers_value, @@ -2114,10 +2366,10 @@ def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=No **locals() ), ) - data__registers_value_any_of_count47 += 1 + data__registers_value_any_of_count53 += 1 except JsonSchemaValueException: pass - if not data__registers_value_any_of_count47: + if not data__registers_value_any_of_count53: try: if not isinstance( data__registers_value, (NoneType) @@ -2130,10 +2382,10 @@ def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=No ) + " must be null" ) - data__registers_value_any_of_count47 += 1 + data__registers_value_any_of_count53 += 1 except JsonSchemaValueException: pass - if not data__registers_value_any_of_count47: + if not data__registers_value_any_of_count53: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2159,10 +2411,10 @@ def validate___definitions_rawstacktrace(data, custom_formats={}, name_prefix=No + str(data_keys) + " properties" ) - data_any_of_count44 += 1 + data_any_of_count50 += 1 except JsonSchemaValueException: pass - if not data_any_of_count44: + if not data_any_of_count50: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -2190,8 +2442,8 @@ def validate___definitions_instructionaddradjustment( def validate___definitions_frame(data, custom_formats={}, name_prefix=None): - data_any_of_count48 = 0 - if not data_any_of_count48: + data_any_of_count54 = 0 + if not data_any_of_count54: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -2203,18 +2455,18 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): if "abs_path" in data_keys: data_keys.remove("abs_path") data__abspath = data["abs_path"] - data__abspath_any_of_count49 = 0 - if not data__abspath_any_of_count49: + data__abspath_any_of_count55 = 0 + if not data__abspath_any_of_count55: try: validate___definitions_nativeimagepath( data__abspath, custom_formats, (name_prefix or "data") + ".abs_path", ) - data__abspath_any_of_count49 += 1 + data__abspath_any_of_count55 += 1 except JsonSchemaValueException: pass - if not data__abspath_any_of_count49: + if not data__abspath_any_of_count55: try: if not isinstance(data__abspath, (NoneType)): raise JsonSchemaValueException( @@ -2222,10 +2474,10 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".abs_path must be null" ) - data__abspath_any_of_count49 += 1 + data__abspath_any_of_count55 += 1 except JsonSchemaValueException: pass - if not data__abspath_any_of_count49: + if not data__abspath_any_of_count55: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2274,18 +2526,18 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): if "filename" in data_keys: data_keys.remove("filename") data__filename = data["filename"] - data__filename_any_of_count50 = 0 - if not data__filename_any_of_count50: + data__filename_any_of_count56 = 0 + if not data__filename_any_of_count56: try: validate___definitions_nativeimagepath( data__filename, custom_formats, (name_prefix or "data") + ".filename", ) - data__filename_any_of_count50 += 1 + data__filename_any_of_count56 += 1 except JsonSchemaValueException: pass - if not data__filename_any_of_count50: + if not data__filename_any_of_count56: try: if not isinstance(data__filename, (NoneType)): raise JsonSchemaValueException( @@ -2293,10 +2545,10 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".filename must be null" ) - data__filename_any_of_count50 += 1 + data__filename_any_of_count56 += 1 except JsonSchemaValueException: pass - if not data__filename_any_of_count50: + if not data__filename_any_of_count56: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2314,18 +2566,18 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): if "function_id" in data_keys: data_keys.remove("function_id") data__functionid = data["function_id"] - data__functionid_any_of_count51 = 0 - if not data__functionid_any_of_count51: + data__functionid_any_of_count57 = 0 + if not data__functionid_any_of_count57: try: validate___definitions_addr( data__functionid, custom_formats, (name_prefix or "data") + ".function_id", ) - data__functionid_any_of_count51 += 1 + data__functionid_any_of_count57 += 1 except JsonSchemaValueException: pass - if not data__functionid_any_of_count51: + if not data__functionid_any_of_count57: try: if not isinstance(data__functionid, (NoneType)): raise JsonSchemaValueException( @@ -2333,10 +2585,10 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".function_id must be null" ) - data__functionid_any_of_count51 += 1 + data__functionid_any_of_count57 += 1 except JsonSchemaValueException: pass - if not data__functionid_any_of_count51: + if not data__functionid_any_of_count57: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2345,18 +2597,18 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): if "image_addr" in data_keys: data_keys.remove("image_addr") data__imageaddr = data["image_addr"] - data__imageaddr_any_of_count52 = 0 - if not data__imageaddr_any_of_count52: + data__imageaddr_any_of_count58 = 0 + if not data__imageaddr_any_of_count58: try: validate___definitions_addr( data__imageaddr, custom_formats, (name_prefix or "data") + ".image_addr", ) - data__imageaddr_any_of_count52 += 1 + data__imageaddr_any_of_count58 += 1 except JsonSchemaValueException: pass - if not data__imageaddr_any_of_count52: + if not data__imageaddr_any_of_count58: try: if not isinstance(data__imageaddr, (NoneType)): raise JsonSchemaValueException( @@ -2364,10 +2616,10 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".image_addr must be null" ) - data__imageaddr_any_of_count52 += 1 + data__imageaddr_any_of_count58 += 1 except JsonSchemaValueException: pass - if not data__imageaddr_any_of_count52: + if not data__imageaddr_any_of_count58: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2385,18 +2637,18 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): if "instruction_addr" in data_keys: data_keys.remove("instruction_addr") data__instructionaddr = data["instruction_addr"] - data__instructionaddr_any_of_count53 = 0 - if not data__instructionaddr_any_of_count53: + data__instructionaddr_any_of_count59 = 0 + if not data__instructionaddr_any_of_count59: try: validate___definitions_addr( data__instructionaddr, custom_formats, (name_prefix or "data") + ".instruction_addr", ) - data__instructionaddr_any_of_count53 += 1 + data__instructionaddr_any_of_count59 += 1 except JsonSchemaValueException: pass - if not data__instructionaddr_any_of_count53: + if not data__instructionaddr_any_of_count59: try: if not isinstance(data__instructionaddr, (NoneType)): raise JsonSchemaValueException( @@ -2404,10 +2656,10 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".instruction_addr must be null" ) - data__instructionaddr_any_of_count53 += 1 + data__instructionaddr_any_of_count59 += 1 except JsonSchemaValueException: pass - if not data__instructionaddr_any_of_count53: + if not data__instructionaddr_any_of_count59: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2439,27 +2691,27 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): if "lock" in data_keys: data_keys.remove("lock") data__lock = data["lock"] - data__lock_any_of_count54 = 0 - if not data__lock_any_of_count54: + data__lock_any_of_count60 = 0 + if not data__lock_any_of_count60: try: validate___definitions_lockreason( data__lock, custom_formats, (name_prefix or "data") + ".lock", ) - data__lock_any_of_count54 += 1 + data__lock_any_of_count60 += 1 except JsonSchemaValueException: pass - if not data__lock_any_of_count54: + if not data__lock_any_of_count60: try: if not isinstance(data__lock, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".lock must be null" ) - data__lock_any_of_count54 += 1 + data__lock_any_of_count60 += 1 except JsonSchemaValueException: pass - if not data__lock_any_of_count54: + if not data__lock_any_of_count60: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2574,18 +2826,18 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): if "symbol_addr" in data_keys: data_keys.remove("symbol_addr") data__symboladdr = data["symbol_addr"] - data__symboladdr_any_of_count55 = 0 - if not data__symboladdr_any_of_count55: + data__symboladdr_any_of_count61 = 0 + if not data__symboladdr_any_of_count61: try: validate___definitions_addr( data__symboladdr, custom_formats, (name_prefix or "data") + ".symbol_addr", ) - data__symboladdr_any_of_count55 += 1 + data__symboladdr_any_of_count61 += 1 except JsonSchemaValueException: pass - if not data__symboladdr_any_of_count55: + if not data__symboladdr_any_of_count61: try: if not isinstance(data__symboladdr, (NoneType)): raise JsonSchemaValueException( @@ -2593,10 +2845,10 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".symbol_addr must be null" ) - data__symboladdr_any_of_count55 += 1 + data__symboladdr_any_of_count61 += 1 except JsonSchemaValueException: pass - if not data__symboladdr_any_of_count55: + if not data__symboladdr_any_of_count61: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2605,27 +2857,27 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): if "vars" in data_keys: data_keys.remove("vars") data__vars = data["vars"] - data__vars_any_of_count56 = 0 - if not data__vars_any_of_count56: + data__vars_any_of_count62 = 0 + if not data__vars_any_of_count62: try: validate___definitions_framevars( data__vars, custom_formats, (name_prefix or "data") + ".vars", ) - data__vars_any_of_count56 += 1 + data__vars_any_of_count62 += 1 except JsonSchemaValueException: pass - if not data__vars_any_of_count56: + if not data__vars_any_of_count62: try: if not isinstance(data__vars, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".vars must be null" ) - data__vars_any_of_count56 += 1 + data__vars_any_of_count62 += 1 except JsonSchemaValueException: pass - if not data__vars_any_of_count56: + if not data__vars_any_of_count62: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2639,10 +2891,10 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count48 += 1 + data_any_of_count54 += 1 except JsonSchemaValueException: pass - if not data_any_of_count48: + if not data_any_of_count54: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -2650,8 +2902,8 @@ def validate___definitions_frame(data, custom_formats={}, name_prefix=None): def validate___definitions_framevars(data, custom_formats={}, name_prefix=None): - data_any_of_count57 = 0 - if not data_any_of_count57: + data_any_of_count63 = 0 + if not data_any_of_count63: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -2660,10 +2912,10 @@ def validate___definitions_framevars(data, custom_formats={}, name_prefix=None): data_is_dict = isinstance(data, dict) if data_is_dict: data_keys = set(data.keys()) - data_any_of_count57 += 1 + data_any_of_count63 += 1 except JsonSchemaValueException: pass - if not data_any_of_count57: + if not data_any_of_count63: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -2677,17 +2929,17 @@ def validate___definitions_addr(data, custom_formats={}, name_prefix=None): def validate___definitions_nativeimagepath(data, custom_formats={}, name_prefix=None): - data_any_of_count58 = 0 - if not data_any_of_count58: + data_any_of_count64 = 0 + if not data_any_of_count64: try: if not isinstance(data, (str)): raise JsonSchemaValueException( "" + (name_prefix or "data") + " must be string" ) - data_any_of_count58 += 1 + data_any_of_count64 += 1 except JsonSchemaValueException: pass - if not data_any_of_count58: + if not data_any_of_count64: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -2695,8 +2947,8 @@ def validate___definitions_nativeimagepath(data, custom_formats={}, name_prefix= def validate___definitions_clientsdkinfo(data, custom_formats={}, name_prefix=None): - data_any_of_count59 = 0 - if not data_any_of_count59: + data_any_of_count65 = 0 + if not data_any_of_count65: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -2782,8 +3034,8 @@ def validate___definitions_clientsdkinfo(data, custom_formats={}, name_prefix=No for data__packages_x, data__packages_item in enumerate( data__packages ): - data__packages_item_any_of_count60 = 0 - if not data__packages_item_any_of_count60: + data__packages_item_any_of_count66 = 0 + if not data__packages_item_any_of_count66: try: validate___definitions_clientsdkpackage( data__packages_item, @@ -2793,10 +3045,10 @@ def validate___definitions_clientsdkinfo(data, custom_formats={}, name_prefix=No **locals() ), ) - data__packages_item_any_of_count60 += 1 + data__packages_item_any_of_count66 += 1 except JsonSchemaValueException: pass - if not data__packages_item_any_of_count60: + if not data__packages_item_any_of_count66: try: if not isinstance(data__packages_item, (NoneType)): raise JsonSchemaValueException( @@ -2807,10 +3059,10 @@ def validate___definitions_clientsdkinfo(data, custom_formats={}, name_prefix=No ) + " must be null" ) - data__packages_item_any_of_count60 += 1 + data__packages_item_any_of_count66 += 1 except JsonSchemaValueException: pass - if not data__packages_item_any_of_count60: + if not data__packages_item_any_of_count66: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2834,10 +3086,10 @@ def validate___definitions_clientsdkinfo(data, custom_formats={}, name_prefix=No + str(data_keys) + " properties" ) - data_any_of_count59 += 1 + data_any_of_count65 += 1 except JsonSchemaValueException: pass - if not data_any_of_count59: + if not data_any_of_count65: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -2845,8 +3097,8 @@ def validate___definitions_clientsdkinfo(data, custom_formats={}, name_prefix=No def validate___definitions_clientsdkpackage(data, custom_formats={}, name_prefix=None): - data_any_of_count61 = 0 - if not data_any_of_count61: + data_any_of_count67 = 0 + if not data_any_of_count67: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -2881,10 +3133,10 @@ def validate___definitions_clientsdkpackage(data, custom_formats={}, name_prefix + str(data_keys) + " properties" ) - data_any_of_count61 += 1 + data_any_of_count67 += 1 except JsonSchemaValueException: pass - if not data_any_of_count61: + if not data_any_of_count67: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -2892,8 +3144,8 @@ def validate___definitions_clientsdkpackage(data, custom_formats={}, name_prefix def validate___definitions_request(data, custom_formats={}, name_prefix=None): - data_any_of_count62 = 0 - if not data_any_of_count62: + data_any_of_count68 = 0 + if not data_any_of_count68: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -2937,18 +3189,18 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): if "cookies" in data_keys: data_keys.remove("cookies") data__cookies = data["cookies"] - data__cookies_any_of_count63 = 0 - if not data__cookies_any_of_count63: + data__cookies_any_of_count69 = 0 + if not data__cookies_any_of_count69: try: validate___definitions_cookies( data__cookies, custom_formats, (name_prefix or "data") + ".cookies", ) - data__cookies_any_of_count63 += 1 + data__cookies_any_of_count69 += 1 except JsonSchemaValueException: pass - if not data__cookies_any_of_count63: + if not data__cookies_any_of_count69: try: if not isinstance(data__cookies, (NoneType)): raise JsonSchemaValueException( @@ -2956,10 +3208,10 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".cookies must be null" ) - data__cookies_any_of_count63 += 1 + data__cookies_any_of_count69 += 1 except JsonSchemaValueException: pass - if not data__cookies_any_of_count63: + if not data__cookies_any_of_count69: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -2990,18 +3242,18 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): if "headers" in data_keys: data_keys.remove("headers") data__headers = data["headers"] - data__headers_any_of_count64 = 0 - if not data__headers_any_of_count64: + data__headers_any_of_count70 = 0 + if not data__headers_any_of_count70: try: validate___definitions_headers( data__headers, custom_formats, (name_prefix or "data") + ".headers", ) - data__headers_any_of_count64 += 1 + data__headers_any_of_count70 += 1 except JsonSchemaValueException: pass - if not data__headers_any_of_count64: + if not data__headers_any_of_count70: try: if not isinstance(data__headers, (NoneType)): raise JsonSchemaValueException( @@ -3009,10 +3261,10 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".headers must be null" ) - data__headers_any_of_count64 += 1 + data__headers_any_of_count70 += 1 except JsonSchemaValueException: pass - if not data__headers_any_of_count64: + if not data__headers_any_of_count70: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -3048,11 +3300,11 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): if "query_string" in data_keys: data_keys.remove("query_string") data__querystring = data["query_string"] - data__querystring_any_of_count65 = 0 - if not data__querystring_any_of_count65: + data__querystring_any_of_count71 = 0 + if not data__querystring_any_of_count71: try: - data__querystring_any_of_count66 = 0 - if not data__querystring_any_of_count66: + data__querystring_any_of_count72 = 0 + if not data__querystring_any_of_count72: try: if not isinstance(data__querystring, (str)): raise JsonSchemaValueException( @@ -3060,13 +3312,13 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".query_string must be string" ) - data__querystring_any_of_count66 += 1 + data__querystring_any_of_count72 += 1 except JsonSchemaValueException: pass - if not data__querystring_any_of_count66: + if not data__querystring_any_of_count72: try: - data__querystring_any_of_count67 = 0 - if not data__querystring_any_of_count67: + data__querystring_any_of_count73 = 0 + if not data__querystring_any_of_count73: try: if not isinstance( data__querystring, (dict) @@ -3107,10 +3359,10 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): ) + " must be string or null" ) - data__querystring_any_of_count67 += 1 + data__querystring_any_of_count73 += 1 except JsonSchemaValueException: pass - if not data__querystring_any_of_count67: + if not data__querystring_any_of_count73: try: if not isinstance( data__querystring, (list, tuple) @@ -3227,28 +3479,28 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): ) + " must be string or null" ) - data__querystring_any_of_count67 += 1 + data__querystring_any_of_count73 += 1 except JsonSchemaValueException: pass - if not data__querystring_any_of_count67: + if not data__querystring_any_of_count73: raise JsonSchemaValueException( "" + (name_prefix or "data") + ".query_string cannot be validated by any definition" ) - data__querystring_any_of_count66 += 1 + data__querystring_any_of_count72 += 1 except JsonSchemaValueException: pass - if not data__querystring_any_of_count66: + if not data__querystring_any_of_count72: raise JsonSchemaValueException( "" + (name_prefix or "data") + ".query_string cannot be validated by any definition" ) - data__querystring_any_of_count65 += 1 + data__querystring_any_of_count71 += 1 except JsonSchemaValueException: pass - if not data__querystring_any_of_count65: + if not data__querystring_any_of_count71: try: if not isinstance(data__querystring, (NoneType)): raise JsonSchemaValueException( @@ -3256,10 +3508,10 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".query_string must be null" ) - data__querystring_any_of_count65 += 1 + data__querystring_any_of_count71 += 1 except JsonSchemaValueException: pass - if not data__querystring_any_of_count65: + if not data__querystring_any_of_count71: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -3280,10 +3532,10 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count62 += 1 + data_any_of_count68 += 1 except JsonSchemaValueException: pass - if not data_any_of_count62: + if not data_any_of_count68: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -3291,11 +3543,11 @@ def validate___definitions_request(data, custom_formats={}, name_prefix=None): def validate___definitions_headers(data, custom_formats={}, name_prefix=None): - data_any_of_count68 = 0 - if not data_any_of_count68: + data_any_of_count74 = 0 + if not data_any_of_count74: try: - data_any_of_count69 = 0 - if not data_any_of_count69: + data_any_of_count75 = 0 + if not data_any_of_count75: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -3307,8 +3559,8 @@ def validate___definitions_headers(data, custom_formats={}, name_prefix=None): for data_key in data_keys: if data_key not in []: data_value = data.get(data_key) - data_value_any_of_count70 = 0 - if not data_value_any_of_count70: + data_value_any_of_count76 = 0 + if not data_value_any_of_count76: try: validate___definitions_headervalue( data_value, @@ -3316,10 +3568,10 @@ def validate___definitions_headers(data, custom_formats={}, name_prefix=None): (name_prefix or "data") + ".{data_key}".format(**locals()), ) - data_value_any_of_count70 += 1 + data_value_any_of_count76 += 1 except JsonSchemaValueException: pass - if not data_value_any_of_count70: + if not data_value_any_of_count76: try: if not isinstance(data_value, (NoneType)): raise JsonSchemaValueException( @@ -3328,20 +3580,20 @@ def validate___definitions_headers(data, custom_formats={}, name_prefix=None): + ".{data_key}".format(**locals()) + " must be null" ) - data_value_any_of_count70 += 1 + data_value_any_of_count76 += 1 except JsonSchemaValueException: pass - if not data_value_any_of_count70: + if not data_value_any_of_count76: raise JsonSchemaValueException( "" + (name_prefix or "data") + ".{data_key}".format(**locals()) + " cannot be validated by any definition" ) - data_any_of_count69 += 1 + data_any_of_count75 += 1 except JsonSchemaValueException: pass - if not data_any_of_count69: + if not data_any_of_count75: try: if not isinstance(data, (list, tuple)): raise JsonSchemaValueException( @@ -3377,8 +3629,8 @@ def validate___definitions_headers(data, custom_formats={}, name_prefix=None): ) if data_item_len > 0: data_item__0 = data_item[0] - data_item__0_any_of_count71 = 0 - if not data_item__0_any_of_count71: + data_item__0_any_of_count77 = 0 + if not data_item__0_any_of_count77: try: validate___definitions_headername( data_item__0, @@ -3386,10 +3638,10 @@ def validate___definitions_headers(data, custom_formats={}, name_prefix=None): (name_prefix or "data") + "[{data_x}][0]".format(**locals()), ) - data_item__0_any_of_count71 += 1 + data_item__0_any_of_count77 += 1 except JsonSchemaValueException: pass - if not data_item__0_any_of_count71: + if not data_item__0_any_of_count77: try: if not isinstance(data_item__0, (NoneType)): raise JsonSchemaValueException( @@ -3398,10 +3650,10 @@ def validate___definitions_headers(data, custom_formats={}, name_prefix=None): + "[{data_x}][0]".format(**locals()) + " must be null" ) - data_item__0_any_of_count71 += 1 + data_item__0_any_of_count77 += 1 except JsonSchemaValueException: pass - if not data_item__0_any_of_count71: + if not data_item__0_any_of_count77: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -3410,8 +3662,8 @@ def validate___definitions_headers(data, custom_formats={}, name_prefix=None): ) if data_item_len > 1: data_item__1 = data_item[1] - data_item__1_any_of_count72 = 0 - if not data_item__1_any_of_count72: + data_item__1_any_of_count78 = 0 + if not data_item__1_any_of_count78: try: validate___definitions_headervalue( data_item__1, @@ -3419,10 +3671,10 @@ def validate___definitions_headers(data, custom_formats={}, name_prefix=None): (name_prefix or "data") + "[{data_x}][1]".format(**locals()), ) - data_item__1_any_of_count72 += 1 + data_item__1_any_of_count78 += 1 except JsonSchemaValueException: pass - if not data_item__1_any_of_count72: + if not data_item__1_any_of_count78: try: if not isinstance(data_item__1, (NoneType)): raise JsonSchemaValueException( @@ -3431,61 +3683,25 @@ def validate___definitions_headers(data, custom_formats={}, name_prefix=None): + "[{data_x}][1]".format(**locals()) + " must be null" ) - data_item__1_any_of_count72 += 1 + data_item__1_any_of_count78 += 1 except JsonSchemaValueException: pass - if not data_item__1_any_of_count72: + if not data_item__1_any_of_count78: raise JsonSchemaValueException( "" + (name_prefix or "data") + "[{data_x}][1]".format(**locals()) + " cannot be validated by any definition" ) - data_any_of_count69 += 1 + data_any_of_count75 += 1 except JsonSchemaValueException: pass - if not data_any_of_count69: + if not data_any_of_count75: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) - data_any_of_count68 += 1 - except JsonSchemaValueException: - pass - if not data_any_of_count68: - raise JsonSchemaValueException( - "" + (name_prefix or "data") + " cannot be validated by any definition" - ) - return data - - -def validate___definitions_headername(data, custom_formats={}, name_prefix=None): - data_any_of_count73 = 0 - if not data_any_of_count73: - try: - if not isinstance(data, (str)): - raise JsonSchemaValueException( - "" + (name_prefix or "data") + " must be string" - ) - data_any_of_count73 += 1 - except JsonSchemaValueException: - pass - if not data_any_of_count73: - raise JsonSchemaValueException( - "" + (name_prefix or "data") + " cannot be validated by any definition" - ) - return data - - -def validate___definitions_headervalue(data, custom_formats={}, name_prefix=None): - data_any_of_count74 = 0 - if not data_any_of_count74: - try: - if not isinstance(data, (str)): - raise JsonSchemaValueException( - "" + (name_prefix or "data") + " must be string" - ) data_any_of_count74 += 1 except JsonSchemaValueException: pass @@ -3496,12 +3712,48 @@ def validate___definitions_headervalue(data, custom_formats={}, name_prefix=None return data -def validate___definitions_cookies(data, custom_formats={}, name_prefix=None): - data_any_of_count75 = 0 - if not data_any_of_count75: +def validate___definitions_headername(data, custom_formats={}, name_prefix=None): + data_any_of_count79 = 0 + if not data_any_of_count79: try: - data_any_of_count76 = 0 - if not data_any_of_count76: + if not isinstance(data, (str)): + raise JsonSchemaValueException( + "" + (name_prefix or "data") + " must be string" + ) + data_any_of_count79 += 1 + except JsonSchemaValueException: + pass + if not data_any_of_count79: + raise JsonSchemaValueException( + "" + (name_prefix or "data") + " cannot be validated by any definition" + ) + return data + + +def validate___definitions_headervalue(data, custom_formats={}, name_prefix=None): + data_any_of_count80 = 0 + if not data_any_of_count80: + try: + if not isinstance(data, (str)): + raise JsonSchemaValueException( + "" + (name_prefix or "data") + " must be string" + ) + data_any_of_count80 += 1 + except JsonSchemaValueException: + pass + if not data_any_of_count80: + raise JsonSchemaValueException( + "" + (name_prefix or "data") + " cannot be validated by any definition" + ) + return data + + +def validate___definitions_cookies(data, custom_formats={}, name_prefix=None): + data_any_of_count81 = 0 + if not data_any_of_count81: + try: + data_any_of_count82 = 0 + if not data_any_of_count82: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -3520,10 +3772,10 @@ def validate___definitions_cookies(data, custom_formats={}, name_prefix=None): + ".{data_key}".format(**locals()) + " must be string or null" ) - data_any_of_count76 += 1 + data_any_of_count82 += 1 except JsonSchemaValueException: pass - if not data_any_of_count76: + if not data_any_of_count82: try: if not isinstance(data, (list, tuple)): raise JsonSchemaValueException( @@ -3575,19 +3827,19 @@ def validate___definitions_cookies(data, custom_formats={}, name_prefix=None): + "[{data_x}][1]".format(**locals()) + " must be string or null" ) - data_any_of_count76 += 1 + data_any_of_count82 += 1 except JsonSchemaValueException: pass - if not data_any_of_count76: + if not data_any_of_count82: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) - data_any_of_count75 += 1 + data_any_of_count81 += 1 except JsonSchemaValueException: pass - if not data_any_of_count75: + if not data_any_of_count81: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -3595,26 +3847,26 @@ def validate___definitions_cookies(data, custom_formats={}, name_prefix=None): def validate___definitions_timestamp(data, custom_formats={}, name_prefix=None): - data_any_of_count77 = 0 - if not data_any_of_count77: + data_any_of_count83 = 0 + if not data_any_of_count83: try: if not isinstance(data, (int, float, Decimal)) or isinstance(data, bool): raise JsonSchemaValueException( "" + (name_prefix or "data") + " must be number" ) - data_any_of_count77 += 1 + data_any_of_count83 += 1 except JsonSchemaValueException: pass - if not data_any_of_count77: + if not data_any_of_count83: try: if not isinstance(data, (str)): raise JsonSchemaValueException( "" + (name_prefix or "data") + " must be string" ) - data_any_of_count77 += 1 + data_any_of_count83 += 1 except JsonSchemaValueException: pass - if not data_any_of_count77: + if not data_any_of_count83: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -3622,8 +3874,8 @@ def validate___definitions_timestamp(data, custom_formats={}, name_prefix=None): def validate___definitions_logentry(data, custom_formats={}, name_prefix=None): - data_any_of_count78 = 0 - if not data_any_of_count78: + data_any_of_count84 = 0 + if not data_any_of_count84: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -3635,18 +3887,18 @@ def validate___definitions_logentry(data, custom_formats={}, name_prefix=None): if "formatted" in data_keys: data_keys.remove("formatted") data__formatted = data["formatted"] - data__formatted_any_of_count79 = 0 - if not data__formatted_any_of_count79: + data__formatted_any_of_count85 = 0 + if not data__formatted_any_of_count85: try: validate___definitions_message( data__formatted, custom_formats, (name_prefix or "data") + ".formatted", ) - data__formatted_any_of_count79 += 1 + data__formatted_any_of_count85 += 1 except JsonSchemaValueException: pass - if not data__formatted_any_of_count79: + if not data__formatted_any_of_count85: try: if not isinstance(data__formatted, (NoneType)): raise JsonSchemaValueException( @@ -3654,10 +3906,10 @@ def validate___definitions_logentry(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".formatted must be null" ) - data__formatted_any_of_count79 += 1 + data__formatted_any_of_count85 += 1 except JsonSchemaValueException: pass - if not data__formatted_any_of_count79: + if not data__formatted_any_of_count85: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -3666,18 +3918,18 @@ def validate___definitions_logentry(data, custom_formats={}, name_prefix=None): if "message" in data_keys: data_keys.remove("message") data__message = data["message"] - data__message_any_of_count80 = 0 - if not data__message_any_of_count80: + data__message_any_of_count86 = 0 + if not data__message_any_of_count86: try: validate___definitions_message( data__message, custom_formats, (name_prefix or "data") + ".message", ) - data__message_any_of_count80 += 1 + data__message_any_of_count86 += 1 except JsonSchemaValueException: pass - if not data__message_any_of_count80: + if not data__message_any_of_count86: try: if not isinstance(data__message, (NoneType)): raise JsonSchemaValueException( @@ -3685,10 +3937,10 @@ def validate___definitions_logentry(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".message must be null" ) - data__message_any_of_count80 += 1 + data__message_any_of_count86 += 1 except JsonSchemaValueException: pass - if not data__message_any_of_count80: + if not data__message_any_of_count86: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -3705,10 +3957,10 @@ def validate___definitions_logentry(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count78 += 1 + data_any_of_count84 += 1 except JsonSchemaValueException: pass - if not data_any_of_count78: + if not data_any_of_count84: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -3716,17 +3968,17 @@ def validate___definitions_logentry(data, custom_formats={}, name_prefix=None): def validate___definitions_message(data, custom_formats={}, name_prefix=None): - data_any_of_count81 = 0 - if not data_any_of_count81: + data_any_of_count87 = 0 + if not data_any_of_count87: try: if not isinstance(data, (str)): raise JsonSchemaValueException( "" + (name_prefix or "data") + " must be string" ) - data_any_of_count81 += 1 + data_any_of_count87 += 1 except JsonSchemaValueException: pass - if not data_any_of_count81: + if not data_any_of_count87: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -3746,8 +3998,8 @@ def validate___definitions_level(data, custom_formats={}, name_prefix=None): def validate___definitions_fingerprint(data, custom_formats={}, name_prefix=None): - data_any_of_count82 = 0 - if not data_any_of_count82: + data_any_of_count88 = 0 + if not data_any_of_count88: try: if not isinstance(data, (list, tuple)): raise JsonSchemaValueException( @@ -3764,10 +4016,10 @@ def validate___definitions_fingerprint(data, custom_formats={}, name_prefix=None + "[{data_x}]".format(**locals()) + " must be string" ) - data_any_of_count82 += 1 + data_any_of_count88 += 1 except JsonSchemaValueException: pass - if not data_any_of_count82: + if not data_any_of_count88: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -3775,8 +4027,8 @@ def validate___definitions_fingerprint(data, custom_formats={}, name_prefix=None def validate___definitions_exception(data, custom_formats={}, name_prefix=None): - data_any_of_count83 = 0 - if not data_any_of_count83: + data_any_of_count89 = 0 + if not data_any_of_count89: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -3788,18 +4040,18 @@ def validate___definitions_exception(data, custom_formats={}, name_prefix=None): if "mechanism" in data_keys: data_keys.remove("mechanism") data__mechanism = data["mechanism"] - data__mechanism_any_of_count84 = 0 - if not data__mechanism_any_of_count84: + data__mechanism_any_of_count90 = 0 + if not data__mechanism_any_of_count90: try: validate___definitions_mechanism( data__mechanism, custom_formats, (name_prefix or "data") + ".mechanism", ) - data__mechanism_any_of_count84 += 1 + data__mechanism_any_of_count90 += 1 except JsonSchemaValueException: pass - if not data__mechanism_any_of_count84: + if not data__mechanism_any_of_count90: try: if not isinstance(data__mechanism, (NoneType)): raise JsonSchemaValueException( @@ -3807,10 +4059,10 @@ def validate___definitions_exception(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".mechanism must be null" ) - data__mechanism_any_of_count84 += 1 + data__mechanism_any_of_count90 += 1 except JsonSchemaValueException: pass - if not data__mechanism_any_of_count84: + if not data__mechanism_any_of_count90: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -3828,18 +4080,18 @@ def validate___definitions_exception(data, custom_formats={}, name_prefix=None): if "stacktrace" in data_keys: data_keys.remove("stacktrace") data__stacktrace = data["stacktrace"] - data__stacktrace_any_of_count85 = 0 - if not data__stacktrace_any_of_count85: + data__stacktrace_any_of_count91 = 0 + if not data__stacktrace_any_of_count91: try: validate___definitions_stacktrace( data__stacktrace, custom_formats, (name_prefix or "data") + ".stacktrace", ) - data__stacktrace_any_of_count85 += 1 + data__stacktrace_any_of_count91 += 1 except JsonSchemaValueException: pass - if not data__stacktrace_any_of_count85: + if not data__stacktrace_any_of_count91: try: if not isinstance(data__stacktrace, (NoneType)): raise JsonSchemaValueException( @@ -3847,10 +4099,10 @@ def validate___definitions_exception(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".stacktrace must be null" ) - data__stacktrace_any_of_count85 += 1 + data__stacktrace_any_of_count91 += 1 except JsonSchemaValueException: pass - if not data__stacktrace_any_of_count85: + if not data__stacktrace_any_of_count91: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -3859,18 +4111,18 @@ def validate___definitions_exception(data, custom_formats={}, name_prefix=None): if "thread_id" in data_keys: data_keys.remove("thread_id") data__threadid = data["thread_id"] - data__threadid_any_of_count86 = 0 - if not data__threadid_any_of_count86: + data__threadid_any_of_count92 = 0 + if not data__threadid_any_of_count92: try: validate___definitions_threadid( data__threadid, custom_formats, (name_prefix or "data") + ".thread_id", ) - data__threadid_any_of_count86 += 1 + data__threadid_any_of_count92 += 1 except JsonSchemaValueException: pass - if not data__threadid_any_of_count86: + if not data__threadid_any_of_count92: try: if not isinstance(data__threadid, (NoneType)): raise JsonSchemaValueException( @@ -3878,10 +4130,10 @@ def validate___definitions_exception(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".thread_id must be null" ) - data__threadid_any_of_count86 += 1 + data__threadid_any_of_count92 += 1 except JsonSchemaValueException: pass - if not data__threadid_any_of_count86: + if not data__threadid_any_of_count92: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -3899,27 +4151,27 @@ def validate___definitions_exception(data, custom_formats={}, name_prefix=None): if "value" in data_keys: data_keys.remove("value") data__value = data["value"] - data__value_any_of_count87 = 0 - if not data__value_any_of_count87: + data__value_any_of_count93 = 0 + if not data__value_any_of_count93: try: validate___definitions_jsonlenientstring( data__value, custom_formats, (name_prefix or "data") + ".value", ) - data__value_any_of_count87 += 1 + data__value_any_of_count93 += 1 except JsonSchemaValueException: pass - if not data__value_any_of_count87: + if not data__value_any_of_count93: try: if not isinstance(data__value, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".value must be null" ) - data__value_any_of_count87 += 1 + data__value_any_of_count93 += 1 except JsonSchemaValueException: pass - if not data__value_any_of_count87: + if not data__value_any_of_count93: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -3933,10 +4185,10 @@ def validate___definitions_exception(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count83 += 1 + data_any_of_count89 += 1 except JsonSchemaValueException: pass - if not data_any_of_count83: + if not data_any_of_count89: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -3944,17 +4196,17 @@ def validate___definitions_exception(data, custom_formats={}, name_prefix=None): def validate___definitions_jsonlenientstring(data, custom_formats={}, name_prefix=None): - data_any_of_count88 = 0 - if not data_any_of_count88: + data_any_of_count94 = 0 + if not data_any_of_count94: try: if not isinstance(data, (str)): raise JsonSchemaValueException( "" + (name_prefix or "data") + " must be string" ) - data_any_of_count88 += 1 + data_any_of_count94 += 1 except JsonSchemaValueException: pass - if not data_any_of_count88: + if not data_any_of_count94: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -3962,8 +4214,8 @@ def validate___definitions_jsonlenientstring(data, custom_formats={}, name_prefi def validate___definitions_mechanism(data, custom_formats={}, name_prefix=None): - data_any_of_count89 = 0 - if not data_any_of_count89: + data_any_of_count95 = 0 + if not data_any_of_count95: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4051,27 +4303,27 @@ def validate___definitions_mechanism(data, custom_formats={}, name_prefix=None): if "meta" in data_keys: data_keys.remove("meta") data__meta = data["meta"] - data__meta_any_of_count90 = 0 - if not data__meta_any_of_count90: + data__meta_any_of_count96 = 0 + if not data__meta_any_of_count96: try: validate___definitions_mechanismmeta( data__meta, custom_formats, (name_prefix or "data") + ".meta", ) - data__meta_any_of_count90 += 1 + data__meta_any_of_count96 += 1 except JsonSchemaValueException: pass - if not data__meta_any_of_count90: + if not data__meta_any_of_count96: try: if not isinstance(data__meta, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".meta must be null" ) - data__meta_any_of_count90 += 1 + data__meta_any_of_count96 += 1 except JsonSchemaValueException: pass - if not data__meta_any_of_count90: + if not data__meta_any_of_count96: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -4135,10 +4387,10 @@ def validate___definitions_mechanism(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count89 += 1 + data_any_of_count95 += 1 except JsonSchemaValueException: pass - if not data_any_of_count89: + if not data_any_of_count95: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4146,8 +4398,8 @@ def validate___definitions_mechanism(data, custom_formats={}, name_prefix=None): def validate___definitions_mechanismmeta(data, custom_formats={}, name_prefix=None): - data_any_of_count91 = 0 - if not data_any_of_count91: + data_any_of_count97 = 0 + if not data_any_of_count97: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4159,27 +4411,27 @@ def validate___definitions_mechanismmeta(data, custom_formats={}, name_prefix=No if "errno" in data_keys: data_keys.remove("errno") data__errno = data["errno"] - data__errno_any_of_count92 = 0 - if not data__errno_any_of_count92: + data__errno_any_of_count98 = 0 + if not data__errno_any_of_count98: try: validate___definitions_cerror( data__errno, custom_formats, (name_prefix or "data") + ".errno", ) - data__errno_any_of_count92 += 1 + data__errno_any_of_count98 += 1 except JsonSchemaValueException: pass - if not data__errno_any_of_count92: + if not data__errno_any_of_count98: try: if not isinstance(data__errno, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".errno must be null" ) - data__errno_any_of_count92 += 1 + data__errno_any_of_count98 += 1 except JsonSchemaValueException: pass - if not data__errno_any_of_count92: + if not data__errno_any_of_count98: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -4188,18 +4440,18 @@ def validate___definitions_mechanismmeta(data, custom_formats={}, name_prefix=No if "mach_exception" in data_keys: data_keys.remove("mach_exception") data__machexception = data["mach_exception"] - data__machexception_any_of_count93 = 0 - if not data__machexception_any_of_count93: + data__machexception_any_of_count99 = 0 + if not data__machexception_any_of_count99: try: validate___definitions_machexception( data__machexception, custom_formats, (name_prefix or "data") + ".mach_exception", ) - data__machexception_any_of_count93 += 1 + data__machexception_any_of_count99 += 1 except JsonSchemaValueException: pass - if not data__machexception_any_of_count93: + if not data__machexception_any_of_count99: try: if not isinstance(data__machexception, (NoneType)): raise JsonSchemaValueException( @@ -4207,10 +4459,10 @@ def validate___definitions_mechanismmeta(data, custom_formats={}, name_prefix=No + (name_prefix or "data") + ".mach_exception must be null" ) - data__machexception_any_of_count93 += 1 + data__machexception_any_of_count99 += 1 except JsonSchemaValueException: pass - if not data__machexception_any_of_count93: + if not data__machexception_any_of_count99: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -4219,18 +4471,18 @@ def validate___definitions_mechanismmeta(data, custom_formats={}, name_prefix=No if "ns_error" in data_keys: data_keys.remove("ns_error") data__nserror = data["ns_error"] - data__nserror_any_of_count94 = 0 - if not data__nserror_any_of_count94: + data__nserror_any_of_count100 = 0 + if not data__nserror_any_of_count100: try: validate___definitions_nserror( data__nserror, custom_formats, (name_prefix or "data") + ".ns_error", ) - data__nserror_any_of_count94 += 1 + data__nserror_any_of_count100 += 1 except JsonSchemaValueException: pass - if not data__nserror_any_of_count94: + if not data__nserror_any_of_count100: try: if not isinstance(data__nserror, (NoneType)): raise JsonSchemaValueException( @@ -4238,10 +4490,10 @@ def validate___definitions_mechanismmeta(data, custom_formats={}, name_prefix=No + (name_prefix or "data") + ".ns_error must be null" ) - data__nserror_any_of_count94 += 1 + data__nserror_any_of_count100 += 1 except JsonSchemaValueException: pass - if not data__nserror_any_of_count94: + if not data__nserror_any_of_count100: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -4250,18 +4502,18 @@ def validate___definitions_mechanismmeta(data, custom_formats={}, name_prefix=No if "signal" in data_keys: data_keys.remove("signal") data__signal = data["signal"] - data__signal_any_of_count95 = 0 - if not data__signal_any_of_count95: + data__signal_any_of_count101 = 0 + if not data__signal_any_of_count101: try: validate___definitions_posixsignal( data__signal, custom_formats, (name_prefix or "data") + ".signal", ) - data__signal_any_of_count95 += 1 + data__signal_any_of_count101 += 1 except JsonSchemaValueException: pass - if not data__signal_any_of_count95: + if not data__signal_any_of_count101: try: if not isinstance(data__signal, (NoneType)): raise JsonSchemaValueException( @@ -4269,10 +4521,10 @@ def validate___definitions_mechanismmeta(data, custom_formats={}, name_prefix=No + (name_prefix or "data") + ".signal must be null" ) - data__signal_any_of_count95 += 1 + data__signal_any_of_count101 += 1 except JsonSchemaValueException: pass - if not data__signal_any_of_count95: + if not data__signal_any_of_count101: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -4286,10 +4538,10 @@ def validate___definitions_mechanismmeta(data, custom_formats={}, name_prefix=No + str(data_keys) + " properties" ) - data_any_of_count91 += 1 + data_any_of_count97 += 1 except JsonSchemaValueException: pass - if not data_any_of_count91: + if not data_any_of_count97: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4297,8 +4549,8 @@ def validate___definitions_mechanismmeta(data, custom_formats={}, name_prefix=No def validate___definitions_posixsignal(data, custom_formats={}, name_prefix=None): - data_any_of_count96 = 0 - if not data_any_of_count96: + data_any_of_count102 = 0 + if not data_any_of_count102: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4364,10 +4616,10 @@ def validate___definitions_posixsignal(data, custom_formats={}, name_prefix=None + str(data_keys) + " properties" ) - data_any_of_count96 += 1 + data_any_of_count102 += 1 except JsonSchemaValueException: pass - if not data_any_of_count96: + if not data_any_of_count102: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4375,8 +4627,8 @@ def validate___definitions_posixsignal(data, custom_formats={}, name_prefix=None def validate___definitions_nserror(data, custom_formats={}, name_prefix=None): - data_any_of_count97 = 0 - if not data_any_of_count97: + data_any_of_count103 = 0 + if not data_any_of_count103: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4417,10 +4669,10 @@ def validate___definitions_nserror(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count97 += 1 + data_any_of_count103 += 1 except JsonSchemaValueException: pass - if not data_any_of_count97: + if not data_any_of_count103: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4428,8 +4680,8 @@ def validate___definitions_nserror(data, custom_formats={}, name_prefix=None): def validate___definitions_machexception(data, custom_formats={}, name_prefix=None): - data_any_of_count98 = 0 - if not data_any_of_count98: + data_any_of_count104 = 0 + if not data_any_of_count104: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4516,10 +4768,10 @@ def validate___definitions_machexception(data, custom_formats={}, name_prefix=No + str(data_keys) + " properties" ) - data_any_of_count98 += 1 + data_any_of_count104 += 1 except JsonSchemaValueException: pass - if not data_any_of_count98: + if not data_any_of_count104: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4527,8 +4779,8 @@ def validate___definitions_machexception(data, custom_formats={}, name_prefix=No def validate___definitions_cerror(data, custom_formats={}, name_prefix=None): - data_any_of_count99 = 0 - if not data_any_of_count99: + data_any_of_count105 = 0 + if not data_any_of_count105: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4570,10 +4822,10 @@ def validate___definitions_cerror(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count99 += 1 + data_any_of_count105 += 1 except JsonSchemaValueException: pass - if not data_any_of_count99: + if not data_any_of_count105: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4581,17 +4833,17 @@ def validate___definitions_cerror(data, custom_formats={}, name_prefix=None): def validate___definitions_eventid(data, custom_formats={}, name_prefix=None): - data_any_of_count100 = 0 - if not data_any_of_count100: + data_any_of_count106 = 0 + if not data_any_of_count106: try: if not isinstance(data, (str)): raise JsonSchemaValueException( "" + (name_prefix or "data") + " must be string" ) - data_any_of_count100 += 1 + data_any_of_count106 += 1 except JsonSchemaValueException: pass - if not data_any_of_count100: + if not data_any_of_count106: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4601,8 +4853,8 @@ def validate___definitions_eventid(data, custom_formats={}, name_prefix=None): def validate___definitions_eventprocessingerror( data, custom_formats={}, name_prefix=None ): - data_any_of_count101 = 0 - if not data_any_of_count101: + data_any_of_count107 = 0 + if not data_any_of_count107: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4645,10 +4897,10 @@ def validate___definitions_eventprocessingerror( + str(data_keys) + " properties" ) - data_any_of_count101 += 1 + data_any_of_count107 += 1 except JsonSchemaValueException: pass - if not data_any_of_count101: + if not data_any_of_count107: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4656,8 +4908,8 @@ def validate___definitions_eventprocessingerror( def validate___definitions_debugmeta(data, custom_formats={}, name_prefix=None): - data_any_of_count102 = 0 - if not data_any_of_count102: + data_any_of_count108 = 0 + if not data_any_of_count108: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4681,8 +4933,8 @@ def validate___definitions_debugmeta(data, custom_formats={}, name_prefix=None): for data__images_x, data__images_item in enumerate( data__images ): - data__images_item_any_of_count103 = 0 - if not data__images_item_any_of_count103: + data__images_item_any_of_count109 = 0 + if not data__images_item_any_of_count109: try: validate___definitions_debugimage( data__images_item, @@ -4692,10 +4944,10 @@ def validate___definitions_debugmeta(data, custom_formats={}, name_prefix=None): **locals() ), ) - data__images_item_any_of_count103 += 1 + data__images_item_any_of_count109 += 1 except JsonSchemaValueException: pass - if not data__images_item_any_of_count103: + if not data__images_item_any_of_count109: try: if not isinstance(data__images_item, (NoneType)): raise JsonSchemaValueException( @@ -4706,10 +4958,10 @@ def validate___definitions_debugmeta(data, custom_formats={}, name_prefix=None): ) + " must be null" ) - data__images_item_any_of_count103 += 1 + data__images_item_any_of_count109 += 1 except JsonSchemaValueException: pass - if not data__images_item_any_of_count103: + if not data__images_item_any_of_count109: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -4719,18 +4971,18 @@ def validate___definitions_debugmeta(data, custom_formats={}, name_prefix=None): if "sdk_info" in data_keys: data_keys.remove("sdk_info") data__sdkinfo = data["sdk_info"] - data__sdkinfo_any_of_count104 = 0 - if not data__sdkinfo_any_of_count104: + data__sdkinfo_any_of_count110 = 0 + if not data__sdkinfo_any_of_count110: try: validate___definitions_systemsdkinfo( data__sdkinfo, custom_formats, (name_prefix or "data") + ".sdk_info", ) - data__sdkinfo_any_of_count104 += 1 + data__sdkinfo_any_of_count110 += 1 except JsonSchemaValueException: pass - if not data__sdkinfo_any_of_count104: + if not data__sdkinfo_any_of_count110: try: if not isinstance(data__sdkinfo, (NoneType)): raise JsonSchemaValueException( @@ -4738,10 +4990,10 @@ def validate___definitions_debugmeta(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".sdk_info must be null" ) - data__sdkinfo_any_of_count104 += 1 + data__sdkinfo_any_of_count110 += 1 except JsonSchemaValueException: pass - if not data__sdkinfo_any_of_count104: + if not data__sdkinfo_any_of_count110: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -4755,10 +5007,10 @@ def validate___definitions_debugmeta(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count102 += 1 + data_any_of_count108 += 1 except JsonSchemaValueException: pass - if not data_any_of_count102: + if not data_any_of_count108: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4766,8 +5018,8 @@ def validate___definitions_debugmeta(data, custom_formats={}, name_prefix=None): def validate___definitions_systemsdkinfo(data, custom_formats={}, name_prefix=None): - data_any_of_count105 = 0 - if not data_any_of_count105: + data_any_of_count111 = 0 + if not data_any_of_count111: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4862,10 +5114,10 @@ def validate___definitions_systemsdkinfo(data, custom_formats={}, name_prefix=No + str(data_keys) + " properties" ) - data_any_of_count105 += 1 + data_any_of_count111 += 1 except JsonSchemaValueException: pass - if not data_any_of_count105: + if not data_any_of_count111: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4873,88 +5125,88 @@ def validate___definitions_systemsdkinfo(data, custom_formats={}, name_prefix=No def validate___definitions_debugimage(data, custom_formats={}, name_prefix=None): - data_any_of_count106 = 0 - if not data_any_of_count106: + data_any_of_count112 = 0 + if not data_any_of_count112: try: validate___definitions_appledebugimage( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: try: validate___definitions_nativedebugimage( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: try: validate___definitions_nativedebugimage( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: try: validate___definitions_nativedebugimage( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: try: validate___definitions_nativedebugimage( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: try: validate___definitions_nativedebugimage( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: try: validate___definitions_proguarddebugimage( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: try: validate___definitions_nativedebugimage( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: try: validate___definitions_sourcemapdebugimage( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: try: validate___definitions_jvmdebugimage( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4963,10 +5215,10 @@ def validate___definitions_debugimage(data, custom_formats={}, name_prefix=None) data_is_dict = isinstance(data, dict) if data_is_dict: data_keys = set(data.keys()) - data_any_of_count106 += 1 + data_any_of_count112 += 1 except JsonSchemaValueException: pass - if not data_any_of_count106: + if not data_any_of_count112: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -4974,8 +5226,8 @@ def validate___definitions_debugimage(data, custom_formats={}, name_prefix=None) def validate___definitions_jvmdebugimage(data, custom_formats={}, name_prefix=None): - data_any_of_count107 = 0 - if not data_any_of_count107: + data_any_of_count113 = 0 + if not data_any_of_count113: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -4992,18 +5244,18 @@ def validate___definitions_jvmdebugimage(data, custom_formats={}, name_prefix=No if "debug_id" in data_keys: data_keys.remove("debug_id") data__debugid = data["debug_id"] - data__debugid_any_of_count108 = 0 - if not data__debugid_any_of_count108: + data__debugid_any_of_count114 = 0 + if not data__debugid_any_of_count114: try: validate___definitions_debugid( data__debugid, custom_formats, (name_prefix or "data") + ".debug_id", ) - data__debugid_any_of_count108 += 1 + data__debugid_any_of_count114 += 1 except JsonSchemaValueException: pass - if not data__debugid_any_of_count108: + if not data__debugid_any_of_count114: try: if not isinstance(data__debugid, (NoneType)): raise JsonSchemaValueException( @@ -5011,10 +5263,10 @@ def validate___definitions_jvmdebugimage(data, custom_formats={}, name_prefix=No + (name_prefix or "data") + ".debug_id must be null" ) - data__debugid_any_of_count108 += 1 + data__debugid_any_of_count114 += 1 except JsonSchemaValueException: pass - if not data__debugid_any_of_count108: + if not data__debugid_any_of_count114: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5028,10 +5280,10 @@ def validate___definitions_jvmdebugimage(data, custom_formats={}, name_prefix=No + str(data_keys) + " properties" ) - data_any_of_count107 += 1 + data_any_of_count113 += 1 except JsonSchemaValueException: pass - if not data_any_of_count107: + if not data_any_of_count113: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -5047,8 +5299,8 @@ def validate___definitions_debugid(data, custom_formats={}, name_prefix=None): def validate___definitions_sourcemapdebugimage( data, custom_formats={}, name_prefix=None ): - data_any_of_count109 = 0 - if not data_any_of_count109: + data_any_of_count115 = 0 + if not data_any_of_count115: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -5083,18 +5335,18 @@ def validate___definitions_sourcemapdebugimage( if "debug_id" in data_keys: data_keys.remove("debug_id") data__debugid = data["debug_id"] - data__debugid_any_of_count110 = 0 - if not data__debugid_any_of_count110: + data__debugid_any_of_count116 = 0 + if not data__debugid_any_of_count116: try: validate___definitions_debugid( data__debugid, custom_formats, (name_prefix or "data") + ".debug_id", ) - data__debugid_any_of_count110 += 1 + data__debugid_any_of_count116 += 1 except JsonSchemaValueException: pass - if not data__debugid_any_of_count110: + if not data__debugid_any_of_count116: try: if not isinstance(data__debugid, (NoneType)): raise JsonSchemaValueException( @@ -5102,10 +5354,10 @@ def validate___definitions_sourcemapdebugimage( + (name_prefix or "data") + ".debug_id must be null" ) - data__debugid_any_of_count110 += 1 + data__debugid_any_of_count116 += 1 except JsonSchemaValueException: pass - if not data__debugid_any_of_count110: + if not data__debugid_any_of_count116: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5119,10 +5371,10 @@ def validate___definitions_sourcemapdebugimage( + str(data_keys) + " properties" ) - data_any_of_count109 += 1 + data_any_of_count115 += 1 except JsonSchemaValueException: pass - if not data_any_of_count109: + if not data_any_of_count115: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -5132,8 +5384,8 @@ def validate___definitions_sourcemapdebugimage( def validate___definitions_proguarddebugimage( data, custom_formats={}, name_prefix=None ): - data_any_of_count111 = 0 - if not data_any_of_count111: + data_any_of_count117 = 0 + if not data_any_of_count117: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -5164,10 +5416,10 @@ def validate___definitions_proguarddebugimage( + str(data_keys) + " properties" ) - data_any_of_count111 += 1 + data_any_of_count117 += 1 except JsonSchemaValueException: pass - if not data_any_of_count111: + if not data_any_of_count117: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -5175,8 +5427,8 @@ def validate___definitions_proguarddebugimage( def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix=None): - data_any_of_count112 = 0 - if not data_any_of_count112: + data_any_of_count118 = 0 + if not data_any_of_count118: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -5202,18 +5454,18 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix if "code_file" in data_keys: data_keys.remove("code_file") data__codefile = data["code_file"] - data__codefile_any_of_count113 = 0 - if not data__codefile_any_of_count113: + data__codefile_any_of_count119 = 0 + if not data__codefile_any_of_count119: try: validate___definitions_nativeimagepath( data__codefile, custom_formats, (name_prefix or "data") + ".code_file", ) - data__codefile_any_of_count113 += 1 + data__codefile_any_of_count119 += 1 except JsonSchemaValueException: pass - if not data__codefile_any_of_count113: + if not data__codefile_any_of_count119: try: if not isinstance(data__codefile, (NoneType)): raise JsonSchemaValueException( @@ -5221,10 +5473,10 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix + (name_prefix or "data") + ".code_file must be null" ) - data__codefile_any_of_count113 += 1 + data__codefile_any_of_count119 += 1 except JsonSchemaValueException: pass - if not data__codefile_any_of_count113: + if not data__codefile_any_of_count119: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5233,18 +5485,18 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix if "code_id" in data_keys: data_keys.remove("code_id") data__codeid = data["code_id"] - data__codeid_any_of_count114 = 0 - if not data__codeid_any_of_count114: + data__codeid_any_of_count120 = 0 + if not data__codeid_any_of_count120: try: validate___definitions_codeid( data__codeid, custom_formats, (name_prefix or "data") + ".code_id", ) - data__codeid_any_of_count114 += 1 + data__codeid_any_of_count120 += 1 except JsonSchemaValueException: pass - if not data__codeid_any_of_count114: + if not data__codeid_any_of_count120: try: if not isinstance(data__codeid, (NoneType)): raise JsonSchemaValueException( @@ -5252,10 +5504,10 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix + (name_prefix or "data") + ".code_id must be null" ) - data__codeid_any_of_count114 += 1 + data__codeid_any_of_count120 += 1 except JsonSchemaValueException: pass - if not data__codeid_any_of_count114: + if not data__codeid_any_of_count120: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5273,18 +5525,18 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix if "debug_file" in data_keys: data_keys.remove("debug_file") data__debugfile = data["debug_file"] - data__debugfile_any_of_count115 = 0 - if not data__debugfile_any_of_count115: + data__debugfile_any_of_count121 = 0 + if not data__debugfile_any_of_count121: try: validate___definitions_nativeimagepath( data__debugfile, custom_formats, (name_prefix or "data") + ".debug_file", ) - data__debugfile_any_of_count115 += 1 + data__debugfile_any_of_count121 += 1 except JsonSchemaValueException: pass - if not data__debugfile_any_of_count115: + if not data__debugfile_any_of_count121: try: if not isinstance(data__debugfile, (NoneType)): raise JsonSchemaValueException( @@ -5292,10 +5544,10 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix + (name_prefix or "data") + ".debug_file must be null" ) - data__debugfile_any_of_count115 += 1 + data__debugfile_any_of_count121 += 1 except JsonSchemaValueException: pass - if not data__debugfile_any_of_count115: + if not data__debugfile_any_of_count121: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5304,18 +5556,18 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix if "debug_id" in data_keys: data_keys.remove("debug_id") data__debugid = data["debug_id"] - data__debugid_any_of_count116 = 0 - if not data__debugid_any_of_count116: + data__debugid_any_of_count122 = 0 + if not data__debugid_any_of_count122: try: validate___definitions_debugid( data__debugid, custom_formats, (name_prefix or "data") + ".debug_id", ) - data__debugid_any_of_count116 += 1 + data__debugid_any_of_count122 += 1 except JsonSchemaValueException: pass - if not data__debugid_any_of_count116: + if not data__debugid_any_of_count122: try: if not isinstance(data__debugid, (NoneType)): raise JsonSchemaValueException( @@ -5323,10 +5575,10 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix + (name_prefix or "data") + ".debug_id must be null" ) - data__debugid_any_of_count116 += 1 + data__debugid_any_of_count122 += 1 except JsonSchemaValueException: pass - if not data__debugid_any_of_count116: + if not data__debugid_any_of_count122: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5335,18 +5587,18 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix if "image_addr" in data_keys: data_keys.remove("image_addr") data__imageaddr = data["image_addr"] - data__imageaddr_any_of_count117 = 0 - if not data__imageaddr_any_of_count117: + data__imageaddr_any_of_count123 = 0 + if not data__imageaddr_any_of_count123: try: validate___definitions_addr( data__imageaddr, custom_formats, (name_prefix or "data") + ".image_addr", ) - data__imageaddr_any_of_count117 += 1 + data__imageaddr_any_of_count123 += 1 except JsonSchemaValueException: pass - if not data__imageaddr_any_of_count117: + if not data__imageaddr_any_of_count123: try: if not isinstance(data__imageaddr, (NoneType)): raise JsonSchemaValueException( @@ -5354,10 +5606,10 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix + (name_prefix or "data") + ".image_addr must be null" ) - data__imageaddr_any_of_count117 += 1 + data__imageaddr_any_of_count123 += 1 except JsonSchemaValueException: pass - if not data__imageaddr_any_of_count117: + if not data__imageaddr_any_of_count123: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5389,18 +5641,18 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix if "image_vmaddr" in data_keys: data_keys.remove("image_vmaddr") data__imagevmaddr = data["image_vmaddr"] - data__imagevmaddr_any_of_count118 = 0 - if not data__imagevmaddr_any_of_count118: + data__imagevmaddr_any_of_count124 = 0 + if not data__imagevmaddr_any_of_count124: try: validate___definitions_addr( data__imagevmaddr, custom_formats, (name_prefix or "data") + ".image_vmaddr", ) - data__imagevmaddr_any_of_count118 += 1 + data__imagevmaddr_any_of_count124 += 1 except JsonSchemaValueException: pass - if not data__imagevmaddr_any_of_count118: + if not data__imagevmaddr_any_of_count124: try: if not isinstance(data__imagevmaddr, (NoneType)): raise JsonSchemaValueException( @@ -5408,10 +5660,10 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix + (name_prefix or "data") + ".image_vmaddr must be null" ) - data__imagevmaddr_any_of_count118 += 1 + data__imagevmaddr_any_of_count124 += 1 except JsonSchemaValueException: pass - if not data__imagevmaddr_any_of_count118: + if not data__imagevmaddr_any_of_count124: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5425,10 +5677,10 @@ def validate___definitions_nativedebugimage(data, custom_formats={}, name_prefix + str(data_keys) + " properties" ) - data_any_of_count112 += 1 + data_any_of_count118 += 1 except JsonSchemaValueException: pass - if not data_any_of_count112: + if not data_any_of_count118: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -5442,8 +5694,8 @@ def validate___definitions_codeid(data, custom_formats={}, name_prefix=None): def validate___definitions_appledebugimage(data, custom_formats={}, name_prefix=None): - data_any_of_count119 = 0 - if not data_any_of_count119: + data_any_of_count125 = 0 + if not data_any_of_count125: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -5517,18 +5769,18 @@ def validate___definitions_appledebugimage(data, custom_formats={}, name_prefix= if "image_addr" in data_keys: data_keys.remove("image_addr") data__imageaddr = data["image_addr"] - data__imageaddr_any_of_count120 = 0 - if not data__imageaddr_any_of_count120: + data__imageaddr_any_of_count126 = 0 + if not data__imageaddr_any_of_count126: try: validate___definitions_addr( data__imageaddr, custom_formats, (name_prefix or "data") + ".image_addr", ) - data__imageaddr_any_of_count120 += 1 + data__imageaddr_any_of_count126 += 1 except JsonSchemaValueException: pass - if not data__imageaddr_any_of_count120: + if not data__imageaddr_any_of_count126: try: if not isinstance(data__imageaddr, (NoneType)): raise JsonSchemaValueException( @@ -5536,10 +5788,10 @@ def validate___definitions_appledebugimage(data, custom_formats={}, name_prefix= + (name_prefix or "data") + ".image_addr must be null" ) - data__imageaddr_any_of_count120 += 1 + data__imageaddr_any_of_count126 += 1 except JsonSchemaValueException: pass - if not data__imageaddr_any_of_count120: + if not data__imageaddr_any_of_count126: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5571,18 +5823,18 @@ def validate___definitions_appledebugimage(data, custom_formats={}, name_prefix= if "image_vmaddr" in data_keys: data_keys.remove("image_vmaddr") data__imagevmaddr = data["image_vmaddr"] - data__imagevmaddr_any_of_count121 = 0 - if not data__imagevmaddr_any_of_count121: + data__imagevmaddr_any_of_count127 = 0 + if not data__imagevmaddr_any_of_count127: try: validate___definitions_addr( data__imagevmaddr, custom_formats, (name_prefix or "data") + ".image_vmaddr", ) - data__imagevmaddr_any_of_count121 += 1 + data__imagevmaddr_any_of_count127 += 1 except JsonSchemaValueException: pass - if not data__imagevmaddr_any_of_count121: + if not data__imagevmaddr_any_of_count127: try: if not isinstance(data__imagevmaddr, (NoneType)): raise JsonSchemaValueException( @@ -5590,10 +5842,10 @@ def validate___definitions_appledebugimage(data, custom_formats={}, name_prefix= + (name_prefix or "data") + ".image_vmaddr must be null" ) - data__imagevmaddr_any_of_count121 += 1 + data__imagevmaddr_any_of_count127 += 1 except JsonSchemaValueException: pass - if not data__imagevmaddr_any_of_count121: + if not data__imagevmaddr_any_of_count127: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5625,10 +5877,10 @@ def validate___definitions_appledebugimage(data, custom_formats={}, name_prefix= + str(data_keys) + " properties" ) - data_any_of_count119 += 1 + data_any_of_count125 += 1 except JsonSchemaValueException: pass - if not data_any_of_count119: + if not data_any_of_count125: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -5636,8 +5888,8 @@ def validate___definitions_appledebugimage(data, custom_formats={}, name_prefix= def validate___definitions_contexts(data, custom_formats={}, name_prefix=None): - data_any_of_count122 = 0 - if not data_any_of_count122: + data_any_of_count128 = 0 + if not data_any_of_count128: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -5649,8 +5901,8 @@ def validate___definitions_contexts(data, custom_formats={}, name_prefix=None): for data_key in data_keys: if data_key not in []: data_value = data.get(data_key) - data_value_any_of_count123 = 0 - if not data_value_any_of_count123: + data_value_any_of_count129 = 0 + if not data_value_any_of_count129: try: validate___definitions_contextinner( data_value, @@ -5658,10 +5910,10 @@ def validate___definitions_contexts(data, custom_formats={}, name_prefix=None): (name_prefix or "data") + ".{data_key}".format(**locals()), ) - data_value_any_of_count123 += 1 + data_value_any_of_count129 += 1 except JsonSchemaValueException: pass - if not data_value_any_of_count123: + if not data_value_any_of_count129: try: if not isinstance(data_value, (NoneType)): raise JsonSchemaValueException( @@ -5670,20 +5922,20 @@ def validate___definitions_contexts(data, custom_formats={}, name_prefix=None): + ".{data_key}".format(**locals()) + " must be null" ) - data_value_any_of_count123 += 1 + data_value_any_of_count129 += 1 except JsonSchemaValueException: pass - if not data_value_any_of_count123: + if not data_value_any_of_count129: raise JsonSchemaValueException( "" + (name_prefix or "data") + ".{data_key}".format(**locals()) + " cannot be validated by any definition" ) - data_any_of_count122 += 1 + data_any_of_count128 += 1 except JsonSchemaValueException: pass - if not data_any_of_count122: + if not data_any_of_count128: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -5691,16 +5943,16 @@ def validate___definitions_contexts(data, custom_formats={}, name_prefix=None): def validate___definitions_contextinner(data, custom_formats={}, name_prefix=None): - data_any_of_count124 = 0 - if not data_any_of_count124: + data_any_of_count130 = 0 + if not data_any_of_count130: try: validate___definitions_context( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count124 += 1 + data_any_of_count130 += 1 except JsonSchemaValueException: pass - if not data_any_of_count124: + if not data_any_of_count130: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -5708,136 +5960,136 @@ def validate___definitions_contextinner(data, custom_formats={}, name_prefix=Non def validate___definitions_context(data, custom_formats={}, name_prefix=None): - data_any_of_count125 = 0 - if not data_any_of_count125: + data_any_of_count131 = 0 + if not data_any_of_count131: try: validate___definitions_devicecontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_oscontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_runtimecontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_appcontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_browsercontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_gpucontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_tracecontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_profilecontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_replaycontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_userreportv2context( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_monitorcontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_responsecontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_otelcontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_cloudresourcecontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_nelcontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: validate___definitions_performancescorecontext( data, custom_formats, (name_prefix or "data") + "" ) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -5846,10 +6098,10 @@ def validate___definitions_context(data, custom_formats={}, name_prefix=None): data_is_dict = isinstance(data, dict) if data_is_dict: data_keys = set(data.keys()) - data_any_of_count125 += 1 + data_any_of_count131 += 1 except JsonSchemaValueException: pass - if not data_any_of_count125: + if not data_any_of_count131: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -5859,8 +6111,8 @@ def validate___definitions_context(data, custom_formats={}, name_prefix=None): def validate___definitions_performancescorecontext( data, custom_formats={}, name_prefix=None ): - data_any_of_count126 = 0 - if not data_any_of_count126: + data_any_of_count132 = 0 + if not data_any_of_count132: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -5886,10 +6138,10 @@ def validate___definitions_performancescorecontext( + str(data_keys) + " properties" ) - data_any_of_count126 += 1 + data_any_of_count132 += 1 except JsonSchemaValueException: pass - if not data_any_of_count126: + if not data_any_of_count132: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -5897,8 +6149,8 @@ def validate___definitions_performancescorecontext( def validate___definitions_nelcontext(data, custom_formats={}, name_prefix=None): - data_any_of_count127 = 0 - if not data_any_of_count127: + data_any_of_count133 = 0 + if not data_any_of_count133: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -5942,27 +6194,27 @@ def validate___definitions_nelcontext(data, custom_formats={}, name_prefix=None) if "phase" in data_keys: data_keys.remove("phase") data__phase = data["phase"] - data__phase_any_of_count128 = 0 - if not data__phase_any_of_count128: + data__phase_any_of_count134 = 0 + if not data__phase_any_of_count134: try: validate___definitions_networkreportphases( data__phase, custom_formats, (name_prefix or "data") + ".phase", ) - data__phase_any_of_count128 += 1 + data__phase_any_of_count134 += 1 except JsonSchemaValueException: pass - if not data__phase_any_of_count128: + if not data__phase_any_of_count134: try: if not isinstance(data__phase, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".phase must be null" ) - data__phase_any_of_count128 += 1 + data__phase_any_of_count134 += 1 except JsonSchemaValueException: pass - if not data__phase_any_of_count128: + if not data__phase_any_of_count134: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -5982,18 +6234,18 @@ def validate___definitions_nelcontext(data, custom_formats={}, name_prefix=None) if "server_ip" in data_keys: data_keys.remove("server_ip") data__serverip = data["server_ip"] - data__serverip_any_of_count129 = 0 - if not data__serverip_any_of_count129: + data__serverip_any_of_count135 = 0 + if not data__serverip_any_of_count135: try: validate___definitions_string( data__serverip, custom_formats, (name_prefix or "data") + ".server_ip", ) - data__serverip_any_of_count129 += 1 + data__serverip_any_of_count135 += 1 except JsonSchemaValueException: pass - if not data__serverip_any_of_count129: + if not data__serverip_any_of_count135: try: if not isinstance(data__serverip, (NoneType)): raise JsonSchemaValueException( @@ -6001,10 +6253,10 @@ def validate___definitions_nelcontext(data, custom_formats={}, name_prefix=None) + (name_prefix or "data") + ".server_ip must be null" ) - data__serverip_any_of_count129 += 1 + data__serverip_any_of_count135 += 1 except JsonSchemaValueException: pass - if not data__serverip_any_of_count129: + if not data__serverip_any_of_count135: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6018,10 +6270,10 @@ def validate___definitions_nelcontext(data, custom_formats={}, name_prefix=None) + str(data_keys) + " properties" ) - data_any_of_count127 += 1 + data_any_of_count133 += 1 except JsonSchemaValueException: pass - if not data_any_of_count127: + if not data_any_of_count133: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6031,8 +6283,8 @@ def validate___definitions_nelcontext(data, custom_formats={}, name_prefix=None) def validate___definitions_networkreportphases( data, custom_formats={}, name_prefix=None ): - data_any_of_count130 = 0 - if not data_any_of_count130: + data_any_of_count136 = 0 + if not data_any_of_count136: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6049,10 +6301,10 @@ def validate___definitions_networkreportphases( + str(data_keys) + " properties" ) - data_any_of_count130 += 1 + data_any_of_count136 += 1 except JsonSchemaValueException: pass - if not data_any_of_count130: + if not data_any_of_count136: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6069,10 +6321,10 @@ def validate___definitions_networkreportphases( + str(data_keys) + " properties" ) - data_any_of_count130 += 1 + data_any_of_count136 += 1 except JsonSchemaValueException: pass - if not data_any_of_count130: + if not data_any_of_count136: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6089,19 +6341,19 @@ def validate___definitions_networkreportphases( + str(data_keys) + " properties" ) - data_any_of_count130 += 1 + data_any_of_count136 += 1 except JsonSchemaValueException: pass - if not data_any_of_count130: + if not data_any_of_count136: try: if not isinstance(data, (str)): raise JsonSchemaValueException( "" + (name_prefix or "data") + " must be string" ) - data_any_of_count130 += 1 + data_any_of_count136 += 1 except JsonSchemaValueException: pass - if not data_any_of_count130: + if not data_any_of_count136: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6111,8 +6363,8 @@ def validate___definitions_networkreportphases( def validate___definitions_cloudresourcecontext( data, custom_formats={}, name_prefix=None ): - data_any_of_count131 = 0 - if not data_any_of_count131: + data_any_of_count137 = 0 + if not data_any_of_count137: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6192,10 +6444,10 @@ def validate___definitions_cloudresourcecontext( + str(data_keys) + " properties" ) - data_any_of_count131 += 1 + data_any_of_count137 += 1 except JsonSchemaValueException: pass - if not data_any_of_count131: + if not data_any_of_count137: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6203,8 +6455,8 @@ def validate___definitions_cloudresourcecontext( def validate___definitions_otelcontext(data, custom_formats={}, name_prefix=None): - data_any_of_count132 = 0 - if not data_any_of_count132: + data_any_of_count138 = 0 + if not data_any_of_count138: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6245,10 +6497,10 @@ def validate___definitions_otelcontext(data, custom_formats={}, name_prefix=None + str(data_keys) + " properties" ) - data_any_of_count132 += 1 + data_any_of_count138 += 1 except JsonSchemaValueException: pass - if not data_any_of_count132: + if not data_any_of_count138: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6256,8 +6508,8 @@ def validate___definitions_otelcontext(data, custom_formats={}, name_prefix=None def validate___definitions_responsecontext(data, custom_formats={}, name_prefix=None): - data_any_of_count133 = 0 - if not data_any_of_count133: + data_any_of_count139 = 0 + if not data_any_of_count139: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6292,18 +6544,18 @@ def validate___definitions_responsecontext(data, custom_formats={}, name_prefix= if "cookies" in data_keys: data_keys.remove("cookies") data__cookies = data["cookies"] - data__cookies_any_of_count134 = 0 - if not data__cookies_any_of_count134: + data__cookies_any_of_count140 = 0 + if not data__cookies_any_of_count140: try: validate___definitions_cookies( data__cookies, custom_formats, (name_prefix or "data") + ".cookies", ) - data__cookies_any_of_count134 += 1 + data__cookies_any_of_count140 += 1 except JsonSchemaValueException: pass - if not data__cookies_any_of_count134: + if not data__cookies_any_of_count140: try: if not isinstance(data__cookies, (NoneType)): raise JsonSchemaValueException( @@ -6311,10 +6563,10 @@ def validate___definitions_responsecontext(data, custom_formats={}, name_prefix= + (name_prefix or "data") + ".cookies must be null" ) - data__cookies_any_of_count134 += 1 + data__cookies_any_of_count140 += 1 except JsonSchemaValueException: pass - if not data__cookies_any_of_count134: + if not data__cookies_any_of_count140: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6326,18 +6578,18 @@ def validate___definitions_responsecontext(data, custom_formats={}, name_prefix= if "headers" in data_keys: data_keys.remove("headers") data__headers = data["headers"] - data__headers_any_of_count135 = 0 - if not data__headers_any_of_count135: + data__headers_any_of_count141 = 0 + if not data__headers_any_of_count141: try: validate___definitions_headers( data__headers, custom_formats, (name_prefix or "data") + ".headers", ) - data__headers_any_of_count135 += 1 + data__headers_any_of_count141 += 1 except JsonSchemaValueException: pass - if not data__headers_any_of_count135: + if not data__headers_any_of_count141: try: if not isinstance(data__headers, (NoneType)): raise JsonSchemaValueException( @@ -6345,10 +6597,10 @@ def validate___definitions_responsecontext(data, custom_formats={}, name_prefix= + (name_prefix or "data") + ".headers must be null" ) - data__headers_any_of_count135 += 1 + data__headers_any_of_count141 += 1 except JsonSchemaValueException: pass - if not data__headers_any_of_count135: + if not data__headers_any_of_count141: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6394,10 +6646,10 @@ def validate___definitions_responsecontext(data, custom_formats={}, name_prefix= + str(data_keys) + " properties" ) - data_any_of_count133 += 1 + data_any_of_count139 += 1 except JsonSchemaValueException: pass - if not data_any_of_count133: + if not data_any_of_count139: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6405,8 +6657,8 @@ def validate___definitions_responsecontext(data, custom_formats={}, name_prefix= def validate___definitions_monitorcontext(data, custom_formats={}, name_prefix=None): - data_any_of_count136 = 0 - if not data_any_of_count136: + data_any_of_count142 = 0 + if not data_any_of_count142: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6415,10 +6667,10 @@ def validate___definitions_monitorcontext(data, custom_formats={}, name_prefix=N data_is_dict = isinstance(data, dict) if data_is_dict: data_keys = set(data.keys()) - data_any_of_count136 += 1 + data_any_of_count142 += 1 except JsonSchemaValueException: pass - if not data_any_of_count136: + if not data_any_of_count142: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6428,8 +6680,8 @@ def validate___definitions_monitorcontext(data, custom_formats={}, name_prefix=N def validate___definitions_userreportv2context( data, custom_formats={}, name_prefix=None ): - data_any_of_count137 = 0 - if not data_any_of_count137: + data_any_of_count143 = 0 + if not data_any_of_count143: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6464,10 +6716,10 @@ def validate___definitions_userreportv2context( + str(data_keys) + " properties" ) - data_any_of_count137 += 1 + data_any_of_count143 += 1 except JsonSchemaValueException: pass - if not data_any_of_count137: + if not data_any_of_count143: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6475,8 +6727,8 @@ def validate___definitions_userreportv2context( def validate___definitions_replaycontext(data, custom_formats={}, name_prefix=None): - data_any_of_count138 = 0 - if not data_any_of_count138: + data_any_of_count144 = 0 + if not data_any_of_count144: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6488,18 +6740,18 @@ def validate___definitions_replaycontext(data, custom_formats={}, name_prefix=No if "replay_id" in data_keys: data_keys.remove("replay_id") data__replayid = data["replay_id"] - data__replayid_any_of_count139 = 0 - if not data__replayid_any_of_count139: + data__replayid_any_of_count145 = 0 + if not data__replayid_any_of_count145: try: validate___definitions_eventid( data__replayid, custom_formats, (name_prefix or "data") + ".replay_id", ) - data__replayid_any_of_count139 += 1 + data__replayid_any_of_count145 += 1 except JsonSchemaValueException: pass - if not data__replayid_any_of_count139: + if not data__replayid_any_of_count145: try: if not isinstance(data__replayid, (NoneType)): raise JsonSchemaValueException( @@ -6507,10 +6759,10 @@ def validate___definitions_replaycontext(data, custom_formats={}, name_prefix=No + (name_prefix or "data") + ".replay_id must be null" ) - data__replayid_any_of_count139 += 1 + data__replayid_any_of_count145 += 1 except JsonSchemaValueException: pass - if not data__replayid_any_of_count139: + if not data__replayid_any_of_count145: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6524,10 +6776,10 @@ def validate___definitions_replaycontext(data, custom_formats={}, name_prefix=No + str(data_keys) + " properties" ) - data_any_of_count138 += 1 + data_any_of_count144 += 1 except JsonSchemaValueException: pass - if not data_any_of_count138: + if not data_any_of_count144: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6535,8 +6787,8 @@ def validate___definitions_replaycontext(data, custom_formats={}, name_prefix=No def validate___definitions_profilecontext(data, custom_formats={}, name_prefix=None): - data_any_of_count140 = 0 - if not data_any_of_count140: + data_any_of_count146 = 0 + if not data_any_of_count146: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6548,18 +6800,18 @@ def validate___definitions_profilecontext(data, custom_formats={}, name_prefix=N if "profile_id" in data_keys: data_keys.remove("profile_id") data__profileid = data["profile_id"] - data__profileid_any_of_count141 = 0 - if not data__profileid_any_of_count141: + data__profileid_any_of_count147 = 0 + if not data__profileid_any_of_count147: try: validate___definitions_eventid( data__profileid, custom_formats, (name_prefix or "data") + ".profile_id", ) - data__profileid_any_of_count141 += 1 + data__profileid_any_of_count147 += 1 except JsonSchemaValueException: pass - if not data__profileid_any_of_count141: + if not data__profileid_any_of_count147: try: if not isinstance(data__profileid, (NoneType)): raise JsonSchemaValueException( @@ -6567,10 +6819,10 @@ def validate___definitions_profilecontext(data, custom_formats={}, name_prefix=N + (name_prefix or "data") + ".profile_id must be null" ) - data__profileid_any_of_count141 += 1 + data__profileid_any_of_count147 += 1 except JsonSchemaValueException: pass - if not data__profileid_any_of_count141: + if not data__profileid_any_of_count147: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6579,18 +6831,18 @@ def validate___definitions_profilecontext(data, custom_formats={}, name_prefix=N if "profiler_id" in data_keys: data_keys.remove("profiler_id") data__profilerid = data["profiler_id"] - data__profilerid_any_of_count142 = 0 - if not data__profilerid_any_of_count142: + data__profilerid_any_of_count148 = 0 + if not data__profilerid_any_of_count148: try: validate___definitions_eventid( data__profilerid, custom_formats, (name_prefix or "data") + ".profiler_id", ) - data__profilerid_any_of_count142 += 1 + data__profilerid_any_of_count148 += 1 except JsonSchemaValueException: pass - if not data__profilerid_any_of_count142: + if not data__profilerid_any_of_count148: try: if not isinstance(data__profilerid, (NoneType)): raise JsonSchemaValueException( @@ -6598,10 +6850,10 @@ def validate___definitions_profilecontext(data, custom_formats={}, name_prefix=N + (name_prefix or "data") + ".profiler_id must be null" ) - data__profilerid_any_of_count142 += 1 + data__profilerid_any_of_count148 += 1 except JsonSchemaValueException: pass - if not data__profilerid_any_of_count142: + if not data__profilerid_any_of_count148: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6615,10 +6867,10 @@ def validate___definitions_profilecontext(data, custom_formats={}, name_prefix=N + str(data_keys) + " properties" ) - data_any_of_count140 += 1 + data_any_of_count146 += 1 except JsonSchemaValueException: pass - if not data_any_of_count140: + if not data_any_of_count146: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6626,8 +6878,8 @@ def validate___definitions_profilecontext(data, custom_formats={}, name_prefix=N def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=None): - data_any_of_count143 = 0 - if not data_any_of_count143: + data_any_of_count149 = 0 + if not data_any_of_count149: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -6655,27 +6907,27 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non if "data" in data_keys: data_keys.remove("data") data__data = data["data"] - data__data_any_of_count144 = 0 - if not data__data_any_of_count144: + data__data_any_of_count150 = 0 + if not data__data_any_of_count150: try: validate___definitions_spandata( data__data, custom_formats, (name_prefix or "data") + ".data", ) - data__data_any_of_count144 += 1 + data__data_any_of_count150 += 1 except JsonSchemaValueException: pass - if not data__data_any_of_count144: + if not data__data_any_of_count150: try: if not isinstance(data__data, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".data must be null" ) - data__data_any_of_count144 += 1 + data__data_any_of_count150 += 1 except JsonSchemaValueException: pass - if not data__data_any_of_count144: + if not data__data_any_of_count150: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6711,18 +6963,18 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non if "parent_span_id" in data_keys: data_keys.remove("parent_span_id") data__parentspanid = data["parent_span_id"] - data__parentspanid_any_of_count145 = 0 - if not data__parentspanid_any_of_count145: + data__parentspanid_any_of_count151 = 0 + if not data__parentspanid_any_of_count151: try: validate___definitions_spanid( data__parentspanid, custom_formats, (name_prefix or "data") + ".parent_span_id", ) - data__parentspanid_any_of_count145 += 1 + data__parentspanid_any_of_count151 += 1 except JsonSchemaValueException: pass - if not data__parentspanid_any_of_count145: + if not data__parentspanid_any_of_count151: try: if not isinstance(data__parentspanid, (NoneType)): raise JsonSchemaValueException( @@ -6730,10 +6982,10 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non + (name_prefix or "data") + ".parent_span_id must be null" ) - data__parentspanid_any_of_count145 += 1 + data__parentspanid_any_of_count151 += 1 except JsonSchemaValueException: pass - if not data__parentspanid_any_of_count145: + if not data__parentspanid_any_of_count151: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6751,18 +7003,18 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non if "span_id" in data_keys: data_keys.remove("span_id") data__spanid = data["span_id"] - data__spanid_any_of_count146 = 0 - if not data__spanid_any_of_count146: + data__spanid_any_of_count152 = 0 + if not data__spanid_any_of_count152: try: validate___definitions_spanid( data__spanid, custom_formats, (name_prefix or "data") + ".span_id", ) - data__spanid_any_of_count146 += 1 + data__spanid_any_of_count152 += 1 except JsonSchemaValueException: pass - if not data__spanid_any_of_count146: + if not data__spanid_any_of_count152: try: if not isinstance(data__spanid, (NoneType)): raise JsonSchemaValueException( @@ -6770,10 +7022,10 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non + (name_prefix or "data") + ".span_id must be null" ) - data__spanid_any_of_count146 += 1 + data__spanid_any_of_count152 += 1 except JsonSchemaValueException: pass - if not data__spanid_any_of_count146: + if not data__spanid_any_of_count152: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6782,18 +7034,18 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non if "status" in data_keys: data_keys.remove("status") data__status = data["status"] - data__status_any_of_count147 = 0 - if not data__status_any_of_count147: + data__status_any_of_count153 = 0 + if not data__status_any_of_count153: try: validate___definitions_spanstatus( data__status, custom_formats, (name_prefix or "data") + ".status", ) - data__status_any_of_count147 += 1 + data__status_any_of_count153 += 1 except JsonSchemaValueException: pass - if not data__status_any_of_count147: + if not data__status_any_of_count153: try: if not isinstance(data__status, (NoneType)): raise JsonSchemaValueException( @@ -6801,10 +7053,10 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non + (name_prefix or "data") + ".status must be null" ) - data__status_any_of_count147 += 1 + data__status_any_of_count153 += 1 except JsonSchemaValueException: pass - if not data__status_any_of_count147: + if not data__status_any_of_count153: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6813,18 +7065,18 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non if "trace_id" in data_keys: data_keys.remove("trace_id") data__traceid = data["trace_id"] - data__traceid_any_of_count148 = 0 - if not data__traceid_any_of_count148: + data__traceid_any_of_count154 = 0 + if not data__traceid_any_of_count154: try: validate___definitions_traceid( data__traceid, custom_formats, (name_prefix or "data") + ".trace_id", ) - data__traceid_any_of_count148 += 1 + data__traceid_any_of_count154 += 1 except JsonSchemaValueException: pass - if not data__traceid_any_of_count148: + if not data__traceid_any_of_count154: try: if not isinstance(data__traceid, (NoneType)): raise JsonSchemaValueException( @@ -6832,10 +7084,10 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non + (name_prefix or "data") + ".trace_id must be null" ) - data__traceid_any_of_count148 += 1 + data__traceid_any_of_count154 += 1 except JsonSchemaValueException: pass - if not data__traceid_any_of_count148: + if not data__traceid_any_of_count154: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -6849,10 +7101,10 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non + str(data_keys) + " properties" ) - data_any_of_count143 += 1 + data_any_of_count149 += 1 except JsonSchemaValueException: pass - if not data_any_of_count143: + if not data_any_of_count149: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6860,17 +7112,17 @@ def validate___definitions_tracecontext(data, custom_formats={}, name_prefix=Non def validate___definitions_traceid(data, custom_formats={}, name_prefix=None): - data_any_of_count149 = 0 - if not data_any_of_count149: + data_any_of_count155 = 0 + if not data_any_of_count155: try: if not isinstance(data, (str)): raise JsonSchemaValueException( "" + (name_prefix or "data") + " must be string" ) - data_any_of_count149 += 1 + data_any_of_count155 += 1 except JsonSchemaValueException: pass - if not data_any_of_count149: + if not data_any_of_count155: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6908,17 +7160,17 @@ def validate___definitions_spanstatus(data, custom_formats={}, name_prefix=None) def validate___definitions_spanid(data, custom_formats={}, name_prefix=None): - data_any_of_count150 = 0 - if not data_any_of_count150: + data_any_of_count156 = 0 + if not data_any_of_count156: try: if not isinstance(data, (str)): raise JsonSchemaValueException( "" + (name_prefix or "data") + " must be string" ) - data_any_of_count150 += 1 + data_any_of_count156 += 1 except JsonSchemaValueException: pass - if not data_any_of_count150: + if not data_any_of_count156: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -6926,8 +7178,8 @@ def validate___definitions_spanid(data, custom_formats={}, name_prefix=None): def validate___definitions_spandata(data, custom_formats={}, name_prefix=None): - data_any_of_count151 = 0 - if not data_any_of_count151: + data_any_of_count157 = 0 + if not data_any_of_count157: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -7061,18 +7313,18 @@ def validate___definitions_spandata(data, custom_formats={}, name_prefix=None): if "previousRoute" in data_keys: data_keys.remove("previousRoute") data__previousRoute = data["previousRoute"] - data__previousRoute_any_of_count152 = 0 - if not data__previousRoute_any_of_count152: + data__previousRoute_any_of_count158 = 0 + if not data__previousRoute_any_of_count158: try: validate___definitions_route( data__previousRoute, custom_formats, (name_prefix or "data") + ".previousRoute", ) - data__previousRoute_any_of_count152 += 1 + data__previousRoute_any_of_count158 += 1 except JsonSchemaValueException: pass - if not data__previousRoute_any_of_count152: + if not data__previousRoute_any_of_count158: try: if not isinstance(data__previousRoute, (NoneType)): raise JsonSchemaValueException( @@ -7080,10 +7332,10 @@ def validate___definitions_spandata(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".previousRoute must be null" ) - data__previousRoute_any_of_count152 += 1 + data__previousRoute_any_of_count158 += 1 except JsonSchemaValueException: pass - if not data__previousRoute_any_of_count152: + if not data__previousRoute_any_of_count158: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -7097,27 +7349,27 @@ def validate___definitions_spandata(data, custom_formats={}, name_prefix=None): if "route" in data_keys: data_keys.remove("route") data__route = data["route"] - data__route_any_of_count153 = 0 - if not data__route_any_of_count153: + data__route_any_of_count159 = 0 + if not data__route_any_of_count159: try: validate___definitions_route( data__route, custom_formats, (name_prefix or "data") + ".route", ) - data__route_any_of_count153 += 1 + data__route_any_of_count159 += 1 except JsonSchemaValueException: pass - if not data__route_any_of_count153: + if not data__route_any_of_count159: try: if not isinstance(data__route, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".route must be null" ) - data__route_any_of_count153 += 1 + data__route_any_of_count159 += 1 except JsonSchemaValueException: pass - if not data__route_any_of_count153: + if not data__route_any_of_count159: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -7186,18 +7438,18 @@ def validate___definitions_spandata(data, custom_formats={}, name_prefix=None): if "thread.id" in data_keys: data_keys.remove("thread.id") data__threadid = data["thread.id"] - data__threadid_any_of_count154 = 0 - if not data__threadid_any_of_count154: + data__threadid_any_of_count160 = 0 + if not data__threadid_any_of_count160: try: validate___definitions_threadid( data__threadid, custom_formats, (name_prefix or "data") + ".thread.id", ) - data__threadid_any_of_count154 += 1 + data__threadid_any_of_count160 += 1 except JsonSchemaValueException: pass - if not data__threadid_any_of_count154: + if not data__threadid_any_of_count160: try: if not isinstance(data__threadid, (NoneType)): raise JsonSchemaValueException( @@ -7205,10 +7457,10 @@ def validate___definitions_spandata(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".thread.id must be null" ) - data__threadid_any_of_count154 += 1 + data__threadid_any_of_count160 += 1 except JsonSchemaValueException: pass - if not data__threadid_any_of_count154: + if not data__threadid_any_of_count160: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -7258,10 +7510,10 @@ def validate___definitions_spandata(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count151 += 1 + data_any_of_count157 += 1 except JsonSchemaValueException: pass - if not data_any_of_count151: + if not data_any_of_count157: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -7269,8 +7521,8 @@ def validate___definitions_spandata(data, custom_formats={}, name_prefix=None): def validate___definitions_route(data, custom_formats={}, name_prefix=None): - data_any_of_count155 = 0 - if not data_any_of_count155: + data_any_of_count161 = 0 + if not data_any_of_count161: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -7308,10 +7560,10 @@ def validate___definitions_route(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count155 += 1 + data_any_of_count161 += 1 except JsonSchemaValueException: pass - if not data_any_of_count155: + if not data_any_of_count161: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -7319,8 +7571,8 @@ def validate___definitions_route(data, custom_formats={}, name_prefix=None): def validate___definitions_gpucontext(data, custom_formats={}, name_prefix=None): - data_any_of_count156 = 0 - if not data_any_of_count156: + data_any_of_count162 = 0 + if not data_any_of_count162: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -7498,10 +7750,10 @@ def validate___definitions_gpucontext(data, custom_formats={}, name_prefix=None) + str(data_keys) + " properties" ) - data_any_of_count156 += 1 + data_any_of_count162 += 1 except JsonSchemaValueException: pass - if not data_any_of_count156: + if not data_any_of_count162: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -7509,8 +7761,8 @@ def validate___definitions_gpucontext(data, custom_formats={}, name_prefix=None) def validate___definitions_browsercontext(data, custom_formats={}, name_prefix=None): - data_any_of_count157 = 0 - if not data_any_of_count157: + data_any_of_count163 = 0 + if not data_any_of_count163: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -7545,10 +7797,10 @@ def validate___definitions_browsercontext(data, custom_formats={}, name_prefix=N + str(data_keys) + " properties" ) - data_any_of_count157 += 1 + data_any_of_count163 += 1 except JsonSchemaValueException: pass - if not data_any_of_count157: + if not data_any_of_count163: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -7556,8 +7808,8 @@ def validate___definitions_browsercontext(data, custom_formats={}, name_prefix=N def validate___definitions_appcontext(data, custom_formats={}, name_prefix=None): - data_any_of_count158 = 0 - if not data_any_of_count158: + data_any_of_count164 = 0 + if not data_any_of_count164: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -7693,10 +7945,10 @@ def validate___definitions_appcontext(data, custom_formats={}, name_prefix=None) + str(data_keys) + " properties" ) - data_any_of_count158 += 1 + data_any_of_count164 += 1 except JsonSchemaValueException: pass - if not data_any_of_count158: + if not data_any_of_count164: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -7704,8 +7956,8 @@ def validate___definitions_appcontext(data, custom_formats={}, name_prefix=None) def validate___definitions_runtimecontext(data, custom_formats={}, name_prefix=None): - data_any_of_count159 = 0 - if not data_any_of_count159: + data_any_of_count165 = 0 + if not data_any_of_count165: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -7758,10 +8010,10 @@ def validate___definitions_runtimecontext(data, custom_formats={}, name_prefix=N + str(data_keys) + " properties" ) - data_any_of_count159 += 1 + data_any_of_count165 += 1 except JsonSchemaValueException: pass - if not data_any_of_count159: + if not data_any_of_count165: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -7769,8 +8021,8 @@ def validate___definitions_runtimecontext(data, custom_formats={}, name_prefix=N def validate___definitions_oscontext(data, custom_formats={}, name_prefix=None): - data_any_of_count160 = 0 - if not data_any_of_count160: + data_any_of_count166 = 0 + if not data_any_of_count166: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -7791,18 +8043,18 @@ def validate___definitions_oscontext(data, custom_formats={}, name_prefix=None): if "distribution" in data_keys: data_keys.remove("distribution") data__distribution = data["distribution"] - data__distribution_any_of_count161 = 0 - if not data__distribution_any_of_count161: + data__distribution_any_of_count167 = 0 + if not data__distribution_any_of_count167: try: validate___definitions_linuxdistribution( data__distribution, custom_formats, (name_prefix or "data") + ".distribution", ) - data__distribution_any_of_count161 += 1 + data__distribution_any_of_count167 += 1 except JsonSchemaValueException: pass - if not data__distribution_any_of_count161: + if not data__distribution_any_of_count167: try: if not isinstance(data__distribution, (NoneType)): raise JsonSchemaValueException( @@ -7810,10 +8062,10 @@ def validate___definitions_oscontext(data, custom_formats={}, name_prefix=None): + (name_prefix or "data") + ".distribution must be null" ) - data__distribution_any_of_count161 += 1 + data__distribution_any_of_count167 += 1 except JsonSchemaValueException: pass - if not data__distribution_any_of_count161: + if not data__distribution_any_of_count167: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -7872,10 +8124,10 @@ def validate___definitions_oscontext(data, custom_formats={}, name_prefix=None): + str(data_keys) + " properties" ) - data_any_of_count160 += 1 + data_any_of_count166 += 1 except JsonSchemaValueException: pass - if not data_any_of_count160: + if not data_any_of_count166: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -7883,8 +8135,8 @@ def validate___definitions_oscontext(data, custom_formats={}, name_prefix=None): def validate___definitions_linuxdistribution(data, custom_formats={}, name_prefix=None): - data_any_of_count162 = 0 - if not data_any_of_count162: + data_any_of_count168 = 0 + if not data_any_of_count168: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -7928,10 +8180,10 @@ def validate___definitions_linuxdistribution(data, custom_formats={}, name_prefi + str(data_keys) + " properties" ) - data_any_of_count162 += 1 + data_any_of_count168 += 1 except JsonSchemaValueException: pass - if not data_any_of_count162: + if not data_any_of_count168: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -7939,8 +8191,8 @@ def validate___definitions_linuxdistribution(data, custom_formats={}, name_prefi def validate___definitions_devicecontext(data, custom_formats={}, name_prefix=None): - data_any_of_count163 = 0 - if not data_any_of_count163: + data_any_of_count169 = 0 + if not data_any_of_count169: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -8489,10 +8741,10 @@ def validate___definitions_devicecontext(data, custom_formats={}, name_prefix=No + str(data_keys) + " properties" ) - data_any_of_count163 += 1 + data_any_of_count169 += 1 except JsonSchemaValueException: pass - if not data_any_of_count163: + if not data_any_of_count169: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) @@ -8500,8 +8752,8 @@ def validate___definitions_devicecontext(data, custom_formats={}, name_prefix=No def validate___definitions_breadcrumb(data, custom_formats={}, name_prefix=None): - data_any_of_count164 = 0 - if not data_any_of_count164: + data_any_of_count170 = 0 + if not data_any_of_count170: try: if not isinstance(data, (dict)): raise JsonSchemaValueException( @@ -8534,18 +8786,18 @@ def validate___definitions_breadcrumb(data, custom_formats={}, name_prefix=None) if "event_id" in data_keys: data_keys.remove("event_id") data__eventid = data["event_id"] - data__eventid_any_of_count165 = 0 - if not data__eventid_any_of_count165: + data__eventid_any_of_count171 = 0 + if not data__eventid_any_of_count171: try: validate___definitions_eventid( data__eventid, custom_formats, (name_prefix or "data") + ".event_id", ) - data__eventid_any_of_count165 += 1 + data__eventid_any_of_count171 += 1 except JsonSchemaValueException: pass - if not data__eventid_any_of_count165: + if not data__eventid_any_of_count171: try: if not isinstance(data__eventid, (NoneType)): raise JsonSchemaValueException( @@ -8553,10 +8805,10 @@ def validate___definitions_breadcrumb(data, custom_formats={}, name_prefix=None) + (name_prefix or "data") + ".event_id must be null" ) - data__eventid_any_of_count165 += 1 + data__eventid_any_of_count171 += 1 except JsonSchemaValueException: pass - if not data__eventid_any_of_count165: + if not data__eventid_any_of_count171: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -8565,27 +8817,27 @@ def validate___definitions_breadcrumb(data, custom_formats={}, name_prefix=None) if "level" in data_keys: data_keys.remove("level") data__level = data["level"] - data__level_any_of_count166 = 0 - if not data__level_any_of_count166: + data__level_any_of_count172 = 0 + if not data__level_any_of_count172: try: validate___definitions_level( data__level, custom_formats, (name_prefix or "data") + ".level", ) - data__level_any_of_count166 += 1 + data__level_any_of_count172 += 1 except JsonSchemaValueException: pass - if not data__level_any_of_count166: + if not data__level_any_of_count172: try: if not isinstance(data__level, (NoneType)): raise JsonSchemaValueException( "" + (name_prefix or "data") + ".level must be null" ) - data__level_any_of_count166 += 1 + data__level_any_of_count172 += 1 except JsonSchemaValueException: pass - if not data__level_any_of_count166: + if not data__level_any_of_count172: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -8603,18 +8855,18 @@ def validate___definitions_breadcrumb(data, custom_formats={}, name_prefix=None) if "timestamp" in data_keys: data_keys.remove("timestamp") data__timestamp = data["timestamp"] - data__timestamp_any_of_count167 = 0 - if not data__timestamp_any_of_count167: + data__timestamp_any_of_count173 = 0 + if not data__timestamp_any_of_count173: try: validate___definitions_timestamp( data__timestamp, custom_formats, (name_prefix or "data") + ".timestamp", ) - data__timestamp_any_of_count167 += 1 + data__timestamp_any_of_count173 += 1 except JsonSchemaValueException: pass - if not data__timestamp_any_of_count167: + if not data__timestamp_any_of_count173: try: if not isinstance(data__timestamp, (NoneType)): raise JsonSchemaValueException( @@ -8622,10 +8874,10 @@ def validate___definitions_breadcrumb(data, custom_formats={}, name_prefix=None) + (name_prefix or "data") + ".timestamp must be null" ) - data__timestamp_any_of_count167 += 1 + data__timestamp_any_of_count173 += 1 except JsonSchemaValueException: pass - if not data__timestamp_any_of_count167: + if not data__timestamp_any_of_count173: raise JsonSchemaValueException( "" + (name_prefix or "data") @@ -8648,10 +8900,10 @@ def validate___definitions_breadcrumb(data, custom_formats={}, name_prefix=None) + str(data_keys) + " properties" ) - data_any_of_count164 += 1 + data_any_of_count170 += 1 except JsonSchemaValueException: pass - if not data_any_of_count164: + if not data_any_of_count170: raise JsonSchemaValueException( "" + (name_prefix or "data") + " cannot be validated by any definition" ) diff --git a/ingest/management/commands/fetch_event_schema_json.py b/ingest/management/commands/fetch_event_schema_json.py index 897a0f3..a6fe33f 100644 --- a/ingest/management/commands/fetch_event_schema_json.py +++ b/ingest/management/commands/fetch_event_schema_json.py @@ -8,6 +8,20 @@ from django.conf import settings from django.core.management.base import BaseCommand +def alter_schema(schema): + # alter the schema to make it fit what the actual documentation says; see issues/utils.py' get_values() for details + + for key in ['exception', 'threads', 'breadcrumbs']: + # no idea why that anyOf[0] is there (since there's no anyOf[1]), but it's there. + definition = schema['anyOf'][0]['properties'][key] + schema['anyOf'][0]['properties'][key] = { + "anyOf": [ + definition, + definition['properties']['values'], + ] + } + + class Command(BaseCommand): help = "Fetches the event schema JSON from the API and saves it to disk. Used during development." @@ -26,7 +40,7 @@ class Command(BaseCommand): license_result.raise_for_status() with open(settings.BASE_DIR / "api/LICENSE", "w") as f: - f.write("""This licence applies to the file: event.schema.json + f.write("""This licence applies to the files: event.schema.json (and its derivatives) The source of this file is: %s """ % license_url) @@ -35,13 +49,17 @@ The source of this file is: %s # Generate fastjsonschema code from the schema and save it to disk schema = json.loads(json_result.text) + alter_schema(schema) + with open(settings.BASE_DIR / "api/event.schema.altered.json", "w") as f: + json.dump(schema, f, indent=2) + # use_formats=False for "uint64", see https://github.com/horejsek/python-fastjsonschema/issues/108 # use_default=False to avoid fastjsonschema adding default values to the data it validates (bwegh) code = fastjsonschema.compile_to_code(schema, use_formats=False, use_default=False, detailed_exceptions=False) with open(settings.BASE_DIR / "bugsink/event_schema.py", "w") as f: f.write("""# This file is generated by fetch_event_schema_json.py -# it is based on api/event.schema.json +# it is based on api/event.schema.altered.json # the following license applies: """) f.write("\n".join("# " + line for line in license_result.text.splitlines()) + "\n\n") diff --git a/ingest/management/commands/send_json.py b/ingest/management/commands/send_json.py index 133d7c1..86bf4fe 100644 --- a/ingest/management/commands/send_json.py +++ b/ingest/management/commands/send_json.py @@ -41,7 +41,7 @@ class Command(BaseCommand): del data["_meta"] try: - schema_filename = settings.BASE_DIR / 'api/event.schema.json' + schema_filename = settings.BASE_DIR / 'api/event.schema.altered.json' with open(schema_filename, 'r') as f: schema = json.loads(f.read()) diff --git a/ingest/views.py b/ingest/views.py index 824d222..bffd029 100644 --- a/ingest/views.py +++ b/ingest/views.py @@ -134,7 +134,7 @@ class BaseIngestAPIView(View): # exceptions, so you'd have that cost-of-rollback anyway. def get_schema(): - schema_filename = settings.BASE_DIR / 'api/event.schema.json' + schema_filename = settings.BASE_DIR / 'api/event.schema.altered.json' with open(schema_filename, 'r') as f: return json.loads(f.read())