diff --git a/stable b/stable
index e88c34ff0..3061e9e5a 120000
--- a/stable
+++ b/stable
@@ -1 +1 @@
-v0.11.0
\ No newline at end of file
+v0.12.0
\ No newline at end of file
diff --git a/v0.12.0/404.html b/v0.12.0/404.html
new file mode 100644
index 000000000..920148e07
--- /dev/null
+++ b/v0.12.0/404.html
@@ -0,0 +1,1909 @@
+
+
+
+
ANTA is a Python library that can be used in user applications. This section describes how you can leverage ANTA Python modules to help you create your own NRFU solution.
A device is represented in ANTA as a instance of a subclass of the AntaDevice abstract class.
+There are few abstract methods that needs to be implemented by child classes:
+
+
The collect() coroutine is in charge of collecting outputs of AntaCommand instances.
+
The refresh() coroutine is in charge of updating attributes of the AntaDevice instance. These attributes are used by AntaInventory to filter out unreachable devices or by AntaTest to skip devices based on their hardware models.
+
+
The copy() coroutine is used to copy files to and from the device. It does not need to be implemented if tests are not using it.
The refresh() coroutine tries to open a TCP connection on the eAPI port and update the is_online attribute accordingly. If the TCP connection succeeds, it sends a show version command to gather the hardware model of the device and updates the established and hw_model attributes.
+
The copy() coroutine copies files to and from the device using the SCP protocol.
The AntaInventory class is a subclass of the standard Python type dict. The keys of this dictionary are the device names, the values are AntaDevice instances.
+
AntaInventory provides methods to interact with the ANTA inventory:
+
+
The add_device() method adds an AntaDevice instance to the inventory. Adding an entry to AntaInventory with a key different from the device name is not allowed.
To parse a YAML inventory file and print the devices connection status:
+
"""
+Example
+"""
+importasyncio
+
+fromanta.inventoryimportAntaInventory
+
+
+asyncdefmain(inv:AntaInventory)->None:
+"""
+ Take an AntaInventory and:
+ 1. try to connect to every device in the inventory
+ 2. print a message for every device connection status
+ """
+ awaitinv.connect_inventory()
+
+ fordeviceininv.values():
+ ifdevice.established:
+ print(f"Device {device.name} is online")
+ else:
+ print(f"Could not connect to device {device.name}")
+
+if__name__=="__main__":
+ # Create the AntaInventory instance
+ inventory=AntaInventory.parse(
+ filename="inv.yml",
+ username="arista",
+ password="@rista123",
+ timeout=15,
+ )
+
+ # Run the main coroutine
+ res=asyncio.run(main(inventory))
+
+
+How to create your inventory file
+
Please visit this dedicated section for how to use inventory and catalog files.
+
+
To run an EOS commands list on the reachable devices from the inventory:
+
"""
+Example
+"""
+# This is needed to run the script for python < 3.10 for typing annotations
+from__future__importannotations
+
+importasyncio
+frompprintimportpprint
+
+fromanta.inventoryimportAntaInventory
+fromanta.modelsimportAntaCommand
+
+
+asyncdefmain(inv:AntaInventory,commands:list[str])->dict[str,list[AntaCommand]]:
+"""
+ Take an AntaInventory and a list of commands as string and:
+ 1. try to connect to every device in the inventory
+ 2. collect the results of the commands from each device
+
+ Returns:
+ a dictionary where key is the device name and the value is the list of AntaCommand ran towards the device
+ """
+ awaitinv.connect_inventory()
+
+ # Make a list of coroutine to run commands towards each connected device
+ coros=[]
+ # dict to keep track of the commands per device
+ result_dict={}
+ forname,deviceininv.get_inventory(established_only=True).items():
+ anta_commands=[AntaCommand(command=command,ofmt="json")forcommandincommands]
+ result_dict[name]=anta_commands
+ coros.append(device.collect_commands(anta_commands))
+
+ # Run the coroutines
+ awaitasyncio.gather(*coros)
+
+ returnresult_dict
+
+
+if__name__=="__main__":
+ # Create the AntaInventory instance
+ inventory=AntaInventory.parse(
+ filename="inv.yml",
+ username="arista",
+ password="@rista123",
+ timeout=15,
+ )
+
+ # Create a list of commands with json output
+ commands=["show version","show ip bgp summary"]
+
+ # Run the main asyncio entry point
+ res=asyncio.run(main(inventory,commands))
+
+ pprint(res)
+
All the test classes inherit from the same abstract Base Class AntaTest. The Class definition indicates which commands are required for the test and the user should focus only on writing the test function with optional keywords argument. The instance of the class upon creation instantiates a TestResult object that can be accessed later on to check the status of the test ([unset, skipped, success, failure, error]).
All tests are built on a class named AntaTest which provides a complete toolset for a test:
+
+
Object creation
+
Test definition
+
TestResult definition
+
Abstracted method to collect data
+
+
This approach means each time you create a test it will be based on this AntaTest class. Besides that, you will have to provide some elements:
+
+
name: Name of the test
+
description: A human readable description of your test
+
categories: a list of categories to sort test.
+
commands: a list of command to run. This list must be a list of AntaCommand which is described in the next part of this document.
+
+
Here is an example of a hardware test related to device temperature:
+
from__future__importannotations
+
+importlogging
+fromtypingimportAny,Dict,List,Optional,cast
+
+fromanta.modelsimportAntaTest,AntaCommand
+
+
+classVerifyTemperature(AntaTest):
+"""
+ Verifies device temparture is currently OK.
+ """
+
+ # The test name
+ name="VerifyTemperature"
+ # A small description of the test, usually the first line of the class docstring
+ description="Verifies device temparture is currently OK"
+ # The category of the test, usually the module name
+ categories=["hardware"]
+ # The command(s) used for the test. Could be a template instead
+ commands=[AntaCommand(command="show system environment temperature",ofmt="json")]
+
+ # Decorator
+ @AntaTest.anta_test
+ # abstract method that must be defined by the child Test class
+ deftest(self)->None:
+"""Run VerifyTemperature validation"""
+ command_output=cast(Dict[str,Dict[Any,Any]],self.instance_commands[0].output)
+ temperature_status=command_output["systemStatus"]if"systemStatus"incommand_output.keys()else""
+ iftemperature_status=="temperatureOk":
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Device temperature is not OK, systemStatus: {temperature_status}")
+
+
When you run the test, object will automatically call its anta.models.AntaTest.collect() method to get device output for each command if no pre-collected data was given to the test. This method does a loop to call anta.inventory.models.InventoryDevice.collect() methods which is in charge of managing device connection and how to get data.
+
+run test offline
+
You can also pass eos data directly to your test if you want to validate data collected in a different workflow. An example is provided below just for information:
The test function is always the same and must be defined with the @AntaTest.anta_test decorator. This function takes at least one argument which is a anta.inventory.models.InventoryDevice object.
+In some cases a test would rely on some additional inputs from the user, for instance the number of expected peers or some expected numbers. All parameters must come with a default value and the test function should validate the parameters values (at this stage this is the only place where validation can be done but there are future plans to make this better).
The test itself does not return any value, but the result is directly availble from your AntaTest object and exposes a anta.result_manager.models.TestResult object with result, name of the test and optional messages:
+
+
name (str): Device name where the test has run.
+
test (str): Test name runs on the device.
+
categories (List[str]): List of categories the TestResult belongs to, by default the AntaTest categories.
+
description (str): TestResult description, by default the AntaTest description.
+
results (str): Result of the test. Can be one of [“unset”, “success”, “failure”, “error”, “skipped”].
+
message (str, optional): Message to report after the test if any.
+
custom_field (str, optional): Custom field to store a string for flexibility in integrating with ANTA
Most of EOS commands return a JSON structure according to a model (some commands may not be modeled hence the necessity to use text outformat sometimes.
+
The model can change across time (adding feature, … ) and when the model is changed in a non backward-compatible way, the revision number is bumped. The initial model starts with revision 1.
+
A revision applies to a particular CLI command whereas a version is global to an eAPI call. The version is internally translated to a specific revision for each CLI command in the RPC call. The currently supported version vaues are 1 and latest.
+
A revision takes precedence over a version (e.g. if a command is run with version=”latest” and revision=1, the first revision of the model is returned)
+
By default eAPI returns the first revision of each model to ensure that when upgrading, intergation with existing tools is not broken. This is done by using by default version=1 in eAPI calls.
+
+
ANTA uses by default version="latest" in AntaCommand. For some commands, you may want to run them with a different revision or version.
+
For instance the VerifyRoutingTableSize test leverages the first revision of show bfd peers:
+
# revision 1 as later revision introduce additional nesting for type
+commands = [AntaCommand(command="show bfd peers", revision=1)]
+
Because some command can require more dynamic than just a command with no parameter provided by user, ANTA supports command template: you define a template in your test class and user provide parameters when creating test object.
+
classRunArbitraryTemplateCommand(AntaTest):
+"""
+ Run an EOS command and return result
+ Based on AntaTest to build relevant output for pytest
+ """
+
+ name="Run aributrary EOS command"
+ description="To be used only with anta debug commands"
+ template=AntaTemplate(template="show interfaces {ifd}")
+ categories=["debug"]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ errdisabled_interfaces=[interfaceforinterface,valueinresponse["interfaceStatuses"].items()ifvalue["linkStatus"]=="errdisabled"]
+ ...
+
+
+params=[{"ifd":"Ethernet2"},{"ifd":"Ethernet49/1"}]
+run_command1=RunArbitraryTemplateCommand(device_anta,params)
+
+
In this example, test waits for interfaces to check from user setup and will only check for interfaces in params
ANTA is a streamlined Python framework designed for efficient interaction with network devices. This section outlines how ANTA incorporates caching mechanisms to collect command outputs from network devices.
By default, ANTA utilizes aiocache’s memory cache backend, also called SimpleMemoryCache. This library aims for simplicity and supports asynchronous operations to go along with Python asyncio used in ANTA.
+
The _init_cache() method of the AntaDevice abstract class initializes the cache. Child classes can override this method to tweak the cache configuration:
+
def_init_cache(self)->None:
+"""
+ Initialize cache for the device, can be overridden by subclasses to manipulate how it works
+ """
+ self.cache=Cache(cache_class=Cache.MEMORY,ttl=60,namespace=self.name,plugins=[HitMissRatioPlugin()])
+ self.cache_locks=defaultdict(asyncio.Lock)
+
+
The cache is also configured with aiocache’s HitMissRatioPlugin plugin to calculate the ratio of hits the cache has and give useful statistics for logging purposes in ANTA.
The cache is initialized per AntaDevice and uses the following cache key design:
+
<device_name>:<uid>
+
The uid is an attribute of AntaCommand, which is a unique identifier generated from the command, version, revision and output format.
+
Each UID has its own asyncio lock. This design allows coroutines that need to access the cache for different UIDs to do so concurrently. The locks are managed by the self.cache_locks dictionary.
By default, once the cache is initialized, it is used in the collect() method of AntaDevice. The collect() method prioritizes retrieving the output of the command from the cache. If the output is not in the cache, the private _collect() method will retrieve and then store it for future access.
Caching can be disabled per device, network or range by setting the disable_cache key to True when definining the ANTA Inventory file:
+
anta_inventory:
+hosts:
+-host:172.20.20.101
+name:DC1-SPINE1
+tags:["SPINE","DC1"]
+disable_cache:True# Set this key to True
+-host:172.20.20.102
+name:DC1-SPINE2
+tags:["SPINE","DC1"]
+disable_cache:False# Optional since it's the default
+
+networks:
+-network:"172.21.21.0/24"
+disable_cache:True
+
+ranges:
+-start:172.22.22.10
+end:172.22.22.19
+disable_cache:True
+
+ This approach effectively disables caching for ALL commands sent to devices targeted by the disable_cache key.
+
+
+
For tests developpers, caching can be disabled for a specific AntaCommand or AntaTemplate by setting the use_cache attribute to False. That means the command output will always be collected on the device and therefore, never use caching.
Since caching is implemented at the AntaDevice abstract class level, all subclasses will inherit that default behavior. As a result, if you need to disable caching in any custom implementation of AntaDevice outside of the ANTA framework, you must initialize AntaDevice with disable_cache set to True:
+
classAnsibleEOSDevice(AntaDevice):
+"""
+ Implementation of an AntaDevice using Ansible HttpApi plugin for EOS.
+ """
+ def__init__(self,name:str,connection:ConnectionBase,tags:list=None)->None:
+ super().__init__(name,tags,disable_cache=True)
+
A test is a Python class where a test function is defined and will be run by the framework.
+
ANTA provides an abstract class AntaTest. This class does the heavy lifting and provide the logic to define, collect and test data. The code below is an example of a simple test in ANTA, which is an AntaTest subclass:
+
fromanta.modelsimportAntaTest,AntaCommand
+fromanta.decoratorsimportskip_on_platforms
+
+
+classVerifyTemperature(AntaTest):
+"""
+ This test verifies if the device temperature is within acceptable limits.
+
+ Expected Results:
+ * success: The test will pass if the device temperature is currently OK: 'temperatureOk'.
+ * failure: The test will fail if the device temperature is NOT OK.
+ """
+
+ name="VerifyTemperature"
+ description="Verifies if the device temperature is within the acceptable range."
+ categories=["hardware"]
+ commands=[AntaCommand(command="show system environment temperature",ofmt="json")]
+
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ temperature_status=command_output["systemStatus"]if"systemStatus"incommand_output.keys()else""
+ iftemperature_status=="temperatureOk":
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Device temperature exceeds acceptable limits. Current system status: '{temperature_status}'")
+
+
AntaTest also provide more advanced capabilities like AntaCommand templating using the AntaTemplate class or test inputs definition and validation using AntaTest.Inputpydantic model. This will be discussed in the sections below.
name (str): Name of the test. Used during reporting.
+
description (str): A human readable description of your test.
+
categories (list[str]): A list of categories in which the test belongs.
+
commands (list[Union[AntaTemplate, AntaCommand]]): A list of command to collect from devices. This list must be a list of AntaCommand or AntaTemplate instances. Rendering AntaTemplate instances will be discussed later.
+
+
+
Info
+
All these class attributes are mandatory. If any attribute is missing, a NotImplementedError exception will be raised during class instantiation.
You can access an instance attribute in your code using the self reference. E.g. you can access the test input values using self.inputs.
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
Logger object
+
ANTA already provides comprehensive logging at every steps of a test execution. The AntaTest class also provides a logger attribute that is a Python logger specific to the test instance. See Python documentation for more information.
+
+
+
AntaDevice object
+
Even if device is not a private attribute, you should not need to access this object in your code.
AntaTest.Input is a pydantic model that allow test developers to define their test inputs. pydantic provides out of the box error handling for test input validation based on the type hints defined by the test developer.
+
The base definition of AntaTest.Input provides common test inputs for all AntaTest instances:
test(self) -> None: This is an abstract method that must be implemented. It contains the test logic that can access the collected command outputs using the instance_commands instance attribute, access the test inputs using the inputs instance attribute and must set the result instance attribute accordingly. It must be implemented using the AntaTest.anta_test decorator that provides logging and will collect commands before executing the test() method.
Below is a high level description of the test execution flow in ANTA:
+
+
+
ANTA will parse the test catalog to get the list of AntaTest subclasses to instantiate and their associated input values. We consider a single AntaTest subclass in the following steps.
+
+
+
ANTA will instantiate the AntaTest subclass and a single device will be provided to the test instance. The Input model defined in the class will also be instantiated at this moment. If any ValidationError is raised, the test execution will be stopped.
+
+
+
If there is any AntaTemplate instance in the commands class attribute, render() will be called for every occurrence. At this moment, the instance_commands attribute has been initialized. If any rendering error occurs, the test execution will be stopped.
+
+
+
The AntaTest.anta_test decorator will collect the commands from the device and update the instance_commands attribute with the outputs. If any collection error occurs, the test execution will be stopped.
Caching can be disabled per AntaCommand or AntaTemplate by setting the use_cache argument to False. For more details about how caching is implemented in ANTA, please refer to Caching in ANTA.
+
+
fromanta.modelsimportAntaTest,AntaCommand,AntaTemplate
+
+
+class<YourTestName>(AntaTest):
+"""
+ <a docstring description of your test>
+ """
+
+ name="YourTestName"# should be your class name
+ description="<test description in human reading format>"
+ categories=["<arbitrary category>","<another arbitrary category>"]
+ commands=[
+ AntaCommand(
+ command="<EOS command to run>",
+ ofmt="<command format output>",
+ version="<eAPI version to use>",
+ revision="<revision to use for the command>",# revision has precedence over version
+ use_cache="<Use cache for the command>",
+ ),
+ AntaTemplate(
+ template="<Python f-string to render an EOS command>",
+ ofmt="<command format output>",
+ version="<eAPI version to use>",
+ revision="<revision to use for the command>",# revision has precedence over version
+ use_cache="<Use cache for the command>",
+ )
+ ]
+
To define an input field type, refer to the pydantic documentation about types.
+You can also leverage anta.custom_types that provides reusable types defined in ANTA tests.
+
Regarding required, optional and nullable fields, refer to this documentation on how to define them.
+
+
Note
+
All the pydantic features are supported. For instance you can define validators for complex input validation.
The logic usually includes the following different stages:
+1. Parse the command outputs using the self.instance_commands instance attribute.
+2. If needed, access the test inputs using the self.inputs instance attribute and write your conditional logic.
+3. Set the result instance attribute to reflect the test result by either calling self.result.is_success() or self.result.is_failure("<FAILURE REASON>"). Sometimes, setting the test result to skipped using self.result.is_skipped("<SKIPPED REASON>") can make sense (e.g. testing the OSPF neighbor states but no neighbor was found). However, you should not need to catch any exception and set the test result to error since the error handling is done by the framework, see below.
classVerifyTemperature(AntaTest):
+ ...
+ @AntaTest.anta_test
+ deftest(self)->None:
+ # Grab output of the collected command
+ command_output=self.instance_commands[0].json_output
+
+ # Do your test: In this example we check a specific field of the JSON output from EOS
+ temperature_status=command_output["systemStatus"]if"systemStatus"incommand_output.keys()else""
+ iftemperature_status=="temperatureOk":
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Device temperature exceeds acceptable limits. Current system status: '{temperature_status}'")
+
+
As you can see there is no error handling to do in your code. Everything is packaged in the AntaTest.anta_tests decorator and below is a simple example of error captured when trying to access a dictionary with an incorrect key:
+
classVerifyTemperature(AntaTest):
+ ...
+ @AntaTest.anta_test
+ deftest(self)->None:
+ # Grab output of the collected command
+ command_output=self.instance_commands[0].json_output
+
+ # Access the dictionary with an incorrect key
+ command_output['incorrectKey']
+
This section is required only if you are not merging your development into ANTA. Otherwise, just follow contribution guide.
+
+
For that, you need to create your own Python package as described in this hitchhiker’s guide to package Python code. We assume it is well known and we won’t focus on this aspect. Thus, your package must be impartable by ANTA hence available in the module search path sys.path (you can use PYTHONPATH for example).
+
It is very similar to what is documented in catalog section but you have to use your own package name.2
+
Let say the custom Python package is anta_titom73 and the test is defined in anta_titom73.dc_project Python module, the test catalog would look like:
Create an AntaCatalog instance from a dictionary data structure.
+See RawCatalogInput type alias for details.
+It is the data structure returned by yaml.load() function of a valid
+YAML Test Catalog file.
+
+
+
+
Parameters:
+
+
+
+
Name
+
Type
+
Description
+
Default
+
+
+
+
+
data
+
+ RawCatalogInput
+
+
+
+
Python dictionary used to instantiate the AntaCatalog instance
@staticmethod
+deffrom_dict(data:RawCatalogInput,filename:str|Path|None=None)->AntaCatalog:
+"""
+ Create an AntaCatalog instance from a dictionary data structure.
+ See RawCatalogInput type alias for details.
+ It is the data structure returned by `yaml.load()` function of a valid
+ YAML Test Catalog file.
+
+ Args:
+ data: Python dictionary used to instantiate the AntaCatalog instance
+ filename: value to be set as AntaCatalog instance attribute
+ """
+ tests:list[AntaTestDefinition]=[]
+ ifdataisNone:
+ logger.warning("Catalog input data is empty")
+ returnAntaCatalog(filename=filename)
+
+ ifnotisinstance(data,dict):
+ raiseValueError(f"Wrong input type for catalog data{f' (from {filename})'iffilenameisnotNoneelse''}, must be a dict, got {type(data).__name__}")
+
+ try:
+ catalog_data=AntaCatalogFile(**data)# type: ignore[arg-type]
+ exceptValidationErrorase:
+ anta_log_exception(e,f"Test catalog is invalid!{f' (from {filename})'iffilenameisnotNoneelse''}",logger)
+ raise
+ fortincatalog_data.root.values():
+ tests.extend(t)
+ returnAntaCatalog(tests,filename=filename)
+
@staticmethod
+deffrom_list(data:ListAntaTestTuples)->AntaCatalog:
+"""
+ Create an AntaCatalog instance from a list data structure.
+ See ListAntaTestTuples type alias for details.
+
+ Args:
+ data: Python list used to instantiate the AntaCatalog instance
+ """
+ tests:list[AntaTestDefinition]=[]
+ try:
+ tests.extend(AntaTestDefinition(test=test,inputs=inputs)fortest,inputsindata)
+ exceptValidationErrorase:
+ anta_log_exception(e,"Test catalog is invalid!",logger)
+ raise
+ returnAntaCatalog(tests)
+
Return all the tests that have matching tags in their input filters.
+If strict=True, returns only tests that match all the tags provided as input.
+If strict=False, return all the tests that match at least one tag provided as input.
defget_tests_by_tags(self,tags:list[str],strict:bool=False)->list[AntaTestDefinition]:
+"""
+ Return all the tests that have matching tags in their input filters.
+ If strict=True, returns only tests that match all the tags provided as input.
+ If strict=False, return all the tests that match at least one tag provided as input.
+ """
+ result:list[AntaTestDefinition]=[]
+ fortestinself.tests:
+ iftest.inputs.filtersand(f:=test.inputs.filters.tags):
+ if(strictandall(tintagsfortinf))or(notstrictandany(tintagsfortinf)):
+ result.append(test)
+ returnresult
+
test: An AntaTest concrete subclass
+inputs: The associated AntaTest.Input subclass instance
+
+
+ Source code in anta/catalog.py
+
45
+46
+47
+48
+49
+50
+51
+52
+53
+54
+55
def__init__(self,**data:Any)->None:
+"""
+ Inject test in the context to allow to instantiate Input in the BeforeValidator
+ https://docs.pydantic.dev/2.0/usage/validators/#using-validation-context-with-basemodel-initialization
+ """
+ self.__pydantic_validator__.validate_python(
+ data,
+ self_instance=self,
+ context={"test":data["test"]},
+ )
+ super(BaseModel,self).__init__()
+
The inputs class attribute needs to be an instance of the AntaTest.Input subclass defined in the class test.
+
+
+ Source code in anta/catalog.py
+
81
+82
+83
+84
+85
+86
+87
+88
@model_validator(mode="after")
+defcheck_inputs(self)->"AntaTestDefinition":
+"""
+ The `inputs` class attribute needs to be an instance of the AntaTest.Input subclass defined in the class `test`.
+ """
+ ifnotisinstance(self.inputs,self.test.Input):
+ raiseValueError(f"Test input has type {self.inputs.__class__.__qualname__} but expected type {self.test.Input.__qualname__}")
+ returnself
+
If the test has no inputs, allow the user to omit providing the inputs field.
+If the test has inputs, allow the user to provide a valid dictionary of the input fields.
+This model validator will instantiate an Input class from the test class field.
@field_validator("inputs",mode="before")
+@classmethod
+definstantiate_inputs(cls,data:AntaTest.Input|dict[str,Any]|None,info:ValidationInfo)->AntaTest.Input:
+"""
+ If the test has no inputs, allow the user to omit providing the `inputs` field.
+ If the test has inputs, allow the user to provide a valid dictionary of the input fields.
+ This model validator will instantiate an Input class from the `test` class field.
+ """
+ ifinfo.contextisNone:
+ raiseValueError("Could not validate inputs as no test class could be identified")
+ # Pydantic guarantees at this stage that test_class is a subclass of AntaTest because of the ordering
+ # of fields in the class definition - so no need to check for this
+ test_class=info.context["test"]
+ ifnot(isclass(test_class)andissubclass(test_class,AntaTest)):
+ raiseValueError(f"Could not validate inputs as no test class {test_class} is not a subclass of AntaTest")
+
+ ifdataisNone:
+ returntest_class.Input()
+ ifisinstance(data,AntaTest.Input):
+ returndata
+ ifisinstance(data,dict):
+ returntest_class.Input(**data)
+ raiseValueError(f"Coud not instantiate inputs as type {type(data).__name__} is not valid")
+
Allow the user to provide a Python data structure that only has string values.
+This validator will try to flatten and import Python modules, check if the tests classes
+are actually defined in their respective Python module and instantiate Input instances
+with provided value to validate test inputs.
@model_validator(mode="before")
+@classmethod
+defcheck_tests(cls,data:Any)->Any:
+"""
+ Allow the user to provide a Python data structure that only has string values.
+ This validator will try to flatten and import Python modules, check if the tests classes
+ are actually defined in their respective Python module and instantiate Input instances
+ with provided value to validate test inputs.
+ """
+
+ defflatten_modules(data:dict[str,Any],package:str|None=None)->dict[ModuleType,list[Any]]:
+"""
+ Allow the user to provide a data structure with nested Python modules.
+
+ Example:
+ ```
+ anta.tests.routing:
+ generic:
+ - <AntaTestDefinition>
+ bgp:
+ - <AntaTestDefinition>
+ ```
+ `anta.tests.routing.generic` and `anta.tests.routing.bgp` are importable Python modules.
+ """
+ modules:dict[ModuleType,list[Any]]={}
+ formodule_name,testsindata.items():
+ ifpackageandnotmodule_name.startswith("."):
+ module_name=f".{module_name}"
+ try:
+ module:ModuleType=importlib.import_module(name=module_name,package=package)
+ exceptExceptionase:# pylint: disable=broad-exception-caught
+ # A test module is potentially user-defined code.
+ # We need to catch everything if we want to have meaningful logs
+ module_str=f"{module_name[1:]ifmodule_name.startswith('.')elsemodule_name}{f' from package {package}'ifpackageelse''}"
+ message=f"Module named {module_str} cannot be imported. Verify that the module exists and there is no Python syntax issues."
+ anta_log_exception(e,message,logger)
+ raiseValueError(message)frome
+ ifisinstance(tests,dict):
+ # This is an inner Python module
+ modules.update(flatten_modules(data=tests,package=module.__name__))
+ else:
+ ifnotisinstance(tests,list):
+ raiseValueError(f"Syntax error when parsing: {tests}\nIt must be a list of ANTA tests. Check the test catalog.")
+ # This is a list of AntaTestDefinition
+ modules[module]=tests
+ returnmodules
+
+ ifisinstance(data,dict):
+ typed_data:dict[ModuleType,list[Any]]=flatten_modules(data)
+ formodule,testsintyped_data.items():
+ test_definitions:list[AntaTestDefinition]=[]
+ fortest_definitionintests:
+ ifnotisinstance(test_definition,dict):
+ raiseValueError(f"Syntax error when parsing: {test_definition}\nIt must be a dictionary. Check the test catalog.")
+ iflen(test_definition)!=1:
+ raiseValueError(
+ f"Syntax error when parsing: {test_definition}\nIt must be a dictionary with a single entry. Check the indentation in the test catalog."
+ )
+ fortest_name,test_inputsintest_definition.copy().items():
+ test:type[AntaTest]|None=getattr(module,test_name,None)
+ iftestisNone:
+ raiseValueError(
+ f"{test_name} is not defined in Python module {module.__name__}{f' (from {module.__file__})'ifmodule.__file__isnotNoneelse''}"
+ )
+ test_definitions.append(AntaTestDefinition(test=test,inputs=test_inputs))
+ typed_data[module]=test_definitions
+ returntyped_data
+
def__init__(self,name:str,tags:Optional[list[str]]=None,disable_cache:bool=False)->None:
+"""
+ Constructor of AntaDevice
+
+ Args:
+ name: Device name
+ tags: List of tags for this device
+ disable_cache: Disable caching for all commands for this device. Defaults to False.
+ """
+ self.name:str=name
+ self.hw_model:Optional[str]=None
+ self.tags:list[str]=tagsiftagsisnotNoneelse[]
+ # A device always has its own name as tag
+ self.tags.append(self.name)
+ self.is_online:bool=False
+ self.established:bool=False
+ self.cache:Optional[Cache]=None
+ self.cache_locks:Optional[defaultdict[str,asyncio.Lock]]=None
+
+ # Initialize cache if not disabled
+ ifnotdisable_cache:
+ self._init_cache()
+
When caching is activated on both the device and the command,
+this method prioritizes retrieving the output from the cache. In cases where the output isn’t cached yet,
+it will be freshly collected and then stored in the cache for future access.
+The method employs asynchronous locks based on the command’s UID to guarantee exclusive access to the cache.
+
When caching is NOT enabled, either at the device or command level, the method directly collects the output
+via the private _collect method without interacting with the cache.
asyncdefcollect(self,command:AntaCommand)->None:
+"""
+ Collects the output for a specified command.
+
+ When caching is activated on both the device and the command,
+ this method prioritizes retrieving the output from the cache. In cases where the output isn't cached yet,
+ it will be freshly collected and then stored in the cache for future access.
+ The method employs asynchronous locks based on the command's UID to guarantee exclusive access to the cache.
+
+ When caching is NOT enabled, either at the device or command level, the method directly collects the output
+ via the private `_collect` method without interacting with the cache.
+
+ Args:
+ command (AntaCommand): The command to process.
+ """
+ # Need to ignore pylint no-member as Cache is a proxy class and pylint is not smart enough
+ # https://github.com/pylint-dev/pylint/issues/7258
+ ifself.cacheisnotNoneandself.cache_locksisnotNoneandcommand.use_cache:
+ asyncwithself.cache_locks[command.uid]:
+ cached_output=awaitself.cache.get(command.uid)# pylint: disable=no-member
+
+ ifcached_outputisnotNone:
+ logger.debug(f"Cache hit for {command.command} on {self.name}")
+ command.output=cached_output
+ else:
+ awaitself._collect(command=command)
+ awaitself.cache.set(command.uid,command.output)# pylint: disable=no-member
+ else:
+ awaitself._collect(command=command)
+
asyncdefcopy(self,sources:list[Path],destination:Path,direction:Literal["to","from"]="from")->None:
+"""
+ Copy files to and from the device, usually through SCP.
+ It is not mandatory to implement this for a valid AntaDevice subclass.
+
+ Args:
+ sources: List of files to copy to or from the device.
+ destination: Local or remote destination when copying the files. Can be a folder.
+ direction: Defines if this coroutine copies files to or from the device.
+ """
+ raiseNotImplementedError(f"copy() method has not been implemented in {self.__class__.__name__} definition")
+
+
+
+ This coroutine must update the following attributes of AntaDevice
+
+
is_online: When the device IP is reachable and a port can be open
+
established: When a command execution succeeds
+
hw_model: The hardware model of the device
+
+
+
+ Source code in anta/device.py
+
183
+184
+185
+186
+187
+188
+189
+190
+191
+192
@abstractmethod
+asyncdefrefresh(self)->None:
+"""
+ Update attributes of an AntaDevice instance.
+
+ This coroutine must update the following attributes of AntaDevice:
+ - `is_online`: When the device IP is reachable and a port can be open
+ - `established`: When a command execution succeeds
+ - `hw_model`: The hardware model of the device
+ """
+
Returns True if the command is supported on the device hardware platform, False otherwise.
+
+
+ Source code in anta/device.py
+
175
+176
+177
+178
+179
+180
+181
defsupports(self,command:AntaCommand)->bool:
+"""Returns True if the command is supported on the device hardware platform, False otherwise."""
+ unsupported=any("not supported on this hardware platform"ineforeincommand.errors)
+ logger.debug(command)
+ ifunsupported:
+ logger.debug(f"{command.command} is not supported on {self.hw_model}")
+ returnnotunsupported
+
def__init__(# pylint: disable=R0913
+ self,
+ host:str,
+ username:str,
+ password:str,
+ name:Optional[str]=None,
+ enable:bool=False,
+ enable_password:Optional[str]=None,
+ port:Optional[int]=None,
+ ssh_port:Optional[int]=22,
+ tags:Optional[list[str]]=None,
+ timeout:Optional[float]=None,
+ insecure:bool=False,
+ proto:Literal["http","https"]="https",
+ disable_cache:bool=False,
+)->None:
+"""
+ Constructor of AsyncEOSDevice
+
+ Args:
+ host: Device FQDN or IP
+ username: Username to connect to eAPI and SSH
+ password: Password to connect to eAPI and SSH
+ name: Device name
+ enable: Device needs privileged access
+ enable_password: Password used to gain privileged access on EOS
+ port: eAPI port. Defaults to 80 is proto is 'http' or 443 if proto is 'https'.
+ ssh_port: SSH port
+ tags: List of tags for this device
+ timeout: Timeout value in seconds for outgoing connections. Default to 10 secs.
+ insecure: Disable SSH Host Key validation
+ proto: eAPI protocol. Value can be 'http' or 'https'
+ disable_cache: Disable caching for all commands for this device. Defaults to False.
+ """
+ ifhostisNone:
+ message="'host' is required to create an AsyncEOSDevice"
+ logger.error(message)
+ raiseValueError(message)
+ ifnameisNone:
+ name=f"{host}{f':{port}'ifportelse''}"
+ super().__init__(name,tags,disable_cache)
+ ifusernameisNone:
+ message=f"'username' is required to instantiate device '{self.name}'"
+ logger.error(message)
+ raiseValueError(message)
+ ifpasswordisNone:
+ message=f"'password' is required to instantiate device '{self.name}'"
+ logger.error(message)
+ raiseValueError(message)
+ self.enable=enable
+ self._enable_password=enable_password
+ self._session:aioeapi.Device=aioeapi.Device(host=host,port=port,username=username,password=password,proto=proto,timeout=timeout)
+ ssh_params:dict[str,Any]={}
+ ifinsecure:
+ ssh_params["known_hosts"]=None
+ self._ssh_opts:SSHClientConnectionOptions=SSHClientConnectionOptions(host=host,port=ssh_port,username=username,password=password,**ssh_params)
+
asyncdefcopy(self,sources:list[Path],destination:Path,direction:Literal["to","from"]="from")->None:
+"""
+ Copy files to and from the device using asyncssh.scp().
+
+ Args:
+ sources: List of files to copy to or from the device.
+ destination: Local or remote destination when copying the files. Can be a folder.
+ direction: Defines if this coroutine copies files to or from the device.
+ """
+ asyncwithasyncssh.connect(
+ host=self._ssh_opts.host,
+ port=self._ssh_opts.port,
+ tunnel=self._ssh_opts.tunnel,
+ family=self._ssh_opts.family,
+ local_addr=self._ssh_opts.local_addr,
+ options=self._ssh_opts,
+ )asconn:
+ src:Union[list[tuple[SSHClientConnection,Path]],list[Path]]
+ dst:Union[tuple[SSHClientConnection,Path],Path]
+ ifdirection=="from":
+ src=[(conn,file)forfileinsources]
+ dst=destination
+ forfileinsources:
+ logger.info(f"Copying '{file}' from device {self.name} to '{destination}' locally")
+
+ elifdirection=="to":
+ src=sources
+ dst=conn,destination
+ forfileinsrc:
+ logger.info(f"Copying '{file}' to device {self.name} to '{destination}' remotely")
+
+ else:
+ logger.critical(f"'direction' argument to copy() fonction is invalid: {direction}")
+
+ return
+ awaitasyncssh.scp(src,dst)
+
This coroutine must update the following attributes of AsyncEOSDevice:
+- is_online: When a device IP is reachable and a port can be open
+- established: When a command execution succeeds
+- hw_model: The hardware model of the device
asyncdefrefresh(self)->None:
+"""
+ Update attributes of an AsyncEOSDevice instance.
+
+ This coroutine must update the following attributes of AsyncEOSDevice:
+ - is_online: When a device IP is reachable and a port can be open
+ - established: When a command execution succeeds
+ - hw_model: The hardware model of the device
+ """
+ logger.debug(f"Refreshing device {self.name}")
+ self.is_online=awaitself._session.check_connection()
+ ifself.is_online:
+ COMMAND:str="show version"
+ HW_MODEL_KEY:str="modelName"
+ try:
+ response=awaitself._session.cli(command=COMMAND)
+ exceptaioeapi.EapiCommandErrorase:
+ logger.warning(f"Cannot get hardware information from device {self.name}: {e.errmsg}")
+
+ except(HTTPError,ConnectError)ase:
+ logger.warning(f"Cannot get hardware information from device {self.name}: {exc_to_str(e)}")
+
+ else:
+ ifHW_MODEL_KEYinresponse:
+ self.hw_model=response[HW_MODEL_KEY]
+ else:
+ logger.warning(f"Cannot get hardware information from device {self.name}: cannot parse '{COMMAND}'")
+
+ else:
+ logger.warning(f"Could not connect to device {self.name}: cannot open eAPI port")
+
+ self.established=bool(self.is_onlineandself.hw_model)
+
defadd_device(self,device:AntaDevice)->None:
+"""Add a device to final inventory.
+
+ Args:
+ device: Device object to be added
+ """
+ self[device.name]=device
+
defget_inventory(self,established_only:bool=False,tags:Optional[list[str]]=None)->AntaInventory:
+"""
+ Returns a filtered inventory.
+
+ Args:
+ established_only: Whether or not to include only established devices. Default False.
+ tags: List of tags to filter devices.
+
+ Returns:
+ AntaInventory: An inventory with filtered AntaDevice objects.
+ """
+
+ def_filter_devices(device:AntaDevice)->bool:
+"""
+ Helper function to select the devices based on the input tags
+ and the requirement for an established connection.
+ """
+ iftagsisnotNoneandall(tagnotintagsfortagindevice.tags):
+ returnFalse
+ returnbool(notestablished_onlyordevice.established)
+
+ devices:list[AntaDevice]=list(filter(_filter_devices,self.values()))
+ result=AntaInventory()
+ fordeviceindevices:
+ result.add_device(device)
+ returnresult
+
The goal of this class is to handle the heavy lifting and make
+writing a test as simple as possible.
+
+
+
+
Examples:
+
+
The following is an example of an AntaTest subclass implementation:
+
classVerifyReachability(AntaTest):
+ name="VerifyReachability"
+ description="Test the network reachability to one or many destination IP(s)."
+ categories=["connectivity"]
+ commands=[AntaTemplate(template="ping vrf {vrf}{dst} source {src} repeat 2")]
+
+ classInput(AntaTest.Input):
+ hosts:list[Host]
+ classHost(BaseModel):
+ dst:IPv4Address
+ src:IPv4Address
+ vrf:str="default"
+
+ defrender(self,template:AntaTemplate)->list[AntaCommand]:
+ return[template.render({"dst":host.dst,"src":host.src,"vrf":host.vrf})forhostinself.inputs.hosts]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ failures=[]
+ forcommandinself.instance_commands:
+ ifcommand.paramsand("src"and"dst")incommand.params:
+ src,dst=command.params["src"],command.params["dst"]
+ if"2 received"notincommand.json_output["messages"][0]:
+ failures.append((str(src),str(dst)))
+ ifnotfailures:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Connectivity test failed for the following source-destination pairs: {failures}")
+
+Attributes:
+ device: AntaDevice instance on which this test is run
+ inputs: AntaTest.Input instance carrying the test inputs
+ instance_commands: List of AntaCommand instances of this test
+ result: TestResult instance representing the result of this test
+ logger: Python logger for this test instance
+
+
+
+
+
Populate outputs of the test commands instead of collecting from devices.
+ This list must have the same length and order than the instance_commands instance attribute.
def__init__(
+ self,
+ device:AntaDevice,
+ inputs:dict[str,Any]|AntaTest.Input|None=None,
+ eos_data:list[dict[Any,Any]|str]|None=None,
+):
+"""AntaTest Constructor
+
+ Args:
+ device: AntaDevice instance on which the test will be run
+ inputs: dictionary of attributes used to instantiate the AntaTest.Input instance
+ eos_data: Populate outputs of the test commands instead of collecting from devices.
+ This list must have the same length and order than the `instance_commands` instance attribute.
+ """
+ self.logger:logging.Logger=logging.getLogger(f"{self.__module__}.{self.__class__.__name__}")
+ self.device:AntaDevice=device
+ self.inputs:AntaTest.Input
+ self.instance_commands:list[AntaCommand]=[]
+ self.result:TestResult=TestResult(name=device.name,test=self.name,categories=self.categories,description=self.description)
+ self._init_inputs(inputs)
+ ifself.result.result=="unset":
+ self._init_commands(eos_data)
+
Implement generic hashing for AntaTest.Input.
+This will work in most cases but this does not consider 2 lists with different ordering as equal.
+
+
+ Source code in anta/models.py
+
264
+265
+266
+267
+268
+269
def__hash__(self)->int:
+"""
+ Implement generic hashing for AntaTest.Input.
+ This will work in most cases but this does not consider 2 lists with different ordering as equal.
+ """
+ returnhash(self.model_dump_json())
+
@staticmethod
+defanta_test(function:F)->Callable[...,Coroutine[Any,Any,TestResult]]:
+"""
+ Decorator for the `test()` method.
+
+ This decorator implements (in this order):
+
+ 1. Instantiate the command outputs if `eos_data` is provided to the `test()` method
+ 2. Collect the commands from the device
+ 3. Run the `test()` method
+ 4. Catches any exception in `test()` user code and set the `result` instance attribute
+ """
+
+ @wraps(function)
+ asyncdefwrapper(
+ self:AntaTest,
+ eos_data:list[dict[Any,Any]|str]|None=None,
+ **kwargs:Any,
+ )->TestResult:
+"""
+ Args:
+ eos_data: Populate outputs of the test commands instead of collecting from devices.
+ This list must have the same length and order than the `instance_commands` instance attribute.
+
+ Returns:
+ result: TestResult instance attribute populated with error status if any
+ """
+
+ defformat_td(seconds:float,digits:int=3)->str:
+ isec,fsec=divmod(round(seconds*10**digits),10**digits)
+ returnf"{timedelta(seconds=isec)}.{fsec:0{digits}.0f}"
+
+ start_time=time.time()
+ ifself.result.result!="unset":
+ returnself.result
+
+ # Data
+ ifeos_dataisnotNone:
+ self.save_commands_data(eos_data)
+ self.logger.debug(f"Test {self.name} initialized with input data {eos_data}")
+
+ # If some data is missing, try to collect
+ ifnotself.collected:
+ awaitself.collect()
+ ifself.result.result!="unset":
+ returnself.result
+
+ ifcmds:=self.failed_commands:
+ self.logger.debug(self.device.supports)
+ unsupported_commands=[f"Skipped because {c.command} is not supported on {self.device.hw_model}"forcincmdsifnotself.device.supports(c)]
+ self.logger.debug(unsupported_commands)
+ ifunsupported_commands:
+ self.logger.warning(f"Test {self.name} has been skipped because it is not supported on {self.device.hw_model}: {GITHUB_SUGGESTION}")
+ self.result.is_skipped("\n".join(unsupported_commands))
+ returnself.result
+ self.result.is_error(message="\n".join([f"{c.command} has failed: {', '.join(c.errors)}"forcincmds]))
+ returnself.result
+
+ try:
+ function(self,**kwargs)
+ exceptExceptionase:# pylint: disable=broad-exception-caught
+ # test() is user-defined code.
+ # We need to catch everything if we want the AntaTest object
+ # to live until the reporting
+ message=f"Exception raised for test {self.name} (on device {self.device.name})"
+ anta_log_exception(e,message,self.logger)
+ self.result.is_error(message=exc_to_str(e))
+
+ test_duration=time.time()-start_time
+ self.logger.debug(f"Executing test {self.name} on device {self.device.name} took {format_td(test_duration)}")
+
+ AntaTest.update_progress()
+ returnself.result
+
+ returnwrapper
+
asyncdefcollect(self)->None:
+"""
+ Method used to collect outputs of all commands of this test class from the device of this test instance.
+ """
+ try:
+ ifself.blockedisFalse:
+ awaitself.device.collect_commands(self.instance_commands)
+ exceptExceptionase:# pylint: disable=broad-exception-caught
+ # device._collect() is user-defined code.
+ # We need to catch everything if we want the AntaTest object
+ # to live until the reporting
+ message=f"Exception raised while collecting commands for test {self.name} (on device {self.device.name})"
+ anta_log_exception(e,message,self.logger)
+ self.result.is_error(message=exc_to_str(e))
+
Render an AntaTemplate instance of this AntaTest using the provided
+ AntaTest.Input instance at self.inputs.
+
This is not an abstract method because it does not need to be implemented if there is
+no AntaTemplate for this test.
+
+
+ Source code in anta/models.py
+
405
+406
+407
+408
+409
+410
+411
defrender(self,template:AntaTemplate)->list[AntaCommand]:
+"""Render an AntaTemplate instance of this AntaTest using the provided
+ AntaTest.Input instance at self.inputs.
+
+ This is not an abstract method because it does not need to be implemented if there is
+ no AntaTemplate for this test."""
+ raiseNotImplementedError(f"AntaTemplate are provided but render() method has not been implemented for {self.__module__}.{self.name}")
+
Populate output of all AntaCommand instances in instance_commands
+
+
+ Source code in anta/models.py
+
377
+378
+379
+380
+381
+382
+383
+384
+385
+386
defsave_commands_data(self,eos_data:list[dict[str,Any]|str])->None:
+"""Populate output of all AntaCommand instances in `instance_commands`"""
+ iflen(eos_data)>len(self.instance_commands):
+ self.result.is_error(message="Test initialization error: Trying to save more data than there are commands for the test")
+ return
+ iflen(eos_data)<len(self.instance_commands):
+ self.result.is_error(message="Test initialization error: Trying to save less data than there are commands for the test")
+ return
+ forindex,datainenumerate(eos_dataor[]):
+ self.instance_commands[index].output=data
+
This abstract method is the core of the test logic.
+It must set the correct status of the result instance attribute
+with the appropriate outcome of the test.
+
+
+
+
Examples:
+
+
It must be implemented using the AntaTest.anta_test decorator:
+
@AntaTest.anta_test
+deftest(self)->None:
+ self.result.is_success()
+ forcommandinself.instance_commands:
+ ifnotself._test_command(command):# _test_command() is an arbitrary test logic
+ self.result.is_failure("Failure reson")
+
@abstractmethod
+deftest(self)->Coroutine[Any,Any,TestResult]:
+"""
+ This abstract method is the core of the test logic.
+ It must set the correct status of the `result` instance attribute
+ with the appropriate outcome of the test.
+
+ Examples:
+ It must be implemented using the `AntaTest.anta_test` decorator:
+ ```python
+ @AntaTest.anta_test
+ def test(self) -> None:
+ self.result.is_success()
+ for command in self.instance_commands:
+ if not self._test_command(command): # _test_command() is an arbitrary test logic
+ self.result.is_failure("Failure reson")
+ ```
+ """
+
eAPI models are revisioned, this means that if a model is modified in a non-backwards compatible way, then its revision will be bumped up
+(revisions are numbers, default value is 1).
+
By default an eAPI request will return revision 1 of the model instance,
+this ensures that older management software will not suddenly stop working when a switch is upgraded.
+A revision applies to a particular CLI command whereas a version is global and is internally
+translated to a specific revision for each CLI command in the RPC.
defrender(self,**params:dict[str,Any])->AntaCommand:
+"""Render an AntaCommand from an AntaTemplate instance.
+ Keep the parameters used in the AntaTemplate instance.
+
+ Args:
+ params: dictionary of variables with string values to render the Python f-string
+
+ Returns:
+ command: The rendered AntaCommand.
+ This AntaCommand instance have a template attribute that references this
+ AntaTemplate instance.
+ """
+ try:
+ returnAntaCommand(
+ command=self.template.format(**params),
+ ofmt=self.ofmt,
+ version=self.version,
+ revision=self.revision,
+ template=self,
+ params=params,
+ use_cache=self.use_cache,
+ )
+ exceptKeyErrorase:
+ raiseAntaTemplateRenderError(self,e.args[0])frome
+
defreport_all(
+ self,
+ result_manager:ResultManager,
+ host:Optional[str]=None,
+ testcase:Optional[str]=None,
+ title:str="All tests results",
+)->Table:
+"""
+ Create a table report with all tests for one or all devices.
+
+ Create table with full output: Host / Test / Status / Message
+
+ Args:
+ result_manager (ResultManager): A manager with a list of tests.
+ host (str, optional): IP Address of a host to search for. Defaults to None.
+ testcase (str, optional): A test name to search for. Defaults to None.
+ title (str, optional): Title for the report. Defaults to 'All tests results'.
+
+ Returns:
+ Table: A fully populated rich Table
+ """
+ table=Table(title=title,show_lines=True)
+ headers=["Device","Test Name","Test Status","Message(s)","Test description","Test category"]
+ table=self._build_headers(headers=headers,table=table)
+
+ forresultinresult_manager.get_results():
+ # pylint: disable=R0916
+ if(hostisNoneandtestcaseisNone)or(hostisnotNoneandstr(result.name)==host)or(testcaseisnotNoneandtestcase==str(result.test)):
+ state=self._color_result(result.result)
+ message=self._split_list_to_txt_list(result.messages)iflen(result.messages)>0else""
+ categories=", ".join(result.categories)
+ table.add_row(str(result.name),result.test,state,message,result.description,categories)
+ returntable
+
defreport_summary_hosts(
+ self,
+ result_manager:ResultManager,
+ host:Optional[str]=None,
+ title:str="Summary per host",
+)->Table:
+"""
+ Create a table report with result agregated per host.
+
+ Create table with full output: Host / Number of success / Number of failure / Number of error / List of nodes in error or failure
+
+ Args:
+ result_manager (ResultManager): A manager with a list of tests.
+ host (str, optional): IP Address of a host to search for. Defaults to None.
+ title (str, optional): Title for the report. Defaults to 'All tests results'.
+
+ Returns:
+ Table: A fully populated rich Table
+ """
+ table=Table(title=title,show_lines=True)
+ headers=[
+ "Device",
+ "# of success",
+ "# of skipped",
+ "# of failure",
+ "# of errors",
+ "List of failed or error test cases",
+ ]
+ table=self._build_headers(headers=headers,table=table)
+ forhost_readinresult_manager.get_hosts():
+ ifhostisNoneorstr(host_read)==host:
+ results=result_manager.get_result_by_host(host_read)
+ logger.debug("data to use for computation")
+ logger.debug(f"{host}: {results}")
+ nb_failure=len([resultforresultinresultsifresult.result=="failure"])
+ nb_error=len([resultforresultinresultsifresult.result=="error"])
+ list_failure=[str(result.test)forresultinresultsifresult.resultin["failure","error"]]
+ nb_success=len([resultforresultinresultsifresult.result=="success"])
+ nb_skipped=len([resultforresultinresultsifresult.result=="skipped"])
+ table.add_row(
+ str(host_read),
+ str(nb_success),
+ str(nb_skipped),
+ str(nb_failure),
+ str(nb_error),
+ str(list_failure),
+ )
+ returntable
+
defreport_summary_tests(
+ self,
+ result_manager:ResultManager,
+ testcase:Optional[str]=None,
+ title:str="Summary per test case",
+)->Table:
+"""
+ Create a table report with result agregated per test.
+
+ Create table with full output: Test / Number of success / Number of failure / Number of error / List of nodes in error or failure
+
+ Args:
+ result_manager (ResultManager): A manager with a list of tests.
+ testcase (str, optional): A test name to search for. Defaults to None.
+ title (str, optional): Title for the report. Defaults to 'All tests results'.
+
+ Returns:
+ Table: A fully populated rich Table
+ """
+ # sourcery skip: class-extract-method
+ table=Table(title=title,show_lines=True)
+ headers=[
+ "Test Case",
+ "# of success",
+ "# of skipped",
+ "# of failure",
+ "# of errors",
+ "List of failed or error nodes",
+ ]
+ table=self._build_headers(headers=headers,table=table)
+ fortestcase_readinresult_manager.get_testcases():
+ iftestcaseisNoneorstr(testcase_read)==testcase:
+ results=result_manager.get_result_by_test(testcase_read)
+ nb_failure=len([resultforresultinresultsifresult.result=="failure"])
+ nb_error=len([resultforresultinresultsifresult.result=="error"])
+ list_failure=[str(result.name)forresultinresultsifresult.resultin["failure","error"]]
+ nb_success=len([resultforresultinresultsifresult.result=="success"])
+ nb_skipped=len([resultforresultinresultsifresult.result=="skipped"])
+ table.add_row(
+ testcase_read,
+ str(nb_success),
+ str(nb_skipped),
+ str(nb_failure),
+ str(nb_error),
+ str(list_failure),
+ )
+ returntable
+
def__init__(self)->None:
+"""
+ Class constructor.
+
+ The status of the class is initialized to "unset"
+
+ Then when adding a test with a status that is NOT 'error' the following
+ table shows the updated status:
+
+ | Current Status | Added test Status | Updated Status |
+ | -------------- | ------------------------------- | -------------- |
+ | unset | Any | Any |
+ | skipped | unset, skipped | skipped |
+ | skipped | success | success |
+ | skipped | failure | failure |
+ | success | unset, skipped, success | success |
+ | success | failure | failure |
+ | failure | unset, skipped success, failure | failure |
+
+ If the status of the added test is error, the status is untouched and the
+ error_status is set to True.
+ """
+ self._result_entries:list[TestResult]=[]
+ # Initialize status
+ self.status:TestStatus="unset"
+ self.error_status=False
+
+
+
+ Source code in anta/result_manager/__init__.py
+
117
+118
+119
+120
+121
+122
+123
+124
+125
defadd_test_result(self,entry:TestResult)->None:
+"""Add a result to the list
+
+ Args:
+ entry (TestResult): TestResult data to add to the report
+ """
+ logger.debug(entry)
+ self._result_entries.append(entry)
+ self._update_status(entry.result)
+
+
+
+ Source code in anta/result_manager/__init__.py
+
127
+128
+129
+130
+131
+132
+133
+134
defadd_test_results(self,entries:list[TestResult])->None:
+"""Add a list of results to the list
+
+ Args:
+ entries (list[TestResult]): List of TestResult data to add to the report
+ """
+ foreinentries:
+ self.add_test_result(e)
+
defget_hosts(self)->list[str]:
+"""
+ Get list of IP addresses in current manager.
+
+ Returns:
+ list[str]: List of IP addresses.
+ """
+ result_list=[]
+ fortestcaseinself._result_entries:
+ ifstr(testcase.name)notinresult_list:
+ result_list.append(str(testcase.name))
+ returnresult_list
+
defget_json_results(self)->str:
+"""
+ Expose list of all test results in JSON
+
+ Returns:
+ str: JSON dumps of the list of results
+ """
+ res=[]
+ fordeviceinself._result_entries:
+ res.append({k:vifisinstance(v,list)elsestr(v)fork,vindevice})
+ returnjson.dumps(res,indent=4)
+
defget_result_by_host(self,host_ip:str)->list[TestResult]:
+"""
+ Get list of test result for a given host.
+
+ Args:
+ host_ip (str): IP Address of the host to use to filter results.
+ output_format (str, optional): format selector. Can be either native/list. Defaults to 'native'.
+
+ Returns:
+ list[TestResult]: List of results related to the host.
+ """
+ return[resultforresultinself._result_entriesifstr(result.name)==host_ip]
+
defget_result_by_test(self,test_name:str)->list[TestResult]:
+"""
+ Get list of test result for a given test.
+
+ Args:
+ test_name (str): Test name to use to filter results
+ output_format (str, optional): format selector. Can be either native/list. Defaults to 'native'.
+
+ Returns:
+ list[TestResult]: List of results related to the test.
+ """
+ return[resultforresultinself._result_entriesifstr(result.test)==test_name]
+
+
+
+ Source code in anta/result_manager/__init__.py
+
142
+143
+144
+145
+146
+147
+148
+149
defget_results(self)->list[TestResult]:
+"""
+ Expose list of all test results in different format
+
+ Returns:
+ any: List of results.
+ """
+ returnself._result_entries
+
Returns the current status including error_status if ignore_error is False
+
+
+ Source code in anta/result_manager/__init__.py
+
136
+137
+138
+139
+140
defget_status(self,ignore_error:bool=False)->str:
+"""
+ Returns the current status including error_status if ignore_error is False
+ """
+ return"error"ifself.error_statusandnotignore_errorelseself.status
+
defget_testcases(self)->list[str]:
+"""
+ Get list of name of all test cases in current manager.
+
+ Returns:
+ list[str]: List of names for all tests.
+ """
+ result_list=[]
+ fortestcaseinself._result_entries:
+ ifstr(testcase.test)notinresult_list:
+ result_list.append(str(testcase.test))
+ returnresult_list
+
+
+
+ Source code in anta/result_manager/models.py
+
46
+47
+48
+49
+50
+51
+52
+53
defis_failure(self,message:str|None=None)->None:
+"""
+ Helper to set status to failure
+
+ Args:
+ message: Optional message related to the test
+ """
+ self._set_status("failure",message)
+
+
+
+ Source code in anta/result_manager/models.py
+
55
+56
+57
+58
+59
+60
+61
+62
defis_skipped(self,message:str|None=None)->None:
+"""
+ Helper to set status to skipped
+
+ Args:
+ message: Optional message related to the test
+ """
+ self._set_status("skipped",message)
+
+
+
+ Source code in anta/result_manager/models.py
+
37
+38
+39
+40
+41
+42
+43
+44
defis_success(self,message:str|None=None)->None:
+"""
+ Helper to set status to success
+
+ Args:
+ message: Optional message related to the test
+ """
+ self._set_status("success",message)
+
classVerifyAcctConsoleMethods(AntaTest):
+"""
+ Verifies the AAA accounting console method lists for different accounting types (system, exec, commands, dot1x).
+
+ Expected Results:
+ * success: The test will pass if the provided AAA accounting console method list is matching in the configured accounting types.
+ * failure: The test will fail if the provided AAA accounting console method list is NOT matching in the configured accounting types.
+ """
+
+ name="VerifyAcctConsoleMethods"
+ description="Verifies the AAA accounting console method lists for different accounting types (system, exec, commands, dot1x)."
+ categories=["aaa"]
+ commands=[AntaCommand(command="show aaa methods accounting")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ methods:List[AAAAuthMethod]
+"""List of AAA accounting console methods. Methods should be in the right order"""
+ types:Set[Literal["commands","exec","system","dot1x"]]
+"""List of accounting console types to verify"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ not_matching=[]
+ not_configured=[]
+ fork,vincommand_output.items():
+ acct_type=k.replace("AcctMethods","")
+ ifacct_typenotinself.inputs.types:
+ # We do not need to verify this accounting type
+ continue
+ formethodsinv.values():
+ if"consoleAction"notinmethods:
+ not_configured.append(acct_type)
+ ifmethods["consoleMethods"]!=self.inputs.methods:
+ not_matching.append(acct_type)
+ ifnot_configured:
+ self.result.is_failure(f"AAA console accounting is not configured for {not_configured}")
+ return
+ ifnotnot_matching:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"AAA accounting console methods {self.inputs.methods} are not matching for {not_matching}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ methods:List[AAAAuthMethod]
+"""List of AAA accounting console methods. Methods should be in the right order"""
+ types:Set[Literal["commands","exec","system","dot1x"]]
+"""List of accounting console types to verify"""
+
classVerifyAcctDefaultMethods(AntaTest):
+"""
+ Verifies the AAA accounting default method lists for different accounting types (system, exec, commands, dot1x).
+
+ Expected Results:
+ * success: The test will pass if the provided AAA accounting default method list is matching in the configured accounting types.
+ * failure: The test will fail if the provided AAA accounting default method list is NOT matching in the configured accounting types.
+ """
+
+ name="VerifyAcctDefaultMethods"
+ description="Verifies the AAA accounting default method lists for different accounting types (system, exec, commands, dot1x)."
+ categories=["aaa"]
+ commands=[AntaCommand(command="show aaa methods accounting")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ methods:List[AAAAuthMethod]
+"""List of AAA accounting methods. Methods should be in the right order"""
+ types:Set[Literal["commands","exec","system","dot1x"]]
+"""List of accounting types to verify"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ not_matching=[]
+ not_configured=[]
+ fork,vincommand_output.items():
+ acct_type=k.replace("AcctMethods","")
+ ifacct_typenotinself.inputs.types:
+ # We do not need to verify this accounting type
+ continue
+ formethodsinv.values():
+ if"defaultAction"notinmethods:
+ not_configured.append(acct_type)
+ ifmethods["defaultMethods"]!=self.inputs.methods:
+ not_matching.append(acct_type)
+ ifnot_configured:
+ self.result.is_failure(f"AAA default accounting is not configured for {not_configured}")
+ return
+ ifnotnot_matching:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"AAA accounting default methods {self.inputs.methods} are not matching for {not_matching}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ methods:List[AAAAuthMethod]
+"""List of AAA accounting methods. Methods should be in the right order"""
+ types:Set[Literal["commands","exec","system","dot1x"]]
+"""List of accounting types to verify"""
+
classVerifyAuthenMethods(AntaTest):
+"""
+ Verifies the AAA authentication method lists for different authentication types (login, enable, dot1x).
+
+ Expected Results:
+ * success: The test will pass if the provided AAA authentication method list is matching in the configured authentication types.
+ * failure: The test will fail if the provided AAA authentication method list is NOT matching in the configured authentication types.
+ """
+
+ name="VerifyAuthenMethods"
+ description="Verifies the AAA authentication method lists for different authentication types (login, enable, dot1x)."
+ categories=["aaa"]
+ commands=[AntaCommand(command="show aaa methods authentication")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ methods:List[AAAAuthMethod]
+"""List of AAA authentication methods. Methods should be in the right order"""
+ types:Set[Literal["login","enable","dot1x"]]
+"""List of authentication types to verify"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ not_matching=[]
+ fork,vincommand_output.items():
+ auth_type=k.replace("AuthenMethods","")
+ ifauth_typenotinself.inputs.types:
+ # We do not need to verify this accounting type
+ continue
+ ifauth_type=="login":
+ if"login"notinv:
+ self.result.is_failure("AAA authentication methods are not configured for login console")
+ return
+ ifv["login"]["methods"]!=self.inputs.methods:
+ self.result.is_failure(f"AAA authentication methods {self.inputs.methods} are not matching for login console")
+ return
+ formethodsinv.values():
+ ifmethods["methods"]!=self.inputs.methods:
+ not_matching.append(auth_type)
+ ifnotnot_matching:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"AAA authentication methods {self.inputs.methods} are not matching for {not_matching}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ methods:List[AAAAuthMethod]
+"""List of AAA authentication methods. Methods should be in the right order"""
+ types:Set[Literal["login","enable","dot1x"]]
+"""List of authentication types to verify"""
+
classVerifyAuthzMethods(AntaTest):
+"""
+ Verifies the AAA authorization method lists for different authorization types (commands, exec).
+
+ Expected Results:
+ * success: The test will pass if the provided AAA authorization method list is matching in the configured authorization types.
+ * failure: The test will fail if the provided AAA authorization method list is NOT matching in the configured authorization types.
+ """
+
+ name="VerifyAuthzMethods"
+ description="Verifies the AAA authorization method lists for different authorization types (commands, exec)."
+ categories=["aaa"]
+ commands=[AntaCommand(command="show aaa methods authorization")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ methods:List[AAAAuthMethod]
+"""List of AAA authorization methods. Methods should be in the right order"""
+ types:Set[Literal["commands","exec"]]
+"""List of authorization types to verify"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ not_matching=[]
+ fork,vincommand_output.items():
+ authz_type=k.replace("AuthzMethods","")
+ ifauthz_typenotinself.inputs.types:
+ # We do not need to verify this accounting type
+ continue
+ formethodsinv.values():
+ ifmethods["methods"]!=self.inputs.methods:
+ not_matching.append(authz_type)
+ ifnotnot_matching:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"AAA authorization methods {self.inputs.methods} are not matching for {not_matching}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ methods:List[AAAAuthMethod]
+"""List of AAA authorization methods. Methods should be in the right order"""
+ types:Set[Literal["commands","exec"]]
+"""List of authorization types to verify"""
+
classVerifyTacacsServerGroups(AntaTest):
+"""
+ Verifies if the provided TACACS server group(s) are configured.
+
+ Expected Results:
+ * success: The test will pass if the provided TACACS server group(s) are configured.
+ * failure: The test will fail if one or all the provided TACACS server group(s) are NOT configured.
+ """
+
+ name="VerifyTacacsServerGroups"
+ description="Verifies if the provided TACACS server group(s) are configured."
+ categories=["aaa"]
+ commands=[AntaCommand(command="show tacacs")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ groups:List[str]
+"""List of TACACS server group"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ tacacs_groups=command_output["groups"]
+ ifnottacacs_groups:
+ self.result.is_failure("No TACACS server group(s) are configured")
+ return
+ not_configured=[groupforgroupinself.inputs.groupsifgroupnotintacacs_groups]
+ ifnotnot_configured:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"TACACS server group(s) {not_configured} are not configured")
+
classVerifyTacacsServers(AntaTest):
+"""
+ Verifies TACACS servers are configured for a specified VRF.
+
+ Expected Results:
+ * success: The test will pass if the provided TACACS servers are configured in the specified VRF.
+ * failure: The test will fail if the provided TACACS servers are NOT configured in the specified VRF.
+ """
+
+ name="VerifyTacacsServers"
+ description="Verifies TACACS servers are configured for a specified VRF."
+ categories=["aaa"]
+ commands=[AntaCommand(command="show tacacs")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ servers:List[IPv4Address]
+"""List of TACACS servers"""
+ vrf:str="default"
+"""The name of the VRF to transport TACACS messages"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ tacacs_servers=command_output["tacacsServers"]
+ ifnottacacs_servers:
+ self.result.is_failure("No TACACS servers are configured")
+ return
+ not_configured=[
+ str(server)
+ forserverinself.inputs.servers
+ ifnotany(
+ str(server)==tacacs_server["serverInfo"]["hostname"]andself.inputs.vrf==tacacs_server["serverInfo"]["vrf"]fortacacs_serverintacacs_servers
+ )
+ ]
+ ifnotnot_configured:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"TACACS servers {not_configured} are not configured in VRF {self.inputs.vrf}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ servers:List[IPv4Address]
+"""List of TACACS servers"""
+ vrf:str="default"
+"""The name of the VRF to transport TACACS messages"""
+
classVerifyTacacsSourceIntf(AntaTest):
+"""
+ Verifies TACACS source-interface for a specified VRF.
+
+ Expected Results:
+ * success: The test will pass if the provided TACACS source-interface is configured in the specified VRF.
+ * failure: The test will fail if the provided TACACS source-interface is NOT configured in the specified VRF.
+ """
+
+ name="VerifyTacacsSourceIntf"
+ description="Verifies TACACS source-interface for a specified VRF."
+ categories=["aaa"]
+ commands=[AntaCommand(command="show tacacs")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ intf:str
+"""Source-interface to use as source IP of TACACS messages"""
+ vrf:str="default"
+"""The name of the VRF to transport TACACS messages"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ try:
+ ifcommand_output["srcIntf"][self.inputs.vrf]==self.inputs.intf:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Wrong source-interface configured in VRF {self.inputs.vrf}")
+ exceptKeyError:
+ self.result.is_failure(f"Source-interface {self.inputs.intf} is not configured in VRF {self.inputs.vrf}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ intf:str
+"""Source-interface to use as source IP of TACACS messages"""
+ vrf:str="default"
+"""The name of the VRF to transport TACACS messages"""
+
classVerifyRunningConfigDiffs(AntaTest):
+"""
+ Verifies there is no difference between the running-config and the startup-config
+ """
+
+ name="VerifyRunningConfigDiffs"
+ description="Verifies there is no difference between the running-config and the startup-config"
+ categories=["configuration"]
+ commands=[AntaCommand(command="show running-config diffs",ofmt="text")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].output
+ ifcommand_outputisNoneorcommand_output=="":
+ self.result.is_success()
+ else:
+ self.result.is_failure()
+ self.result.is_failure(str(command_output))
+
classVerifyLLDPNeighbors(AntaTest):
+"""
+ This test verifies that the provided LLDP neighbors are present and connected with the correct configuration.
+
+ Expected Results:
+ * success: The test will pass if each of the provided LLDP neighbors is present and connected to the specified port and device.
+ * failure: The test will fail if any of the following conditions are met:
+ - The provided LLDP neighbor is not found.
+ - The system name or port of the LLDP neighbor does not match the provided information.
+ """
+
+ name="VerifyLLDPNeighbors"
+ description="Verifies that the provided LLDP neighbors are present and connected with the correct configuration."
+ categories=["connectivity"]
+ commands=[AntaCommand(command="show lldp neighbors detail")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ neighbors:List[Neighbor]
+"""List of LLDP neighbors"""
+
+ classNeighbor(BaseModel):
+"""LLDP neighbor"""
+
+ port:Interface
+"""LLDP port"""
+ neighbor_device:str
+"""LLDP neighbor device"""
+ neighbor_port:Interface
+"""LLDP neighbor port"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+
+ failures:dict[str,list[str]]={}
+
+ forneighborinself.inputs.neighbors:
+ ifneighbor.portnotincommand_output["lldpNeighbors"]:
+ failures.setdefault("port_not_configured",[]).append(neighbor.port)
+ eliflen(lldp_neighbor_info:=command_output["lldpNeighbors"][neighbor.port]["lldpNeighborInfo"])==0:
+ failures.setdefault("no_lldp_neighbor",[]).append(neighbor.port)
+ elif(
+ lldp_neighbor_info[0]["systemName"]!=neighbor.neighbor_device
+ orlldp_neighbor_info[0]["neighborInterfaceInfo"]["interfaceId_v2"]!=neighbor.neighbor_port
+ ):
+ failures.setdefault("wrong_lldp_neighbor",[]).append(neighbor.port)
+
+ ifnotfailures:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"The following port(s) have issues: {failures}")
+
classVerifyReachability(AntaTest):
+"""
+ Test network reachability to one or many destination IP(s).
+
+ Expected Results:
+ * success: The test will pass if all destination IP(s) are reachable.
+ * failure: The test will fail if one or many destination IP(s) are unreachable.
+ """
+
+ name="VerifyReachability"
+ description="Test the network reachability to one or many destination IP(s)."
+ categories=["connectivity"]
+ commands=[AntaTemplate(template="ping vrf {vrf}{destination} source {source} repeat {repeat}")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ hosts:List[Host]
+"""List of hosts to ping"""
+
+ classHost(BaseModel):
+"""Remote host to ping"""
+
+ destination:IPv4Address
+"""IPv4 address to ping"""
+ source:Union[IPv4Address,Interface]
+"""IPv4 address source IP or Egress interface to use"""
+ vrf:str="default"
+"""VRF context"""
+ repeat:int=2
+"""Number of ping repetition (default=2)"""
+
+ defrender(self,template:AntaTemplate)->list[AntaCommand]:
+ return[template.render(destination=host.destination,source=host.source,vrf=host.vrf,repeat=host.repeat)forhostinself.inputs.hosts]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ failures=[]
+ forcommandinself.instance_commands:
+ src=command.params.get("source")
+ dst=command.params.get("destination")
+ repeat=command.params.get("repeat")
+
+ ifany(elemisNoneforelemin(src,dst,repeat)):
+ raiseAntaMissingParamException(f"A parameter is missing to execute the test for command {command}")
+
+ iff"{repeat} received"notincommand.json_output["messages"][0]:
+ failures.append((str(src),str(dst)))
+
+ ifnotfailures:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Connectivity test failed for the following source-destination pairs: {failures}")
+
Verifies the device is using an Aboot version that fix the bug discussed
+in the field notice 44 (Aboot manages system settings prior to EOS initialization).
classVerifyFieldNotice72Resolution(AntaTest):
+"""
+ Checks if the device is potentially exposed to Field Notice 72, and if the issue has been mitigated.
+
+ https://www.arista.com/en/support/advisories-notices/field-notice/17410-field-notice-0072
+ """
+
+ name="VerifyFieldNotice72Resolution"
+ description="Verifies if the device has exposeure to FN72, and if the issue has been mitigated"
+ categories=["field notices","software"]
+ commands=[AntaCommand(command="show version detail")]
+
+ # TODO maybe implement ONLY ON PLATFORMS instead
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+
+ devices=["DCS-7280SR3-48YC8","DCS-7280SR3K-48YC8"]
+ variants=["-SSD-F","-SSD-R","-M-F","-M-R","-F","-R"]
+ model=command_output["modelName"]
+
+ forvariantinvariants:
+ model=model.replace(variant,"")
+ ifmodelnotindevices:
+ self.result.is_skipped("Platform is not impacted by FN072")
+ return
+
+ serial=command_output["serialNumber"]
+ number=int(serial[3:7])
+
+ if"JPE"notinserialand"JAS"notinserial:
+ self.result.is_skipped("Device not exposed")
+ return
+
+ ifmodel=="DCS-7280SR3-48YC8"and"JPE"inserialandnumber>=2131:
+ self.result.is_skipped("Device not exposed")
+ return
+
+ ifmodel=="DCS-7280SR3-48YC8"and"JAS"inserialandnumber>=2041:
+ self.result.is_skipped("Device not exposed")
+ return
+
+ ifmodel=="DCS-7280SR3K-48YC8"and"JPE"inserialandnumber>=2134:
+ self.result.is_skipped("Device not exposed")
+ return
+
+ ifmodel=="DCS-7280SR3K-48YC8"and"JAS"inserialandnumber>=2041:
+ self.result.is_skipped("Device not exposed")
+ return
+
+ # Because each of the if checks above will return if taken, we only run the long
+ # check if we get this far
+ forentryincommand_output["details"]["components"]:
+ ifentry["name"]=="FixedSystemvrm1":
+ ifint(entry["version"])<7:
+ self.result.is_failure("Device is exposed to FN72")
+ else:
+ self.result.is_success("FN72 is mitigated")
+ return
+ # We should never hit this point
+ self.result.is_error(message="Error in running test - FixedSystemvrm1 not found")
+ return
+
classVerifyAdverseDrops(AntaTest):
+"""
+ This test verifies if there are no adverse drops on DCS7280E and DCS7500E.
+
+ Expected Results:
+ * success: The test will pass if there are no adverse drops.
+ * failure: The test will fail if there are adverse drops.
+ """
+
+ name="VerifyAdverseDrops"
+ description="Verifies there are no adverse drops on DCS7280E and DCS7500E"
+ categories=["hardware"]
+ commands=[AntaCommand(command="show hardware counter drop",ofmt="json")]
+
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ total_adverse_drop=command_output["totalAdverseDrops"]if"totalAdverseDrops"incommand_output.keys()else""
+ iftotal_adverse_drop==0:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Device totalAdverseDrops counter is: '{total_adverse_drop}'")
+
classVerifyEnvironmentCooling(AntaTest):
+"""
+ This test verifies the fans status.
+
+ Expected Results:
+ * success: The test will pass if the fans status are within the accepted states list.
+ * failure: The test will fail if some fans status is not within the accepted states list.
+ """
+
+ name="VerifyEnvironmentCooling"
+ description="Verifies if the fans status are within the accepted states list."
+ categories=["hardware"]
+ commands=[AntaCommand(command="show system environment cooling",ofmt="json")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ states:List[str]
+"""Accepted states list for fan status"""
+
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ self.result.is_success()
+ # First go through power supplies fans
+ forpower_supplyincommand_output.get("powerSupplySlots",[]):
+ forfaninpower_supply.get("fans",[]):
+ if(state:=fan["status"])notinself.inputs.states:
+ self.result.is_failure(f"Fan {fan['label']} on PowerSupply {power_supply['label']} is: '{state}'")
+ # Then go through fan trays
+ forfan_trayincommand_output.get("fanTraySlots",[]):
+ forfaninfan_tray.get("fans",[]):
+ if(state:=fan["status"])notinself.inputs.states:
+ self.result.is_failure(f"Fan {fan['label']} on Fan Tray {fan_tray['label']} is: '{state}'")
+
classVerifyEnvironmentPower(AntaTest):
+"""
+ This test verifies the power supplies status.
+
+ Expected Results:
+ * success: The test will pass if the power supplies status are within the accepted states list.
+ * failure: The test will fail if some power supplies status is not within the accepted states list.
+ """
+
+ name="VerifyEnvironmentPower"
+ description="Verifies if the power supplies status are within the accepted states list."
+ categories=["hardware"]
+ commands=[AntaCommand(command="show system environment power",ofmt="json")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ states:List[str]
+"""Accepted states list for power supplies status"""
+
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ power_supplies=command_output["powerSupplies"]if"powerSupplies"incommand_output.keys()else"{}"
+ wrong_power_supplies={
+ powersupply:{"state":value["state"]}forpowersupply,valueindict(power_supplies).items()ifvalue["state"]notinself.inputs.states
+ }
+ ifnotwrong_power_supplies:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"The following power supplies status are not in the accepted states list: {wrong_power_supplies}")
+
classVerifyEnvironmentSystemCooling(AntaTest):
+"""
+ This test verifies the device's system cooling.
+
+ Expected Results:
+ * success: The test will pass if the system cooling status is OK: 'coolingOk'.
+ * failure: The test will fail if the system cooling status is NOT OK.
+ """
+
+ name="VerifyEnvironmentSystemCooling"
+ description="Verifies the system cooling status."
+ categories=["hardware"]
+ commands=[AntaCommand(command="show system environment cooling",ofmt="json")]
+
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ sys_status=command_output["systemStatus"]if"systemStatus"incommand_output.keys()else""
+ self.result.is_success()
+ ifsys_status!="coolingOk":
+ self.result.is_failure(f"Device system cooling is not OK: '{sys_status}'")
+
classVerifyTemperature(AntaTest):
+"""
+ This test verifies if the device temperature is within acceptable limits.
+
+ Expected Results:
+ * success: The test will pass if the device temperature is currently OK: 'temperatureOk'.
+ * failure: The test will fail if the device temperature is NOT OK.
+ """
+
+ name="VerifyTemperature"
+ description="Verifies if the device temperature is within the acceptable range."
+ categories=["hardware"]
+ commands=[AntaCommand(command="show system environment temperature",ofmt="json")]
+
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ temperature_status=command_output["systemStatus"]if"systemStatus"incommand_output.keys()else""
+ iftemperature_status=="temperatureOk":
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Device temperature exceeds acceptable limits. Current system status: '{temperature_status}'")
+
classVerifyTransceiversManufacturers(AntaTest):
+"""
+ This test verifies if all the transceivers come from approved manufacturers.
+
+ Expected Results:
+ * success: The test will pass if all transceivers are from approved manufacturers.
+ * failure: The test will fail if some transceivers are from unapproved manufacturers.
+ """
+
+ name="VerifyTransceiversManufacturers"
+ description="Verifies the transceiver's manufacturer against a list of approved manufacturers."
+ categories=["hardware"]
+ commands=[AntaCommand(command="show inventory",ofmt="json")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ manufacturers:List[str]
+"""List of approved transceivers manufacturers"""
+
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ wrong_manufacturers={
+ interface:value["mfgName"]forinterface,valueincommand_output["xcvrSlots"].items()ifvalue["mfgName"]notinself.inputs.manufacturers
+ }
+ ifnotwrong_manufacturers:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Some transceivers are from unapproved manufacturers: {wrong_manufacturers}")
+
classVerifyTransceiversTemperature(AntaTest):
+"""
+ This test verifies if all the transceivers are operating at an acceptable temperature.
+
+ Expected Results:
+ * success: The test will pass if all transceivers status are OK: 'ok'.
+ * failure: The test will fail if some transceivers are NOT OK.
+ """
+
+ name="VerifyTransceiversTemperature"
+ description="Verifies that all transceivers are operating within the acceptable temperature range."
+ categories=["hardware"]
+ commands=[AntaCommand(command="show system environment temperature transceiver",ofmt="json")]
+
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ sensors=command_output["tempSensors"]if"tempSensors"incommand_output.keys()else""
+ wrong_sensors={
+ sensor["name"]:{
+ "hwStatus":sensor["hwStatus"],
+ "alertCount":sensor["alertCount"],
+ }
+ forsensorinsensors
+ ifsensor["hwStatus"]!="ok"orsensor["alertCount"]!=0
+ }
+ ifnotwrong_sensors:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"The following sensors are operating outside the acceptable temperature range or have raised alerts: {wrong_sensors}")
+
classVerifyIPProxyARP(AntaTest):
+"""
+ Verifies if Proxy-ARP is enabled for the provided list of interface(s).
+
+ Expected Results:
+ * success: The test will pass if Proxy-ARP is enabled on the specified interface(s).
+ * failure: The test will fail if Proxy-ARP is disabled on the specified interface(s).
+ """
+
+ name="VerifyIPProxyARP"
+ description="Verifies if Proxy-ARP is enabled for the provided list of interface(s)."
+ categories=["interfaces"]
+ commands=[AntaTemplate(template="show ip interface {intf}")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ interfaces:List[str]
+"""list of interfaces to be tested"""
+
+ defrender(self,template:AntaTemplate)->list[AntaCommand]:
+ return[template.render(intf=intf)forintfinself.inputs.interfaces]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ disabled_intf=[]
+ forcommandinself.instance_commands:
+ if"intf"incommand.params:
+ intf=command.params["intf"]
+ ifnotcommand.json_output["interfaces"][intf]["proxyArp"]:
+ disabled_intf.append(intf)
+ ifdisabled_intf:
+ self.result.is_failure(f"The following interface(s) have Proxy-ARP disabled: {disabled_intf}")
+ else:
+ self.result.is_success()
+
classVerifyInterfaceErrors(AntaTest):
+"""
+ This test verifies that interfaces error counters are equal to zero.
+
+ Expected Results:
+ * success: The test will pass if all interfaces have error counters equal to zero.
+ * failure: The test will fail if one or more interfaces have non-zero error counters.
+ """
+
+ name="VerifyInterfaceErrors"
+ description="Verifies that interfaces error counters are equal to zero."
+ categories=["interfaces"]
+ commands=[AntaCommand(command="show interfaces counters errors")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ wrong_interfaces:list[dict[str,dict[str,int]]]=[]
+ forinterface,countersincommand_output["interfaceErrorCounters"].items():
+ ifany(value>0forvalueincounters.values())andall(interfacenotinwrong_interfaceforwrong_interfaceinwrong_interfaces):
+ wrong_interfaces.append({interface:counters})
+ ifnotwrong_interfaces:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"The following interface(s) have non-zero error counters: {wrong_interfaces}")
+
classVerifyInterfacesStatus(AntaTest):
+"""
+ This test verifies if the provided list of interfaces are all in the expected state.
+
+ Expected Results:
+ * success: The test will pass if the provided interfaces are all in the expected state.
+ * failure: The test will fail if any interface is not in the expected state.
+ """
+
+ name="VerifyInterfacesStatus"
+ description="Verifies if the provided list of interfaces are all in the expected state."
+ categories=["interfaces"]
+ commands=[AntaCommand(command="show interfaces description")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ interfaces:List[InterfaceStatus]
+"""List of interfaces to validate with the expected state"""
+
+ classInterfaceStatus(BaseModel):# pylint: disable=missing-class-docstring
+ interface:Interface
+ state:Literal["up","adminDown"]
+ protocol_status:Literal["up","down"]="up"
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+
+ self.result.is_success()
+
+ intf_not_configured=[]
+ intf_wrong_state=[]
+
+ forinterface_statusinself.inputs.interfaces:
+ intf_status=get_value(command_output["interfaceDescriptions"],interface_status.interface,separator=";")
+ ifintf_statusisNone:
+ intf_not_configured.append(interface_status.interface)
+ continue
+
+ proto=intf_status["lineProtocolStatus"]
+ status=intf_status["interfaceStatus"]
+
+ ifinterface_status.state=="up"andnot(re.match(r"connected|up",proto)andre.match(r"connected|up",status)):
+ intf_wrong_state.append(f"{interface_status.interface} is {proto}/{status} expected {interface_status.protocol_status}/{interface_status.state}")
+ elifinterface_status.state=="adminDown":
+ ifinterface_status.protocol_status=="up"andnot(re.match(r"up",proto)andre.match(r"adminDown",status)):
+ intf_wrong_state.append(f"{interface_status.interface} is {proto}/{status} expected {interface_status.protocol_status}/{interface_status.state}")
+ elifinterface_status.protocol_status=="down"andnot(re.match(r"down",proto)andre.match(r"adminDown",status)):
+ intf_wrong_state.append(f"{interface_status.interface} is {proto}/{status} expected {interface_status.protocol_status}/{interface_status.state}")
+
+ ifintf_not_configured:
+ self.result.is_failure(f"The following interface(s) are not configured: {intf_not_configured}")
+
+ ifintf_wrong_state:
+ self.result.is_failure(f"The following interface(s) are not in the expected state: {intf_wrong_state}")
+
Verifies the global layer 2 Maximum Transfer Unit (MTU) for all L2 interfaces.
+
Test that L2 interfaces are configured with the correct MTU. It supports Ethernet, Port Channel and VLAN interfaces.
+You can define a global MTU to check and also an MTU per interface and also ignored some interfaces.
+
+
+ Expected Results
+
+
success: The test will pass if all layer 2 interfaces have the proper MTU configured.
+
failure: The test will fail if one or many layer 2 interfaces have the wrong MTU configured.
classVerifyL2MTU(AntaTest):
+"""
+ Verifies the global layer 2 Maximum Transfer Unit (MTU) for all L2 interfaces.
+
+ Test that L2 interfaces are configured with the correct MTU. It supports Ethernet, Port Channel and VLAN interfaces.
+ You can define a global MTU to check and also an MTU per interface and also ignored some interfaces.
+
+ Expected Results:
+ * success: The test will pass if all layer 2 interfaces have the proper MTU configured.
+ * failure: The test will fail if one or many layer 2 interfaces have the wrong MTU configured.
+ """
+
+ name="VerifyL2MTU"
+ description="Verifies the global layer 2 Maximum Transfer Unit (MTU) for all layer 2 interfaces."
+ categories=["interfaces"]
+ commands=[AntaCommand(command="show interfaces")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ mtu:int=9214
+"""Default MTU we should have configured on all non-excluded interfaces"""
+ ignored_interfaces:List[str]=["Management","Loopback","Vxlan","Tunnel"]
+"""A list of L2 interfaces to ignore"""
+ specific_mtu:List[Dict[str,int]]=[]
+"""A list of dictionary of L2 interfaces with their specific MTU configured"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ # Parameter to save incorrect interface settings
+ wrong_l2mtu_intf:list[dict[str,int]]=[]
+ command_output=self.instance_commands[0].json_output
+ # Set list of interfaces with specific settings
+ specific_interfaces:list[str]=[]
+ ifself.inputs.specific_mtu:
+ fordinself.inputs.specific_mtu:
+ specific_interfaces.extend(d)
+ forinterface,valuesincommand_output["interfaces"].items():
+ ifre.findall(r"[a-z]+",interface,re.IGNORECASE)[0]notinself.inputs.ignored_interfacesandvalues["forwardingModel"]=="bridged":
+ ifinterfaceinspecific_interfaces:
+ wrong_l2mtu_intf.extend({interface:values["mtu"]}forcustom_datainself.inputs.specific_mtuifvalues["mtu"]!=custom_data[interface])
+ # Comparison with generic setting
+ elifvalues["mtu"]!=self.inputs.mtu:
+ wrong_l2mtu_intf.append({interface:values["mtu"]})
+ ifwrong_l2mtu_intf:
+ self.result.is_failure(f"Some L2 interfaces do not have correct MTU configured:\n{wrong_l2mtu_intf}")
+ else:
+ self.result.is_success()
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ mtu:int=9214
+"""Default MTU we should have configured on all non-excluded interfaces"""
+ ignored_interfaces:List[str]=["Management","Loopback","Vxlan","Tunnel"]
+"""A list of L2 interfaces to ignore"""
+ specific_mtu:List[Dict[str,int]]=[]
+"""A list of dictionary of L2 interfaces with their specific MTU configured"""
+
Verifies the global layer 3 Maximum Transfer Unit (MTU) for all L3 interfaces.
+
Test that L3 interfaces are configured with the correct MTU. It supports Ethernet, Port Channel and VLAN interfaces.
+You can define a global MTU to check and also an MTU per interface and also ignored some interfaces.
+
+
+ Expected Results
+
+
success: The test will pass if all layer 3 interfaces have the proper MTU configured.
+
failure: The test will fail if one or many layer 3 interfaces have the wrong MTU configured.
classVerifyL3MTU(AntaTest):
+"""
+ Verifies the global layer 3 Maximum Transfer Unit (MTU) for all L3 interfaces.
+
+ Test that L3 interfaces are configured with the correct MTU. It supports Ethernet, Port Channel and VLAN interfaces.
+ You can define a global MTU to check and also an MTU per interface and also ignored some interfaces.
+
+ Expected Results:
+ * success: The test will pass if all layer 3 interfaces have the proper MTU configured.
+ * failure: The test will fail if one or many layer 3 interfaces have the wrong MTU configured.
+ """
+
+ name="VerifyL3MTU"
+ description="Verifies the global layer 3 Maximum Transfer Unit (MTU) for all layer 3 interfaces."
+ categories=["interfaces"]
+ commands=[AntaCommand(command="show interfaces")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ mtu:int=1500
+"""Default MTU we should have configured on all non-excluded interfaces"""
+ ignored_interfaces:List[str]=["Management","Loopback","Vxlan","Tunnel"]
+"""A list of L3 interfaces to ignore"""
+ specific_mtu:List[Dict[str,int]]=[]
+"""A list of dictionary of L3 interfaces with their specific MTU configured"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ # Parameter to save incorrect interface settings
+ wrong_l3mtu_intf:list[dict[str,int]]=[]
+ command_output=self.instance_commands[0].json_output
+ # Set list of interfaces with specific settings
+ specific_interfaces:list[str]=[]
+ ifself.inputs.specific_mtu:
+ fordinself.inputs.specific_mtu:
+ specific_interfaces.extend(d)
+ forinterface,valuesincommand_output["interfaces"].items():
+ ifre.findall(r"[a-z]+",interface,re.IGNORECASE)[0]notinself.inputs.ignored_interfacesandvalues["forwardingModel"]=="routed":
+ ifinterfaceinspecific_interfaces:
+ wrong_l3mtu_intf.extend({interface:values["mtu"]}forcustom_datainself.inputs.specific_mtuifvalues["mtu"]!=custom_data[interface])
+ # Comparison with generic setting
+ elifvalues["mtu"]!=self.inputs.mtu:
+ wrong_l3mtu_intf.append({interface:values["mtu"]})
+ ifwrong_l3mtu_intf:
+ self.result.is_failure(f"Some interfaces do not have correct MTU configured:\n{wrong_l3mtu_intf}")
+ else:
+ self.result.is_success()
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ mtu:int=1500
+"""Default MTU we should have configured on all non-excluded interfaces"""
+ ignored_interfaces:List[str]=["Management","Loopback","Vxlan","Tunnel"]
+"""A list of L3 interfaces to ignore"""
+ specific_mtu:List[Dict[str,int]]=[]
+"""A list of dictionary of L3 interfaces with their specific MTU configured"""
+
classVerifyLoopbackCount(AntaTest):
+"""
+ Verifies the number of loopback interfaces on the device is the one we expect and if none of the loopback is down.
+ """
+
+ name="VerifyLoopbackCount"
+ description="Verifies the number of loopback interfaces on the device is the one we expect and if none of the loopback is down."
+ categories=["interfaces"]
+ commands=[AntaCommand(command="show ip interface brief")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type: ignore
+"""Number of loopback interfaces expected to be present"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ loopback_count=0
+ down_loopback_interfaces=[]
+ forinterfaceincommand_output["interfaces"]:
+ interface_dict=command_output["interfaces"][interface]
+ if"Loopback"ininterface:
+ loopback_count+=1
+ ifnot(interface_dict["lineProtocolStatus"]=="up"andinterface_dict["interfaceStatus"]=="connected"):
+ down_loopback_interfaces.append(interface)
+ ifloopback_count==self.inputs.numberandlen(down_loopback_interfaces)==0:
+ self.result.is_success()
+ else:
+ self.result.is_failure()
+ ifloopback_count!=self.inputs.number:
+ self.result.is_failure(f"Found {loopback_count} Loopbacks when expecting {self.inputs.number}")
+ eliflen(down_loopback_interfaces)!=0:
+ self.result.is_failure(f"The following Loopbacks are not up: {down_loopback_interfaces}")
+
classVerifyPortChannels(AntaTest):
+"""
+ Verifies there is no inactive port in port channels.
+ """
+
+ name="VerifyPortChannels"
+ description="Verifies there is no inactive port in port channels."
+ categories=["interfaces"]
+ commands=[AntaCommand(command="show port-channel")]
+
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ po_with_invactive_ports:list[dict[str,str]]=[]
+ forportchannel,portchannel_dictincommand_output["portChannels"].items():
+ iflen(portchannel_dict["inactivePorts"])!=0:
+ po_with_invactive_ports.extend({portchannel:portchannel_dict["inactivePorts"]})
+ ifnotpo_with_invactive_ports:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"The following port-channels have inactive port(s): {po_with_invactive_ports}")
+
classVerifyStormControlDrops(AntaTest):
+"""
+ Verifies the device did not drop packets due its to storm-control configuration.
+ """
+
+ name="VerifyStormControlDrops"
+ description="Verifies the device did not drop packets due its to storm-control configuration."
+ categories=["interfaces"]
+ commands=[AntaCommand(command="show storm-control")]
+
+ @skip_on_platforms(["cEOSLab","vEOS-lab"])
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ storm_controlled_interfaces:dict[str,dict[str,Any]]={}
+ forinterface,interface_dictincommand_output["interfaces"].items():
+ fortraffic_type,traffic_type_dictininterface_dict["trafficTypes"].items():
+ if"drop"intraffic_type_dictandtraffic_type_dict["drop"]!=0:
+ storm_controlled_interface_dict=storm_controlled_interfaces.setdefault(interface,{})
+ storm_controlled_interface_dict.update({traffic_type:traffic_type_dict["drop"]})
+ ifnotstorm_controlled_interfaces:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"The following interfaces have none 0 storm-control drop counters {storm_controlled_interfaces}")
+
classVerifyLoggingErrors(AntaTest):
+"""
+ This test verifies there are no syslog messages with a severity of ERRORS or higher.
+
+ Expected Results:
+ * success: The test will pass if there are NO syslog messages with a severity of ERRORS or higher.
+ * failure: The test will fail if ERRORS or higher syslog messages are present.
+ """
+
+ name="VerifyLoggingWarning"
+ description="This test verifies there are no syslog messages with a severity of ERRORS or higher."
+ categories=["logging"]
+ commands=[AntaCommand(command="show logging threshold errors",ofmt="text")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+"""
+ Run VerifyLoggingWarning validation
+ """
+ command_output=self.instance_commands[0].text_output
+
+ iflen(command_output)==0:
+ self.result.is_success()
+ else:
+ self.result.is_failure("Device has reported syslog messages with a severity of ERRORS or higher")
+
classVerifyLoggingHostname(AntaTest):
+"""
+ Verifies if logs are generated with the device FQDN.
+
+ Expected Results:
+ * success: The test will pass if logs are generated with the device FQDN.
+ * failure: The test will fail if logs are NOT generated with the device FQDN.
+ """
+
+ name="VerifyLoggingHostname"
+ description="Verifies if logs are generated with the device FQDN."
+ categories=["logging"]
+ commands=[
+ AntaCommand(command="show hostname"),
+ AntaCommand(command="send log level informational message ANTA VerifyLoggingHostname validation"),
+ AntaCommand(command="show logging informational last 30 seconds | grep ANTA",ofmt="text",use_cache=False),
+ ]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ output_hostname=self.instance_commands[0].json_output
+ output_logging=self.instance_commands[2].text_output
+ fqdn=output_hostname["fqdn"]
+ lines=output_logging.strip().split("\n")[::-1]
+ log_pattern=r"ANTA VerifyLoggingHostname validation"
+ last_line_with_pattern=""
+ forlineinlines:
+ ifre.search(log_pattern,line):
+ last_line_with_pattern=line
+ break
+ iffqdninlast_line_with_pattern:
+ self.result.is_success()
+ else:
+ self.result.is_failure("Logs are not generated with the device FQDN")
+
classVerifyLoggingHosts(AntaTest):
+"""
+ Verifies logging hosts (syslog servers) for a specified VRF.
+
+ Expected Results:
+ * success: The test will pass if the provided syslog servers are configured in the specified VRF.
+ * failure: The test will fail if the provided syslog servers are NOT configured in the specified VRF.
+ """
+
+ name="VerifyLoggingHosts"
+ description="Verifies logging hosts (syslog servers) for a specified VRF."
+ categories=["logging"]
+ commands=[AntaCommand(command="show logging",ofmt="text")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ hosts:List[IPv4Address]
+"""List of hosts (syslog servers) IP addresses"""
+ vrf:str="default"
+"""The name of the VRF to transport log messages"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ output=self.instance_commands[0].text_output
+ not_configured=[]
+ forhostinself.inputs.hosts:
+ pattern=rf"Logging to '{str(host)}'.*VRF {self.inputs.vrf}"
+ ifnotre.search(pattern,_get_logging_states(self.logger,output)):
+ not_configured.append(str(host))
+
+ ifnotnot_configured:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Syslog servers {not_configured} are not configured in VRF {self.inputs.vrf}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ hosts:List[IPv4Address]
+"""List of hosts (syslog servers) IP addresses"""
+ vrf:str="default"
+"""The name of the VRF to transport log messages"""
+
classVerifyLoggingPersistent(AntaTest):
+"""
+ Verifies if logging persistent is enabled and logs are saved in flash.
+
+ Expected Results:
+ * success: The test will pass if logging persistent is enabled and logs are in flash.
+ * failure: The test will fail if logging persistent is disabled or no logs are saved in flash.
+ """
+
+ name="VerifyLoggingPersistent"
+ description="Verifies if logging persistent is enabled and logs are saved in flash."
+ categories=["logging"]
+ commands=[
+ AntaCommand(command="show logging",ofmt="text"),
+ AntaCommand(command="dir flash:/persist/messages",ofmt="text"),
+ ]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ self.result.is_success()
+ log_output=self.instance_commands[0].text_output
+ dir_flash_output=self.instance_commands[1].text_output
+ if"Persistent logging: disabled"in_get_logging_states(self.logger,log_output):
+ self.result.is_failure("Persistent logging is disabled")
+ return
+ pattern=r"-rw-\s+(\d+)"
+ persist_logs=re.search(pattern,dir_flash_output)
+ ifnotpersist_logsorint(persist_logs.group(1))==0:
+ self.result.is_failure("No persistent logs are saved in flash")
+
classVerifyLoggingSourceIntf(AntaTest):
+"""
+ Verifies logging source-interface for a specified VRF.
+
+ Expected Results:
+ * success: The test will pass if the provided logging source-interface is configured in the specified VRF.
+ * failure: The test will fail if the provided logging source-interface is NOT configured in the specified VRF.
+ """
+
+ name="VerifyLoggingSourceInt"
+ description="Verifies logging source-interface for a specified VRF."
+ categories=["logging"]
+ commands=[AntaCommand(command="show logging",ofmt="text")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ interface:str
+"""Source-interface to use as source IP of log messages"""
+ vrf:str="default"
+"""The name of the VRF to transport log messages"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ output=self.instance_commands[0].text_output
+ pattern=rf"Logging source-interface '{self.inputs.interface}'.*VRF {self.inputs.vrf}"
+ ifre.search(pattern,_get_logging_states(self.logger,output)):
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Source-interface '{self.inputs.interface}' is not configured in VRF {self.inputs.vrf}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ interface:str
+"""Source-interface to use as source IP of log messages"""
+ vrf:str="default"
+"""The name of the VRF to transport log messages"""
+
classVerifyLoggingTimestamp(AntaTest):
+"""
+ Verifies if logs are generated with the approprate timestamp.
+
+ Expected Results:
+ * success: The test will pass if logs are generated with the appropriated timestamp.
+ * failure: The test will fail if logs are NOT generated with the appropriated timestamp.
+ """
+
+ name="VerifyLoggingTimestamp"
+ description="Verifies if logs are generated with the appropriate timestamp."
+ categories=["logging"]
+ commands=[
+ AntaCommand(command="send log level informational message ANTA VerifyLoggingTimestamp validation"),
+ AntaCommand(command="show logging informational last 30 seconds | grep ANTA",ofmt="text",use_cache=False),
+ ]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ log_pattern=r"ANTA VerifyLoggingTimestamp validation"
+ timestamp_pattern=r"\d{4}-\d{2}-\d{2}T\d{2}:\d{2}:\d{2}\.\d{6}-\d{2}:\d{2}"
+ output=self.instance_commands[1].text_output
+ lines=output.strip().split("\n")[::-1]
+ last_line_with_pattern=""
+ forlineinlines:
+ ifre.search(log_pattern,line):
+ last_line_with_pattern=line
+ break
+ ifre.search(timestamp_pattern,last_line_with_pattern):
+ self.result.is_success()
+ else:
+ self.result.is_failure("Logs are not generated with the appropriate timestamp format")
+
classVerifyMlagConfigSanity(AntaTest):
+"""
+ This test verifies there are no MLAG config-sanity inconsistencies.
+
+ Expected Results:
+ * success: The test will pass if there are NO MLAG config-sanity inconsistencies.
+ * failure: The test will fail if there are MLAG config-sanity inconsistencies.
+ * skipped: The test will be skipped if MLAG is 'disabled'.
+ * error: The test will give an error if 'mlagActive' is not found in the JSON response.
+ """
+
+ name="VerifyMlagConfigSanity"
+ description="This test verifies there are no MLAG config-sanity inconsistencies."
+ categories=["mlag"]
+ commands=[AntaCommand(command="show mlag config-sanity",ofmt="json")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ if(mlag_status:=get_value(command_output,"mlagActive"))isNone:
+ self.result.is_error(message="Incorrect JSON response - 'mlagActive' state was not found")
+ return
+ ifmlag_statusisFalse:
+ self.result.is_skipped("MLAG is disabled")
+ return
+ keys_to_verify=["globalConfiguration","interfaceConfiguration"]
+ verified_output={key:get_value(command_output,key)forkeyinkeys_to_verify}
+ ifnotany(verified_output.values()):
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"MLAG config-sanity returned inconsistencies: {verified_output}")
+
classVerifyMlagDualPrimary(AntaTest):
+"""
+ This test verifies the dual-primary detection and its parameters of the MLAG configuration.
+
+ Expected Results:
+ * success: The test will pass if the dual-primary detection is enabled and its parameters are configured properly.
+ * failure: The test will fail if the dual-primary detection is NOT enabled or its parameters are NOT configured properly.
+ * skipped: The test will be skipped if MLAG is 'disabled'.
+ """
+
+ name="VerifyMlagDualPrimary"
+ description="This test verifies the dual-primary detection and its parameters of the MLAG configuration."
+ categories=["mlag"]
+ commands=[AntaCommand(command="show mlag detail",ofmt="json")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ detection_delay:conint(ge=0)# type: ignore
+"""Delay detection (seconds)"""
+ errdisabled:bool=False
+"""Errdisabled all interfaces when dual-primary is detected"""
+ recovery_delay:conint(ge=0)# type: ignore
+"""Delay (seconds) after dual-primary detection resolves until non peer-link ports that are part of an MLAG are enabled"""
+ recovery_delay_non_mlag:conint(ge=0)# type: ignore
+"""Delay (seconds) after dual-primary detection resolves until ports that are not part of an MLAG are enabled"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ errdisabled_action="errdisableAllInterfaces"ifself.inputs.errdisabledelse"none"
+ command_output=self.instance_commands[0].json_output
+ ifcommand_output["state"]=="disabled":
+ self.result.is_skipped("MLAG is disabled")
+ return
+ ifcommand_output["dualPrimaryDetectionState"]=="disabled":
+ self.result.is_failure("Dual-primary detection is disabled")
+ return
+ keys_to_verify=["detail.dualPrimaryDetectionDelay","detail.dualPrimaryAction","dualPrimaryMlagRecoveryDelay","dualPrimaryNonMlagRecoveryDelay"]
+ verified_output={key:get_value(command_output,key)forkeyinkeys_to_verify}
+ if(
+ verified_output["detail.dualPrimaryDetectionDelay"]==self.inputs.detection_delay
+ andverified_output["detail.dualPrimaryAction"]==errdisabled_action
+ andverified_output["dualPrimaryMlagRecoveryDelay"]==self.inputs.recovery_delay
+ andverified_output["dualPrimaryNonMlagRecoveryDelay"]==self.inputs.recovery_delay_non_mlag
+ ):
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"The dual-primary parameters are not configured properly: {verified_output}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ detection_delay:conint(ge=0)# type: ignore
+"""Delay detection (seconds)"""
+ errdisabled:bool=False
+"""Errdisabled all interfaces when dual-primary is detected"""
+ recovery_delay:conint(ge=0)# type: ignore
+"""Delay (seconds) after dual-primary detection resolves until non peer-link ports that are part of an MLAG are enabled"""
+ recovery_delay_non_mlag:conint(ge=0)# type: ignore
+"""Delay (seconds) after dual-primary detection resolves until ports that are not part of an MLAG are enabled"""
+
classVerifyMlagInterfaces(AntaTest):
+"""
+ This test verifies there are no inactive or active-partial MLAG ports.
+
+ Expected Results:
+ * success: The test will pass if there are NO inactive or active-partial MLAG ports.
+ * failure: The test will fail if there are inactive or active-partial MLAG ports.
+ * skipped: The test will be skipped if MLAG is 'disabled'.
+ """
+
+ name="VerifyMlagInterfaces"
+ description="This test verifies there are no inactive or active-partial MLAG ports."
+ categories=["mlag"]
+ commands=[AntaCommand(command="show mlag",ofmt="json")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ifcommand_output["state"]=="disabled":
+ self.result.is_skipped("MLAG is disabled")
+ return
+ ifcommand_output["mlagPorts"]["Inactive"]==0andcommand_output["mlagPorts"]["Active-partial"]==0:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"MLAG status is not OK: {command_output['mlagPorts']}")
+
classVerifyMlagReloadDelay(AntaTest):
+"""
+ This test verifies the reload-delay parameters of the MLAG configuration.
+
+ Expected Results:
+ * success: The test will pass if the reload-delay parameters are configured properly.
+ * failure: The test will fail if the reload-delay parameters are NOT configured properly.
+ * skipped: The test will be skipped if MLAG is 'disabled'.
+ """
+
+ name="VerifyMlagReloadDelay"
+ description="This test verifies the reload-delay parameters of the MLAG configuration."
+ categories=["mlag"]
+ commands=[AntaCommand(command="show mlag",ofmt="json")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ reload_delay:conint(ge=0)# type: ignore
+"""Delay (seconds) after reboot until non peer-link ports that are part of an MLAG are enabled"""
+ reload_delay_non_mlag:conint(ge=0)# type: ignore
+"""Delay (seconds) after reboot until ports that are not part of an MLAG are enabled"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ifcommand_output["state"]=="disabled":
+ self.result.is_skipped("MLAG is disabled")
+ return
+ keys_to_verify=["reloadDelay","reloadDelayNonMlag"]
+ verified_output={key:get_value(command_output,key)forkeyinkeys_to_verify}
+ ifverified_output["reloadDelay"]==self.inputs.reload_delayandverified_output["reloadDelayNonMlag"]==self.inputs.reload_delay_non_mlag:
+ self.result.is_success()
+
+ else:
+ self.result.is_failure(f"The reload-delay parameters are not configured properly: {verified_output}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ reload_delay:conint(ge=0)# type: ignore
+"""Delay (seconds) after reboot until non peer-link ports that are part of an MLAG are enabled"""
+ reload_delay_non_mlag:conint(ge=0)# type: ignore
+"""Delay (seconds) after reboot until ports that are not part of an MLAG are enabled"""
+
This test verifies the health status of the MLAG configuration.
+
+
+ Expected Results
+
+
success: The test will pass if the MLAG state is ‘active’, negotiation status is ‘connected’,
+ peer-link status and local interface status are ‘up’.
+
failure: The test will fail if the MLAG state is not ‘active’, negotiation status is not ‘connected’,
+ peer-link status or local interface status are not ‘up’.
+
skipped: The test will be skipped if MLAG is ‘disabled’.
classVerifyMlagStatus(AntaTest):
+"""
+ This test verifies the health status of the MLAG configuration.
+
+ Expected Results:
+ * success: The test will pass if the MLAG state is 'active', negotiation status is 'connected',
+ peer-link status and local interface status are 'up'.
+ * failure: The test will fail if the MLAG state is not 'active', negotiation status is not 'connected',
+ peer-link status or local interface status are not 'up'.
+ * skipped: The test will be skipped if MLAG is 'disabled'.
+ """
+
+ name="VerifyMlagStatus"
+ description="This test verifies the health status of the MLAG configuration."
+ categories=["mlag"]
+ commands=[AntaCommand(command="show mlag",ofmt="json")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ifcommand_output["state"]=="disabled":
+ self.result.is_skipped("MLAG is disabled")
+ return
+ keys_to_verify=["state","negStatus","localIntfStatus","peerLinkStatus"]
+ verified_output={key:get_value(command_output,key)forkeyinkeys_to_verify}
+ if(
+ verified_output["state"]=="active"
+ andverified_output["negStatus"]=="connected"
+ andverified_output["localIntfStatus"]=="up"
+ andverified_output["peerLinkStatus"]=="up"
+ ):
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"MLAG status is not OK: {verified_output}")
+
This test verifies the count of BGP peers for a given address family.
+
It supports multiple types of address families (AFI) and subsequent service families (SAFI).
+Please refer to the Input class attributes below for details.
+
+
+ Expected Results
+
+
success: If the count of BGP peers matches the expected count for each address family and VRF.
+
failure: If the count of BGP peers does not match the expected count, or if BGP is not configured for an expected VRF or address family.
classVerifyBGPPeerCount(AntaTest):
+"""
+ This test verifies the count of BGP peers for a given address family.
+
+ It supports multiple types of address families (AFI) and subsequent service families (SAFI).
+ Please refer to the Input class attributes below for details.
+
+ Expected Results:
+ * success: If the count of BGP peers matches the expected count for each address family and VRF.
+ * failure: If the count of BGP peers does not match the expected count, or if BGP is not configured for an expected VRF or address family.
+ """
+
+ name="VerifyBGPPeerCount"
+ description="Verifies the count of BGP peers."
+ categories=["routing","bgp"]
+ commands=[
+ AntaTemplate(template="show bgp {afi}{safi} summary vrf {vrf}"),
+ AntaTemplate(template="show bgp {afi} summary"),
+ ]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ address_families:List[BgpAfi]
+"""
+ List of BGP address families (BgpAfi)
+ """
+
+ classBgpAfi(BaseModel):# pylint: disable=missing-class-docstring
+ afi:Afi
+"""BGP address family (AFI)"""
+ safi:Optional[Safi]=None
+"""Optional BGP subsequent service family (SAFI).
+
+ If the input `afi` is `ipv4` or `ipv6`, a valid `safi` must be provided.
+ """
+ vrf:str="default"
+"""
+ Optional VRF for IPv4 and IPv6. If not provided, it defaults to `default`.
+
+ If the input `afi` is not `ipv4` or `ipv6`, e.g. `evpn`, `vrf` must be `default`.
+ """
+ num_peers:PositiveInt
+"""Number of expected BGP peer(s)"""
+
+ @model_validator(mode="after")
+ defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
+ defrender(self,template:AntaTemplate)->list[AntaCommand]:
+ commands=[]
+ forafiinself.inputs.address_families:
+ iftemplate==VerifyBGPPeerCount.commands[0]andafi.afiin["ipv4","ipv6"]:
+ commands.append(template.render(afi=afi.afi,safi=afi.safi,vrf=afi.vrf,num_peers=afi.num_peers))
+ eliftemplate==VerifyBGPPeerCount.commands[1]andafi.afinotin["ipv4","ipv6"]:
+ commands.append(template.render(afi=afi.afi,vrf=afi.vrf,num_peers=afi.num_peers))
+ returncommands
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ self.result.is_success()
+
+ failures:dict[tuple[str,Any],dict[str,Any]]={}
+
+ forcommandinself.instance_commands:
+ peer_count=0
+ command_output=command.json_output
+
+ afi=cast(Afi,command.params.get("afi"))
+ safi=cast(Optional[Safi],command.params.get("safi"))
+ afi_vrf=cast(str,command.params.get("vrf"))
+ num_peers=cast(PositiveInt,command.params.get("num_peers"))
+
+ ifnot(vrfs:=command_output.get("vrfs")):
+ _add_bgp_failures(failures=failures,afi=afi,safi=safi,vrf=afi_vrf,issue="Not Configured")
+ continue
+
+ ifafi_vrf=="all":
+ forvrf_datainvrfs.values():
+ peer_count+=len(vrf_data["peers"])
+ else:
+ peer_count+=len(command_output["vrfs"][afi_vrf]["peers"])
+
+ ifpeer_count!=num_peers:
+ _add_bgp_failures(failures=failures,afi=afi,safi=safi,vrf=afi_vrf,issue=f"Expected: {num_peers}, Actual: {peer_count}")
+
+ iffailures:
+ self.result.is_failure(f"Failures: {list(failures.values())}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ address_families:List[BgpAfi]
+"""
+ List of BGP address families (BgpAfi)
+ """
+
+ classBgpAfi(BaseModel):# pylint: disable=missing-class-docstring
+ afi:Afi
+"""BGP address family (AFI)"""
+ safi:Optional[Safi]=None
+"""Optional BGP subsequent service family (SAFI).
+
+ If the input `afi` is `ipv4` or `ipv6`, a valid `safi` must be provided.
+ """
+ vrf:str="default"
+"""
+ Optional VRF for IPv4 and IPv6. If not provided, it defaults to `default`.
+
+ If the input `afi` is not `ipv4` or `ipv6`, e.g. `evpn`, `vrf` must be `default`.
+ """
+ num_peers:PositiveInt
+"""Number of expected BGP peer(s)"""
+
+ @model_validator(mode="after")
+ defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
classBgpAfi(BaseModel):# pylint: disable=missing-class-docstring
+ afi:Afi
+"""BGP address family (AFI)"""
+ safi:Optional[Safi]=None
+"""Optional BGP subsequent service family (SAFI).
+
+ If the input `afi` is `ipv4` or `ipv6`, a valid `safi` must be provided.
+ """
+ vrf:str="default"
+"""
+ Optional VRF for IPv4 and IPv6. If not provided, it defaults to `default`.
+
+ If the input `afi` is not `ipv4` or `ipv6`, e.g. `evpn`, `vrf` must be `default`.
+ """
+ num_peers:PositiveInt
+"""Number of expected BGP peer(s)"""
+
+ @model_validator(mode="after")
+ defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
@model_validator(mode="after")
+defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
It will validate that all BGP sessions are established and all message queues for these BGP sessions are empty for a given address family.
+
It supports multiple types of address families (AFI) and subsequent service families (SAFI).
+Please refer to the Input class attributes below for details.
+
+
+ Expected Results
+
+
success: If all BGP sessions are established and all messages queues are empty for each address family and VRF.
+
failure: If there are issues with any of the BGP sessions, or if BGP is not configured for an expected VRF or address family.
classVerifyBGPPeersHealth(AntaTest):
+"""
+ This test verifies the health of BGP peers.
+
+ It will validate that all BGP sessions are established and all message queues for these BGP sessions are empty for a given address family.
+
+ It supports multiple types of address families (AFI) and subsequent service families (SAFI).
+ Please refer to the Input class attributes below for details.
+
+ Expected Results:
+ * success: If all BGP sessions are established and all messages queues are empty for each address family and VRF.
+ * failure: If there are issues with any of the BGP sessions, or if BGP is not configured for an expected VRF or address family.
+ """
+
+ name="VerifyBGPPeersHealth"
+ description="Verifies the health of BGP peers"
+ categories=["routing","bgp"]
+ commands=[
+ AntaTemplate(template="show bgp {afi}{safi} summary vrf {vrf}"),
+ AntaTemplate(template="show bgp {afi} summary"),
+ ]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ address_families:List[BgpAfi]
+"""
+ List of BGP address families (BgpAfi)
+ """
+
+ classBgpAfi(BaseModel):# pylint: disable=missing-class-docstring
+ afi:Afi
+"""BGP address family (AFI)"""
+ safi:Optional[Safi]=None
+"""Optional BGP subsequent service family (SAFI).
+
+ If the input `afi` is `ipv4` or `ipv6`, a valid `safi` must be provided.
+ """
+ vrf:str="default"
+"""
+ Optional VRF for IPv4 and IPv6. If not provided, it defaults to `default`.
+
+ If the input `afi` is not `ipv4` or `ipv6`, e.g. `evpn`, `vrf` must be `default`.
+ """
+
+ @model_validator(mode="after")
+ defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
+ defrender(self,template:AntaTemplate)->list[AntaCommand]:
+ commands=[]
+ forafiinself.inputs.address_families:
+ iftemplate==VerifyBGPPeersHealth.commands[0]andafi.afiin["ipv4","ipv6"]:
+ commands.append(template.render(afi=afi.afi,safi=afi.safi,vrf=afi.vrf))
+ eliftemplate==VerifyBGPPeersHealth.commands[1]andafi.afinotin["ipv4","ipv6"]:
+ commands.append(template.render(afi=afi.afi,vrf=afi.vrf))
+ returncommands
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ self.result.is_success()
+
+ failures:dict[tuple[str,Any],dict[str,Any]]={}
+
+ forcommandinself.instance_commands:
+ command_output=command.json_output
+
+ afi=cast(Afi,command.params.get("afi"))
+ safi=cast(Optional[Safi],command.params.get("safi"))
+ afi_vrf=cast(str,command.params.get("vrf"))
+
+ ifnot(vrfs:=command_output.get("vrfs")):
+ _add_bgp_failures(failures=failures,afi=afi,safi=safi,vrf=afi_vrf,issue="Not Configured")
+ continue
+
+ forvrf,vrf_datainvrfs.items():
+ ifnot(peers:=vrf_data.get("peers")):
+ _add_bgp_failures(failures=failures,afi=afi,safi=safi,vrf=afi_vrf,issue="No Peers")
+ continue
+
+ peer_issues={}
+ forpeer,peer_datainpeers.items():
+ issues=_check_peer_issues(peer_data)
+
+ ifissues:
+ peer_issues[peer]=issues
+
+ ifpeer_issues:
+ _add_bgp_failures(failures=failures,afi=afi,safi=safi,vrf=vrf,issue=peer_issues)
+
+ iffailures:
+ self.result.is_failure(f"Failures: {list(failures.values())}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ address_families:List[BgpAfi]
+"""
+ List of BGP address families (BgpAfi)
+ """
+
+ classBgpAfi(BaseModel):# pylint: disable=missing-class-docstring
+ afi:Afi
+"""BGP address family (AFI)"""
+ safi:Optional[Safi]=None
+"""Optional BGP subsequent service family (SAFI).
+
+ If the input `afi` is `ipv4` or `ipv6`, a valid `safi` must be provided.
+ """
+ vrf:str="default"
+"""
+ Optional VRF for IPv4 and IPv6. If not provided, it defaults to `default`.
+
+ If the input `afi` is not `ipv4` or `ipv6`, e.g. `evpn`, `vrf` must be `default`.
+ """
+
+ @model_validator(mode="after")
+ defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
classBgpAfi(BaseModel):# pylint: disable=missing-class-docstring
+ afi:Afi
+"""BGP address family (AFI)"""
+ safi:Optional[Safi]=None
+"""Optional BGP subsequent service family (SAFI).
+
+ If the input `afi` is `ipv4` or `ipv6`, a valid `safi` must be provided.
+ """
+ vrf:str="default"
+"""
+ Optional VRF for IPv4 and IPv6. If not provided, it defaults to `default`.
+
+ If the input `afi` is not `ipv4` or `ipv6`, e.g. `evpn`, `vrf` must be `default`.
+ """
+
+ @model_validator(mode="after")
+ defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
@model_validator(mode="after")
+defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
This test verifies the health of specific BGP peer(s).
+
It will validate that the BGP session is established and all message queues for this BGP session are empty for the given peer(s).
+
It supports multiple types of address families (AFI) and subsequent service families (SAFI).
+Please refer to the Input class attributes below for details.
+
+
+ Expected Results
+
+
success: If the BGP session is established and all messages queues are empty for each given peer.
+
failure: If the BGP session has issues or is not configured, or if BGP is not configured for an expected VRF or address family.
classVerifyBGPSpecificPeers(AntaTest):
+"""
+ This test verifies the health of specific BGP peer(s).
+
+ It will validate that the BGP session is established and all message queues for this BGP session are empty for the given peer(s).
+
+ It supports multiple types of address families (AFI) and subsequent service families (SAFI).
+ Please refer to the Input class attributes below for details.
+
+ Expected Results:
+ * success: If the BGP session is established and all messages queues are empty for each given peer.
+ * failure: If the BGP session has issues or is not configured, or if BGP is not configured for an expected VRF or address family.
+ """
+
+ name="VerifyBGPSpecificPeers"
+ description="Verifies the health of specific BGP peer(s)."
+ categories=["routing","bgp"]
+ commands=[
+ AntaTemplate(template="show bgp {afi}{safi} summary vrf {vrf}"),
+ AntaTemplate(template="show bgp {afi} summary"),
+ ]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ address_families:List[BgpAfi]
+"""
+ List of BGP address families (BgpAfi)
+ """
+
+ classBgpAfi(BaseModel):# pylint: disable=missing-class-docstring
+ afi:Afi
+"""BGP address family (AFI)"""
+ safi:Optional[Safi]=None
+"""Optional BGP subsequent service family (SAFI).
+
+ If the input `afi` is `ipv4` or `ipv6`, a valid `safi` must be provided.
+ """
+ vrf:str="default"
+"""
+ Optional VRF for IPv4 and IPv6. If not provided, it defaults to `default`.
+
+ `all` is NOT supported.
+
+ If the input `afi` is not `ipv4` or `ipv6`, e.g. `evpn`, `vrf` must be `default`.
+ """
+ peers:List[Union[IPv4Address,IPv6Address]]
+"""List of BGP IPv4 or IPv6 peer"""
+
+ @model_validator(mode="after")
+ defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided and vrf must NOT be all.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ ifself.vrf=="all":
+ raiseValueError("'all' is not supported in this test. Use VerifyBGPPeersHealth test instead.")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
+ defrender(self,template:AntaTemplate)->list[AntaCommand]:
+ commands=[]
+ forafiinself.inputs.address_families:
+ iftemplate==VerifyBGPSpecificPeers.commands[0]andafi.afiin["ipv4","ipv6"]:
+ commands.append(template.render(afi=afi.afi,safi=afi.safi,vrf=afi.vrf,peers=afi.peers))
+ eliftemplate==VerifyBGPSpecificPeers.commands[1]andafi.afinotin["ipv4","ipv6"]:
+ commands.append(template.render(afi=afi.afi,vrf=afi.vrf,peers=afi.peers))
+ returncommands
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ self.result.is_success()
+
+ failures:dict[tuple[str,Any],dict[str,Any]]={}
+
+ forcommandinself.instance_commands:
+ command_output=command.json_output
+
+ afi=cast(Afi,command.params.get("afi"))
+ safi=cast(Optional[Safi],command.params.get("safi"))
+ afi_vrf=cast(str,command.params.get("vrf"))
+ afi_peers=cast(List[Union[IPv4Address,IPv6Address]],command.params.get("peers",[]))
+
+ ifnot(vrfs:=command_output.get("vrfs")):
+ _add_bgp_failures(failures=failures,afi=afi,safi=safi,vrf=afi_vrf,issue="Not Configured")
+ continue
+
+ peer_issues={}
+ forpeerinafi_peers:
+ peer_ip=str(peer)
+ peer_data=get_value(dictionary=vrfs,key=f"{afi_vrf}_peers_{peer_ip}",separator="_")
+ issues=_check_peer_issues(peer_data)
+ ifissues:
+ peer_issues[peer_ip]=issues
+
+ ifpeer_issues:
+ _add_bgp_failures(failures=failures,afi=afi,safi=safi,vrf=afi_vrf,issue=peer_issues)
+
+ iffailures:
+ self.result.is_failure(f"Failures: {list(failures.values())}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ address_families:List[BgpAfi]
+"""
+ List of BGP address families (BgpAfi)
+ """
+
+ classBgpAfi(BaseModel):# pylint: disable=missing-class-docstring
+ afi:Afi
+"""BGP address family (AFI)"""
+ safi:Optional[Safi]=None
+"""Optional BGP subsequent service family (SAFI).
+
+ If the input `afi` is `ipv4` or `ipv6`, a valid `safi` must be provided.
+ """
+ vrf:str="default"
+"""
+ Optional VRF for IPv4 and IPv6. If not provided, it defaults to `default`.
+
+ `all` is NOT supported.
+
+ If the input `afi` is not `ipv4` or `ipv6`, e.g. `evpn`, `vrf` must be `default`.
+ """
+ peers:List[Union[IPv4Address,IPv6Address]]
+"""List of BGP IPv4 or IPv6 peer"""
+
+ @model_validator(mode="after")
+ defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided and vrf must NOT be all.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ ifself.vrf=="all":
+ raiseValueError("'all' is not supported in this test. Use VerifyBGPPeersHealth test instead.")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
classBgpAfi(BaseModel):# pylint: disable=missing-class-docstring
+ afi:Afi
+"""BGP address family (AFI)"""
+ safi:Optional[Safi]=None
+"""Optional BGP subsequent service family (SAFI).
+
+ If the input `afi` is `ipv4` or `ipv6`, a valid `safi` must be provided.
+ """
+ vrf:str="default"
+"""
+ Optional VRF for IPv4 and IPv6. If not provided, it defaults to `default`.
+
+ `all` is NOT supported.
+
+ If the input `afi` is not `ipv4` or `ipv6`, e.g. `evpn`, `vrf` must be `default`.
+ """
+ peers:List[Union[IPv4Address,IPv6Address]]
+"""List of BGP IPv4 or IPv6 peer"""
+
+ @model_validator(mode="after")
+ defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided and vrf must NOT be all.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ ifself.vrf=="all":
+ raiseValueError("'all' is not supported in this test. Use VerifyBGPPeersHealth test instead.")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
@model_validator(mode="after")
+defvalidate_inputs(self:BaseModel)->BaseModel:
+"""
+ Validate the inputs provided to the BgpAfi class.
+
+ If afi is either ipv4 or ipv6, safi must be provided and vrf must NOT be all.
+
+ If afi is not ipv4 or ipv6, safi must not be provided and vrf must be default.
+ """
+ ifself.afiin["ipv4","ipv6"]:
+ ifself.safiisNone:
+ raiseValueError("'safi' must be provided when afi is ipv4 or ipv6")
+ ifself.vrf=="all":
+ raiseValueError("'all' is not supported in this test. Use VerifyBGPPeersHealth test instead.")
+ elifself.safiisnotNone:
+ raiseValueError("'safi' must not be provided when afi is not ipv4 or ipv6")
+ elifself.vrf!="default":
+ raiseValueError("'vrf' must be default when afi is not ipv4 or ipv6")
+ returnself
+
classVerifyBFD(AntaTest):
+"""
+ Verifies there is no BFD peer in down state (all VRF, IPv4 neighbors).
+ """
+
+ name="VerifyBFD"
+ description="Verifies there is no BFD peer in down state (all VRF, IPv4 neighbors)."
+ categories=["routing","generic"]
+ # revision 1 as later revision introduce additional nesting for type
+ commands=[AntaCommand(command="show bfd peers",revision=1)]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ self.result.is_success()
+ for_,vrf_dataincommand_output["vrfs"].items():
+ for_,neighbor_datainvrf_data["ipv4Neighbors"].items():
+ forpeer,peer_datainneighbor_data["peerStats"].items():
+ if(peer_status:=peer_data["status"])!="up":
+ failure_message=f"bfd state for peer '{peer}' is {peer_status} (expected up)."
+ if(peer_l3intf:=peer_data.get("l3intf"))isnotNoneandpeer_l3intf!="":
+ failure_message+=f" Interface: {peer_l3intf}."
+ self.result.is_failure(failure_message)
+
Verifies the configured routing protocol model is the one we expect.
+And if there is no mismatch between the configured and operating routing protocol model.
+
+
+ Source code in anta/tests/routing/generic.py
+
classVerifyRoutingProtocolModel(AntaTest):
+"""
+ Verifies the configured routing protocol model is the one we expect.
+ And if there is no mismatch between the configured and operating routing protocol model.
+ """
+
+ name="VerifyRoutingProtocolModel"
+ description=(
+ "Verifies the configured routing protocol model is the expected one and if there is no mismatch between the configured and operating routing protocol model."
+ )
+ categories=["routing","generic"]
+ commands=[AntaCommand(command="show ip route summary",revision=3)]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ model:Literal["multi-agent","ribd"]="multi-agent"
+"""Expected routing protocol model"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ configured_model=command_output["protoModelStatus"]["configuredProtoModel"]
+ operating_model=command_output["protoModelStatus"]["operatingProtoModel"]
+ ifconfigured_model==operating_model==self.inputs.model:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"routing model is misconfigured: configured: {configured_model} - operating: {operating_model} - expected: {self.inputs.model}")
+
classVerifyRoutingTableEntry(AntaTest):
+"""
+ This test verifies that the provided routes are present in the routing table of a specified VRF.
+
+ Expected Results:
+ * success: The test will pass if the provided routes are present in the routing table.
+ * failure: The test will fail if one or many provided routes are missing from the routing table.
+ """
+
+ name="VerifyRoutingTableEntry"
+ description="Verifies that the provided routes are present in the routing table of a specified VRF."
+ categories=["routing","generic"]
+ commands=[AntaTemplate(template="show ip route vrf {vrf}{route}")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ vrf:str="default"
+"""VRF context"""
+ routes:List[IPv4Address]
+"""Routes to verify"""
+
+ defrender(self,template:AntaTemplate)->list[AntaCommand]:
+ return[template.render(vrf=self.inputs.vrf,route=route)forrouteinself.inputs.routes]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ missing_routes=[]
+
+ forcommandinself.instance_commands:
+ if"vrf"incommand.paramsand"route"incommand.params:
+ vrf,route=command.params["vrf"],command.params["route"]
+ iflen(routes:=command.json_output["vrfs"][vrf]["routes"])==0orroute!=ip_interface(list(routes)[0]).ip:
+ missing_routes.append(str(route))
+
+ ifnotmissing_routes:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"The following route(s) are missing from the routing table of VRF {self.inputs.vrf}: {missing_routes}")
+
classVerifyRoutingTableSize(AntaTest):
+"""
+ Verifies the size of the IP routing table (default VRF).
+ Should be between the two provided thresholds.
+ """
+
+ name="VerifyRoutingTableSize"
+ description="Verifies the size of the IP routing table (default VRF). Should be between the two provided thresholds."
+ categories=["routing","generic"]
+ commands=[AntaCommand(command="show ip route summary",revision=3)]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ minimum:int
+"""Expected minimum routing table (default VRF) size"""
+ maximum:int
+"""Expected maximum routing table (default VRF) size"""
+
+ @model_validator(mode="after")# type: ignore
+ defcheck_min_max(self)->AntaTest.Input:
+"""Validate that maximum is greater than minimum"""
+ ifself.minimum>self.maximum:
+ raiseValueError(f"Minimum {self.minimum} is greater than maximum {self.maximum}")
+ returnself
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ total_routes=int(command_output["vrfs"]["default"]["totalRoutes"])
+ ifself.inputs.minimum<=total_routes<=self.inputs.maximum:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"routing-table has {total_routes} routes and not between min ({self.inputs.minimum}) and maximum ({self.inputs.maximum})")
+
+
+
+ Source code in anta/tests/routing/generic.py
+
67
+68
+69
+70
+71
+72
@model_validator(mode="after")# type: ignore
+defcheck_min_max(self)->AntaTest.Input:
+"""Validate that maximum is greater than minimum"""
+ ifself.minimum>self.maximum:
+ raiseValueError(f"Minimum {self.minimum} is greater than maximum {self.maximum}")
+ returnself
+
classVerifyOSPFNeighborCount(AntaTest):
+"""
+ Verifies the number of OSPF neighbors in FULL state is the one we expect.
+ """
+
+ name="VerifyOSPFNeighborCount"
+ description="Verifies the number of OSPF neighbors in FULL state is the one we expect."
+ categories=["routing","ospf"]
+ commands=[AntaCommand(command="show ip ospf neighbor")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:int
+"""The expected number of OSPF neighbors in FULL state"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ if(neighbor_count:=_count_ospf_neighbor(command_output))==0:
+ self.result.is_skipped("no OSPF neighbor found")
+ return
+ self.result.is_success()
+ ifneighbor_count!=self.inputs.number:
+ self.result.is_failure(f"device has {neighbor_count} neighbors (expected {self.inputs.number})")
+ not_full_neighbors=_get_not_full_ospf_neighbors(command_output)
+ print(not_full_neighbors)
+ ifnot_full_neighbors:
+ self.result.is_failure(f"Some neighbors are not correctly configured: {not_full_neighbors}.")
+
classVerifyAPIHttpStatus(AntaTest):
+"""
+ Verifies if eAPI HTTP server is disabled globally.
+
+ Expected Results:
+ * success: The test will pass if eAPI HTTP server is disabled globally.
+ * failure: The test will fail if eAPI HTTP server is NOT disabled globally.
+ """
+
+ name="VerifyAPIHttpStatus"
+ description="Verifies if eAPI HTTP server is disabled globally."
+ categories=["security"]
+ commands=[AntaCommand(command="show management api http-commands")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ifcommand_output["enabled"]andnotcommand_output["httpServer"]["running"]:
+ self.result.is_success()
+ else:
+ self.result.is_failure("eAPI HTTP server is enabled globally")
+
classVerifyAPIHttpsSSL(AntaTest):
+"""
+ Verifies if eAPI HTTPS server SSL profile is configured and valid.
+
+ Expected results:
+ * success: The test will pass if the eAPI HTTPS server SSL profile is configured and valid.
+ * failure: The test will fail if the eAPI HTTPS server SSL profile is NOT configured, misconfigured or invalid.
+ """
+
+ name="VerifyAPIHttpsSSL"
+ description="Verifies if eAPI HTTPS server SSL profile is configured and valid."
+ categories=["security"]
+ commands=[AntaCommand(command="show management api http-commands")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ profile:str
+"""SSL profile to verify"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ try:
+ ifcommand_output["sslProfile"]["name"]==self.inputs.profileandcommand_output["sslProfile"]["state"]=="valid":
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"eAPI HTTPS server SSL profile ({self.inputs.profile}) is misconfigured or invalid")
+
+ exceptKeyError:
+ self.result.is_failure(f"eAPI HTTPS server SSL profile ({self.inputs.profile}) is not configured")
+
classVerifyAPIIPv4Acl(AntaTest):
+"""
+ Verifies if eAPI has the right number IPv4 ACL(s) configured for a specified VRF.
+
+ Expected results:
+ * success: The test will pass if eAPI has the provided number of IPv4 ACL(s) in the specified VRF.
+ * failure: The test will fail if eAPI has not the right number of IPv4 ACL(s) in the specified VRF.
+ """
+
+ name="VerifyAPIIPv4Acl"
+ description="Verifies if eAPI has the right number IPv4 ACL(s) configured for a specified VRF."
+ categories=["security"]
+ commands=[AntaCommand(command="show management api http-commands ip access-list summary")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv4 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for eAPI"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ipv4_acl_list=command_output["ipAclList"]["aclList"]
+ ipv4_acl_number=len(ipv4_acl_list)
+ not_configured_acl_list=[]
+ ifipv4_acl_number!=self.inputs.number:
+ self.result.is_failure(f"Expected {self.inputs.number} eAPI IPv4 ACL(s) in vrf {self.inputs.vrf} but got {ipv4_acl_number}")
+ return
+ foripv4_aclinipv4_acl_list:
+ ifself.inputs.vrfnotinipv4_acl["configuredVrfs"]orself.inputs.vrfnotinipv4_acl["activeVrfs"]:
+ not_configured_acl_list.append(ipv4_acl["name"])
+ ifnot_configured_acl_list:
+ self.result.is_failure(f"eAPI IPv4 ACL(s) not configured or active in vrf {self.inputs.vrf}: {not_configured_acl_list}")
+ else:
+ self.result.is_success()
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv4 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for eAPI"""
+
classVerifyAPIIPv6Acl(AntaTest):
+"""
+ Verifies if eAPI has the right number IPv6 ACL(s) configured for a specified VRF.
+
+ Expected results:
+ * success: The test will pass if eAPI has the provided number of IPv6 ACL(s) in the specified VRF.
+ * failure: The test will fail if eAPI has not the right number of IPv6 ACL(s) in the specified VRF.
+ * skipped: The test will be skipped if the number of IPv6 ACL(s) or VRF parameter is not provided.
+ """
+
+ name="VerifyAPIIPv6Acl"
+ description="Verifies if eAPI has the right number IPv6 ACL(s) configured for a specified VRF."
+ categories=["security"]
+ commands=[AntaCommand(command="show management api http-commands ipv6 access-list summary")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv6 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for eAPI"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ipv6_acl_list=command_output["ipv6AclList"]["aclList"]
+ ipv6_acl_number=len(ipv6_acl_list)
+ not_configured_acl_list=[]
+ ifipv6_acl_number!=self.inputs.number:
+ self.result.is_failure(f"Expected {self.inputs.number} eAPI IPv6 ACL(s) in vrf {self.inputs.vrf} but got {ipv6_acl_number}")
+ return
+ foripv6_aclinipv6_acl_list:
+ ifself.inputs.vrfnotinipv6_acl["configuredVrfs"]orself.inputs.vrfnotinipv6_acl["activeVrfs"]:
+ not_configured_acl_list.append(ipv6_acl["name"])
+ ifnot_configured_acl_list:
+ self.result.is_failure(f"eAPI IPv6 ACL(s) not configured or active in vrf {self.inputs.vrf}: {not_configured_acl_list}")
+ else:
+ self.result.is_success()
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv6 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for eAPI"""
+
classVerifySSHIPv4Acl(AntaTest):
+"""
+ Verifies if the SSHD agent has the right number IPv4 ACL(s) configured for a specified VRF.
+
+ Expected results:
+ * success: The test will pass if the SSHD agent has the provided number of IPv4 ACL(s) in the specified VRF.
+ * failure: The test will fail if the SSHD agent has not the right number of IPv4 ACL(s) in the specified VRF.
+ """
+
+ name="VerifySSHIPv4Acl"
+ description="Verifies if the SSHD agent has IPv4 ACL(s) configured."
+ categories=["security"]
+ commands=[AntaCommand(command="show management ssh ip access-list summary")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv4 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for the SSHD agent"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ipv4_acl_list=command_output["ipAclList"]["aclList"]
+ ipv4_acl_number=len(ipv4_acl_list)
+ not_configured_acl_list=[]
+ ifipv4_acl_number!=self.inputs.number:
+ self.result.is_failure(f"Expected {self.inputs.number} SSH IPv4 ACL(s) in vrf {self.inputs.vrf} but got {ipv4_acl_number}")
+ return
+ foripv4_aclinipv4_acl_list:
+ ifself.inputs.vrfnotinipv4_acl["configuredVrfs"]orself.inputs.vrfnotinipv4_acl["activeVrfs"]:
+ not_configured_acl_list.append(ipv4_acl["name"])
+ ifnot_configured_acl_list:
+ self.result.is_failure(f"SSH IPv4 ACL(s) not configured or active in vrf {self.inputs.vrf}: {not_configured_acl_list}")
+ else:
+ self.result.is_success()
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv4 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for the SSHD agent"""
+
classVerifySSHIPv6Acl(AntaTest):
+"""
+ Verifies if the SSHD agent has the right number IPv6 ACL(s) configured for a specified VRF.
+
+ Expected results:
+ * success: The test will pass if the SSHD agent has the provided number of IPv6 ACL(s) in the specified VRF.
+ * failure: The test will fail if the SSHD agent has not the right number of IPv6 ACL(s) in the specified VRF.
+ """
+
+ name="VerifySSHIPv6Acl"
+ description="Verifies if the SSHD agent has IPv6 ACL(s) configured."
+ categories=["security"]
+ commands=[AntaCommand(command="show management ssh ipv6 access-list summary")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv6 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for the SSHD agent"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ipv6_acl_list=command_output["ipv6AclList"]["aclList"]
+ ipv6_acl_number=len(ipv6_acl_list)
+ not_configured_acl_list=[]
+ ifipv6_acl_number!=self.inputs.number:
+ self.result.is_failure(f"Expected {self.inputs.number} SSH IPv6 ACL(s) in vrf {self.inputs.vrf} but got {ipv6_acl_number}")
+ return
+ foripv6_aclinipv6_acl_list:
+ ifself.inputs.vrfnotinipv6_acl["configuredVrfs"]orself.inputs.vrfnotinipv6_acl["activeVrfs"]:
+ not_configured_acl_list.append(ipv6_acl["name"])
+ ifnot_configured_acl_list:
+ self.result.is_failure(f"SSH IPv6 ACL(s) not configured or active in vrf {self.inputs.vrf}: {not_configured_acl_list}")
+ else:
+ self.result.is_success()
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv6 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for the SSHD agent"""
+
classVerifySSHStatus(AntaTest):
+"""
+ Verifies if the SSHD agent is disabled in the default VRF.
+
+ Expected Results:
+ * success: The test will pass if the SSHD agent is disabled in the default VRF.
+ * failure: The test will fail if the SSHD agent is NOT disabled in the default VRF.
+ """
+
+ name="VerifySSHStatus"
+ description="Verifies if the SSHD agent is disabled in the default VRF."
+ categories=["security"]
+ commands=[AntaCommand(command="show management ssh",ofmt="text")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].text_output
+
+ line=[lineforlineincommand_output.split("\n")ifline.startswith("SSHD status")][0]
+ status=line.split("is ")[1]
+
+ ifstatus=="disabled":
+ self.result.is_success()
+ else:
+ self.result.is_failure(line)
+
classVerifyTelnetStatus(AntaTest):
+"""
+ Verifies if Telnet is disabled in the default VRF.
+
+ Expected Results:
+ * success: The test will pass if Telnet is disabled in the default VRF.
+ * failure: The test will fail if Telnet is NOT disabled in the default VRF.
+ """
+
+ name="VerifyTelnetStatus"
+ description="Verifies if Telnet is disabled in the default VRF."
+ categories=["security"]
+ commands=[AntaCommand(command="show management telnet")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ifcommand_output["serverState"]=="disabled":
+ self.result.is_success()
+ else:
+ self.result.is_failure("Telnet status for Default VRF is enabled")
+
classVerifySnmpIPv4Acl(AntaTest):
+"""
+ Verifies if the SNMP agent has the right number IPv4 ACL(s) configured for a specified VRF.
+
+ Expected results:
+ * success: The test will pass if the SNMP agent has the provided number of IPv4 ACL(s) in the specified VRF.
+ * failure: The test will fail if the SNMP agent has not the right number of IPv4 ACL(s) in the specified VRF.
+ """
+
+ name="VerifySnmpIPv4Acl"
+ description="Verifies if the SNMP agent has IPv4 ACL(s) configured."
+ categories=["snmp"]
+ commands=[AntaCommand(command="show snmp ipv4 access-list summary")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv4 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for the SNMP agent"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ipv4_acl_list=command_output["ipAclList"]["aclList"]
+ ipv4_acl_number=len(ipv4_acl_list)
+ not_configured_acl_list=[]
+ ifipv4_acl_number!=self.inputs.number:
+ self.result.is_failure(f"Expected {self.inputs.number} SNMP IPv4 ACL(s) in vrf {self.inputs.vrf} but got {ipv4_acl_number}")
+ return
+ foripv4_aclinipv4_acl_list:
+ ifself.inputs.vrfnotinipv4_acl["configuredVrfs"]orself.inputs.vrfnotinipv4_acl["activeVrfs"]:
+ not_configured_acl_list.append(ipv4_acl["name"])
+ ifnot_configured_acl_list:
+ self.result.is_failure(f"SNMP IPv4 ACL(s) not configured or active in vrf {self.inputs.vrf}: {not_configured_acl_list}")
+ else:
+ self.result.is_success()
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv4 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for the SNMP agent"""
+
classVerifySnmpIPv6Acl(AntaTest):
+"""
+ Verifies if the SNMP agent has the right number IPv6 ACL(s) configured for a specified VRF.
+
+ Expected results:
+ * success: The test will pass if the SNMP agent has the provided number of IPv6 ACL(s) in the specified VRF.
+ * failure: The test will fail if the SNMP agent has not the right number of IPv6 ACL(s) in the specified VRF.
+ """
+
+ name="VerifySnmpIPv6Acl"
+ description="Verifies if the SNMP agent has IPv6 ACL(s) configured."
+ categories=["snmp"]
+ commands=[AntaCommand(command="show snmp ipv6 access-list summary")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv6 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for the SNMP agent"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ipv6_acl_list=command_output["ipv6AclList"]["aclList"]
+ ipv6_acl_number=len(ipv6_acl_list)
+ not_configured_acl_list=[]
+ ifipv6_acl_number!=self.inputs.number:
+ self.result.is_failure(f"Expected {self.inputs.number} SNMP IPv6 ACL(s) in vrf {self.inputs.vrf} but got {ipv6_acl_number}")
+ return
+ foripv6_aclinipv6_acl_list:
+ ifself.inputs.vrfnotinipv6_acl["configuredVrfs"]orself.inputs.vrfnotinipv6_acl["activeVrfs"]:
+ not_configured_acl_list.append(ipv6_acl["name"])
+ ifnot_configured_acl_list:
+ self.result.is_failure(f"SNMP IPv6 ACL(s) not configured or active in vrf {self.inputs.vrf}: {not_configured_acl_list}")
+ else:
+ self.result.is_success()
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ number:conint(ge=0)# type:ignore
+"""The number of expected IPv6 ACL(s)"""
+ vrf:str="default"
+"""The name of the VRF in which to check for the SNMP agent"""
+
classVerifySnmpStatus(AntaTest):
+"""
+ Verifies whether the SNMP agent is enabled in a specified VRF.
+
+ Expected Results:
+ * success: The test will pass if the SNMP agent is enabled in the specified VRF.
+ * failure: The test will fail if the SNMP agent is disabled in the specified VRF.
+ """
+
+ name="VerifySnmpStatus"
+ description="Verifies if the SNMP agent is enabled."
+ categories=["snmp"]
+ commands=[AntaCommand(command="show snmp")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ vrf:str="default"
+"""The name of the VRF in which to check for the SNMP agent"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ifcommand_output["enabled"]andself.inputs.vrfincommand_output["vrfs"]["snmpVrfs"]:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"SNMP agent disabled in vrf {self.inputs.vrf}")
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ vrf:str="default"
+"""The name of the VRF in which to check for the SNMP agent"""
+
classVerifyEOSVersion(AntaTest):
+"""
+ Verifies the device is running one of the allowed EOS version.
+ """
+
+ name="VerifyEOSVersion"
+ description="Verifies the device is running one of the allowed EOS version."
+ categories=["software"]
+ commands=[AntaCommand(command="show version")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ versions:List[str]
+"""List of allowed EOS versions"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ifcommand_output["version"]inself.inputs.versions:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f'device is running version "{command_output["version"]}" not in expected versions: {self.inputs.versions}')
+
classVerifyTerminAttrVersion(AntaTest):
+"""
+ Verifies the device is running one of the allowed TerminAttr version.
+ """
+
+ name="VerifyTerminAttrVersion"
+ description="Verifies the device is running one of the allowed TerminAttr version."
+ categories=["software"]
+ commands=[AntaCommand(command="show version detail")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ versions:List[str]
+"""List of allowed TerminAttr versions"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ command_output_data=command_output["details"]["packages"]["TerminAttr-core"]["version"]
+ ifcommand_output_datainself.inputs.versions:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"device is running TerminAttr version {command_output_data} and is not in the allowed list: {self.inputs.versions}")
+
classVerifySTPBlockedPorts(AntaTest):
+"""
+ Verifies there is no STP blocked ports.
+
+ Expected Results:
+ * success: The test will pass if there are NO ports blocked by STP.
+ * failure: The test will fail if there are ports blocked by STP.
+ """
+
+ name="VerifySTPBlockedPorts"
+ description="Verifies there is no STP blocked ports."
+ categories=["stp"]
+ commands=[AntaCommand(command="show spanning-tree blockedports")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ifnot(stp_instances:=command_output["spanningTreeInstances"]):
+ self.result.is_success()
+ else:
+ forkey,valueinstp_instances.items():
+ stp_instances[key]=value.pop("spanningTreeBlockedPorts")
+ self.result.is_failure(f"The following ports are blocked by STP: {stp_instances}")
+
classVerifySTPCounters(AntaTest):
+"""
+ Verifies there is no errors in STP BPDU packets.
+
+ Expected Results:
+ * success: The test will pass if there are NO STP BPDU packet errors under all interfaces participating in STP.
+ * failure: The test will fail if there are STP BPDU packet errors on one or many interface(s).
+ """
+
+ name="VerifySTPCounters"
+ description="Verifies there is no errors in STP BPDU packets."
+ categories=["stp"]
+ commands=[AntaCommand(command="show spanning-tree counters")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ interfaces_with_errors=[
+ interfaceforinterface,countersincommand_output["interfaces"].items()ifcounters["bpduTaggedError"]orcounters["bpduOtherError"]!=0
+ ]
+ ifinterfaces_with_errors:
+ self.result.is_failure(f"The following interfaces have STP BPDU packet errors: {interfaces_with_errors}")
+ else:
+ self.result.is_success()
+
classVerifySTPForwardingPorts(AntaTest):
+"""
+ Verifies that all interfaces are in a forwarding state for a provided list of VLAN(s).
+
+ Expected Results:
+ * success: The test will pass if all interfaces are in a forwarding state for the specified VLAN(s).
+ * failure: The test will fail if one or many interfaces are NOT in a forwarding state in the specified VLAN(s).
+ """
+
+ name="VerifySTPForwardingPorts"
+ description="Verifies that all interfaces are forwarding for a provided list of VLAN(s)."
+ categories=["stp"]
+ commands=[AntaTemplate(template="show spanning-tree topology vlan {vlan} status")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ vlans:List[Vlan]
+"""List of VLAN on which to verify forwarding states"""
+
+ defrender(self,template:AntaTemplate)->list[AntaCommand]:
+ return[template.render(vlan=vlan)forvlaninself.inputs.vlans]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ not_configured=[]
+ not_forwarding=[]
+ forcommandinself.instance_commands:
+ if"vlan"incommand.params:
+ vlan_id=command.params["vlan"]
+ ifnot(topologies:=get_value(command.json_output,"topologies")):
+ not_configured.append(vlan_id)
+ else:
+ forvalueintopologies.values():
+ ifint(vlan_id)invalue["vlans"]:
+ interfaces_not_forwarding=[interfaceforinterface,stateinvalue["interfaces"].items()ifstate["state"]!="forwarding"]
+ ifinterfaces_not_forwarding:
+ not_forwarding.append({f"VLAN {vlan_id}":interfaces_not_forwarding})
+ ifnot_configured:
+ self.result.is_failure(f"STP instance is not configured for the following VLAN(s): {not_configured}")
+ ifnot_forwarding:
+ self.result.is_failure(f"The following VLAN(s) have interface(s) that are not in a fowarding state: {not_forwarding}")
+ ifnotnot_configuredandnotinterfaces_not_forwarding:
+ self.result.is_success()
+
classVerifySTPMode(AntaTest):
+"""
+ Verifies the configured STP mode for a provided list of VLAN(s).
+
+ Expected Results:
+ * success: The test will pass if the STP mode is configured properly in the specified VLAN(s).
+ * failure: The test will fail if the STP mode is NOT configured properly for one or more specified VLAN(s).
+ """
+
+ name="VerifySTPMode"
+ description="Verifies the configured STP mode for a provided list of VLAN(s)."
+ categories=["stp"]
+ commands=[AntaTemplate(template="show spanning-tree vlan {vlan}")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ mode:Literal["mstp","rstp","rapidPvst"]="mstp"
+"""STP mode to verify"""
+ vlans:List[Vlan]
+"""List of VLAN on which to verify STP mode"""
+
+ defrender(self,template:AntaTemplate)->list[AntaCommand]:
+ return[template.render(vlan=vlan)forvlaninself.inputs.vlans]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ not_configured=[]
+ wrong_stp_mode=[]
+ forcommandinself.instance_commands:
+ if"vlan"incommand.params:
+ vlan_id=command.params["vlan"]
+ ifnot(stp_mode:=get_value(command.json_output,f"spanningTreeVlanInstances.{vlan_id}.spanningTreeVlanInstance.protocol")):
+ not_configured.append(vlan_id)
+ elifstp_mode!=self.inputs.mode:
+ wrong_stp_mode.append(vlan_id)
+ ifnot_configured:
+ self.result.is_failure(f"STP mode '{self.inputs.mode}' not configured for the following VLAN(s): {not_configured}")
+ ifwrong_stp_mode:
+ self.result.is_failure(f"Wrong STP mode configured for the following VLAN(s): {wrong_stp_mode}")
+ ifnotnot_configuredandnotwrong_stp_mode:
+ self.result.is_success()
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ mode:Literal["mstp","rstp","rapidPvst"]="mstp"
+"""STP mode to verify"""
+ vlans:List[Vlan]
+"""List of VLAN on which to verify STP mode"""
+
classVerifySTPRootPriority(AntaTest):
+"""
+ Verifies the STP root priority for a provided list of VLAN or MST instance ID(s).
+
+ Expected Results:
+ * success: The test will pass if the STP root priority is configured properly for the specified VLAN or MST instance ID(s).
+ * failure: The test will fail if the STP root priority is NOT configured properly for the specified VLAN or MST instance ID(s).
+ """
+
+ name="VerifySTPRootPriority"
+ description="Verifies the STP root priority for a provided list of VLAN or MST instance ID(s)."
+ categories=["stp"]
+ commands=[AntaCommand(command="show spanning-tree root detail")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ priority:int
+"""STP root priority to verify"""
+ instances:List[Vlan]=[]
+"""List of VLAN or MST instance ID(s). If empty, ALL VLAN or MST instance ID(s) will be verified."""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ifnot(stp_instances:=command_output["instances"]):
+ self.result.is_failure("No STP instances configured")
+ return
+ # Checking the type of instances based on first instance
+ first_name=list(stp_instances)[0]
+ iffirst_name.startswith("MST"):
+ prefix="MST"
+ eliffirst_name.startswith("VL"):
+ prefix="VL"
+ else:
+ self.result.is_failure(f"Unsupported STP instance type: {first_name}")
+ return
+ check_instances=[f"{prefix}{instance_id}"forinstance_idinself.inputs.instances]ifself.inputs.instanceselsecommand_output["instances"].keys()
+ wrong_priority_instances=[
+ instanceforinstanceincheck_instancesifget_value(command_output,f"instances.{instance}.rootBridge.priority")!=self.inputs.priority
+ ]
+ ifwrong_priority_instances:
+ self.result.is_failure(f"The following instance(s) have the wrong STP root priority configured: {wrong_priority_instances}")
+ else:
+ self.result.is_success()
+
classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ priority:int
+"""STP root priority to verify"""
+ instances:List[Vlan]=[]
+"""List of VLAN or MST instance ID(s). If empty, ALL VLAN or MST instance ID(s) will be verified."""
+
classVerifyAgentLogs(AntaTest):
+"""
+ This test verifies that no agent crash reports are present on the device.
+
+ Expected Results:
+ * success: The test will pass if there is NO agent crash reported.
+ * failure: The test will fail if any agent crashes are reported.
+ """
+
+ name="VerifyAgentLogs"
+ description="This test verifies that no agent crash reports are present on the device."
+ categories=["system"]
+ commands=[AntaCommand(command="show agent logs crash",ofmt="text")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].text_output
+ iflen(command_output)==0:
+ self.result.is_success()
+ else:
+ pattern=re.compile(r"^===> (.*?) <===$",re.MULTILINE)
+ agents="\n * ".join(pattern.findall(command_output))
+ self.result.is_failure(f"Device has reported agent crashes:\n * {agents}")
+
classVerifyCPUUtilization(AntaTest):
+"""
+ This test verifies whether the CPU utilization is below 75%.
+
+ Expected Results:
+ * success: The test will pass if the CPU utilization is below 75%.
+ * failure: The test will fail if the CPU utilization is over 75%.
+ """
+
+ name="VerifyCPUUtilization"
+ description="This test verifies whether the CPU utilization is below 75%."
+ categories=["system"]
+ commands=[AntaCommand(command="show processes top once")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ command_output_data=command_output["cpuInfo"]["%Cpu(s)"]["idle"]
+ ifcommand_output_data>25:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Device has reported a high CPU utilization: {100-command_output_data}%")
+
classVerifyCoredump(AntaTest):
+"""
+ This test verifies if there are core dump files in the /var/core directory.
+
+ Expected Results:
+ * success: The test will pass if there are NO core dump(s) in /var/core.
+ * failure: The test will fail if there are core dump(s) in /var/core.
+
+ Note:
+ * This test will NOT check for minidump(s) generated by certain agents in /var/core/minidump.
+ """
+
+ name="VerifyCoredump"
+ description="This test verifies if there are core dump files in the /var/core directory."
+ categories=["system"]
+ commands=[AntaCommand(command="show system coredump",ofmt="json")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ core_files=command_output["coreFiles"]
+ if"minidump"incore_files:
+ core_files.remove("minidump")
+ ifnotcore_files:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Core dump(s) have been found: {core_files}")
+
classVerifyFileSystemUtilization(AntaTest):
+"""
+ This test verifies that no partition is utilizing more than 75% of its disk space.
+
+ Expected Results:
+ * success: The test will pass if all partitions are using less than 75% of its disk space.
+ * failure: The test will fail if any partitions are using more than 75% of its disk space.
+ """
+
+ name="VerifyFileSystemUtilization"
+ description="This test verifies that no partition is utilizing more than 75% of its disk space."
+ categories=["system"]
+ commands=[AntaCommand(command="bash timeout 10 df -h",ofmt="text")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].text_output
+ self.result.is_success()
+ forlineincommand_output.split("\n")[1:]:
+ if"loop"notinlineandlen(line)>0and(percentage:=int(line.split()[4].replace("%","")))>75:
+ self.result.is_failure(f"Mount point {line} is higher than 75%: reported {percentage}%")
+
classVerifyMemoryUtilization(AntaTest):
+"""
+ This test verifies whether the memory utilization is below 75%.
+
+ Expected Results:
+ * success: The test will pass if the memory utilization is below 75%.
+ * failure: The test will fail if the memory utilization is over 75%.
+ """
+
+ name="VerifyMemoryUtilization"
+ description="This test verifies whether the memory utilization is below 75%."
+ categories=["system"]
+ commands=[AntaCommand(command="show version")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ memory_usage=command_output["memFree"]/command_output["memTotal"]
+ ifmemory_usage>0.25:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Device has reported a high memory usage: {(1-memory_usage)*100:.2f}%")
+
classVerifyNTP(AntaTest):
+"""
+ This test verifies that the Network Time Protocol (NTP) is synchronized.
+
+ Expected Results:
+ * success: The test will pass if the NTP is synchronised.
+ * failure: The test will fail if the NTP is NOT synchronised.
+ """
+
+ name="VerifyNTP"
+ description="This test verifies if NTP is synchronised."
+ categories=["system"]
+ commands=[AntaCommand(command="show ntp status",ofmt="text")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].text_output
+ ifcommand_output.split("\n")[0].split(" ")[0]=="synchronised":
+ self.result.is_success()
+ else:
+ data=command_output.split("\n")[0]
+ self.result.is_failure(f"The device is not synchronized with the configured NTP server(s): '{data}'")
+
classVerifyReloadCause(AntaTest):
+"""
+ This test verifies the last reload cause of the device.
+
+ Expected results:
+ * success: The test will pass if there are NO reload causes or if the last reload was caused by the user or after an FPGA upgrade.
+ * failure: The test will fail if the last reload was NOT caused by the user or after an FPGA upgrade.
+ * error: The test will report an error if the reload cause is NOT available.
+ """
+
+ name="VerifyReloadCause"
+ description="This test verifies the last reload cause of the device."
+ categories=["system"]
+ commands=[AntaCommand(command="show reload cause")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ if"resetCauses"notincommand_output.keys():
+ self.result.is_error(message="No reload causes available")
+ return
+ iflen(command_output["resetCauses"])==0:
+ # No reload causes
+ self.result.is_success()
+ return
+ reset_causes=command_output["resetCauses"]
+ command_output_data=reset_causes[0].get("description")
+ ifcommand_output_datain[
+ "Reload requested by the user.",
+ "Reload requested after FPGA upgrade",
+ ]:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Reload cause is: '{command_output_data}'")
+
classVerifyUptime(AntaTest):
+"""
+ This test verifies if the device uptime is higher than the provided minimum uptime value.
+
+ Expected Results:
+ * success: The test will pass if the device uptime is higher than the provided value.
+ * failure: The test will fail if the device uptime is lower than the provided value.
+ """
+
+ name="VerifyUptime"
+ description="This test verifies if the device uptime is higher than the provided minimum uptime value."
+ categories=["system"]
+ commands=[AntaCommand(command="show uptime")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ minimum:conint(ge=0)# type: ignore
+"""Minimum uptime in seconds"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ ifcommand_output["upTime"]>self.inputs.minimum:
+ self.result.is_success()
+ else:
+ self.result.is_failure(f"Device uptime is {command_output['upTime']} seconds")
+
classVerifyVxlan1Interface(AntaTest):
+"""
+ This test verifies if the Vxlan1 interface is configured and 'up/up'.
+
+ !!! warning
+ The name of this test has been updated from 'VerifyVxlan' for better representation.
+
+ Expected Results:
+ * success: The test will pass if the Vxlan1 interface is configured with line protocol status and interface status 'up'.
+ * failure: The test will fail if the Vxlan1 interface line protocol status or interface status are not 'up'.
+ * skipped: The test will be skipped if the Vxlan1 interface is not configured.
+ """
+
+ name="VerifyVxlan1Interface"
+ description="This test verifies if the Vxlan1 interface is configured and 'up/up'."
+ categories=["vxlan"]
+ commands=[AntaCommand(command="show interfaces description",ofmt="json")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ if"Vxlan1"notincommand_output["interfaceDescriptions"]:
+ self.result.is_skipped("Vxlan1 interface is not configured")
+ elif(
+ command_output["interfaceDescriptions"]["Vxlan1"]["lineProtocolStatus"]=="up"
+ andcommand_output["interfaceDescriptions"]["Vxlan1"]["interfaceStatus"]=="up"
+ ):
+ self.result.is_success()
+ else:
+ self.result.is_failure(
+ f"Vxlan1 interface is {command_output['interfaceDescriptions']['Vxlan1']['lineProtocolStatus']}"
+ f"/{command_output['interfaceDescriptions']['Vxlan1']['interfaceStatus']}"
+ )
+
classVerifyVxlanConfigSanity(AntaTest):
+"""
+ This test verifies that no issues are detected with the VXLAN configuration.
+
+ Expected Results:
+ * success: The test will pass if no issues are detected with the VXLAN configuration.
+ * failure: The test will fail if issues are detected with the VXLAN configuration.
+ * skipped: The test will be skipped if VXLAN is not configured on the device.
+ """
+
+ name="VerifyVxlanConfigSanity"
+ description="This test verifies that no issues are detected with the VXLAN configuration."
+ categories=["vxlan"]
+ commands=[AntaCommand(command="show vxlan config-sanity",ofmt="json")]
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ command_output=self.instance_commands[0].json_output
+ if"categories"notincommand_outputorlen(command_output["categories"])==0:
+ self.result.is_skipped("VXLAN is not configured")
+ return
+ failed_categories={
+ category:content
+ forcategory,contentincommand_output["categories"].items()
+ ifcategoryin["localVtep","mlag","pd"]andcontent["allCheckPass"]isnotTrue
+ }
+ iflen(failed_categories)>0:
+ self.result.is_failure(f"VXLAN config sanity check is not passing: {failed_categories}")
+ else:
+ self.result.is_success()
+
classVerifyVxlanVniBinding(AntaTest):
+"""
+ This test verifies the VNI-VLAN bindings of the Vxlan1 interface.
+
+ Expected Results:
+ * success: The test will pass if the VNI-VLAN bindings provided are properly configured.
+ * failure: The test will fail if any VNI lacks bindings or if any bindings are incorrect.
+ * skipped: The test will be skipped if the Vxlan1 interface is not configured.
+ """
+
+ name="VerifyVxlanVniBinding"
+ description="Verifies the VNI-VLAN bindings of the Vxlan1 interface"
+ categories=["vxlan"]
+ commands=[AntaCommand(command="show vxlan vni",ofmt="json")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ bindings:Dict[Vni,Vlan]
+"""VNI to VLAN bindings to verify"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ self.result.is_success()
+
+ no_binding=[]
+ wrong_binding=[]
+
+ if(vxlan1:=get_value(self.instance_commands[0].json_output,"vxlanIntfs.Vxlan1"))isNone:
+ self.result.is_skipped("Vxlan1 interface is not configured")
+ return
+
+ forvni,vlaninself.inputs.bindings.items():
+ vni=str(vni)
+ ifvniinvxlan1["vniBindings"]:
+ retrieved_vlan=vxlan1["vniBindings"][vni]["vlan"]
+ elifvniinvxlan1["vniBindingsToVrf"]:
+ retrieved_vlan=vxlan1["vniBindingsToVrf"][vni]["vlan"]
+ else:
+ no_binding.append(vni)
+ retrieved_vlan=None
+
+ ifretrieved_vlanandvlan!=retrieved_vlan:
+ wrong_binding.append({vni:retrieved_vlan})
+
+ ifno_binding:
+ self.result.is_failure(f"The following VNI(s) have no binding: {no_binding}")
+
+ ifwrong_binding:
+ self.result.is_failure(f"The following VNI(s) have the wrong VLAN binding: {wrong_binding}")
+
classVerifyVxlanVtep(AntaTest):
+"""
+ This test verifies the VTEP peers of the Vxlan1 interface.
+
+ Expected Results:
+ * success: The test will pass if all provided VTEP peers are identified and matching.
+ * failure: The test will fail if any VTEP peer is missing or there are unexpected VTEP peers.
+ * skipped: The test will be skipped if the Vxlan1 interface is not configured.
+ """
+
+ name="VerifyVxlanVtep"
+ description="Verifies the VTEP peers of the Vxlan1 interface"
+ categories=["vxlan"]
+ commands=[AntaCommand(command="show vxlan vtep",ofmt="json")]
+
+ classInput(AntaTest.Input):# pylint: disable=missing-class-docstring
+ vteps:List[IPv4Address]
+"""List of VTEP peers to verify"""
+
+ @AntaTest.anta_test
+ deftest(self)->None:
+ self.result.is_success()
+
+ inputs_vteps=[str(input_vtep)forinput_vtepinself.inputs.vteps]
+
+ if(vxlan1:=get_value(self.instance_commands[0].json_output,"interfaces.Vxlan1"))isNone:
+ self.result.is_skipped("Vxlan1 interface is not configured")
+ return
+
+ difference1=set(inputs_vteps).difference(set(vxlan1["vteps"]))
+ difference2=set(vxlan1["vteps"]).difference(set(inputs_vteps))
+
+ ifdifference1:
+ self.result.is_failure(f"The following VTEP peer(s) are missing from the Vxlan1 interface: {sorted(difference1)}")
+
+ ifdifference2:
+ self.result.is_failure(f"Unexpected VTEP peer(s) on Vxlan1 interface: {sorted(difference2)}")
+
defaaa_group_prefix(v:str)->str:
+"""Prefix the AAA method with 'group' if it is known"""
+ built_in_methods=["local","none","logging"]
+ returnf"group {v}"ifvnotinbuilt_in_methodsandnotv.startswith("group ")elsev
+
definterface_autocomplete(v:str)->str:
+"""Allow the user to only provide the beginning of an interface name.
+
+ Supported alias:
+ - `et`, `eth` will be changed to `Ethernet`
+ - `po` will be changed to `Port-Channel`"""
+ intf_id_re=re.compile(r"[0-9]+(\/[0-9]+)*(\.[0-9]+)?")
+ m=intf_id_re.search(v)
+ ifmisNone:
+ raiseValueError(f"Could not parse interface ID in interface '{v}'")
+ intf_id=m[0]
+ ifany(v.lower().startswith(p)forpin["et","eth"]):
+ returnf"Ethernet{intf_id}"
+ ifv.lower().startswith("po"):
+ returnf"Port-Channel{intf_id}"
+ returnv
+
{"use strict";/*!
+ * escape-html
+ * Copyright(c) 2012-2013 TJ Holowaychuk
+ * Copyright(c) 2015 Andreas Lubbe
+ * Copyright(c) 2015 Tiancheng "Timothy" Gu
+ * MIT Licensed
+ */var Wa=/["'&<>]/;Vn.exports=Ua;function Ua(e){var t=""+e,r=Wa.exec(t);if(!r)return t;var o,n="",i=0,s=0;for(i=r.index;i0&&i[i.length-1])&&(p[0]===6||p[0]===2)){r=0;continue}if(p[0]===3&&(!i||p[1]>i[0]&&p[1]=e.length&&(e=void 0),{value:e&&e[o++],done:!e}}};throw new TypeError(t?"Object is not iterable.":"Symbol.iterator is not defined.")}function z(e,t){var r=typeof Symbol=="function"&&e[Symbol.iterator];if(!r)return e;var o=r.call(e),n,i=[],s;try{for(;(t===void 0||t-- >0)&&!(n=o.next()).done;)i.push(n.value)}catch(a){s={error:a}}finally{try{n&&!n.done&&(r=o.return)&&r.call(o)}finally{if(s)throw s.error}}return i}function K(e,t,r){if(r||arguments.length===2)for(var o=0,n=t.length,i;o1||a(u,d)})})}function a(u,d){try{c(o[u](d))}catch(y){f(i[0][3],y)}}function c(u){u.value instanceof ot?Promise.resolve(u.value.v).then(p,l):f(i[0][2],u)}function p(u){a("next",u)}function l(u){a("throw",u)}function f(u,d){u(d),i.shift(),i.length&&a(i[0][0],i[0][1])}}function po(e){if(!Symbol.asyncIterator)throw new TypeError("Symbol.asyncIterator is not defined.");var t=e[Symbol.asyncIterator],r;return t?t.call(e):(e=typeof be=="function"?be(e):e[Symbol.iterator](),r={},o("next"),o("throw"),o("return"),r[Symbol.asyncIterator]=function(){return this},r);function o(i){r[i]=e[i]&&function(s){return new Promise(function(a,c){s=e[i](s),n(a,c,s.done,s.value)})}}function n(i,s,a,c){Promise.resolve(c).then(function(p){i({value:p,done:a})},s)}}function k(e){return typeof e=="function"}function pt(e){var t=function(o){Error.call(o),o.stack=new Error().stack},r=e(t);return r.prototype=Object.create(Error.prototype),r.prototype.constructor=r,r}var Ut=pt(function(e){return function(r){e(this),this.message=r?r.length+` errors occurred during unsubscription:
+`+r.map(function(o,n){return n+1+") "+o.toString()}).join(`
+ `):"",this.name="UnsubscriptionError",this.errors=r}});function ze(e,t){if(e){var r=e.indexOf(t);0<=r&&e.splice(r,1)}}var je=function(){function e(t){this.initialTeardown=t,this.closed=!1,this._parentage=null,this._finalizers=null}return e.prototype.unsubscribe=function(){var t,r,o,n,i;if(!this.closed){this.closed=!0;var s=this._parentage;if(s)if(this._parentage=null,Array.isArray(s))try{for(var a=be(s),c=a.next();!c.done;c=a.next()){var p=c.value;p.remove(this)}}catch(b){t={error:b}}finally{try{c&&!c.done&&(r=a.return)&&r.call(a)}finally{if(t)throw t.error}}else s.remove(this);var l=this.initialTeardown;if(k(l))try{l()}catch(b){i=b instanceof Ut?b.errors:[b]}var f=this._finalizers;if(f){this._finalizers=null;try{for(var u=be(f),d=u.next();!d.done;d=u.next()){var y=d.value;try{lo(y)}catch(b){i=i!=null?i:[],b instanceof Ut?i=K(K([],z(i)),z(b.errors)):i.push(b)}}}catch(b){o={error:b}}finally{try{d&&!d.done&&(n=u.return)&&n.call(u)}finally{if(o)throw o.error}}}if(i)throw new Ut(i)}},e.prototype.add=function(t){var r;if(t&&t!==this)if(this.closed)lo(t);else{if(t instanceof e){if(t.closed||t._hasParent(this))return;t._addParent(this)}(this._finalizers=(r=this._finalizers)!==null&&r!==void 0?r:[]).push(t)}},e.prototype._hasParent=function(t){var r=this._parentage;return r===t||Array.isArray(r)&&r.includes(t)},e.prototype._addParent=function(t){var r=this._parentage;this._parentage=Array.isArray(r)?(r.push(t),r):r?[r,t]:t},e.prototype._removeParent=function(t){var r=this._parentage;r===t?this._parentage=null:Array.isArray(r)&&ze(r,t)},e.prototype.remove=function(t){var r=this._finalizers;r&&ze(r,t),t instanceof e&&t._removeParent(this)},e.EMPTY=function(){var t=new e;return t.closed=!0,t}(),e}();var Tr=je.EMPTY;function Nt(e){return e instanceof je||e&&"closed"in e&&k(e.remove)&&k(e.add)&&k(e.unsubscribe)}function lo(e){k(e)?e():e.unsubscribe()}var He={onUnhandledError:null,onStoppedNotification:null,Promise:void 0,useDeprecatedSynchronousErrorHandling:!1,useDeprecatedNextContext:!1};var lt={setTimeout:function(e,t){for(var r=[],o=2;o0},enumerable:!1,configurable:!0}),t.prototype._trySubscribe=function(r){return this._throwIfClosed(),e.prototype._trySubscribe.call(this,r)},t.prototype._subscribe=function(r){return this._throwIfClosed(),this._checkFinalizedStatuses(r),this._innerSubscribe(r)},t.prototype._innerSubscribe=function(r){var o=this,n=this,i=n.hasError,s=n.isStopped,a=n.observers;return i||s?Tr:(this.currentObservers=null,a.push(r),new je(function(){o.currentObservers=null,ze(a,r)}))},t.prototype._checkFinalizedStatuses=function(r){var o=this,n=o.hasError,i=o.thrownError,s=o.isStopped;n?r.error(i):s&&r.complete()},t.prototype.asObservable=function(){var r=new I;return r.source=this,r},t.create=function(r,o){return new xo(r,o)},t}(I);var xo=function(e){se(t,e);function t(r,o){var n=e.call(this)||this;return n.destination=r,n.source=o,n}return t.prototype.next=function(r){var o,n;(n=(o=this.destination)===null||o===void 0?void 0:o.next)===null||n===void 0||n.call(o,r)},t.prototype.error=function(r){var o,n;(n=(o=this.destination)===null||o===void 0?void 0:o.error)===null||n===void 0||n.call(o,r)},t.prototype.complete=function(){var r,o;(o=(r=this.destination)===null||r===void 0?void 0:r.complete)===null||o===void 0||o.call(r)},t.prototype._subscribe=function(r){var o,n;return(n=(o=this.source)===null||o===void 0?void 0:o.subscribe(r))!==null&&n!==void 0?n:Tr},t}(x);var St={now:function(){return(St.delegate||Date).now()},delegate:void 0};var Ot=function(e){se(t,e);function t(r,o,n){r===void 0&&(r=1/0),o===void 0&&(o=1/0),n===void 0&&(n=St);var i=e.call(this)||this;return i._bufferSize=r,i._windowTime=o,i._timestampProvider=n,i._buffer=[],i._infiniteTimeWindow=!0,i._infiniteTimeWindow=o===1/0,i._bufferSize=Math.max(1,r),i._windowTime=Math.max(1,o),i}return t.prototype.next=function(r){var o=this,n=o.isStopped,i=o._buffer,s=o._infiniteTimeWindow,a=o._timestampProvider,c=o._windowTime;n||(i.push(r),!s&&i.push(a.now()+c)),this._trimBuffer(),e.prototype.next.call(this,r)},t.prototype._subscribe=function(r){this._throwIfClosed(),this._trimBuffer();for(var o=this._innerSubscribe(r),n=this,i=n._infiniteTimeWindow,s=n._buffer,a=s.slice(),c=0;c0?e.prototype.requestAsyncId.call(this,r,o,n):(r.actions.push(this),r._scheduled||(r._scheduled=ut.requestAnimationFrame(function(){return r.flush(void 0)})))},t.prototype.recycleAsyncId=function(r,o,n){var i;if(n===void 0&&(n=0),n!=null?n>0:this.delay>0)return e.prototype.recycleAsyncId.call(this,r,o,n);var s=r.actions;o!=null&&((i=s[s.length-1])===null||i===void 0?void 0:i.id)!==o&&(ut.cancelAnimationFrame(o),r._scheduled=void 0)},t}(zt);var wo=function(e){se(t,e);function t(){return e!==null&&e.apply(this,arguments)||this}return t.prototype.flush=function(r){this._active=!0;var o=this._scheduled;this._scheduled=void 0;var n=this.actions,i;r=r||n.shift();do if(i=r.execute(r.state,r.delay))break;while((r=n[0])&&r.id===o&&n.shift());if(this._active=!1,i){for(;(r=n[0])&&r.id===o&&n.shift();)r.unsubscribe();throw i}},t}(qt);var ge=new wo(Eo);var M=new I(function(e){return e.complete()});function Kt(e){return e&&k(e.schedule)}function Cr(e){return e[e.length-1]}function Ge(e){return k(Cr(e))?e.pop():void 0}function Ae(e){return Kt(Cr(e))?e.pop():void 0}function Qt(e,t){return typeof Cr(e)=="number"?e.pop():t}var dt=function(e){return e&&typeof e.length=="number"&&typeof e!="function"};function Yt(e){return k(e==null?void 0:e.then)}function Bt(e){return k(e[ft])}function Gt(e){return Symbol.asyncIterator&&k(e==null?void 0:e[Symbol.asyncIterator])}function Jt(e){return new TypeError("You provided "+(e!==null&&typeof e=="object"?"an invalid object":"'"+e+"'")+" where a stream was expected. You can provide an Observable, Promise, ReadableStream, Array, AsyncIterable, or Iterable.")}function Wi(){return typeof Symbol!="function"||!Symbol.iterator?"@@iterator":Symbol.iterator}var Xt=Wi();function Zt(e){return k(e==null?void 0:e[Xt])}function er(e){return co(this,arguments,function(){var r,o,n,i;return Wt(this,function(s){switch(s.label){case 0:r=e.getReader(),s.label=1;case 1:s.trys.push([1,,9,10]),s.label=2;case 2:return[4,ot(r.read())];case 3:return o=s.sent(),n=o.value,i=o.done,i?[4,ot(void 0)]:[3,5];case 4:return[2,s.sent()];case 5:return[4,ot(n)];case 6:return[4,s.sent()];case 7:return s.sent(),[3,2];case 8:return[3,10];case 9:return r.releaseLock(),[7];case 10:return[2]}})})}function tr(e){return k(e==null?void 0:e.getReader)}function F(e){if(e instanceof I)return e;if(e!=null){if(Bt(e))return Ui(e);if(dt(e))return Ni(e);if(Yt(e))return Di(e);if(Gt(e))return To(e);if(Zt(e))return Vi(e);if(tr(e))return zi(e)}throw Jt(e)}function Ui(e){return new I(function(t){var r=e[ft]();if(k(r.subscribe))return r.subscribe(t);throw new TypeError("Provided object does not correctly implement Symbol.observable")})}function Ni(e){return new I(function(t){for(var r=0;r=2;return function(o){return o.pipe(e?v(function(n,i){return e(n,i,o)}):pe,ue(1),r?$e(t):Uo(function(){return new or}))}}function Rr(e){return e<=0?function(){return M}:g(function(t,r){var o=[];t.subscribe(E(r,function(n){o.push(n),e=2,!0))}function de(e){e===void 0&&(e={});var t=e.connector,r=t===void 0?function(){return new x}:t,o=e.resetOnError,n=o===void 0?!0:o,i=e.resetOnComplete,s=i===void 0?!0:i,a=e.resetOnRefCountZero,c=a===void 0?!0:a;return function(p){var l,f,u,d=0,y=!1,b=!1,D=function(){f==null||f.unsubscribe(),f=void 0},Q=function(){D(),l=u=void 0,y=b=!1},J=function(){var C=l;Q(),C==null||C.unsubscribe()};return g(function(C,ct){d++,!b&&!y&&D();var Ve=u=u!=null?u:r();ct.add(function(){d--,d===0&&!b&&!y&&(f=jr(J,c))}),Ve.subscribe(ct),!l&&d>0&&(l=new it({next:function(Fe){return Ve.next(Fe)},error:function(Fe){b=!0,D(),f=jr(Q,n,Fe),Ve.error(Fe)},complete:function(){y=!0,D(),f=jr(Q,s),Ve.complete()}}),F(C).subscribe(l))})(p)}}function jr(e,t){for(var r=[],o=2;oe.next(document)),e}function W(e,t=document){return Array.from(t.querySelectorAll(e))}function U(e,t=document){let r=ce(e,t);if(typeof r=="undefined")throw new ReferenceError(`Missing element: expected "${e}" to be present`);return r}function ce(e,t=document){return t.querySelector(e)||void 0}function Ie(){return document.activeElement instanceof HTMLElement&&document.activeElement||void 0}var ca=L(h(document.body,"focusin"),h(document.body,"focusout")).pipe(ye(1),q(void 0),m(()=>Ie()||document.body),Z(1));function vt(e){return ca.pipe(m(t=>e.contains(t)),X())}function qo(e,t){return L(h(e,"mouseenter").pipe(m(()=>!0)),h(e,"mouseleave").pipe(m(()=>!1))).pipe(t?ye(t):pe,q(!1))}function Ue(e){return{x:e.offsetLeft,y:e.offsetTop}}function Ko(e){return L(h(window,"load"),h(window,"resize")).pipe(Le(0,ge),m(()=>Ue(e)),q(Ue(e)))}function ir(e){return{x:e.scrollLeft,y:e.scrollTop}}function et(e){return L(h(e,"scroll"),h(window,"resize")).pipe(Le(0,ge),m(()=>ir(e)),q(ir(e)))}function Qo(e,t){if(typeof t=="string"||typeof t=="number")e.innerHTML+=t.toString();else if(t instanceof Node)e.appendChild(t);else if(Array.isArray(t))for(let r of t)Qo(e,r)}function S(e,t,...r){let o=document.createElement(e);if(t)for(let n of Object.keys(t))typeof t[n]!="undefined"&&(typeof t[n]!="boolean"?o.setAttribute(n,t[n]):o.setAttribute(n,""));for(let n of r)Qo(o,n);return o}function ar(e){if(e>999){let t=+((e-950)%1e3>99);return`${((e+1e-6)/1e3).toFixed(t)}k`}else return e.toString()}function gt(e){let t=S("script",{src:e});return H(()=>(document.head.appendChild(t),L(h(t,"load"),h(t,"error").pipe(w(()=>kr(()=>new ReferenceError(`Invalid script: ${e}`))))).pipe(m(()=>{}),A(()=>document.head.removeChild(t)),ue(1))))}var Yo=new x,pa=H(()=>typeof ResizeObserver=="undefined"?gt("https://unpkg.com/resize-observer-polyfill"):R(void 0)).pipe(m(()=>new ResizeObserver(e=>{for(let t of e)Yo.next(t)})),w(e=>L(Ke,R(e)).pipe(A(()=>e.disconnect()))),Z(1));function le(e){return{width:e.offsetWidth,height:e.offsetHeight}}function Se(e){return pa.pipe(T(t=>t.observe(e)),w(t=>Yo.pipe(v(({target:r})=>r===e),A(()=>t.unobserve(e)),m(()=>le(e)))),q(le(e)))}function xt(e){return{width:e.scrollWidth,height:e.scrollHeight}}function sr(e){let t=e.parentElement;for(;t&&(e.scrollWidth<=t.scrollWidth&&e.scrollHeight<=t.scrollHeight);)t=(e=t).parentElement;return t?e:void 0}var Bo=new x,la=H(()=>R(new IntersectionObserver(e=>{for(let t of e)Bo.next(t)},{threshold:0}))).pipe(w(e=>L(Ke,R(e)).pipe(A(()=>e.disconnect()))),Z(1));function yt(e){return la.pipe(T(t=>t.observe(e)),w(t=>Bo.pipe(v(({target:r})=>r===e),A(()=>t.unobserve(e)),m(({isIntersecting:r})=>r))))}function Go(e,t=16){return et(e).pipe(m(({y:r})=>{let o=le(e),n=xt(e);return r>=n.height-o.height-t}),X())}var cr={drawer:U("[data-md-toggle=drawer]"),search:U("[data-md-toggle=search]")};function Jo(e){return cr[e].checked}function Ye(e,t){cr[e].checked!==t&&cr[e].click()}function Ne(e){let t=cr[e];return h(t,"change").pipe(m(()=>t.checked),q(t.checked))}function ma(e,t){switch(e.constructor){case HTMLInputElement:return e.type==="radio"?/^Arrow/.test(t):!0;case HTMLSelectElement:case HTMLTextAreaElement:return!0;default:return e.isContentEditable}}function fa(){return L(h(window,"compositionstart").pipe(m(()=>!0)),h(window,"compositionend").pipe(m(()=>!1))).pipe(q(!1))}function Xo(){let e=h(window,"keydown").pipe(v(t=>!(t.metaKey||t.ctrlKey)),m(t=>({mode:Jo("search")?"search":"global",type:t.key,claim(){t.preventDefault(),t.stopPropagation()}})),v(({mode:t,type:r})=>{if(t==="global"){let o=Ie();if(typeof o!="undefined")return!ma(o,r)}return!0}),de());return fa().pipe(w(t=>t?M:e))}function me(){return new URL(location.href)}function st(e,t=!1){if(G("navigation.instant")&&!t){let r=S("a",{href:e.href});document.body.appendChild(r),r.click(),r.remove()}else location.href=e.href}function Zo(){return new x}function en(){return location.hash.slice(1)}function pr(e){let t=S("a",{href:e});t.addEventListener("click",r=>r.stopPropagation()),t.click()}function ua(e){return L(h(window,"hashchange"),e).pipe(m(en),q(en()),v(t=>t.length>0),Z(1))}function tn(e){return ua(e).pipe(m(t=>ce(`[id="${t}"]`)),v(t=>typeof t!="undefined"))}function At(e){let t=matchMedia(e);return nr(r=>t.addListener(()=>r(t.matches))).pipe(q(t.matches))}function rn(){let e=matchMedia("print");return L(h(window,"beforeprint").pipe(m(()=>!0)),h(window,"afterprint").pipe(m(()=>!1))).pipe(q(e.matches))}function Dr(e,t){return e.pipe(w(r=>r?t():M))}function lr(e,t){return new I(r=>{let o=new XMLHttpRequest;o.open("GET",`${e}`),o.responseType="blob",o.addEventListener("load",()=>{o.status>=200&&o.status<300?(r.next(o.response),r.complete()):r.error(new Error(o.statusText))}),o.addEventListener("error",()=>{r.error(new Error("Network Error"))}),o.addEventListener("abort",()=>{r.error(new Error("Request aborted"))}),typeof(t==null?void 0:t.progress$)!="undefined"&&(o.addEventListener("progress",n=>{if(n.lengthComputable)t.progress$.next(n.loaded/n.total*100);else{let i=Number(o.getResponseHeader("Content-Length"))||0;t.progress$.next(n.loaded/i*100)}}),t.progress$.next(5)),o.send()})}function De(e,t){return lr(e,t).pipe(w(r=>r.text()),m(r=>JSON.parse(r)),Z(1))}function on(e,t){let r=new DOMParser;return lr(e,t).pipe(w(o=>o.text()),m(o=>r.parseFromString(o,"text/xml")),Z(1))}function nn(){return{x:Math.max(0,scrollX),y:Math.max(0,scrollY)}}function an(){return L(h(window,"scroll",{passive:!0}),h(window,"resize",{passive:!0})).pipe(m(nn),q(nn()))}function sn(){return{width:innerWidth,height:innerHeight}}function cn(){return h(window,"resize",{passive:!0}).pipe(m(sn),q(sn()))}function pn(){return B([an(),cn()]).pipe(m(([e,t])=>({offset:e,size:t})),Z(1))}function mr(e,{viewport$:t,header$:r}){let o=t.pipe(te("size")),n=B([o,r]).pipe(m(()=>Ue(e)));return B([r,t,n]).pipe(m(([{height:i},{offset:s,size:a},{x:c,y:p}])=>({offset:{x:s.x-c,y:s.y-p+i},size:a})))}function da(e){return h(e,"message",t=>t.data)}function ha(e){let t=new x;return t.subscribe(r=>e.postMessage(r)),t}function ln(e,t=new Worker(e)){let r=da(t),o=ha(t),n=new x;n.subscribe(o);let i=o.pipe(ee(),oe(!0));return n.pipe(ee(),Re(r.pipe(j(i))),de())}var ba=U("#__config"),Et=JSON.parse(ba.textContent);Et.base=`${new URL(Et.base,me())}`;function he(){return Et}function G(e){return Et.features.includes(e)}function we(e,t){return typeof t!="undefined"?Et.translations[e].replace("#",t.toString()):Et.translations[e]}function Oe(e,t=document){return U(`[data-md-component=${e}]`,t)}function ne(e,t=document){return W(`[data-md-component=${e}]`,t)}function va(e){let t=U(".md-typeset > :first-child",e);return h(t,"click",{once:!0}).pipe(m(()=>U(".md-typeset",e)),m(r=>({hash:__md_hash(r.innerHTML)})))}function mn(e){if(!G("announce.dismiss")||!e.childElementCount)return M;if(!e.hidden){let t=U(".md-typeset",e);__md_hash(t.innerHTML)===__md_get("__announce")&&(e.hidden=!0)}return H(()=>{let t=new x;return t.subscribe(({hash:r})=>{e.hidden=!0,__md_set("__announce",r)}),va(e).pipe(T(r=>t.next(r)),A(()=>t.complete()),m(r=>P({ref:e},r)))})}function ga(e,{target$:t}){return t.pipe(m(r=>({hidden:r!==e})))}function fn(e,t){let r=new x;return r.subscribe(({hidden:o})=>{e.hidden=o}),ga(e,t).pipe(T(o=>r.next(o)),A(()=>r.complete()),m(o=>P({ref:e},o)))}function Ct(e,t){return t==="inline"?S("div",{class:"md-tooltip md-tooltip--inline",id:e,role:"tooltip"},S("div",{class:"md-tooltip__inner md-typeset"})):S("div",{class:"md-tooltip",id:e,role:"tooltip"},S("div",{class:"md-tooltip__inner md-typeset"}))}function un(e,t){if(t=t?`${t}_annotation_${e}`:void 0,t){let r=t?`#${t}`:void 0;return S("aside",{class:"md-annotation",tabIndex:0},Ct(t),S("a",{href:r,class:"md-annotation__index",tabIndex:-1},S("span",{"data-md-annotation-id":e})))}else return S("aside",{class:"md-annotation",tabIndex:0},Ct(t),S("span",{class:"md-annotation__index",tabIndex:-1},S("span",{"data-md-annotation-id":e})))}function dn(e){return S("button",{class:"md-clipboard md-icon",title:we("clipboard.copy"),"data-clipboard-target":`#${e} > code`})}function Vr(e,t){let r=t&2,o=t&1,n=Object.keys(e.terms).filter(c=>!e.terms[c]).reduce((c,p)=>[...c,S("del",null,p)," "],[]).slice(0,-1),i=he(),s=new URL(e.location,i.base);G("search.highlight")&&s.searchParams.set("h",Object.entries(e.terms).filter(([,c])=>c).reduce((c,[p])=>`${c} ${p}`.trim(),""));let{tags:a}=he();return S("a",{href:`${s}`,class:"md-search-result__link",tabIndex:-1},S("article",{class:"md-search-result__article md-typeset","data-md-score":e.score.toFixed(2)},r>0&&S("div",{class:"md-search-result__icon md-icon"}),r>0&&S("h1",null,e.title),r<=0&&S("h2",null,e.title),o>0&&e.text.length>0&&e.text,e.tags&&e.tags.map(c=>{let p=a?c in a?`md-tag-icon md-tag--${a[c]}`:"md-tag-icon":"";return S("span",{class:`md-tag ${p}`},c)}),o>0&&n.length>0&&S("p",{class:"md-search-result__terms"},we("search.result.term.missing"),": ",...n)))}function hn(e){let t=e[0].score,r=[...e],o=he(),n=r.findIndex(l=>!`${new URL(l.location,o.base)}`.includes("#")),[i]=r.splice(n,1),s=r.findIndex(l=>l.scoreVr(l,1)),...c.length?[S("details",{class:"md-search-result__more"},S("summary",{tabIndex:-1},S("div",null,c.length>0&&c.length===1?we("search.result.more.one"):we("search.result.more.other",c.length))),...c.map(l=>Vr(l,1)))]:[]];return S("li",{class:"md-search-result__item"},p)}function bn(e){return S("ul",{class:"md-source__facts"},Object.entries(e).map(([t,r])=>S("li",{class:`md-source__fact md-source__fact--${t}`},typeof r=="number"?ar(r):r)))}function zr(e){let t=`tabbed-control tabbed-control--${e}`;return S("div",{class:t,hidden:!0},S("button",{class:"tabbed-button",tabIndex:-1,"aria-hidden":"true"}))}function vn(e){return S("div",{class:"md-typeset__scrollwrap"},S("div",{class:"md-typeset__table"},e))}function xa(e){let t=he(),r=new URL(`../${e.version}/`,t.base);return S("li",{class:"md-version__item"},S("a",{href:`${r}`,class:"md-version__link"},e.title))}function gn(e,t){return S("div",{class:"md-version"},S("button",{class:"md-version__current","aria-label":we("select.version")},t.title),S("ul",{class:"md-version__list"},e.map(xa)))}var ya=0;function Ea(e,t){document.body.append(e);let{width:r}=le(e);e.style.setProperty("--md-tooltip-width",`${r}px`),e.remove();let o=sr(t),n=typeof o!="undefined"?et(o):R({x:0,y:0}),i=L(vt(t),qo(t)).pipe(X());return B([i,n]).pipe(m(([s,a])=>{let{x:c,y:p}=Ue(t),l=le(t),f=t.closest("table");return f&&t.parentElement&&(c+=f.offsetLeft+t.parentElement.offsetLeft,p+=f.offsetTop+t.parentElement.offsetTop),{active:s,offset:{x:c-a.x+l.width/2-r/2,y:p-a.y+l.height+8}}}))}function Be(e){let t=e.title;if(!t.length)return M;let r=`__tooltip_${ya++}`,o=Ct(r,"inline"),n=U(".md-typeset",o);return n.innerHTML=t,H(()=>{let i=new x;return i.subscribe({next({offset:s}){o.style.setProperty("--md-tooltip-x",`${s.x}px`),o.style.setProperty("--md-tooltip-y",`${s.y}px`)},complete(){o.style.removeProperty("--md-tooltip-x"),o.style.removeProperty("--md-tooltip-y")}}),L(i.pipe(v(({active:s})=>s)),i.pipe(ye(250),v(({active:s})=>!s))).subscribe({next({active:s}){s?(e.insertAdjacentElement("afterend",o),e.setAttribute("aria-describedby",r),e.removeAttribute("title")):(o.remove(),e.removeAttribute("aria-describedby"),e.setAttribute("title",t))},complete(){o.remove(),e.removeAttribute("aria-describedby"),e.setAttribute("title",t)}}),i.pipe(Le(16,ge)).subscribe(({active:s})=>{o.classList.toggle("md-tooltip--active",s)}),i.pipe(_t(125,ge),v(()=>!!e.offsetParent),m(()=>e.offsetParent.getBoundingClientRect()),m(({x:s})=>s)).subscribe({next(s){s?o.style.setProperty("--md-tooltip-0",`${-s}px`):o.style.removeProperty("--md-tooltip-0")},complete(){o.style.removeProperty("--md-tooltip-0")}}),Ea(o,e).pipe(T(s=>i.next(s)),A(()=>i.complete()),m(s=>P({ref:e},s)))}).pipe(qe(ie))}function wa(e,t){let r=H(()=>B([Ko(e),et(t)])).pipe(m(([{x:o,y:n},i])=>{let{width:s,height:a}=le(e);return{x:o-i.x+s/2,y:n-i.y+a/2}}));return vt(e).pipe(w(o=>r.pipe(m(n=>({active:o,offset:n})),ue(+!o||1/0))))}function xn(e,t,{target$:r}){let[o,n]=Array.from(e.children);return H(()=>{let i=new x,s=i.pipe(ee(),oe(!0));return i.subscribe({next({offset:a}){e.style.setProperty("--md-tooltip-x",`${a.x}px`),e.style.setProperty("--md-tooltip-y",`${a.y}px`)},complete(){e.style.removeProperty("--md-tooltip-x"),e.style.removeProperty("--md-tooltip-y")}}),yt(e).pipe(j(s)).subscribe(a=>{e.toggleAttribute("data-md-visible",a)}),L(i.pipe(v(({active:a})=>a)),i.pipe(ye(250),v(({active:a})=>!a))).subscribe({next({active:a}){a?e.prepend(o):o.remove()},complete(){e.prepend(o)}}),i.pipe(Le(16,ge)).subscribe(({active:a})=>{o.classList.toggle("md-tooltip--active",a)}),i.pipe(_t(125,ge),v(()=>!!e.offsetParent),m(()=>e.offsetParent.getBoundingClientRect()),m(({x:a})=>a)).subscribe({next(a){a?e.style.setProperty("--md-tooltip-0",`${-a}px`):e.style.removeProperty("--md-tooltip-0")},complete(){e.style.removeProperty("--md-tooltip-0")}}),h(n,"click").pipe(j(s),v(a=>!(a.metaKey||a.ctrlKey))).subscribe(a=>{a.stopPropagation(),a.preventDefault()}),h(n,"mousedown").pipe(j(s),ae(i)).subscribe(([a,{active:c}])=>{var p;if(a.button!==0||a.metaKey||a.ctrlKey)a.preventDefault();else if(c){a.preventDefault();let l=e.parentElement.closest(".md-annotation");l instanceof HTMLElement?l.focus():(p=Ie())==null||p.blur()}}),r.pipe(j(s),v(a=>a===o),Qe(125)).subscribe(()=>e.focus()),wa(e,t).pipe(T(a=>i.next(a)),A(()=>i.complete()),m(a=>P({ref:e},a)))})}function Ta(e){return e.tagName==="CODE"?W(".c, .c1, .cm",e):[e]}function Sa(e){let t=[];for(let r of Ta(e)){let o=[],n=document.createNodeIterator(r,NodeFilter.SHOW_TEXT);for(let i=n.nextNode();i;i=n.nextNode())o.push(i);for(let i of o){let s;for(;s=/(\(\d+\))(!)?/.exec(i.textContent);){let[,a,c]=s;if(typeof c=="undefined"){let p=i.splitText(s.index);i=p.splitText(a.length),t.push(p)}else{i.textContent=a,t.push(i);break}}}}return t}function yn(e,t){t.append(...Array.from(e.childNodes))}function fr(e,t,{target$:r,print$:o}){let n=t.closest("[id]"),i=n==null?void 0:n.id,s=new Map;for(let a of Sa(t)){let[,c]=a.textContent.match(/\((\d+)\)/);ce(`:scope > li:nth-child(${c})`,e)&&(s.set(c,un(c,i)),a.replaceWith(s.get(c)))}return s.size===0?M:H(()=>{let a=new x,c=a.pipe(ee(),oe(!0)),p=[];for(let[l,f]of s)p.push([U(".md-typeset",f),U(`:scope > li:nth-child(${l})`,e)]);return o.pipe(j(c)).subscribe(l=>{e.hidden=!l,e.classList.toggle("md-annotation-list",l);for(let[f,u]of p)l?yn(f,u):yn(u,f)}),L(...[...s].map(([,l])=>xn(l,t,{target$:r}))).pipe(A(()=>a.complete()),de())})}function En(e){if(e.nextElementSibling){let t=e.nextElementSibling;if(t.tagName==="OL")return t;if(t.tagName==="P"&&!t.children.length)return En(t)}}function wn(e,t){return H(()=>{let r=En(e);return typeof r!="undefined"?fr(r,e,t):M})}var Tn=jt(Kr());var Oa=0;function Sn(e){if(e.nextElementSibling){let t=e.nextElementSibling;if(t.tagName==="OL")return t;if(t.tagName==="P"&&!t.children.length)return Sn(t)}}function Ma(e){return Se(e).pipe(m(({width:t})=>({scrollable:xt(e).width>t})),te("scrollable"))}function On(e,t){let{matches:r}=matchMedia("(hover)"),o=H(()=>{let n=new x,i=n.pipe(Rr(1));n.subscribe(({scrollable:c})=>{c&&r?e.setAttribute("tabindex","0"):e.removeAttribute("tabindex")});let s=[];if(Tn.default.isSupported()&&(e.closest(".copy")||G("content.code.copy")&&!e.closest(".no-copy"))){let c=e.closest("pre");c.id=`__code_${Oa++}`;let p=dn(c.id);c.insertBefore(p,e),G("content.tooltips")&&s.push(Be(p))}let a=e.closest(".highlight");if(a instanceof HTMLElement){let c=Sn(a);if(typeof c!="undefined"&&(a.classList.contains("annotate")||G("content.code.annotate"))){let p=fr(c,e,t);s.push(Se(a).pipe(j(i),m(({width:l,height:f})=>l&&f),X(),w(l=>l?p:M)))}}return Ma(e).pipe(T(c=>n.next(c)),A(()=>n.complete()),m(c=>P({ref:e},c)),Re(...s))});return G("content.lazy")?yt(e).pipe(v(n=>n),ue(1),w(()=>o)):o}function La(e,{target$:t,print$:r}){let o=!0;return L(t.pipe(m(n=>n.closest("details:not([open])")),v(n=>e===n),m(()=>({action:"open",reveal:!0}))),r.pipe(v(n=>n||!o),T(()=>o=e.open),m(n=>({action:n?"open":"close"}))))}function Mn(e,t){return H(()=>{let r=new x;return r.subscribe(({action:o,reveal:n})=>{e.toggleAttribute("open",o==="open"),n&&e.scrollIntoView()}),La(e,t).pipe(T(o=>r.next(o)),A(()=>r.complete()),m(o=>P({ref:e},o)))})}var Ln=".node circle,.node ellipse,.node path,.node polygon,.node rect{fill:var(--md-mermaid-node-bg-color);stroke:var(--md-mermaid-node-fg-color)}marker{fill:var(--md-mermaid-edge-color)!important}.edgeLabel .label rect{fill:#0000}.label{color:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.label foreignObject{line-height:normal;overflow:visible}.label div .edgeLabel{color:var(--md-mermaid-label-fg-color)}.edgeLabel,.edgeLabel rect,.label div .edgeLabel{background-color:var(--md-mermaid-label-bg-color)}.edgeLabel,.edgeLabel rect{fill:var(--md-mermaid-label-bg-color);color:var(--md-mermaid-edge-color)}.edgePath .path,.flowchart-link{stroke:var(--md-mermaid-edge-color);stroke-width:.05rem}.edgePath .arrowheadPath{fill:var(--md-mermaid-edge-color);stroke:none}.cluster rect{fill:var(--md-default-fg-color--lightest);stroke:var(--md-default-fg-color--lighter)}.cluster span{color:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}g #flowchart-circleEnd,g #flowchart-circleStart,g #flowchart-crossEnd,g #flowchart-crossStart,g #flowchart-pointEnd,g #flowchart-pointStart{stroke:none}g.classGroup line,g.classGroup rect{fill:var(--md-mermaid-node-bg-color);stroke:var(--md-mermaid-node-fg-color)}g.classGroup text{fill:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.classLabel .box{fill:var(--md-mermaid-label-bg-color);background-color:var(--md-mermaid-label-bg-color);opacity:1}.classLabel .label{fill:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.node .divider{stroke:var(--md-mermaid-node-fg-color)}.relation{stroke:var(--md-mermaid-edge-color)}.cardinality{fill:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.cardinality text{fill:inherit!important}defs #classDiagram-compositionEnd,defs #classDiagram-compositionStart,defs #classDiagram-dependencyEnd,defs #classDiagram-dependencyStart,defs #classDiagram-extensionEnd,defs #classDiagram-extensionStart{fill:var(--md-mermaid-edge-color)!important;stroke:var(--md-mermaid-edge-color)!important}defs #classDiagram-aggregationEnd,defs #classDiagram-aggregationStart{fill:var(--md-mermaid-label-bg-color)!important;stroke:var(--md-mermaid-edge-color)!important}g.stateGroup rect{fill:var(--md-mermaid-node-bg-color);stroke:var(--md-mermaid-node-fg-color)}g.stateGroup .state-title{fill:var(--md-mermaid-label-fg-color)!important;font-family:var(--md-mermaid-font-family)}g.stateGroup .composit{fill:var(--md-mermaid-label-bg-color)}.nodeLabel{color:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.node circle.state-end,.node circle.state-start,.start-state{fill:var(--md-mermaid-edge-color);stroke:none}.end-state-inner,.end-state-outer{fill:var(--md-mermaid-edge-color)}.end-state-inner,.node circle.state-end{stroke:var(--md-mermaid-label-bg-color)}.transition{stroke:var(--md-mermaid-edge-color)}[id^=state-fork] rect,[id^=state-join] rect{fill:var(--md-mermaid-edge-color)!important;stroke:none!important}.statediagram-cluster.statediagram-cluster .inner{fill:var(--md-default-bg-color)}.statediagram-cluster rect{fill:var(--md-mermaid-node-bg-color);stroke:var(--md-mermaid-node-fg-color)}.statediagram-state rect.divider{fill:var(--md-default-fg-color--lightest);stroke:var(--md-default-fg-color--lighter)}defs #statediagram-barbEnd{stroke:var(--md-mermaid-edge-color)}.attributeBoxEven,.attributeBoxOdd{fill:var(--md-mermaid-node-bg-color);stroke:var(--md-mermaid-node-fg-color)}.entityBox{fill:var(--md-mermaid-label-bg-color);stroke:var(--md-mermaid-node-fg-color)}.entityLabel{fill:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.relationshipLabelBox{fill:var(--md-mermaid-label-bg-color);fill-opacity:1;background-color:var(--md-mermaid-label-bg-color);opacity:1}.relationshipLabel{fill:var(--md-mermaid-label-fg-color)}.relationshipLine{stroke:var(--md-mermaid-edge-color)}defs #ONE_OR_MORE_END *,defs #ONE_OR_MORE_START *,defs #ONLY_ONE_END *,defs #ONLY_ONE_START *,defs #ZERO_OR_MORE_END *,defs #ZERO_OR_MORE_START *,defs #ZERO_OR_ONE_END *,defs #ZERO_OR_ONE_START *{stroke:var(--md-mermaid-edge-color)!important}defs #ZERO_OR_MORE_END circle,defs #ZERO_OR_MORE_START circle{fill:var(--md-mermaid-label-bg-color)}.actor{fill:var(--md-mermaid-sequence-actor-bg-color);stroke:var(--md-mermaid-sequence-actor-border-color)}text.actor>tspan{fill:var(--md-mermaid-sequence-actor-fg-color);font-family:var(--md-mermaid-font-family)}line{stroke:var(--md-mermaid-sequence-actor-line-color)}.actor-man circle,.actor-man line{fill:var(--md-mermaid-sequence-actorman-bg-color);stroke:var(--md-mermaid-sequence-actorman-line-color)}.messageLine0,.messageLine1{stroke:var(--md-mermaid-sequence-message-line-color)}.note{fill:var(--md-mermaid-sequence-note-bg-color);stroke:var(--md-mermaid-sequence-note-border-color)}.loopText,.loopText>tspan,.messageText,.noteText>tspan{stroke:none;font-family:var(--md-mermaid-font-family)!important}.messageText{fill:var(--md-mermaid-sequence-message-fg-color)}.loopText,.loopText>tspan{fill:var(--md-mermaid-sequence-loop-fg-color)}.noteText>tspan{fill:var(--md-mermaid-sequence-note-fg-color)}#arrowhead path{fill:var(--md-mermaid-sequence-message-line-color);stroke:none}.loopLine{fill:var(--md-mermaid-sequence-loop-bg-color);stroke:var(--md-mermaid-sequence-loop-border-color)}.labelBox{fill:var(--md-mermaid-sequence-label-bg-color);stroke:none}.labelText,.labelText>span{fill:var(--md-mermaid-sequence-label-fg-color);font-family:var(--md-mermaid-font-family)}.sequenceNumber{fill:var(--md-mermaid-sequence-number-fg-color)}rect.rect{fill:var(--md-mermaid-sequence-box-bg-color);stroke:none}rect.rect+text.text{fill:var(--md-mermaid-sequence-box-fg-color)}defs #sequencenumber{fill:var(--md-mermaid-sequence-number-bg-color)!important}";var Qr,Aa=0;function Ca(){return typeof mermaid=="undefined"||mermaid instanceof Element?gt("https://unpkg.com/mermaid@10.6.1/dist/mermaid.min.js"):R(void 0)}function _n(e){return e.classList.remove("mermaid"),Qr||(Qr=Ca().pipe(T(()=>mermaid.initialize({startOnLoad:!1,themeCSS:Ln,sequence:{actorFontSize:"16px",messageFontSize:"16px",noteFontSize:"16px"}})),m(()=>{}),Z(1))),Qr.subscribe(()=>no(this,null,function*(){e.classList.add("mermaid");let t=`__mermaid_${Aa++}`,r=S("div",{class:"mermaid"}),o=e.textContent,{svg:n,fn:i}=yield mermaid.render(t,o),s=r.attachShadow({mode:"closed"});s.innerHTML=n,e.replaceWith(r),i==null||i(s)})),Qr.pipe(m(()=>({ref:e})))}var An=S("table");function Cn(e){return e.replaceWith(An),An.replaceWith(vn(e)),R({ref:e})}function ka(e){let t=e.find(r=>r.checked)||e[0];return L(...e.map(r=>h(r,"change").pipe(m(()=>U(`label[for="${r.id}"]`))))).pipe(q(U(`label[for="${t.id}"]`)),m(r=>({active:r})))}function kn(e,{viewport$:t,target$:r}){let o=U(".tabbed-labels",e),n=W(":scope > input",e),i=zr("prev");e.append(i);let s=zr("next");return e.append(s),H(()=>{let a=new x,c=a.pipe(ee(),oe(!0));B([a,Se(e)]).pipe(j(c),Le(1,ge)).subscribe({next([{active:p},l]){let f=Ue(p),{width:u}=le(p);e.style.setProperty("--md-indicator-x",`${f.x}px`),e.style.setProperty("--md-indicator-width",`${u}px`);let d=ir(o);(f.xd.x+l.width)&&o.scrollTo({left:Math.max(0,f.x-16),behavior:"smooth"})},complete(){e.style.removeProperty("--md-indicator-x"),e.style.removeProperty("--md-indicator-width")}}),B([et(o),Se(o)]).pipe(j(c)).subscribe(([p,l])=>{let f=xt(o);i.hidden=p.x<16,s.hidden=p.x>f.width-l.width-16}),L(h(i,"click").pipe(m(()=>-1)),h(s,"click").pipe(m(()=>1))).pipe(j(c)).subscribe(p=>{let{width:l}=le(o);o.scrollBy({left:l*p,behavior:"smooth"})}),r.pipe(j(c),v(p=>n.includes(p))).subscribe(p=>p.click()),o.classList.add("tabbed-labels--linked");for(let p of n){let l=U(`label[for="${p.id}"]`);l.replaceChildren(S("a",{href:`#${l.htmlFor}`,tabIndex:-1},...Array.from(l.childNodes))),h(l.firstElementChild,"click").pipe(j(c),v(f=>!(f.metaKey||f.ctrlKey)),T(f=>{f.preventDefault(),f.stopPropagation()})).subscribe(()=>{history.replaceState({},"",`#${l.htmlFor}`),l.click()})}return G("content.tabs.link")&&a.pipe(Ee(1),ae(t)).subscribe(([{active:p},{offset:l}])=>{let f=p.innerText.trim();if(p.hasAttribute("data-md-switching"))p.removeAttribute("data-md-switching");else{let u=e.offsetTop-l.y;for(let y of W("[data-tabs]"))for(let b of W(":scope > input",y)){let D=U(`label[for="${b.id}"]`);if(D!==p&&D.innerText.trim()===f){D.setAttribute("data-md-switching",""),b.click();break}}window.scrollTo({top:e.offsetTop-u});let d=__md_get("__tabs")||[];__md_set("__tabs",[...new Set([f,...d])])}}),a.pipe(j(c)).subscribe(()=>{for(let p of W("audio, video",e))p.pause()}),ka(n).pipe(T(p=>a.next(p)),A(()=>a.complete()),m(p=>P({ref:e},p)))}).pipe(qe(ie))}function Hn(e,{viewport$:t,target$:r,print$:o}){return L(...W(".annotate:not(.highlight)",e).map(n=>wn(n,{target$:r,print$:o})),...W("pre:not(.mermaid) > code",e).map(n=>On(n,{target$:r,print$:o})),...W("pre.mermaid",e).map(n=>_n(n)),...W("table:not([class])",e).map(n=>Cn(n)),...W("details",e).map(n=>Mn(n,{target$:r,print$:o})),...W("[data-tabs]",e).map(n=>kn(n,{viewport$:t,target$:r})),...W("[title]",e).filter(()=>G("content.tooltips")).map(n=>Be(n)))}function Ha(e,{alert$:t}){return t.pipe(w(r=>L(R(!0),R(!1).pipe(Qe(2e3))).pipe(m(o=>({message:r,active:o})))))}function $n(e,t){let r=U(".md-typeset",e);return H(()=>{let o=new x;return o.subscribe(({message:n,active:i})=>{e.classList.toggle("md-dialog--active",i),r.textContent=n}),Ha(e,t).pipe(T(n=>o.next(n)),A(()=>o.complete()),m(n=>P({ref:e},n)))})}function $a({viewport$:e}){if(!G("header.autohide"))return R(!1);let t=e.pipe(m(({offset:{y:n}})=>n),Ce(2,1),m(([n,i])=>[nMath.abs(i-n.y)>100),m(([,[n]])=>n),X()),o=Ne("search");return B([e,o]).pipe(m(([{offset:n},i])=>n.y>400&&!i),X(),w(n=>n?r:R(!1)),q(!1))}function Pn(e,t){return H(()=>B([Se(e),$a(t)])).pipe(m(([{height:r},o])=>({height:r,hidden:o})),X((r,o)=>r.height===o.height&&r.hidden===o.hidden),Z(1))}function Rn(e,{header$:t,main$:r}){return H(()=>{let o=new x,n=o.pipe(ee(),oe(!0));o.pipe(te("active"),Ze(t)).subscribe(([{active:s},{hidden:a}])=>{e.classList.toggle("md-header--shadow",s&&!a),e.hidden=a});let i=fe(W("[title]",e)).pipe(v(()=>G("content.tooltips")),re(s=>Be(s)));return r.subscribe(o),t.pipe(j(n),m(s=>P({ref:e},s)),Re(i.pipe(j(n))))})}function Pa(e,{viewport$:t,header$:r}){return mr(e,{viewport$:t,header$:r}).pipe(m(({offset:{y:o}})=>{let{height:n}=le(e);return{active:o>=n}}),te("active"))}function In(e,t){return H(()=>{let r=new x;r.subscribe({next({active:n}){e.classList.toggle("md-header__title--active",n)},complete(){e.classList.remove("md-header__title--active")}});let o=ce(".md-content h1");return typeof o=="undefined"?M:Pa(o,t).pipe(T(n=>r.next(n)),A(()=>r.complete()),m(n=>P({ref:e},n)))})}function Fn(e,{viewport$:t,header$:r}){let o=r.pipe(m(({height:i})=>i),X()),n=o.pipe(w(()=>Se(e).pipe(m(({height:i})=>({top:e.offsetTop,bottom:e.offsetTop+i})),te("bottom"))));return B([o,n,t]).pipe(m(([i,{top:s,bottom:a},{offset:{y:c},size:{height:p}}])=>(p=Math.max(0,p-Math.max(0,s-c,i)-Math.max(0,p+c-a)),{offset:s-i,height:p,active:s-i<=c})),X((i,s)=>i.offset===s.offset&&i.height===s.height&&i.active===s.active))}function Ra(e){let t=__md_get("__palette")||{index:e.findIndex(r=>matchMedia(r.getAttribute("data-md-color-media")).matches)};return R(...e).pipe(re(r=>h(r,"change").pipe(m(()=>r))),q(e[Math.max(0,t.index)]),m(r=>({index:e.indexOf(r),color:{media:r.getAttribute("data-md-color-media"),scheme:r.getAttribute("data-md-color-scheme"),primary:r.getAttribute("data-md-color-primary"),accent:r.getAttribute("data-md-color-accent")}})),Z(1))}function jn(e){let t=W("input",e),r=S("meta",{name:"theme-color"});document.head.appendChild(r);let o=S("meta",{name:"color-scheme"});document.head.appendChild(o);let n=At("(prefers-color-scheme: light)");return H(()=>{let i=new x;return i.subscribe(s=>{if(document.body.setAttribute("data-md-color-switching",""),s.color.media==="(prefers-color-scheme)"){let a=matchMedia("(prefers-color-scheme: light)"),c=document.querySelector(a.matches?"[data-md-color-media='(prefers-color-scheme: light)']":"[data-md-color-media='(prefers-color-scheme: dark)']");s.color.scheme=c.getAttribute("data-md-color-scheme"),s.color.primary=c.getAttribute("data-md-color-primary"),s.color.accent=c.getAttribute("data-md-color-accent")}for(let[a,c]of Object.entries(s.color))document.body.setAttribute(`data-md-color-${a}`,c);for(let a=0;a{let s=Oe("header"),a=window.getComputedStyle(s);return o.content=a.colorScheme,a.backgroundColor.match(/\d+/g).map(c=>(+c).toString(16).padStart(2,"0")).join("")})).subscribe(s=>r.content=`#${s}`),i.pipe(Me(ie)).subscribe(()=>{document.body.removeAttribute("data-md-color-switching")}),Ra(t).pipe(j(n.pipe(Ee(1))),at(),T(s=>i.next(s)),A(()=>i.complete()),m(s=>P({ref:e},s)))})}function Wn(e,{progress$:t}){return H(()=>{let r=new x;return r.subscribe(({value:o})=>{e.style.setProperty("--md-progress-value",`${o}`)}),t.pipe(T(o=>r.next({value:o})),A(()=>r.complete()),m(o=>({ref:e,value:o})))})}var Yr=jt(Kr());function Ia(e){e.setAttribute("data-md-copying","");let t=e.closest("[data-copy]"),r=t?t.getAttribute("data-copy"):e.innerText;return e.removeAttribute("data-md-copying"),r.trimEnd()}function Un({alert$:e}){Yr.default.isSupported()&&new I(t=>{new Yr.default("[data-clipboard-target], [data-clipboard-text]",{text:r=>r.getAttribute("data-clipboard-text")||Ia(U(r.getAttribute("data-clipboard-target")))}).on("success",r=>t.next(r))}).pipe(T(t=>{t.trigger.focus()}),m(()=>we("clipboard.copied"))).subscribe(e)}function Fa(e){if(e.length<2)return[""];let[t,r]=[...e].sort((n,i)=>n.length-i.length).map(n=>n.replace(/[^/]+$/,"")),o=0;if(t===r)o=t.length;else for(;t.charCodeAt(o)===r.charCodeAt(o);)o++;return e.map(n=>n.replace(t.slice(0,o),""))}function ur(e){let t=__md_get("__sitemap",sessionStorage,e);if(t)return R(t);{let r=he();return on(new URL("sitemap.xml",e||r.base)).pipe(m(o=>Fa(W("loc",o).map(n=>n.textContent))),xe(()=>M),$e([]),T(o=>__md_set("__sitemap",o,sessionStorage,e)))}}function Nn(e){let t=ce("[rel=canonical]",e);typeof t!="undefined"&&(t.href=t.href.replace("//localhost:","//127.0.0.1:"));let r=new Map;for(let o of W(":scope > *",e)){let n=o.outerHTML;for(let i of["href","src"]){let s=o.getAttribute(i);if(s===null)continue;let a=new URL(s,t==null?void 0:t.href),c=o.cloneNode();c.setAttribute(i,`${a}`),n=c.outerHTML;break}r.set(n,o)}return r}function Dn({location$:e,viewport$:t,progress$:r}){let o=he();if(location.protocol==="file:")return M;let n=ur().pipe(m(l=>l.map(f=>`${new URL(f,o.base)}`))),i=h(document.body,"click").pipe(ae(n),w(([l,f])=>{if(!(l.target instanceof Element))return M;let u=l.target.closest("a");if(u===null)return M;if(u.target||l.metaKey||l.ctrlKey)return M;let d=new URL(u.href);return d.search=d.hash="",f.includes(`${d}`)?(l.preventDefault(),R(new URL(u.href))):M}),de());i.pipe(ue(1)).subscribe(()=>{let l=ce("link[rel=icon]");typeof l!="undefined"&&(l.href=l.href)}),h(window,"beforeunload").subscribe(()=>{history.scrollRestoration="auto"}),i.pipe(ae(t)).subscribe(([l,{offset:f}])=>{history.scrollRestoration="manual",history.replaceState(f,""),history.pushState(null,"",l)}),i.subscribe(e);let s=e.pipe(q(me()),te("pathname"),Ee(1),w(l=>lr(l,{progress$:r}).pipe(xe(()=>(st(l,!0),M))))),a=new DOMParser,c=s.pipe(w(l=>l.text()),w(l=>{let f=a.parseFromString(l,"text/html");for(let b of["[data-md-component=announce]","[data-md-component=container]","[data-md-component=header-topic]","[data-md-component=outdated]","[data-md-component=logo]","[data-md-component=skip]",...G("navigation.tabs.sticky")?["[data-md-component=tabs]"]:[]]){let D=ce(b),Q=ce(b,f);typeof D!="undefined"&&typeof Q!="undefined"&&D.replaceWith(Q)}let u=Nn(document.head),d=Nn(f.head);for(let[b,D]of d)D.getAttribute("rel")==="stylesheet"||D.hasAttribute("src")||(u.has(b)?u.delete(b):document.head.appendChild(D));for(let b of u.values())b.getAttribute("rel")==="stylesheet"||b.hasAttribute("src")||b.remove();let y=Oe("container");return We(W("script",y)).pipe(w(b=>{let D=f.createElement("script");if(b.src){for(let Q of b.getAttributeNames())D.setAttribute(Q,b.getAttribute(Q));return b.replaceWith(D),new I(Q=>{D.onload=()=>Q.complete()})}else return D.textContent=b.textContent,b.replaceWith(D),M}),ee(),oe(f))}),de());return h(window,"popstate").pipe(m(me)).subscribe(e),e.pipe(q(me()),Ce(2,1),v(([l,f])=>l.pathname===f.pathname&&l.hash!==f.hash),m(([,l])=>l)).subscribe(l=>{var f,u;history.state!==null||!l.hash?window.scrollTo(0,(u=(f=history.state)==null?void 0:f.y)!=null?u:0):(history.scrollRestoration="auto",pr(l.hash),history.scrollRestoration="manual")}),e.pipe(Ir(i),q(me()),Ce(2,1),v(([l,f])=>l.pathname===f.pathname&&l.hash===f.hash),m(([,l])=>l)).subscribe(l=>{history.scrollRestoration="auto",pr(l.hash),history.scrollRestoration="manual",history.back()}),c.pipe(ae(e)).subscribe(([,l])=>{var f,u;history.state!==null||!l.hash?window.scrollTo(0,(u=(f=history.state)==null?void 0:f.y)!=null?u:0):pr(l.hash)}),t.pipe(te("offset"),ye(100)).subscribe(({offset:l})=>{history.replaceState(l,"")}),c}var qn=jt(zn());function Kn(e){let t=e.separator.split("|").map(n=>n.replace(/(\(\?[!=<][^)]+\))/g,"").length===0?"\uFFFD":n).join("|"),r=new RegExp(t,"img"),o=(n,i,s)=>`${i}${s}`;return n=>{n=n.replace(/[\s*+\-:~^]+/g," ").trim();let i=new RegExp(`(^|${e.separator}|)(${n.replace(/[|\\{}()[\]^$+*?.-]/g,"\\$&").replace(r,"|")})`,"img");return s=>(0,qn.default)(s).replace(i,o).replace(/<\/mark>(\s+)]*>/img,"$1")}}function Ht(e){return e.type===1}function dr(e){return e.type===3}function Qn(e,t){let r=ln(e);return L(R(location.protocol!=="file:"),Ne("search")).pipe(Pe(o=>o),w(()=>t)).subscribe(({config:o,docs:n})=>r.next({type:0,data:{config:o,docs:n,options:{suggest:G("search.suggest")}}})),r}function Yn({document$:e}){let t=he(),r=De(new URL("../versions.json",t.base)).pipe(xe(()=>M)),o=r.pipe(m(n=>{let[,i]=t.base.match(/([^/]+)\/?$/);return n.find(({version:s,aliases:a})=>s===i||a.includes(i))||n[0]}));r.pipe(m(n=>new Map(n.map(i=>[`${new URL(`../${i.version}/`,t.base)}`,i]))),w(n=>h(document.body,"click").pipe(v(i=>!i.metaKey&&!i.ctrlKey),ae(o),w(([i,s])=>{if(i.target instanceof Element){let a=i.target.closest("a");if(a&&!a.target&&n.has(a.href)){let c=a.href;return!i.target.closest(".md-version")&&n.get(c)===s?M:(i.preventDefault(),R(c))}}return M}),w(i=>{let{version:s}=n.get(i);return ur(new URL(i)).pipe(m(a=>{let p=me().href.replace(t.base,"");return a.includes(p.split("#")[0])?new URL(`../${s}/${p}`,t.base):new URL(i)}))})))).subscribe(n=>st(n,!0)),B([r,o]).subscribe(([n,i])=>{U(".md-header__topic").appendChild(gn(n,i))}),e.pipe(w(()=>o)).subscribe(n=>{var s;let i=__md_get("__outdated",sessionStorage);if(i===null){i=!0;let a=((s=t.version)==null?void 0:s.default)||"latest";Array.isArray(a)||(a=[a]);e:for(let c of a)for(let p of n.aliases.concat(n.version))if(new RegExp(c,"i").test(p)){i=!1;break e}__md_set("__outdated",i,sessionStorage)}if(i)for(let a of ne("outdated"))a.hidden=!1})}function Da(e,{worker$:t}){let{searchParams:r}=me();r.has("q")&&(Ye("search",!0),e.value=r.get("q"),e.focus(),Ne("search").pipe(Pe(i=>!i)).subscribe(()=>{let i=me();i.searchParams.delete("q"),history.replaceState({},"",`${i}`)}));let o=vt(e),n=L(t.pipe(Pe(Ht)),h(e,"keyup"),o).pipe(m(()=>e.value),X());return B([n,o]).pipe(m(([i,s])=>({value:i,focus:s})),Z(1))}function Bn(e,{worker$:t}){let r=new x,o=r.pipe(ee(),oe(!0));B([t.pipe(Pe(Ht)),r],(i,s)=>s).pipe(te("value")).subscribe(({value:i})=>t.next({type:2,data:i})),r.pipe(te("focus")).subscribe(({focus:i})=>{i&&Ye("search",i)}),h(e.form,"reset").pipe(j(o)).subscribe(()=>e.focus());let n=U("header [for=__search]");return h(n,"click").subscribe(()=>e.focus()),Da(e,{worker$:t}).pipe(T(i=>r.next(i)),A(()=>r.complete()),m(i=>P({ref:e},i)),Z(1))}function Gn(e,{worker$:t,query$:r}){let o=new x,n=Go(e.parentElement).pipe(v(Boolean)),i=e.parentElement,s=U(":scope > :first-child",e),a=U(":scope > :last-child",e);Ne("search").subscribe(l=>a.setAttribute("role",l?"list":"presentation")),o.pipe(ae(r),Wr(t.pipe(Pe(Ht)))).subscribe(([{items:l},{value:f}])=>{switch(l.length){case 0:s.textContent=f.length?we("search.result.none"):we("search.result.placeholder");break;case 1:s.textContent=we("search.result.one");break;default:let u=ar(l.length);s.textContent=we("search.result.other",u)}});let c=o.pipe(T(()=>a.innerHTML=""),w(({items:l})=>L(R(...l.slice(0,10)),R(...l.slice(10)).pipe(Ce(4),Nr(n),w(([f])=>f)))),m(hn),de());return c.subscribe(l=>a.appendChild(l)),c.pipe(re(l=>{let f=ce("details",l);return typeof f=="undefined"?M:h(f,"toggle").pipe(j(o),m(()=>f))})).subscribe(l=>{l.open===!1&&l.offsetTop<=i.scrollTop&&i.scrollTo({top:l.offsetTop})}),t.pipe(v(dr),m(({data:l})=>l)).pipe(T(l=>o.next(l)),A(()=>o.complete()),m(l=>P({ref:e},l)))}function Va(e,{query$:t}){return t.pipe(m(({value:r})=>{let o=me();return o.hash="",r=r.replace(/\s+/g,"+").replace(/&/g,"%26").replace(/=/g,"%3D"),o.search=`q=${r}`,{url:o}}))}function Jn(e,t){let r=new x,o=r.pipe(ee(),oe(!0));return r.subscribe(({url:n})=>{e.setAttribute("data-clipboard-text",e.href),e.href=`${n}`}),h(e,"click").pipe(j(o)).subscribe(n=>n.preventDefault()),Va(e,t).pipe(T(n=>r.next(n)),A(()=>r.complete()),m(n=>P({ref:e},n)))}function Xn(e,{worker$:t,keyboard$:r}){let o=new x,n=Oe("search-query"),i=L(h(n,"keydown"),h(n,"focus")).pipe(Me(ie),m(()=>n.value),X());return o.pipe(Ze(i),m(([{suggest:a},c])=>{let p=c.split(/([\s-]+)/);if(a!=null&&a.length&&p[p.length-1]){let l=a[a.length-1];l.startsWith(p[p.length-1])&&(p[p.length-1]=l)}else p.length=0;return p})).subscribe(a=>e.innerHTML=a.join("").replace(/\s/g," ")),r.pipe(v(({mode:a})=>a==="search")).subscribe(a=>{switch(a.type){case"ArrowRight":e.innerText.length&&n.selectionStart===n.value.length&&(n.value=e.innerText);break}}),t.pipe(v(dr),m(({data:a})=>a)).pipe(T(a=>o.next(a)),A(()=>o.complete()),m(()=>({ref:e})))}function Zn(e,{index$:t,keyboard$:r}){let o=he();try{let n=Qn(o.search,t),i=Oe("search-query",e),s=Oe("search-result",e);h(e,"click").pipe(v(({target:c})=>c instanceof Element&&!!c.closest("a"))).subscribe(()=>Ye("search",!1)),r.pipe(v(({mode:c})=>c==="search")).subscribe(c=>{let p=Ie();switch(c.type){case"Enter":if(p===i){let l=new Map;for(let f of W(":first-child [href]",s)){let u=f.firstElementChild;l.set(f,parseFloat(u.getAttribute("data-md-score")))}if(l.size){let[[f]]=[...l].sort(([,u],[,d])=>d-u);f.click()}c.claim()}break;case"Escape":case"Tab":Ye("search",!1),i.blur();break;case"ArrowUp":case"ArrowDown":if(typeof p=="undefined")i.focus();else{let l=[i,...W(":not(details) > [href], summary, details[open] [href]",s)],f=Math.max(0,(Math.max(0,l.indexOf(p))+l.length+(c.type==="ArrowUp"?-1:1))%l.length);l[f].focus()}c.claim();break;default:i!==Ie()&&i.focus()}}),r.pipe(v(({mode:c})=>c==="global")).subscribe(c=>{switch(c.type){case"f":case"s":case"/":i.focus(),i.select(),c.claim();break}});let a=Bn(i,{worker$:n});return L(a,Gn(s,{worker$:n,query$:a})).pipe(Re(...ne("search-share",e).map(c=>Jn(c,{query$:a})),...ne("search-suggest",e).map(c=>Xn(c,{worker$:n,keyboard$:r}))))}catch(n){return e.hidden=!0,Ke}}function ei(e,{index$:t,location$:r}){return B([t,r.pipe(q(me()),v(o=>!!o.searchParams.get("h")))]).pipe(m(([o,n])=>Kn(o.config)(n.searchParams.get("h"))),m(o=>{var s;let n=new Map,i=document.createNodeIterator(e,NodeFilter.SHOW_TEXT);for(let a=i.nextNode();a;a=i.nextNode())if((s=a.parentElement)!=null&&s.offsetHeight){let c=a.textContent,p=o(c);p.length>c.length&&n.set(a,p)}for(let[a,c]of n){let{childNodes:p}=S("span",null,c);a.replaceWith(...Array.from(p))}return{ref:e,nodes:n}}))}function za(e,{viewport$:t,main$:r}){let o=e.closest(".md-grid"),n=o.offsetTop-o.parentElement.offsetTop;return B([r,t]).pipe(m(([{offset:i,height:s},{offset:{y:a}}])=>(s=s+Math.min(n,Math.max(0,a-i))-n,{height:s,locked:a>=i+n})),X((i,s)=>i.height===s.height&&i.locked===s.locked))}function Br(e,o){var n=o,{header$:t}=n,r=oo(n,["header$"]);let i=U(".md-sidebar__scrollwrap",e),{y:s}=Ue(i);return H(()=>{let a=new x,c=a.pipe(ee(),oe(!0)),p=a.pipe(Le(0,ge));return p.pipe(ae(t)).subscribe({next([{height:l},{height:f}]){i.style.height=`${l-2*s}px`,e.style.top=`${f}px`},complete(){i.style.height="",e.style.top=""}}),p.pipe(Pe()).subscribe(()=>{for(let l of W(".md-nav__link--active[href]",e)){if(!l.clientHeight)continue;let f=l.closest(".md-sidebar__scrollwrap");if(typeof f!="undefined"){let u=l.offsetTop-f.offsetTop,{height:d}=le(f);f.scrollTo({top:u-d/2})}}}),fe(W("label[tabindex]",e)).pipe(re(l=>h(l,"click").pipe(Me(ie),m(()=>l),j(c)))).subscribe(l=>{let f=U(`[id="${l.htmlFor}"]`);U(`[aria-labelledby="${l.id}"]`).setAttribute("aria-expanded",`${f.checked}`)}),za(e,r).pipe(T(l=>a.next(l)),A(()=>a.complete()),m(l=>P({ref:e},l)))})}function ti(e,t){if(typeof t!="undefined"){let r=`https://api.github.com/repos/${e}/${t}`;return Lt(De(`${r}/releases/latest`).pipe(xe(()=>M),m(o=>({version:o.tag_name})),$e({})),De(r).pipe(xe(()=>M),m(o=>({stars:o.stargazers_count,forks:o.forks_count})),$e({}))).pipe(m(([o,n])=>P(P({},o),n)))}else{let r=`https://api.github.com/users/${e}`;return De(r).pipe(m(o=>({repositories:o.public_repos})),$e({}))}}function ri(e,t){let r=`https://${e}/api/v4/projects/${encodeURIComponent(t)}`;return De(r).pipe(xe(()=>M),m(({star_count:o,forks_count:n})=>({stars:o,forks:n})),$e({}))}function oi(e){let t=e.match(/^.+github\.com\/([^/]+)\/?([^/]+)?/i);if(t){let[,r,o]=t;return ti(r,o)}if(t=e.match(/^.+?([^/]*gitlab[^/]+)\/(.+?)\/?$/i),t){let[,r,o]=t;return ri(r,o)}return M}var qa;function Ka(e){return qa||(qa=H(()=>{let t=__md_get("__source",sessionStorage);if(t)return R(t);if(ne("consent").length){let o=__md_get("__consent");if(!(o&&o.github))return M}return oi(e.href).pipe(T(o=>__md_set("__source",o,sessionStorage)))}).pipe(xe(()=>M),v(t=>Object.keys(t).length>0),m(t=>({facts:t})),Z(1)))}function ni(e){let t=U(":scope > :last-child",e);return H(()=>{let r=new x;return r.subscribe(({facts:o})=>{t.appendChild(bn(o)),t.classList.add("md-source__repository--active")}),Ka(e).pipe(T(o=>r.next(o)),A(()=>r.complete()),m(o=>P({ref:e},o)))})}function Qa(e,{viewport$:t,header$:r}){return Se(document.body).pipe(w(()=>mr(e,{header$:r,viewport$:t})),m(({offset:{y:o}})=>({hidden:o>=10})),te("hidden"))}function ii(e,t){return H(()=>{let r=new x;return r.subscribe({next({hidden:o}){e.hidden=o},complete(){e.hidden=!1}}),(G("navigation.tabs.sticky")?R({hidden:!1}):Qa(e,t)).pipe(T(o=>r.next(o)),A(()=>r.complete()),m(o=>P({ref:e},o)))})}function Ya(e,{viewport$:t,header$:r}){let o=new Map,n=W("[href^=\\#]",e);for(let a of n){let c=decodeURIComponent(a.hash.substring(1)),p=ce(`[id="${c}"]`);typeof p!="undefined"&&o.set(a,p)}let i=r.pipe(te("height"),m(({height:a})=>{let c=Oe("main"),p=U(":scope > :first-child",c);return a+.8*(p.offsetTop-c.offsetTop)}),de());return Se(document.body).pipe(te("height"),w(a=>H(()=>{let c=[];return R([...o].reduce((p,[l,f])=>{for(;c.length&&o.get(c[c.length-1]).tagName>=f.tagName;)c.pop();let u=f.offsetTop;for(;!u&&f.parentElement;)f=f.parentElement,u=f.offsetTop;let d=f.offsetParent;for(;d;d=d.offsetParent)u+=d.offsetTop;return p.set([...c=[...c,l]].reverse(),u)},new Map))}).pipe(m(c=>new Map([...c].sort(([,p],[,l])=>p-l))),Ze(i),w(([c,p])=>t.pipe(Fr(([l,f],{offset:{y:u},size:d})=>{let y=u+d.height>=Math.floor(a.height);for(;f.length;){let[,b]=f[0];if(b-p=u&&!y)f=[l.pop(),...f];else break}return[l,f]},[[],[...c]]),X((l,f)=>l[0]===f[0]&&l[1]===f[1])))))).pipe(m(([a,c])=>({prev:a.map(([p])=>p),next:c.map(([p])=>p)})),q({prev:[],next:[]}),Ce(2,1),m(([a,c])=>a.prev.length{let i=new x,s=i.pipe(ee(),oe(!0));if(i.subscribe(({prev:a,next:c})=>{for(let[p]of c)p.classList.remove("md-nav__link--passed"),p.classList.remove("md-nav__link--active");for(let[p,[l]]of a.entries())l.classList.add("md-nav__link--passed"),l.classList.toggle("md-nav__link--active",p===a.length-1)}),G("toc.follow")){let a=L(t.pipe(ye(1),m(()=>{})),t.pipe(ye(250),m(()=>"smooth")));i.pipe(v(({prev:c})=>c.length>0),Ze(o.pipe(Me(ie))),ae(a)).subscribe(([[{prev:c}],p])=>{let[l]=c[c.length-1];if(l.offsetHeight){let f=sr(l);if(typeof f!="undefined"){let u=l.offsetTop-f.offsetTop,{height:d}=le(f);f.scrollTo({top:u-d/2,behavior:p})}}})}return G("navigation.tracking")&&t.pipe(j(s),te("offset"),ye(250),Ee(1),j(n.pipe(Ee(1))),at({delay:250}),ae(i)).subscribe(([,{prev:a}])=>{let c=me(),p=a[a.length-1];if(p&&p.length){let[l]=p,{hash:f}=new URL(l.href);c.hash!==f&&(c.hash=f,history.replaceState({},"",`${c}`))}else c.hash="",history.replaceState({},"",`${c}`)}),Ya(e,{viewport$:t,header$:r}).pipe(T(a=>i.next(a)),A(()=>i.complete()),m(a=>P({ref:e},a)))})}function Ba(e,{viewport$:t,main$:r,target$:o}){let n=t.pipe(m(({offset:{y:s}})=>s),Ce(2,1),m(([s,a])=>s>a&&a>0),X()),i=r.pipe(m(({active:s})=>s));return B([i,n]).pipe(m(([s,a])=>!(s&&a)),X(),j(o.pipe(Ee(1))),oe(!0),at({delay:250}),m(s=>({hidden:s})))}function si(e,{viewport$:t,header$:r,main$:o,target$:n}){let i=new x,s=i.pipe(ee(),oe(!0));return i.subscribe({next({hidden:a}){e.hidden=a,a?(e.setAttribute("tabindex","-1"),e.blur()):e.removeAttribute("tabindex")},complete(){e.style.top="",e.hidden=!0,e.removeAttribute("tabindex")}}),r.pipe(j(s),te("height")).subscribe(({height:a})=>{e.style.top=`${a+16}px`}),h(e,"click").subscribe(a=>{a.preventDefault(),window.scrollTo({top:0})}),Ba(e,{viewport$:t,main$:o,target$:n}).pipe(T(a=>i.next(a)),A(()=>i.complete()),m(a=>P({ref:e},a)))}function ci({document$:e}){e.pipe(w(()=>W(".md-ellipsis")),re(t=>yt(t).pipe(j(e.pipe(Ee(1))),v(r=>r),m(()=>t),ue(1))),v(t=>t.offsetWidth{let r=t.innerText,o=t.closest("a")||t;return o.title=r,Be(o).pipe(j(e.pipe(Ee(1))),A(()=>o.removeAttribute("title")))})).subscribe(),e.pipe(w(()=>W(".md-status")),re(t=>Be(t))).subscribe()}function pi({document$:e,tablet$:t}){e.pipe(w(()=>W(".md-toggle--indeterminate")),T(r=>{r.indeterminate=!0,r.checked=!1}),re(r=>h(r,"change").pipe(Ur(()=>r.classList.contains("md-toggle--indeterminate")),m(()=>r))),ae(t)).subscribe(([r,o])=>{r.classList.remove("md-toggle--indeterminate"),o&&(r.checked=!1)})}function Ga(){return/(iPad|iPhone|iPod)/.test(navigator.userAgent)}function li({document$:e}){e.pipe(w(()=>W("[data-md-scrollfix]")),T(t=>t.removeAttribute("data-md-scrollfix")),v(Ga),re(t=>h(t,"touchstart").pipe(m(()=>t)))).subscribe(t=>{let r=t.scrollTop;r===0?t.scrollTop=1:r+t.offsetHeight===t.scrollHeight&&(t.scrollTop=r-1)})}function mi({viewport$:e,tablet$:t}){B([Ne("search"),t]).pipe(m(([r,o])=>r&&!o),w(r=>R(r).pipe(Qe(r?400:100))),ae(e)).subscribe(([r,{offset:{y:o}}])=>{if(r)document.body.setAttribute("data-md-scrolllock",""),document.body.style.top=`-${o}px`;else{let n=-1*parseInt(document.body.style.top,10);document.body.removeAttribute("data-md-scrolllock"),document.body.style.top="",n&&window.scrollTo(0,n)}})}Object.entries||(Object.entries=function(e){let t=[];for(let r of Object.keys(e))t.push([r,e[r]]);return t});Object.values||(Object.values=function(e){let t=[];for(let r of Object.keys(e))t.push(e[r]);return t});typeof Element!="undefined"&&(Element.prototype.scrollTo||(Element.prototype.scrollTo=function(e,t){typeof e=="object"?(this.scrollLeft=e.left,this.scrollTop=e.top):(this.scrollLeft=e,this.scrollTop=t)}),Element.prototype.replaceWith||(Element.prototype.replaceWith=function(...e){let t=this.parentNode;if(t){e.length===0&&t.removeChild(this);for(let r=e.length-1;r>=0;r--){let o=e[r];typeof o=="string"?o=document.createTextNode(o):o.parentNode&&o.parentNode.removeChild(o),r?t.insertBefore(this.previousSibling,o):t.replaceChild(o,this)}}}));function Ja(){return location.protocol==="file:"?gt(`${new URL("search/search_index.js",Gr.base)}`).pipe(m(()=>__index),Z(1)):De(new URL("search/search_index.json",Gr.base))}document.documentElement.classList.remove("no-js");document.documentElement.classList.add("js");var rt=zo(),Pt=Zo(),wt=tn(Pt),Jr=Xo(),_e=pn(),hr=At("(min-width: 960px)"),ui=At("(min-width: 1220px)"),di=rn(),Gr=he(),hi=document.forms.namedItem("search")?Ja():Ke,Xr=new x;Un({alert$:Xr});var Zr=new x;G("navigation.instant")&&Dn({location$:Pt,viewport$:_e,progress$:Zr}).subscribe(rt);var fi;((fi=Gr.version)==null?void 0:fi.provider)==="mike"&&Yn({document$:rt});L(Pt,wt).pipe(Qe(125)).subscribe(()=>{Ye("drawer",!1),Ye("search",!1)});Jr.pipe(v(({mode:e})=>e==="global")).subscribe(e=>{switch(e.type){case"p":case",":let t=ce("link[rel=prev]");typeof t!="undefined"&&st(t);break;case"n":case".":let r=ce("link[rel=next]");typeof r!="undefined"&&st(r);break;case"Enter":let o=Ie();o instanceof HTMLLabelElement&&o.click()}});ci({document$:rt});pi({document$:rt,tablet$:hr});li({document$:rt});mi({viewport$:_e,tablet$:hr});var tt=Pn(Oe("header"),{viewport$:_e}),$t=rt.pipe(m(()=>Oe("main")),w(e=>Fn(e,{viewport$:_e,header$:tt})),Z(1)),Xa=L(...ne("consent").map(e=>fn(e,{target$:wt})),...ne("dialog").map(e=>$n(e,{alert$:Xr})),...ne("header").map(e=>Rn(e,{viewport$:_e,header$:tt,main$:$t})),...ne("palette").map(e=>jn(e)),...ne("progress").map(e=>Wn(e,{progress$:Zr})),...ne("search").map(e=>Zn(e,{index$:hi,keyboard$:Jr})),...ne("source").map(e=>ni(e))),Za=H(()=>L(...ne("announce").map(e=>mn(e)),...ne("content").map(e=>Hn(e,{viewport$:_e,target$:wt,print$:di})),...ne("content").map(e=>G("search.highlight")?ei(e,{index$:hi,location$:Pt}):M),...ne("header-title").map(e=>In(e,{viewport$:_e,header$:tt})),...ne("sidebar").map(e=>e.getAttribute("data-md-type")==="navigation"?Dr(ui,()=>Br(e,{viewport$:_e,header$:tt,main$:$t})):Dr(hr,()=>Br(e,{viewport$:_e,header$:tt,main$:$t}))),...ne("tabs").map(e=>ii(e,{viewport$:_e,header$:tt})),...ne("toc").map(e=>ai(e,{viewport$:_e,header$:tt,main$:$t,target$:wt})),...ne("top").map(e=>si(e,{viewport$:_e,header$:tt,main$:$t,target$:wt})))),bi=rt.pipe(w(()=>Za),Re(Xa),Z(1));bi.subscribe();window.document$=rt;window.location$=Pt;window.target$=wt;window.keyboard$=Jr;window.viewport$=_e;window.tablet$=hr;window.screen$=ui;window.print$=di;window.alert$=Xr;window.progress$=Zr;window.component$=bi;})();
+//# sourceMappingURL=bundle.d7c377c4.min.js.map
+
diff --git a/v0.12.0/assets/javascripts/bundle.d7c377c4.min.js.map b/v0.12.0/assets/javascripts/bundle.d7c377c4.min.js.map
new file mode 100644
index 000000000..a57d388af
--- /dev/null
+++ b/v0.12.0/assets/javascripts/bundle.d7c377c4.min.js.map
@@ -0,0 +1,7 @@
+{
+ "version": 3,
+ "sources": ["node_modules/focus-visible/dist/focus-visible.js", "node_modules/clipboard/dist/clipboard.js", "node_modules/escape-html/index.js", "src/templates/assets/javascripts/bundle.ts", "node_modules/rxjs/node_modules/tslib/tslib.es6.js", "node_modules/rxjs/src/internal/util/isFunction.ts", "node_modules/rxjs/src/internal/util/createErrorClass.ts", "node_modules/rxjs/src/internal/util/UnsubscriptionError.ts", "node_modules/rxjs/src/internal/util/arrRemove.ts", "node_modules/rxjs/src/internal/Subscription.ts", "node_modules/rxjs/src/internal/config.ts", "node_modules/rxjs/src/internal/scheduler/timeoutProvider.ts", "node_modules/rxjs/src/internal/util/reportUnhandledError.ts", "node_modules/rxjs/src/internal/util/noop.ts", "node_modules/rxjs/src/internal/NotificationFactories.ts", "node_modules/rxjs/src/internal/util/errorContext.ts", "node_modules/rxjs/src/internal/Subscriber.ts", "node_modules/rxjs/src/internal/symbol/observable.ts", "node_modules/rxjs/src/internal/util/identity.ts", "node_modules/rxjs/src/internal/util/pipe.ts", "node_modules/rxjs/src/internal/Observable.ts", "node_modules/rxjs/src/internal/util/lift.ts", "node_modules/rxjs/src/internal/operators/OperatorSubscriber.ts", "node_modules/rxjs/src/internal/scheduler/animationFrameProvider.ts", "node_modules/rxjs/src/internal/util/ObjectUnsubscribedError.ts", "node_modules/rxjs/src/internal/Subject.ts", "node_modules/rxjs/src/internal/scheduler/dateTimestampProvider.ts", "node_modules/rxjs/src/internal/ReplaySubject.ts", "node_modules/rxjs/src/internal/scheduler/Action.ts", "node_modules/rxjs/src/internal/scheduler/intervalProvider.ts", "node_modules/rxjs/src/internal/scheduler/AsyncAction.ts", "node_modules/rxjs/src/internal/Scheduler.ts", "node_modules/rxjs/src/internal/scheduler/AsyncScheduler.ts", "node_modules/rxjs/src/internal/scheduler/async.ts", "node_modules/rxjs/src/internal/scheduler/AnimationFrameAction.ts", "node_modules/rxjs/src/internal/scheduler/AnimationFrameScheduler.ts", "node_modules/rxjs/src/internal/scheduler/animationFrame.ts", "node_modules/rxjs/src/internal/observable/empty.ts", "node_modules/rxjs/src/internal/util/isScheduler.ts", "node_modules/rxjs/src/internal/util/args.ts", "node_modules/rxjs/src/internal/util/isArrayLike.ts", "node_modules/rxjs/src/internal/util/isPromise.ts", "node_modules/rxjs/src/internal/util/isInteropObservable.ts", "node_modules/rxjs/src/internal/util/isAsyncIterable.ts", "node_modules/rxjs/src/internal/util/throwUnobservableError.ts", "node_modules/rxjs/src/internal/symbol/iterator.ts", "node_modules/rxjs/src/internal/util/isIterable.ts", "node_modules/rxjs/src/internal/util/isReadableStreamLike.ts", "node_modules/rxjs/src/internal/observable/innerFrom.ts", "node_modules/rxjs/src/internal/util/executeSchedule.ts", "node_modules/rxjs/src/internal/operators/observeOn.ts", "node_modules/rxjs/src/internal/operators/subscribeOn.ts", "node_modules/rxjs/src/internal/scheduled/scheduleObservable.ts", "node_modules/rxjs/src/internal/scheduled/schedulePromise.ts", "node_modules/rxjs/src/internal/scheduled/scheduleArray.ts", "node_modules/rxjs/src/internal/scheduled/scheduleIterable.ts", "node_modules/rxjs/src/internal/scheduled/scheduleAsyncIterable.ts", "node_modules/rxjs/src/internal/scheduled/scheduleReadableStreamLike.ts", "node_modules/rxjs/src/internal/scheduled/scheduled.ts", "node_modules/rxjs/src/internal/observable/from.ts", "node_modules/rxjs/src/internal/observable/of.ts", "node_modules/rxjs/src/internal/observable/throwError.ts", "node_modules/rxjs/src/internal/util/EmptyError.ts", "node_modules/rxjs/src/internal/util/isDate.ts", "node_modules/rxjs/src/internal/operators/map.ts", "node_modules/rxjs/src/internal/util/mapOneOrManyArgs.ts", "node_modules/rxjs/src/internal/util/argsArgArrayOrObject.ts", "node_modules/rxjs/src/internal/util/createObject.ts", "node_modules/rxjs/src/internal/observable/combineLatest.ts", "node_modules/rxjs/src/internal/operators/mergeInternals.ts", "node_modules/rxjs/src/internal/operators/mergeMap.ts", "node_modules/rxjs/src/internal/operators/mergeAll.ts", "node_modules/rxjs/src/internal/operators/concatAll.ts", "node_modules/rxjs/src/internal/observable/concat.ts", "node_modules/rxjs/src/internal/observable/defer.ts", "node_modules/rxjs/src/internal/observable/fromEvent.ts", "node_modules/rxjs/src/internal/observable/fromEventPattern.ts", "node_modules/rxjs/src/internal/observable/timer.ts", "node_modules/rxjs/src/internal/observable/merge.ts", "node_modules/rxjs/src/internal/observable/never.ts", "node_modules/rxjs/src/internal/util/argsOrArgArray.ts", "node_modules/rxjs/src/internal/operators/filter.ts", "node_modules/rxjs/src/internal/observable/zip.ts", "node_modules/rxjs/src/internal/operators/audit.ts", "node_modules/rxjs/src/internal/operators/auditTime.ts", "node_modules/rxjs/src/internal/operators/bufferCount.ts", "node_modules/rxjs/src/internal/operators/catchError.ts", "node_modules/rxjs/src/internal/operators/scanInternals.ts", "node_modules/rxjs/src/internal/operators/combineLatest.ts", "node_modules/rxjs/src/internal/operators/combineLatestWith.ts", "node_modules/rxjs/src/internal/operators/debounceTime.ts", "node_modules/rxjs/src/internal/operators/defaultIfEmpty.ts", "node_modules/rxjs/src/internal/operators/take.ts", "node_modules/rxjs/src/internal/operators/ignoreElements.ts", "node_modules/rxjs/src/internal/operators/mapTo.ts", "node_modules/rxjs/src/internal/operators/delayWhen.ts", "node_modules/rxjs/src/internal/operators/delay.ts", "node_modules/rxjs/src/internal/operators/distinctUntilChanged.ts", "node_modules/rxjs/src/internal/operators/distinctUntilKeyChanged.ts", "node_modules/rxjs/src/internal/operators/throwIfEmpty.ts", "node_modules/rxjs/src/internal/operators/endWith.ts", "node_modules/rxjs/src/internal/operators/finalize.ts", "node_modules/rxjs/src/internal/operators/first.ts", "node_modules/rxjs/src/internal/operators/takeLast.ts", "node_modules/rxjs/src/internal/operators/merge.ts", "node_modules/rxjs/src/internal/operators/mergeWith.ts", "node_modules/rxjs/src/internal/operators/repeat.ts", "node_modules/rxjs/src/internal/operators/sample.ts", "node_modules/rxjs/src/internal/operators/scan.ts", "node_modules/rxjs/src/internal/operators/share.ts", "node_modules/rxjs/src/internal/operators/shareReplay.ts", "node_modules/rxjs/src/internal/operators/skip.ts", "node_modules/rxjs/src/internal/operators/skipUntil.ts", "node_modules/rxjs/src/internal/operators/startWith.ts", "node_modules/rxjs/src/internal/operators/switchMap.ts", "node_modules/rxjs/src/internal/operators/takeUntil.ts", "node_modules/rxjs/src/internal/operators/takeWhile.ts", "node_modules/rxjs/src/internal/operators/tap.ts", "node_modules/rxjs/src/internal/operators/throttle.ts", "node_modules/rxjs/src/internal/operators/throttleTime.ts", "node_modules/rxjs/src/internal/operators/withLatestFrom.ts", "node_modules/rxjs/src/internal/operators/zip.ts", "node_modules/rxjs/src/internal/operators/zipWith.ts", "src/templates/assets/javascripts/browser/document/index.ts", "src/templates/assets/javascripts/browser/element/_/index.ts", "src/templates/assets/javascripts/browser/element/focus/index.ts", "src/templates/assets/javascripts/browser/element/hover/index.ts", "src/templates/assets/javascripts/browser/element/offset/_/index.ts", "src/templates/assets/javascripts/browser/element/offset/content/index.ts", "src/templates/assets/javascripts/utilities/h/index.ts", "src/templates/assets/javascripts/utilities/round/index.ts", "src/templates/assets/javascripts/browser/script/index.ts", "src/templates/assets/javascripts/browser/element/size/_/index.ts", "src/templates/assets/javascripts/browser/element/size/content/index.ts", "src/templates/assets/javascripts/browser/element/visibility/index.ts", "src/templates/assets/javascripts/browser/toggle/index.ts", "src/templates/assets/javascripts/browser/keyboard/index.ts", "src/templates/assets/javascripts/browser/location/_/index.ts", "src/templates/assets/javascripts/browser/location/hash/index.ts", "src/templates/assets/javascripts/browser/media/index.ts", "src/templates/assets/javascripts/browser/request/index.ts", "src/templates/assets/javascripts/browser/viewport/offset/index.ts", "src/templates/assets/javascripts/browser/viewport/size/index.ts", "src/templates/assets/javascripts/browser/viewport/_/index.ts", "src/templates/assets/javascripts/browser/viewport/at/index.ts", "src/templates/assets/javascripts/browser/worker/index.ts", "src/templates/assets/javascripts/_/index.ts", "src/templates/assets/javascripts/components/_/index.ts", "src/templates/assets/javascripts/components/announce/index.ts", "src/templates/assets/javascripts/components/consent/index.ts", "src/templates/assets/javascripts/templates/tooltip/index.tsx", "src/templates/assets/javascripts/templates/annotation/index.tsx", "src/templates/assets/javascripts/templates/clipboard/index.tsx", "src/templates/assets/javascripts/templates/search/index.tsx", "src/templates/assets/javascripts/templates/source/index.tsx", "src/templates/assets/javascripts/templates/tabbed/index.tsx", "src/templates/assets/javascripts/templates/table/index.tsx", "src/templates/assets/javascripts/templates/version/index.tsx", "src/templates/assets/javascripts/components/tooltip/index.ts", "src/templates/assets/javascripts/components/content/annotation/_/index.ts", "src/templates/assets/javascripts/components/content/annotation/list/index.ts", "src/templates/assets/javascripts/components/content/annotation/block/index.ts", "src/templates/assets/javascripts/components/content/code/_/index.ts", "src/templates/assets/javascripts/components/content/details/index.ts", "src/templates/assets/javascripts/components/content/mermaid/index.css", "src/templates/assets/javascripts/components/content/mermaid/index.ts", "src/templates/assets/javascripts/components/content/table/index.ts", "src/templates/assets/javascripts/components/content/tabs/index.ts", "src/templates/assets/javascripts/components/content/_/index.ts", "src/templates/assets/javascripts/components/dialog/index.ts", "src/templates/assets/javascripts/components/header/_/index.ts", "src/templates/assets/javascripts/components/header/title/index.ts", "src/templates/assets/javascripts/components/main/index.ts", "src/templates/assets/javascripts/components/palette/index.ts", "src/templates/assets/javascripts/components/progress/index.ts", "src/templates/assets/javascripts/integrations/clipboard/index.ts", "src/templates/assets/javascripts/integrations/sitemap/index.ts", "src/templates/assets/javascripts/integrations/instant/index.ts", "src/templates/assets/javascripts/integrations/search/highlighter/index.ts", "src/templates/assets/javascripts/integrations/search/worker/message/index.ts", "src/templates/assets/javascripts/integrations/search/worker/_/index.ts", "src/templates/assets/javascripts/integrations/version/index.ts", "src/templates/assets/javascripts/components/search/query/index.ts", "src/templates/assets/javascripts/components/search/result/index.ts", "src/templates/assets/javascripts/components/search/share/index.ts", "src/templates/assets/javascripts/components/search/suggest/index.ts", "src/templates/assets/javascripts/components/search/_/index.ts", "src/templates/assets/javascripts/components/search/highlight/index.ts", "src/templates/assets/javascripts/components/sidebar/index.ts", "src/templates/assets/javascripts/components/source/facts/github/index.ts", "src/templates/assets/javascripts/components/source/facts/gitlab/index.ts", "src/templates/assets/javascripts/components/source/facts/_/index.ts", "src/templates/assets/javascripts/components/source/_/index.ts", "src/templates/assets/javascripts/components/tabs/index.ts", "src/templates/assets/javascripts/components/toc/index.ts", "src/templates/assets/javascripts/components/top/index.ts", "src/templates/assets/javascripts/patches/ellipsis/index.ts", "src/templates/assets/javascripts/patches/indeterminate/index.ts", "src/templates/assets/javascripts/patches/scrollfix/index.ts", "src/templates/assets/javascripts/patches/scrolllock/index.ts", "src/templates/assets/javascripts/polyfills/index.ts"],
+ "sourcesContent": ["(function (global, factory) {\n typeof exports === 'object' && typeof module !== 'undefined' ? factory() :\n typeof define === 'function' && define.amd ? define(factory) :\n (factory());\n}(this, (function () { 'use strict';\n\n /**\n * Applies the :focus-visible polyfill at the given scope.\n * A scope in this case is either the top-level Document or a Shadow Root.\n *\n * @param {(Document|ShadowRoot)} scope\n * @see https://github.com/WICG/focus-visible\n */\n function applyFocusVisiblePolyfill(scope) {\n var hadKeyboardEvent = true;\n var hadFocusVisibleRecently = false;\n var hadFocusVisibleRecentlyTimeout = null;\n\n var inputTypesAllowlist = {\n text: true,\n search: true,\n url: true,\n tel: true,\n email: true,\n password: true,\n number: true,\n date: true,\n month: true,\n week: true,\n time: true,\n datetime: true,\n 'datetime-local': true\n };\n\n /**\n * Helper function for legacy browsers and iframes which sometimes focus\n * elements like document, body, and non-interactive SVG.\n * @param {Element} el\n */\n function isValidFocusTarget(el) {\n if (\n el &&\n el !== document &&\n el.nodeName !== 'HTML' &&\n el.nodeName !== 'BODY' &&\n 'classList' in el &&\n 'contains' in el.classList\n ) {\n return true;\n }\n return false;\n }\n\n /**\n * Computes whether the given element should automatically trigger the\n * `focus-visible` class being added, i.e. whether it should always match\n * `:focus-visible` when focused.\n * @param {Element} el\n * @return {boolean}\n */\n function focusTriggersKeyboardModality(el) {\n var type = el.type;\n var tagName = el.tagName;\n\n if (tagName === 'INPUT' && inputTypesAllowlist[type] && !el.readOnly) {\n return true;\n }\n\n if (tagName === 'TEXTAREA' && !el.readOnly) {\n return true;\n }\n\n if (el.isContentEditable) {\n return true;\n }\n\n return false;\n }\n\n /**\n * Add the `focus-visible` class to the given element if it was not added by\n * the author.\n * @param {Element} el\n */\n function addFocusVisibleClass(el) {\n if (el.classList.contains('focus-visible')) {\n return;\n }\n el.classList.add('focus-visible');\n el.setAttribute('data-focus-visible-added', '');\n }\n\n /**\n * Remove the `focus-visible` class from the given element if it was not\n * originally added by the author.\n * @param {Element} el\n */\n function removeFocusVisibleClass(el) {\n if (!el.hasAttribute('data-focus-visible-added')) {\n return;\n }\n el.classList.remove('focus-visible');\n el.removeAttribute('data-focus-visible-added');\n }\n\n /**\n * If the most recent user interaction was via the keyboard;\n * and the key press did not include a meta, alt/option, or control key;\n * then the modality is keyboard. Otherwise, the modality is not keyboard.\n * Apply `focus-visible` to any current active element and keep track\n * of our keyboard modality state with `hadKeyboardEvent`.\n * @param {KeyboardEvent} e\n */\n function onKeyDown(e) {\n if (e.metaKey || e.altKey || e.ctrlKey) {\n return;\n }\n\n if (isValidFocusTarget(scope.activeElement)) {\n addFocusVisibleClass(scope.activeElement);\n }\n\n hadKeyboardEvent = true;\n }\n\n /**\n * If at any point a user clicks with a pointing device, ensure that we change\n * the modality away from keyboard.\n * This avoids the situation where a user presses a key on an already focused\n * element, and then clicks on a different element, focusing it with a\n * pointing device, while we still think we're in keyboard modality.\n * @param {Event} e\n */\n function onPointerDown(e) {\n hadKeyboardEvent = false;\n }\n\n /**\n * On `focus`, add the `focus-visible` class to the target if:\n * - the target received focus as a result of keyboard navigation, or\n * - the event target is an element that will likely require interaction\n * via the keyboard (e.g. a text box)\n * @param {Event} e\n */\n function onFocus(e) {\n // Prevent IE from focusing the document or HTML element.\n if (!isValidFocusTarget(e.target)) {\n return;\n }\n\n if (hadKeyboardEvent || focusTriggersKeyboardModality(e.target)) {\n addFocusVisibleClass(e.target);\n }\n }\n\n /**\n * On `blur`, remove the `focus-visible` class from the target.\n * @param {Event} e\n */\n function onBlur(e) {\n if (!isValidFocusTarget(e.target)) {\n return;\n }\n\n if (\n e.target.classList.contains('focus-visible') ||\n e.target.hasAttribute('data-focus-visible-added')\n ) {\n // To detect a tab/window switch, we look for a blur event followed\n // rapidly by a visibility change.\n // If we don't see a visibility change within 100ms, it's probably a\n // regular focus change.\n hadFocusVisibleRecently = true;\n window.clearTimeout(hadFocusVisibleRecentlyTimeout);\n hadFocusVisibleRecentlyTimeout = window.setTimeout(function() {\n hadFocusVisibleRecently = false;\n }, 100);\n removeFocusVisibleClass(e.target);\n }\n }\n\n /**\n * If the user changes tabs, keep track of whether or not the previously\n * focused element had .focus-visible.\n * @param {Event} e\n */\n function onVisibilityChange(e) {\n if (document.visibilityState === 'hidden') {\n // If the tab becomes active again, the browser will handle calling focus\n // on the element (Safari actually calls it twice).\n // If this tab change caused a blur on an element with focus-visible,\n // re-apply the class when the user switches back to the tab.\n if (hadFocusVisibleRecently) {\n hadKeyboardEvent = true;\n }\n addInitialPointerMoveListeners();\n }\n }\n\n /**\n * Add a group of listeners to detect usage of any pointing devices.\n * These listeners will be added when the polyfill first loads, and anytime\n * the window is blurred, so that they are active when the window regains\n * focus.\n */\n function addInitialPointerMoveListeners() {\n document.addEventListener('mousemove', onInitialPointerMove);\n document.addEventListener('mousedown', onInitialPointerMove);\n document.addEventListener('mouseup', onInitialPointerMove);\n document.addEventListener('pointermove', onInitialPointerMove);\n document.addEventListener('pointerdown', onInitialPointerMove);\n document.addEventListener('pointerup', onInitialPointerMove);\n document.addEventListener('touchmove', onInitialPointerMove);\n document.addEventListener('touchstart', onInitialPointerMove);\n document.addEventListener('touchend', onInitialPointerMove);\n }\n\n function removeInitialPointerMoveListeners() {\n document.removeEventListener('mousemove', onInitialPointerMove);\n document.removeEventListener('mousedown', onInitialPointerMove);\n document.removeEventListener('mouseup', onInitialPointerMove);\n document.removeEventListener('pointermove', onInitialPointerMove);\n document.removeEventListener('pointerdown', onInitialPointerMove);\n document.removeEventListener('pointerup', onInitialPointerMove);\n document.removeEventListener('touchmove', onInitialPointerMove);\n document.removeEventListener('touchstart', onInitialPointerMove);\n document.removeEventListener('touchend', onInitialPointerMove);\n }\n\n /**\n * When the polfyill first loads, assume the user is in keyboard modality.\n * If any event is received from a pointing device (e.g. mouse, pointer,\n * touch), turn off keyboard modality.\n * This accounts for situations where focus enters the page from the URL bar.\n * @param {Event} e\n */\n function onInitialPointerMove(e) {\n // Work around a Safari quirk that fires a mousemove on whenever the\n // window blurs, even if you're tabbing out of the page. \u00AF\\_(\u30C4)_/\u00AF\n if (e.target.nodeName && e.target.nodeName.toLowerCase() === 'html') {\n return;\n }\n\n hadKeyboardEvent = false;\n removeInitialPointerMoveListeners();\n }\n\n // For some kinds of state, we are interested in changes at the global scope\n // only. For example, global pointer input, global key presses and global\n // visibility change should affect the state at every scope:\n document.addEventListener('keydown', onKeyDown, true);\n document.addEventListener('mousedown', onPointerDown, true);\n document.addEventListener('pointerdown', onPointerDown, true);\n document.addEventListener('touchstart', onPointerDown, true);\n document.addEventListener('visibilitychange', onVisibilityChange, true);\n\n addInitialPointerMoveListeners();\n\n // For focus and blur, we specifically care about state changes in the local\n // scope. This is because focus / blur events that originate from within a\n // shadow root are not re-dispatched from the host element if it was already\n // the active element in its own scope:\n scope.addEventListener('focus', onFocus, true);\n scope.addEventListener('blur', onBlur, true);\n\n // We detect that a node is a ShadowRoot by ensuring that it is a\n // DocumentFragment and also has a host property. This check covers native\n // implementation and polyfill implementation transparently. If we only cared\n // about the native implementation, we could just check if the scope was\n // an instance of a ShadowRoot.\n if (scope.nodeType === Node.DOCUMENT_FRAGMENT_NODE && scope.host) {\n // Since a ShadowRoot is a special kind of DocumentFragment, it does not\n // have a root element to add a class to. So, we add this attribute to the\n // host element instead:\n scope.host.setAttribute('data-js-focus-visible', '');\n } else if (scope.nodeType === Node.DOCUMENT_NODE) {\n document.documentElement.classList.add('js-focus-visible');\n document.documentElement.setAttribute('data-js-focus-visible', '');\n }\n }\n\n // It is important to wrap all references to global window and document in\n // these checks to support server-side rendering use cases\n // @see https://github.com/WICG/focus-visible/issues/199\n if (typeof window !== 'undefined' && typeof document !== 'undefined') {\n // Make the polyfill helper globally available. This can be used as a signal\n // to interested libraries that wish to coordinate with the polyfill for e.g.,\n // applying the polyfill to a shadow root:\n window.applyFocusVisiblePolyfill = applyFocusVisiblePolyfill;\n\n // Notify interested libraries of the polyfill's presence, in case the\n // polyfill was loaded lazily:\n var event;\n\n try {\n event = new CustomEvent('focus-visible-polyfill-ready');\n } catch (error) {\n // IE11 does not support using CustomEvent as a constructor directly:\n event = document.createEvent('CustomEvent');\n event.initCustomEvent('focus-visible-polyfill-ready', false, false, {});\n }\n\n window.dispatchEvent(event);\n }\n\n if (typeof document !== 'undefined') {\n // Apply the polyfill to the global document, so that no JavaScript\n // coordination is required to use the polyfill in the top-level document:\n applyFocusVisiblePolyfill(document);\n }\n\n})));\n", "/*!\n * clipboard.js v2.0.11\n * https://clipboardjs.com/\n *\n * Licensed MIT \u00A9 Zeno Rocha\n */\n(function webpackUniversalModuleDefinition(root, factory) {\n\tif(typeof exports === 'object' && typeof module === 'object')\n\t\tmodule.exports = factory();\n\telse if(typeof define === 'function' && define.amd)\n\t\tdefine([], factory);\n\telse if(typeof exports === 'object')\n\t\texports[\"ClipboardJS\"] = factory();\n\telse\n\t\troot[\"ClipboardJS\"] = factory();\n})(this, function() {\nreturn /******/ (function() { // webpackBootstrap\n/******/ \tvar __webpack_modules__ = ({\n\n/***/ 686:\n/***/ (function(__unused_webpack_module, __webpack_exports__, __webpack_require__) {\n\n\"use strict\";\n\n// EXPORTS\n__webpack_require__.d(__webpack_exports__, {\n \"default\": function() { return /* binding */ clipboard; }\n});\n\n// EXTERNAL MODULE: ./node_modules/tiny-emitter/index.js\nvar tiny_emitter = __webpack_require__(279);\nvar tiny_emitter_default = /*#__PURE__*/__webpack_require__.n(tiny_emitter);\n// EXTERNAL MODULE: ./node_modules/good-listener/src/listen.js\nvar listen = __webpack_require__(370);\nvar listen_default = /*#__PURE__*/__webpack_require__.n(listen);\n// EXTERNAL MODULE: ./node_modules/select/src/select.js\nvar src_select = __webpack_require__(817);\nvar select_default = /*#__PURE__*/__webpack_require__.n(src_select);\n;// CONCATENATED MODULE: ./src/common/command.js\n/**\n * Executes a given operation type.\n * @param {String} type\n * @return {Boolean}\n */\nfunction command(type) {\n try {\n return document.execCommand(type);\n } catch (err) {\n return false;\n }\n}\n;// CONCATENATED MODULE: ./src/actions/cut.js\n\n\n/**\n * Cut action wrapper.\n * @param {String|HTMLElement} target\n * @return {String}\n */\n\nvar ClipboardActionCut = function ClipboardActionCut(target) {\n var selectedText = select_default()(target);\n command('cut');\n return selectedText;\n};\n\n/* harmony default export */ var actions_cut = (ClipboardActionCut);\n;// CONCATENATED MODULE: ./src/common/create-fake-element.js\n/**\n * Creates a fake textarea element with a value.\n * @param {String} value\n * @return {HTMLElement}\n */\nfunction createFakeElement(value) {\n var isRTL = document.documentElement.getAttribute('dir') === 'rtl';\n var fakeElement = document.createElement('textarea'); // Prevent zooming on iOS\n\n fakeElement.style.fontSize = '12pt'; // Reset box model\n\n fakeElement.style.border = '0';\n fakeElement.style.padding = '0';\n fakeElement.style.margin = '0'; // Move element out of screen horizontally\n\n fakeElement.style.position = 'absolute';\n fakeElement.style[isRTL ? 'right' : 'left'] = '-9999px'; // Move element to the same position vertically\n\n var yPosition = window.pageYOffset || document.documentElement.scrollTop;\n fakeElement.style.top = \"\".concat(yPosition, \"px\");\n fakeElement.setAttribute('readonly', '');\n fakeElement.value = value;\n return fakeElement;\n}\n;// CONCATENATED MODULE: ./src/actions/copy.js\n\n\n\n/**\n * Create fake copy action wrapper using a fake element.\n * @param {String} target\n * @param {Object} options\n * @return {String}\n */\n\nvar fakeCopyAction = function fakeCopyAction(value, options) {\n var fakeElement = createFakeElement(value);\n options.container.appendChild(fakeElement);\n var selectedText = select_default()(fakeElement);\n command('copy');\n fakeElement.remove();\n return selectedText;\n};\n/**\n * Copy action wrapper.\n * @param {String|HTMLElement} target\n * @param {Object} options\n * @return {String}\n */\n\n\nvar ClipboardActionCopy = function ClipboardActionCopy(target) {\n var options = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : {\n container: document.body\n };\n var selectedText = '';\n\n if (typeof target === 'string') {\n selectedText = fakeCopyAction(target, options);\n } else if (target instanceof HTMLInputElement && !['text', 'search', 'url', 'tel', 'password'].includes(target === null || target === void 0 ? void 0 : target.type)) {\n // If input type doesn't support `setSelectionRange`. Simulate it. https://developer.mozilla.org/en-US/docs/Web/API/HTMLInputElement/setSelectionRange\n selectedText = fakeCopyAction(target.value, options);\n } else {\n selectedText = select_default()(target);\n command('copy');\n }\n\n return selectedText;\n};\n\n/* harmony default export */ var actions_copy = (ClipboardActionCopy);\n;// CONCATENATED MODULE: ./src/actions/default.js\nfunction _typeof(obj) { \"@babel/helpers - typeof\"; if (typeof Symbol === \"function\" && typeof Symbol.iterator === \"symbol\") { _typeof = function _typeof(obj) { return typeof obj; }; } else { _typeof = function _typeof(obj) { return obj && typeof Symbol === \"function\" && obj.constructor === Symbol && obj !== Symbol.prototype ? \"symbol\" : typeof obj; }; } return _typeof(obj); }\n\n\n\n/**\n * Inner function which performs selection from either `text` or `target`\n * properties and then executes copy or cut operations.\n * @param {Object} options\n */\n\nvar ClipboardActionDefault = function ClipboardActionDefault() {\n var options = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : {};\n // Defines base properties passed from constructor.\n var _options$action = options.action,\n action = _options$action === void 0 ? 'copy' : _options$action,\n container = options.container,\n target = options.target,\n text = options.text; // Sets the `action` to be performed which can be either 'copy' or 'cut'.\n\n if (action !== 'copy' && action !== 'cut') {\n throw new Error('Invalid \"action\" value, use either \"copy\" or \"cut\"');\n } // Sets the `target` property using an element that will be have its content copied.\n\n\n if (target !== undefined) {\n if (target && _typeof(target) === 'object' && target.nodeType === 1) {\n if (action === 'copy' && target.hasAttribute('disabled')) {\n throw new Error('Invalid \"target\" attribute. Please use \"readonly\" instead of \"disabled\" attribute');\n }\n\n if (action === 'cut' && (target.hasAttribute('readonly') || target.hasAttribute('disabled'))) {\n throw new Error('Invalid \"target\" attribute. You can\\'t cut text from elements with \"readonly\" or \"disabled\" attributes');\n }\n } else {\n throw new Error('Invalid \"target\" value, use a valid Element');\n }\n } // Define selection strategy based on `text` property.\n\n\n if (text) {\n return actions_copy(text, {\n container: container\n });\n } // Defines which selection strategy based on `target` property.\n\n\n if (target) {\n return action === 'cut' ? actions_cut(target) : actions_copy(target, {\n container: container\n });\n }\n};\n\n/* harmony default export */ var actions_default = (ClipboardActionDefault);\n;// CONCATENATED MODULE: ./src/clipboard.js\nfunction clipboard_typeof(obj) { \"@babel/helpers - typeof\"; if (typeof Symbol === \"function\" && typeof Symbol.iterator === \"symbol\") { clipboard_typeof = function _typeof(obj) { return typeof obj; }; } else { clipboard_typeof = function _typeof(obj) { return obj && typeof Symbol === \"function\" && obj.constructor === Symbol && obj !== Symbol.prototype ? \"symbol\" : typeof obj; }; } return clipboard_typeof(obj); }\n\nfunction _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError(\"Cannot call a class as a function\"); } }\n\nfunction _defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if (\"value\" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } }\n\nfunction _createClass(Constructor, protoProps, staticProps) { if (protoProps) _defineProperties(Constructor.prototype, protoProps); if (staticProps) _defineProperties(Constructor, staticProps); return Constructor; }\n\nfunction _inherits(subClass, superClass) { if (typeof superClass !== \"function\" && superClass !== null) { throw new TypeError(\"Super expression must either be null or a function\"); } subClass.prototype = Object.create(superClass && superClass.prototype, { constructor: { value: subClass, writable: true, configurable: true } }); if (superClass) _setPrototypeOf(subClass, superClass); }\n\nfunction _setPrototypeOf(o, p) { _setPrototypeOf = Object.setPrototypeOf || function _setPrototypeOf(o, p) { o.__proto__ = p; return o; }; return _setPrototypeOf(o, p); }\n\nfunction _createSuper(Derived) { var hasNativeReflectConstruct = _isNativeReflectConstruct(); return function _createSuperInternal() { var Super = _getPrototypeOf(Derived), result; if (hasNativeReflectConstruct) { var NewTarget = _getPrototypeOf(this).constructor; result = Reflect.construct(Super, arguments, NewTarget); } else { result = Super.apply(this, arguments); } return _possibleConstructorReturn(this, result); }; }\n\nfunction _possibleConstructorReturn(self, call) { if (call && (clipboard_typeof(call) === \"object\" || typeof call === \"function\")) { return call; } return _assertThisInitialized(self); }\n\nfunction _assertThisInitialized(self) { if (self === void 0) { throw new ReferenceError(\"this hasn't been initialised - super() hasn't been called\"); } return self; }\n\nfunction _isNativeReflectConstruct() { if (typeof Reflect === \"undefined\" || !Reflect.construct) return false; if (Reflect.construct.sham) return false; if (typeof Proxy === \"function\") return true; try { Date.prototype.toString.call(Reflect.construct(Date, [], function () {})); return true; } catch (e) { return false; } }\n\nfunction _getPrototypeOf(o) { _getPrototypeOf = Object.setPrototypeOf ? Object.getPrototypeOf : function _getPrototypeOf(o) { return o.__proto__ || Object.getPrototypeOf(o); }; return _getPrototypeOf(o); }\n\n\n\n\n\n\n/**\n * Helper function to retrieve attribute value.\n * @param {String} suffix\n * @param {Element} element\n */\n\nfunction getAttributeValue(suffix, element) {\n var attribute = \"data-clipboard-\".concat(suffix);\n\n if (!element.hasAttribute(attribute)) {\n return;\n }\n\n return element.getAttribute(attribute);\n}\n/**\n * Base class which takes one or more elements, adds event listeners to them,\n * and instantiates a new `ClipboardAction` on each click.\n */\n\n\nvar Clipboard = /*#__PURE__*/function (_Emitter) {\n _inherits(Clipboard, _Emitter);\n\n var _super = _createSuper(Clipboard);\n\n /**\n * @param {String|HTMLElement|HTMLCollection|NodeList} trigger\n * @param {Object} options\n */\n function Clipboard(trigger, options) {\n var _this;\n\n _classCallCheck(this, Clipboard);\n\n _this = _super.call(this);\n\n _this.resolveOptions(options);\n\n _this.listenClick(trigger);\n\n return _this;\n }\n /**\n * Defines if attributes would be resolved using internal setter functions\n * or custom functions that were passed in the constructor.\n * @param {Object} options\n */\n\n\n _createClass(Clipboard, [{\n key: \"resolveOptions\",\n value: function resolveOptions() {\n var options = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : {};\n this.action = typeof options.action === 'function' ? options.action : this.defaultAction;\n this.target = typeof options.target === 'function' ? options.target : this.defaultTarget;\n this.text = typeof options.text === 'function' ? options.text : this.defaultText;\n this.container = clipboard_typeof(options.container) === 'object' ? options.container : document.body;\n }\n /**\n * Adds a click event listener to the passed trigger.\n * @param {String|HTMLElement|HTMLCollection|NodeList} trigger\n */\n\n }, {\n key: \"listenClick\",\n value: function listenClick(trigger) {\n var _this2 = this;\n\n this.listener = listen_default()(trigger, 'click', function (e) {\n return _this2.onClick(e);\n });\n }\n /**\n * Defines a new `ClipboardAction` on each click event.\n * @param {Event} e\n */\n\n }, {\n key: \"onClick\",\n value: function onClick(e) {\n var trigger = e.delegateTarget || e.currentTarget;\n var action = this.action(trigger) || 'copy';\n var text = actions_default({\n action: action,\n container: this.container,\n target: this.target(trigger),\n text: this.text(trigger)\n }); // Fires an event based on the copy operation result.\n\n this.emit(text ? 'success' : 'error', {\n action: action,\n text: text,\n trigger: trigger,\n clearSelection: function clearSelection() {\n if (trigger) {\n trigger.focus();\n }\n\n window.getSelection().removeAllRanges();\n }\n });\n }\n /**\n * Default `action` lookup function.\n * @param {Element} trigger\n */\n\n }, {\n key: \"defaultAction\",\n value: function defaultAction(trigger) {\n return getAttributeValue('action', trigger);\n }\n /**\n * Default `target` lookup function.\n * @param {Element} trigger\n */\n\n }, {\n key: \"defaultTarget\",\n value: function defaultTarget(trigger) {\n var selector = getAttributeValue('target', trigger);\n\n if (selector) {\n return document.querySelector(selector);\n }\n }\n /**\n * Allow fire programmatically a copy action\n * @param {String|HTMLElement} target\n * @param {Object} options\n * @returns Text copied.\n */\n\n }, {\n key: \"defaultText\",\n\n /**\n * Default `text` lookup function.\n * @param {Element} trigger\n */\n value: function defaultText(trigger) {\n return getAttributeValue('text', trigger);\n }\n /**\n * Destroy lifecycle.\n */\n\n }, {\n key: \"destroy\",\n value: function destroy() {\n this.listener.destroy();\n }\n }], [{\n key: \"copy\",\n value: function copy(target) {\n var options = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : {\n container: document.body\n };\n return actions_copy(target, options);\n }\n /**\n * Allow fire programmatically a cut action\n * @param {String|HTMLElement} target\n * @returns Text cutted.\n */\n\n }, {\n key: \"cut\",\n value: function cut(target) {\n return actions_cut(target);\n }\n /**\n * Returns the support of the given action, or all actions if no action is\n * given.\n * @param {String} [action]\n */\n\n }, {\n key: \"isSupported\",\n value: function isSupported() {\n var action = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : ['copy', 'cut'];\n var actions = typeof action === 'string' ? [action] : action;\n var support = !!document.queryCommandSupported;\n actions.forEach(function (action) {\n support = support && !!document.queryCommandSupported(action);\n });\n return support;\n }\n }]);\n\n return Clipboard;\n}((tiny_emitter_default()));\n\n/* harmony default export */ var clipboard = (Clipboard);\n\n/***/ }),\n\n/***/ 828:\n/***/ (function(module) {\n\nvar DOCUMENT_NODE_TYPE = 9;\n\n/**\n * A polyfill for Element.matches()\n */\nif (typeof Element !== 'undefined' && !Element.prototype.matches) {\n var proto = Element.prototype;\n\n proto.matches = proto.matchesSelector ||\n proto.mozMatchesSelector ||\n proto.msMatchesSelector ||\n proto.oMatchesSelector ||\n proto.webkitMatchesSelector;\n}\n\n/**\n * Finds the closest parent that matches a selector.\n *\n * @param {Element} element\n * @param {String} selector\n * @return {Function}\n */\nfunction closest (element, selector) {\n while (element && element.nodeType !== DOCUMENT_NODE_TYPE) {\n if (typeof element.matches === 'function' &&\n element.matches(selector)) {\n return element;\n }\n element = element.parentNode;\n }\n}\n\nmodule.exports = closest;\n\n\n/***/ }),\n\n/***/ 438:\n/***/ (function(module, __unused_webpack_exports, __webpack_require__) {\n\nvar closest = __webpack_require__(828);\n\n/**\n * Delegates event to a selector.\n *\n * @param {Element} element\n * @param {String} selector\n * @param {String} type\n * @param {Function} callback\n * @param {Boolean} useCapture\n * @return {Object}\n */\nfunction _delegate(element, selector, type, callback, useCapture) {\n var listenerFn = listener.apply(this, arguments);\n\n element.addEventListener(type, listenerFn, useCapture);\n\n return {\n destroy: function() {\n element.removeEventListener(type, listenerFn, useCapture);\n }\n }\n}\n\n/**\n * Delegates event to a selector.\n *\n * @param {Element|String|Array} [elements]\n * @param {String} selector\n * @param {String} type\n * @param {Function} callback\n * @param {Boolean} useCapture\n * @return {Object}\n */\nfunction delegate(elements, selector, type, callback, useCapture) {\n // Handle the regular Element usage\n if (typeof elements.addEventListener === 'function') {\n return _delegate.apply(null, arguments);\n }\n\n // Handle Element-less usage, it defaults to global delegation\n if (typeof type === 'function') {\n // Use `document` as the first parameter, then apply arguments\n // This is a short way to .unshift `arguments` without running into deoptimizations\n return _delegate.bind(null, document).apply(null, arguments);\n }\n\n // Handle Selector-based usage\n if (typeof elements === 'string') {\n elements = document.querySelectorAll(elements);\n }\n\n // Handle Array-like based usage\n return Array.prototype.map.call(elements, function (element) {\n return _delegate(element, selector, type, callback, useCapture);\n });\n}\n\n/**\n * Finds closest match and invokes callback.\n *\n * @param {Element} element\n * @param {String} selector\n * @param {String} type\n * @param {Function} callback\n * @return {Function}\n */\nfunction listener(element, selector, type, callback) {\n return function(e) {\n e.delegateTarget = closest(e.target, selector);\n\n if (e.delegateTarget) {\n callback.call(element, e);\n }\n }\n}\n\nmodule.exports = delegate;\n\n\n/***/ }),\n\n/***/ 879:\n/***/ (function(__unused_webpack_module, exports) {\n\n/**\n * Check if argument is a HTML element.\n *\n * @param {Object} value\n * @return {Boolean}\n */\nexports.node = function(value) {\n return value !== undefined\n && value instanceof HTMLElement\n && value.nodeType === 1;\n};\n\n/**\n * Check if argument is a list of HTML elements.\n *\n * @param {Object} value\n * @return {Boolean}\n */\nexports.nodeList = function(value) {\n var type = Object.prototype.toString.call(value);\n\n return value !== undefined\n && (type === '[object NodeList]' || type === '[object HTMLCollection]')\n && ('length' in value)\n && (value.length === 0 || exports.node(value[0]));\n};\n\n/**\n * Check if argument is a string.\n *\n * @param {Object} value\n * @return {Boolean}\n */\nexports.string = function(value) {\n return typeof value === 'string'\n || value instanceof String;\n};\n\n/**\n * Check if argument is a function.\n *\n * @param {Object} value\n * @return {Boolean}\n */\nexports.fn = function(value) {\n var type = Object.prototype.toString.call(value);\n\n return type === '[object Function]';\n};\n\n\n/***/ }),\n\n/***/ 370:\n/***/ (function(module, __unused_webpack_exports, __webpack_require__) {\n\nvar is = __webpack_require__(879);\nvar delegate = __webpack_require__(438);\n\n/**\n * Validates all params and calls the right\n * listener function based on its target type.\n *\n * @param {String|HTMLElement|HTMLCollection|NodeList} target\n * @param {String} type\n * @param {Function} callback\n * @return {Object}\n */\nfunction listen(target, type, callback) {\n if (!target && !type && !callback) {\n throw new Error('Missing required arguments');\n }\n\n if (!is.string(type)) {\n throw new TypeError('Second argument must be a String');\n }\n\n if (!is.fn(callback)) {\n throw new TypeError('Third argument must be a Function');\n }\n\n if (is.node(target)) {\n return listenNode(target, type, callback);\n }\n else if (is.nodeList(target)) {\n return listenNodeList(target, type, callback);\n }\n else if (is.string(target)) {\n return listenSelector(target, type, callback);\n }\n else {\n throw new TypeError('First argument must be a String, HTMLElement, HTMLCollection, or NodeList');\n }\n}\n\n/**\n * Adds an event listener to a HTML element\n * and returns a remove listener function.\n *\n * @param {HTMLElement} node\n * @param {String} type\n * @param {Function} callback\n * @return {Object}\n */\nfunction listenNode(node, type, callback) {\n node.addEventListener(type, callback);\n\n return {\n destroy: function() {\n node.removeEventListener(type, callback);\n }\n }\n}\n\n/**\n * Add an event listener to a list of HTML elements\n * and returns a remove listener function.\n *\n * @param {NodeList|HTMLCollection} nodeList\n * @param {String} type\n * @param {Function} callback\n * @return {Object}\n */\nfunction listenNodeList(nodeList, type, callback) {\n Array.prototype.forEach.call(nodeList, function(node) {\n node.addEventListener(type, callback);\n });\n\n return {\n destroy: function() {\n Array.prototype.forEach.call(nodeList, function(node) {\n node.removeEventListener(type, callback);\n });\n }\n }\n}\n\n/**\n * Add an event listener to a selector\n * and returns a remove listener function.\n *\n * @param {String} selector\n * @param {String} type\n * @param {Function} callback\n * @return {Object}\n */\nfunction listenSelector(selector, type, callback) {\n return delegate(document.body, selector, type, callback);\n}\n\nmodule.exports = listen;\n\n\n/***/ }),\n\n/***/ 817:\n/***/ (function(module) {\n\nfunction select(element) {\n var selectedText;\n\n if (element.nodeName === 'SELECT') {\n element.focus();\n\n selectedText = element.value;\n }\n else if (element.nodeName === 'INPUT' || element.nodeName === 'TEXTAREA') {\n var isReadOnly = element.hasAttribute('readonly');\n\n if (!isReadOnly) {\n element.setAttribute('readonly', '');\n }\n\n element.select();\n element.setSelectionRange(0, element.value.length);\n\n if (!isReadOnly) {\n element.removeAttribute('readonly');\n }\n\n selectedText = element.value;\n }\n else {\n if (element.hasAttribute('contenteditable')) {\n element.focus();\n }\n\n var selection = window.getSelection();\n var range = document.createRange();\n\n range.selectNodeContents(element);\n selection.removeAllRanges();\n selection.addRange(range);\n\n selectedText = selection.toString();\n }\n\n return selectedText;\n}\n\nmodule.exports = select;\n\n\n/***/ }),\n\n/***/ 279:\n/***/ (function(module) {\n\nfunction E () {\n // Keep this empty so it's easier to inherit from\n // (via https://github.com/lipsmack from https://github.com/scottcorgan/tiny-emitter/issues/3)\n}\n\nE.prototype = {\n on: function (name, callback, ctx) {\n var e = this.e || (this.e = {});\n\n (e[name] || (e[name] = [])).push({\n fn: callback,\n ctx: ctx\n });\n\n return this;\n },\n\n once: function (name, callback, ctx) {\n var self = this;\n function listener () {\n self.off(name, listener);\n callback.apply(ctx, arguments);\n };\n\n listener._ = callback\n return this.on(name, listener, ctx);\n },\n\n emit: function (name) {\n var data = [].slice.call(arguments, 1);\n var evtArr = ((this.e || (this.e = {}))[name] || []).slice();\n var i = 0;\n var len = evtArr.length;\n\n for (i; i < len; i++) {\n evtArr[i].fn.apply(evtArr[i].ctx, data);\n }\n\n return this;\n },\n\n off: function (name, callback) {\n var e = this.e || (this.e = {});\n var evts = e[name];\n var liveEvents = [];\n\n if (evts && callback) {\n for (var i = 0, len = evts.length; i < len; i++) {\n if (evts[i].fn !== callback && evts[i].fn._ !== callback)\n liveEvents.push(evts[i]);\n }\n }\n\n // Remove event from queue to prevent memory leak\n // Suggested by https://github.com/lazd\n // Ref: https://github.com/scottcorgan/tiny-emitter/commit/c6ebfaa9bc973b33d110a84a307742b7cf94c953#commitcomment-5024910\n\n (liveEvents.length)\n ? e[name] = liveEvents\n : delete e[name];\n\n return this;\n }\n};\n\nmodule.exports = E;\nmodule.exports.TinyEmitter = E;\n\n\n/***/ })\n\n/******/ \t});\n/************************************************************************/\n/******/ \t// The module cache\n/******/ \tvar __webpack_module_cache__ = {};\n/******/ \t\n/******/ \t// The require function\n/******/ \tfunction __webpack_require__(moduleId) {\n/******/ \t\t// Check if module is in cache\n/******/ \t\tif(__webpack_module_cache__[moduleId]) {\n/******/ \t\t\treturn __webpack_module_cache__[moduleId].exports;\n/******/ \t\t}\n/******/ \t\t// Create a new module (and put it into the cache)\n/******/ \t\tvar module = __webpack_module_cache__[moduleId] = {\n/******/ \t\t\t// no module.id needed\n/******/ \t\t\t// no module.loaded needed\n/******/ \t\t\texports: {}\n/******/ \t\t};\n/******/ \t\n/******/ \t\t// Execute the module function\n/******/ \t\t__webpack_modules__[moduleId](module, module.exports, __webpack_require__);\n/******/ \t\n/******/ \t\t// Return the exports of the module\n/******/ \t\treturn module.exports;\n/******/ \t}\n/******/ \t\n/************************************************************************/\n/******/ \t/* webpack/runtime/compat get default export */\n/******/ \t!function() {\n/******/ \t\t// getDefaultExport function for compatibility with non-harmony modules\n/******/ \t\t__webpack_require__.n = function(module) {\n/******/ \t\t\tvar getter = module && module.__esModule ?\n/******/ \t\t\t\tfunction() { return module['default']; } :\n/******/ \t\t\t\tfunction() { return module; };\n/******/ \t\t\t__webpack_require__.d(getter, { a: getter });\n/******/ \t\t\treturn getter;\n/******/ \t\t};\n/******/ \t}();\n/******/ \t\n/******/ \t/* webpack/runtime/define property getters */\n/******/ \t!function() {\n/******/ \t\t// define getter functions for harmony exports\n/******/ \t\t__webpack_require__.d = function(exports, definition) {\n/******/ \t\t\tfor(var key in definition) {\n/******/ \t\t\t\tif(__webpack_require__.o(definition, key) && !__webpack_require__.o(exports, key)) {\n/******/ \t\t\t\t\tObject.defineProperty(exports, key, { enumerable: true, get: definition[key] });\n/******/ \t\t\t\t}\n/******/ \t\t\t}\n/******/ \t\t};\n/******/ \t}();\n/******/ \t\n/******/ \t/* webpack/runtime/hasOwnProperty shorthand */\n/******/ \t!function() {\n/******/ \t\t__webpack_require__.o = function(obj, prop) { return Object.prototype.hasOwnProperty.call(obj, prop); }\n/******/ \t}();\n/******/ \t\n/************************************************************************/\n/******/ \t// module exports must be returned from runtime so entry inlining is disabled\n/******/ \t// startup\n/******/ \t// Load entry module and return exports\n/******/ \treturn __webpack_require__(686);\n/******/ })()\n.default;\n});", "/*!\n * escape-html\n * Copyright(c) 2012-2013 TJ Holowaychuk\n * Copyright(c) 2015 Andreas Lubbe\n * Copyright(c) 2015 Tiancheng \"Timothy\" Gu\n * MIT Licensed\n */\n\n'use strict';\n\n/**\n * Module variables.\n * @private\n */\n\nvar matchHtmlRegExp = /[\"'&<>]/;\n\n/**\n * Module exports.\n * @public\n */\n\nmodule.exports = escapeHtml;\n\n/**\n * Escape special characters in the given string of html.\n *\n * @param {string} string The string to escape for inserting into HTML\n * @return {string}\n * @public\n */\n\nfunction escapeHtml(string) {\n var str = '' + string;\n var match = matchHtmlRegExp.exec(str);\n\n if (!match) {\n return str;\n }\n\n var escape;\n var html = '';\n var index = 0;\n var lastIndex = 0;\n\n for (index = match.index; index < str.length; index++) {\n switch (str.charCodeAt(index)) {\n case 34: // \"\n escape = '"';\n break;\n case 38: // &\n escape = '&';\n break;\n case 39: // '\n escape = ''';\n break;\n case 60: // <\n escape = '<';\n break;\n case 62: // >\n escape = '>';\n break;\n default:\n continue;\n }\n\n if (lastIndex !== index) {\n html += str.substring(lastIndex, index);\n }\n\n lastIndex = index + 1;\n html += escape;\n }\n\n return lastIndex !== index\n ? html + str.substring(lastIndex, index)\n : html;\n}\n", "/*\n * Copyright (c) 2016-2023 Martin Donath \n *\n * Permission is hereby granted, free of charge, to any person obtaining a copy\n * of this software and associated documentation files (the \"Software\"), to\n * deal in the Software without restriction, including without limitation the\n * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or\n * sell copies of the Software, and to permit persons to whom the Software is\n * furnished to do so, subject to the following conditions:\n *\n * The above copyright notice and this permission notice shall be included in\n * all copies or substantial portions of the Software.\n *\n * THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL THE\n * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING\n * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS\n * IN THE SOFTWARE.\n */\n\nimport \"focus-visible\"\n\nimport {\n EMPTY,\n NEVER,\n Observable,\n Subject,\n defer,\n delay,\n filter,\n map,\n merge,\n mergeWith,\n shareReplay,\n switchMap\n} from \"rxjs\"\n\nimport { configuration, feature } from \"./_\"\nimport {\n at,\n getActiveElement,\n getOptionalElement,\n requestJSON,\n setLocation,\n setToggle,\n watchDocument,\n watchKeyboard,\n watchLocation,\n watchLocationTarget,\n watchMedia,\n watchPrint,\n watchScript,\n watchViewport\n} from \"./browser\"\nimport {\n getComponentElement,\n getComponentElements,\n mountAnnounce,\n mountBackToTop,\n mountConsent,\n mountContent,\n mountDialog,\n mountHeader,\n mountHeaderTitle,\n mountPalette,\n mountProgress,\n mountSearch,\n mountSearchHiglight,\n mountSidebar,\n mountSource,\n mountTableOfContents,\n mountTabs,\n watchHeader,\n watchMain\n} from \"./components\"\nimport {\n SearchIndex,\n setupClipboardJS,\n setupInstantNavigation,\n setupVersionSelector\n} from \"./integrations\"\nimport {\n patchEllipsis,\n patchIndeterminate,\n patchScrollfix,\n patchScrolllock\n} from \"./patches\"\nimport \"./polyfills\"\n\n/* ----------------------------------------------------------------------------\n * Functions - @todo refactor\n * ------------------------------------------------------------------------- */\n\n/**\n * Fetch search index\n *\n * @returns Search index observable\n */\nfunction fetchSearchIndex(): Observable {\n if (location.protocol === \"file:\") {\n return watchScript(\n `${new URL(\"search/search_index.js\", config.base)}`\n )\n .pipe(\n // @ts-ignore - @todo fix typings\n map(() => __index),\n shareReplay(1)\n )\n } else {\n return requestJSON(\n new URL(\"search/search_index.json\", config.base)\n )\n }\n}\n\n/* ----------------------------------------------------------------------------\n * Application\n * ------------------------------------------------------------------------- */\n\n/* Yay, JavaScript is available */\ndocument.documentElement.classList.remove(\"no-js\")\ndocument.documentElement.classList.add(\"js\")\n\n/* Set up navigation observables and subjects */\nconst document$ = watchDocument()\nconst location$ = watchLocation()\nconst target$ = watchLocationTarget(location$)\nconst keyboard$ = watchKeyboard()\n\n/* Set up media observables */\nconst viewport$ = watchViewport()\nconst tablet$ = watchMedia(\"(min-width: 960px)\")\nconst screen$ = watchMedia(\"(min-width: 1220px)\")\nconst print$ = watchPrint()\n\n/* Retrieve search index, if search is enabled */\nconst config = configuration()\nconst index$ = document.forms.namedItem(\"search\")\n ? fetchSearchIndex()\n : NEVER\n\n/* Set up Clipboard.js integration */\nconst alert$ = new Subject()\nsetupClipboardJS({ alert$ })\n\n/* Set up progress indicator */\nconst progress$ = new Subject()\n\n/* Set up instant navigation, if enabled */\nif (feature(\"navigation.instant\"))\n setupInstantNavigation({ location$, viewport$, progress$ })\n .subscribe(document$)\n\n/* Set up version selector */\nif (config.version?.provider === \"mike\")\n setupVersionSelector({ document$ })\n\n/* Always close drawer and search on navigation */\nmerge(location$, target$)\n .pipe(\n delay(125)\n )\n .subscribe(() => {\n setToggle(\"drawer\", false)\n setToggle(\"search\", false)\n })\n\n/* Set up global keyboard handlers */\nkeyboard$\n .pipe(\n filter(({ mode }) => mode === \"global\")\n )\n .subscribe(key => {\n switch (key.type) {\n\n /* Go to previous page */\n case \"p\":\n case \",\":\n const prev = getOptionalElement(\"link[rel=prev]\")\n if (typeof prev !== \"undefined\")\n setLocation(prev)\n break\n\n /* Go to next page */\n case \"n\":\n case \".\":\n const next = getOptionalElement(\"link[rel=next]\")\n if (typeof next !== \"undefined\")\n setLocation(next)\n break\n\n /* Expand navigation, see https://bit.ly/3ZjG5io */\n case \"Enter\":\n const active = getActiveElement()\n if (active instanceof HTMLLabelElement)\n active.click()\n }\n })\n\n/* Set up patches */\npatchEllipsis({ document$ })\npatchIndeterminate({ document$, tablet$ })\npatchScrollfix({ document$ })\npatchScrolllock({ viewport$, tablet$ })\n\n/* Set up header and main area observable */\nconst header$ = watchHeader(getComponentElement(\"header\"), { viewport$ })\nconst main$ = document$\n .pipe(\n map(() => getComponentElement(\"main\")),\n switchMap(el => watchMain(el, { viewport$, header$ })),\n shareReplay(1)\n )\n\n/* Set up control component observables */\nconst control$ = merge(\n\n /* Consent */\n ...getComponentElements(\"consent\")\n .map(el => mountConsent(el, { target$ })),\n\n /* Dialog */\n ...getComponentElements(\"dialog\")\n .map(el => mountDialog(el, { alert$ })),\n\n /* Header */\n ...getComponentElements(\"header\")\n .map(el => mountHeader(el, { viewport$, header$, main$ })),\n\n /* Color palette */\n ...getComponentElements(\"palette\")\n .map(el => mountPalette(el)),\n\n /* Progress bar */\n ...getComponentElements(\"progress\")\n .map(el => mountProgress(el, { progress$ })),\n\n /* Search */\n ...getComponentElements(\"search\")\n .map(el => mountSearch(el, { index$, keyboard$ })),\n\n /* Repository information */\n ...getComponentElements(\"source\")\n .map(el => mountSource(el))\n)\n\n/* Set up content component observables */\nconst content$ = defer(() => merge(\n\n /* Announcement bar */\n ...getComponentElements(\"announce\")\n .map(el => mountAnnounce(el)),\n\n /* Content */\n ...getComponentElements(\"content\")\n .map(el => mountContent(el, { viewport$, target$, print$ })),\n\n /* Search highlighting */\n ...getComponentElements(\"content\")\n .map(el => feature(\"search.highlight\")\n ? mountSearchHiglight(el, { index$, location$ })\n : EMPTY\n ),\n\n /* Header title */\n ...getComponentElements(\"header-title\")\n .map(el => mountHeaderTitle(el, { viewport$, header$ })),\n\n /* Sidebar */\n ...getComponentElements(\"sidebar\")\n .map(el => el.getAttribute(\"data-md-type\") === \"navigation\"\n ? at(screen$, () => mountSidebar(el, { viewport$, header$, main$ }))\n : at(tablet$, () => mountSidebar(el, { viewport$, header$, main$ }))\n ),\n\n /* Navigation tabs */\n ...getComponentElements(\"tabs\")\n .map(el => mountTabs(el, { viewport$, header$ })),\n\n /* Table of contents */\n ...getComponentElements(\"toc\")\n .map(el => mountTableOfContents(el, {\n viewport$, header$, main$, target$\n })),\n\n /* Back-to-top button */\n ...getComponentElements(\"top\")\n .map(el => mountBackToTop(el, { viewport$, header$, main$, target$ }))\n))\n\n/* Set up component observables */\nconst component$ = document$\n .pipe(\n switchMap(() => content$),\n mergeWith(control$),\n shareReplay(1)\n )\n\n/* Subscribe to all components */\ncomponent$.subscribe()\n\n/* ----------------------------------------------------------------------------\n * Exports\n * ------------------------------------------------------------------------- */\n\nwindow.document$ = document$ /* Document observable */\nwindow.location$ = location$ /* Location subject */\nwindow.target$ = target$ /* Location target observable */\nwindow.keyboard$ = keyboard$ /* Keyboard observable */\nwindow.viewport$ = viewport$ /* Viewport observable */\nwindow.tablet$ = tablet$ /* Media tablet observable */\nwindow.screen$ = screen$ /* Media screen observable */\nwindow.print$ = print$ /* Media print observable */\nwindow.alert$ = alert$ /* Alert subject */\nwindow.progress$ = progress$ /* Progress indicator subject */\nwindow.component$ = component$ /* Component observable */\n", "/*! *****************************************************************************\r\nCopyright (c) Microsoft Corporation.\r\n\r\nPermission to use, copy, modify, and/or distribute this software for any\r\npurpose with or without fee is hereby granted.\r\n\r\nTHE SOFTWARE IS PROVIDED \"AS IS\" AND THE AUTHOR DISCLAIMS ALL WARRANTIES WITH\r\nREGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY\r\nAND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT,\r\nINDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM\r\nLOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR\r\nOTHER TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR\r\nPERFORMANCE OF THIS SOFTWARE.\r\n***************************************************************************** */\r\n/* global Reflect, Promise */\r\n\r\nvar extendStatics = function(d, b) {\r\n extendStatics = Object.setPrototypeOf ||\r\n ({ __proto__: [] } instanceof Array && function (d, b) { d.__proto__ = b; }) ||\r\n function (d, b) { for (var p in b) if (Object.prototype.hasOwnProperty.call(b, p)) d[p] = b[p]; };\r\n return extendStatics(d, b);\r\n};\r\n\r\nexport function __extends(d, b) {\r\n if (typeof b !== \"function\" && b !== null)\r\n throw new TypeError(\"Class extends value \" + String(b) + \" is not a constructor or null\");\r\n extendStatics(d, b);\r\n function __() { this.constructor = d; }\r\n d.prototype = b === null ? Object.create(b) : (__.prototype = b.prototype, new __());\r\n}\r\n\r\nexport var __assign = function() {\r\n __assign = Object.assign || function __assign(t) {\r\n for (var s, i = 1, n = arguments.length; i < n; i++) {\r\n s = arguments[i];\r\n for (var p in s) if (Object.prototype.hasOwnProperty.call(s, p)) t[p] = s[p];\r\n }\r\n return t;\r\n }\r\n return __assign.apply(this, arguments);\r\n}\r\n\r\nexport function __rest(s, e) {\r\n var t = {};\r\n for (var p in s) if (Object.prototype.hasOwnProperty.call(s, p) && e.indexOf(p) < 0)\r\n t[p] = s[p];\r\n if (s != null && typeof Object.getOwnPropertySymbols === \"function\")\r\n for (var i = 0, p = Object.getOwnPropertySymbols(s); i < p.length; i++) {\r\n if (e.indexOf(p[i]) < 0 && Object.prototype.propertyIsEnumerable.call(s, p[i]))\r\n t[p[i]] = s[p[i]];\r\n }\r\n return t;\r\n}\r\n\r\nexport function __decorate(decorators, target, key, desc) {\r\n var c = arguments.length, r = c < 3 ? target : desc === null ? desc = Object.getOwnPropertyDescriptor(target, key) : desc, d;\r\n if (typeof Reflect === \"object\" && typeof Reflect.decorate === \"function\") r = Reflect.decorate(decorators, target, key, desc);\r\n else for (var i = decorators.length - 1; i >= 0; i--) if (d = decorators[i]) r = (c < 3 ? d(r) : c > 3 ? d(target, key, r) : d(target, key)) || r;\r\n return c > 3 && r && Object.defineProperty(target, key, r), r;\r\n}\r\n\r\nexport function __param(paramIndex, decorator) {\r\n return function (target, key) { decorator(target, key, paramIndex); }\r\n}\r\n\r\nexport function __metadata(metadataKey, metadataValue) {\r\n if (typeof Reflect === \"object\" && typeof Reflect.metadata === \"function\") return Reflect.metadata(metadataKey, metadataValue);\r\n}\r\n\r\nexport function __awaiter(thisArg, _arguments, P, generator) {\r\n function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }\r\n return new (P || (P = Promise))(function (resolve, reject) {\r\n function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }\r\n function rejected(value) { try { step(generator[\"throw\"](value)); } catch (e) { reject(e); } }\r\n function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }\r\n step((generator = generator.apply(thisArg, _arguments || [])).next());\r\n });\r\n}\r\n\r\nexport function __generator(thisArg, body) {\r\n var _ = { label: 0, sent: function() { if (t[0] & 1) throw t[1]; return t[1]; }, trys: [], ops: [] }, f, y, t, g;\r\n return g = { next: verb(0), \"throw\": verb(1), \"return\": verb(2) }, typeof Symbol === \"function\" && (g[Symbol.iterator] = function() { return this; }), g;\r\n function verb(n) { return function (v) { return step([n, v]); }; }\r\n function step(op) {\r\n if (f) throw new TypeError(\"Generator is already executing.\");\r\n while (_) try {\r\n if (f = 1, y && (t = op[0] & 2 ? y[\"return\"] : op[0] ? y[\"throw\"] || ((t = y[\"return\"]) && t.call(y), 0) : y.next) && !(t = t.call(y, op[1])).done) return t;\r\n if (y = 0, t) op = [op[0] & 2, t.value];\r\n switch (op[0]) {\r\n case 0: case 1: t = op; break;\r\n case 4: _.label++; return { value: op[1], done: false };\r\n case 5: _.label++; y = op[1]; op = [0]; continue;\r\n case 7: op = _.ops.pop(); _.trys.pop(); continue;\r\n default:\r\n if (!(t = _.trys, t = t.length > 0 && t[t.length - 1]) && (op[0] === 6 || op[0] === 2)) { _ = 0; continue; }\r\n if (op[0] === 3 && (!t || (op[1] > t[0] && op[1] < t[3]))) { _.label = op[1]; break; }\r\n if (op[0] === 6 && _.label < t[1]) { _.label = t[1]; t = op; break; }\r\n if (t && _.label < t[2]) { _.label = t[2]; _.ops.push(op); break; }\r\n if (t[2]) _.ops.pop();\r\n _.trys.pop(); continue;\r\n }\r\n op = body.call(thisArg, _);\r\n } catch (e) { op = [6, e]; y = 0; } finally { f = t = 0; }\r\n if (op[0] & 5) throw op[1]; return { value: op[0] ? op[1] : void 0, done: true };\r\n }\r\n}\r\n\r\nexport var __createBinding = Object.create ? (function(o, m, k, k2) {\r\n if (k2 === undefined) k2 = k;\r\n Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });\r\n}) : (function(o, m, k, k2) {\r\n if (k2 === undefined) k2 = k;\r\n o[k2] = m[k];\r\n});\r\n\r\nexport function __exportStar(m, o) {\r\n for (var p in m) if (p !== \"default\" && !Object.prototype.hasOwnProperty.call(o, p)) __createBinding(o, m, p);\r\n}\r\n\r\nexport function __values(o) {\r\n var s = typeof Symbol === \"function\" && Symbol.iterator, m = s && o[s], i = 0;\r\n if (m) return m.call(o);\r\n if (o && typeof o.length === \"number\") return {\r\n next: function () {\r\n if (o && i >= o.length) o = void 0;\r\n return { value: o && o[i++], done: !o };\r\n }\r\n };\r\n throw new TypeError(s ? \"Object is not iterable.\" : \"Symbol.iterator is not defined.\");\r\n}\r\n\r\nexport function __read(o, n) {\r\n var m = typeof Symbol === \"function\" && o[Symbol.iterator];\r\n if (!m) return o;\r\n var i = m.call(o), r, ar = [], e;\r\n try {\r\n while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value);\r\n }\r\n catch (error) { e = { error: error }; }\r\n finally {\r\n try {\r\n if (r && !r.done && (m = i[\"return\"])) m.call(i);\r\n }\r\n finally { if (e) throw e.error; }\r\n }\r\n return ar;\r\n}\r\n\r\n/** @deprecated */\r\nexport function __spread() {\r\n for (var ar = [], i = 0; i < arguments.length; i++)\r\n ar = ar.concat(__read(arguments[i]));\r\n return ar;\r\n}\r\n\r\n/** @deprecated */\r\nexport function __spreadArrays() {\r\n for (var s = 0, i = 0, il = arguments.length; i < il; i++) s += arguments[i].length;\r\n for (var r = Array(s), k = 0, i = 0; i < il; i++)\r\n for (var a = arguments[i], j = 0, jl = a.length; j < jl; j++, k++)\r\n r[k] = a[j];\r\n return r;\r\n}\r\n\r\nexport function __spreadArray(to, from, pack) {\r\n if (pack || arguments.length === 2) for (var i = 0, l = from.length, ar; i < l; i++) {\r\n if (ar || !(i in from)) {\r\n if (!ar) ar = Array.prototype.slice.call(from, 0, i);\r\n ar[i] = from[i];\r\n }\r\n }\r\n return to.concat(ar || Array.prototype.slice.call(from));\r\n}\r\n\r\nexport function __await(v) {\r\n return this instanceof __await ? (this.v = v, this) : new __await(v);\r\n}\r\n\r\nexport function __asyncGenerator(thisArg, _arguments, generator) {\r\n if (!Symbol.asyncIterator) throw new TypeError(\"Symbol.asyncIterator is not defined.\");\r\n var g = generator.apply(thisArg, _arguments || []), i, q = [];\r\n return i = {}, verb(\"next\"), verb(\"throw\"), verb(\"return\"), i[Symbol.asyncIterator] = function () { return this; }, i;\r\n function verb(n) { if (g[n]) i[n] = function (v) { return new Promise(function (a, b) { q.push([n, v, a, b]) > 1 || resume(n, v); }); }; }\r\n function resume(n, v) { try { step(g[n](v)); } catch (e) { settle(q[0][3], e); } }\r\n function step(r) { r.value instanceof __await ? Promise.resolve(r.value.v).then(fulfill, reject) : settle(q[0][2], r); }\r\n function fulfill(value) { resume(\"next\", value); }\r\n function reject(value) { resume(\"throw\", value); }\r\n function settle(f, v) { if (f(v), q.shift(), q.length) resume(q[0][0], q[0][1]); }\r\n}\r\n\r\nexport function __asyncDelegator(o) {\r\n var i, p;\r\n return i = {}, verb(\"next\"), verb(\"throw\", function (e) { throw e; }), verb(\"return\"), i[Symbol.iterator] = function () { return this; }, i;\r\n function verb(n, f) { i[n] = o[n] ? function (v) { return (p = !p) ? { value: __await(o[n](v)), done: n === \"return\" } : f ? f(v) : v; } : f; }\r\n}\r\n\r\nexport function __asyncValues(o) {\r\n if (!Symbol.asyncIterator) throw new TypeError(\"Symbol.asyncIterator is not defined.\");\r\n var m = o[Symbol.asyncIterator], i;\r\n return m ? m.call(o) : (o = typeof __values === \"function\" ? __values(o) : o[Symbol.iterator](), i = {}, verb(\"next\"), verb(\"throw\"), verb(\"return\"), i[Symbol.asyncIterator] = function () { return this; }, i);\r\n function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; }\r\n function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); }\r\n}\r\n\r\nexport function __makeTemplateObject(cooked, raw) {\r\n if (Object.defineProperty) { Object.defineProperty(cooked, \"raw\", { value: raw }); } else { cooked.raw = raw; }\r\n return cooked;\r\n};\r\n\r\nvar __setModuleDefault = Object.create ? (function(o, v) {\r\n Object.defineProperty(o, \"default\", { enumerable: true, value: v });\r\n}) : function(o, v) {\r\n o[\"default\"] = v;\r\n};\r\n\r\nexport function __importStar(mod) {\r\n if (mod && mod.__esModule) return mod;\r\n var result = {};\r\n if (mod != null) for (var k in mod) if (k !== \"default\" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);\r\n __setModuleDefault(result, mod);\r\n return result;\r\n}\r\n\r\nexport function __importDefault(mod) {\r\n return (mod && mod.__esModule) ? mod : { default: mod };\r\n}\r\n\r\nexport function __classPrivateFieldGet(receiver, state, kind, f) {\r\n if (kind === \"a\" && !f) throw new TypeError(\"Private accessor was defined without a getter\");\r\n if (typeof state === \"function\" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError(\"Cannot read private member from an object whose class did not declare it\");\r\n return kind === \"m\" ? f : kind === \"a\" ? f.call(receiver) : f ? f.value : state.get(receiver);\r\n}\r\n\r\nexport function __classPrivateFieldSet(receiver, state, value, kind, f) {\r\n if (kind === \"m\") throw new TypeError(\"Private method is not writable\");\r\n if (kind === \"a\" && !f) throw new TypeError(\"Private accessor was defined without a setter\");\r\n if (typeof state === \"function\" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError(\"Cannot write private member to an object whose class did not declare it\");\r\n return (kind === \"a\" ? f.call(receiver, value) : f ? f.value = value : state.set(receiver, value)), value;\r\n}\r\n", "/**\n * Returns true if the object is a function.\n * @param value The value to check\n */\nexport function isFunction(value: any): value is (...args: any[]) => any {\n return typeof value === 'function';\n}\n", "/**\n * Used to create Error subclasses until the community moves away from ES5.\n *\n * This is because compiling from TypeScript down to ES5 has issues with subclassing Errors\n * as well as other built-in types: https://github.com/Microsoft/TypeScript/issues/12123\n *\n * @param createImpl A factory function to create the actual constructor implementation. The returned\n * function should be a named function that calls `_super` internally.\n */\nexport function createErrorClass(createImpl: (_super: any) => any): T {\n const _super = (instance: any) => {\n Error.call(instance);\n instance.stack = new Error().stack;\n };\n\n const ctorFunc = createImpl(_super);\n ctorFunc.prototype = Object.create(Error.prototype);\n ctorFunc.prototype.constructor = ctorFunc;\n return ctorFunc;\n}\n", "import { createErrorClass } from './createErrorClass';\n\nexport interface UnsubscriptionError extends Error {\n readonly errors: any[];\n}\n\nexport interface UnsubscriptionErrorCtor {\n /**\n * @deprecated Internal implementation detail. Do not construct error instances.\n * Cannot be tagged as internal: https://github.com/ReactiveX/rxjs/issues/6269\n */\n new (errors: any[]): UnsubscriptionError;\n}\n\n/**\n * An error thrown when one or more errors have occurred during the\n * `unsubscribe` of a {@link Subscription}.\n */\nexport const UnsubscriptionError: UnsubscriptionErrorCtor = createErrorClass(\n (_super) =>\n function UnsubscriptionErrorImpl(this: any, errors: (Error | string)[]) {\n _super(this);\n this.message = errors\n ? `${errors.length} errors occurred during unsubscription:\n${errors.map((err, i) => `${i + 1}) ${err.toString()}`).join('\\n ')}`\n : '';\n this.name = 'UnsubscriptionError';\n this.errors = errors;\n }\n);\n", "/**\n * Removes an item from an array, mutating it.\n * @param arr The array to remove the item from\n * @param item The item to remove\n */\nexport function arrRemove(arr: T[] | undefined | null, item: T) {\n if (arr) {\n const index = arr.indexOf(item);\n 0 <= index && arr.splice(index, 1);\n }\n}\n", "import { isFunction } from './util/isFunction';\nimport { UnsubscriptionError } from './util/UnsubscriptionError';\nimport { SubscriptionLike, TeardownLogic, Unsubscribable } from './types';\nimport { arrRemove } from './util/arrRemove';\n\n/**\n * Represents a disposable resource, such as the execution of an Observable. A\n * Subscription has one important method, `unsubscribe`, that takes no argument\n * and just disposes the resource held by the subscription.\n *\n * Additionally, subscriptions may be grouped together through the `add()`\n * method, which will attach a child Subscription to the current Subscription.\n * When a Subscription is unsubscribed, all its children (and its grandchildren)\n * will be unsubscribed as well.\n *\n * @class Subscription\n */\nexport class Subscription implements SubscriptionLike {\n /** @nocollapse */\n public static EMPTY = (() => {\n const empty = new Subscription();\n empty.closed = true;\n return empty;\n })();\n\n /**\n * A flag to indicate whether this Subscription has already been unsubscribed.\n */\n public closed = false;\n\n private _parentage: Subscription[] | Subscription | null = null;\n\n /**\n * The list of registered finalizers to execute upon unsubscription. Adding and removing from this\n * list occurs in the {@link #add} and {@link #remove} methods.\n */\n private _finalizers: Exclude[] | null = null;\n\n /**\n * @param initialTeardown A function executed first as part of the finalization\n * process that is kicked off when {@link #unsubscribe} is called.\n */\n constructor(private initialTeardown?: () => void) {}\n\n /**\n * Disposes the resources held by the subscription. May, for instance, cancel\n * an ongoing Observable execution or cancel any other type of work that\n * started when the Subscription was created.\n * @return {void}\n */\n unsubscribe(): void {\n let errors: any[] | undefined;\n\n if (!this.closed) {\n this.closed = true;\n\n // Remove this from it's parents.\n const { _parentage } = this;\n if (_parentage) {\n this._parentage = null;\n if (Array.isArray(_parentage)) {\n for (const parent of _parentage) {\n parent.remove(this);\n }\n } else {\n _parentage.remove(this);\n }\n }\n\n const { initialTeardown: initialFinalizer } = this;\n if (isFunction(initialFinalizer)) {\n try {\n initialFinalizer();\n } catch (e) {\n errors = e instanceof UnsubscriptionError ? e.errors : [e];\n }\n }\n\n const { _finalizers } = this;\n if (_finalizers) {\n this._finalizers = null;\n for (const finalizer of _finalizers) {\n try {\n execFinalizer(finalizer);\n } catch (err) {\n errors = errors ?? [];\n if (err instanceof UnsubscriptionError) {\n errors = [...errors, ...err.errors];\n } else {\n errors.push(err);\n }\n }\n }\n }\n\n if (errors) {\n throw new UnsubscriptionError(errors);\n }\n }\n }\n\n /**\n * Adds a finalizer to this subscription, so that finalization will be unsubscribed/called\n * when this subscription is unsubscribed. If this subscription is already {@link #closed},\n * because it has already been unsubscribed, then whatever finalizer is passed to it\n * will automatically be executed (unless the finalizer itself is also a closed subscription).\n *\n * Closed Subscriptions cannot be added as finalizers to any subscription. Adding a closed\n * subscription to a any subscription will result in no operation. (A noop).\n *\n * Adding a subscription to itself, or adding `null` or `undefined` will not perform any\n * operation at all. (A noop).\n *\n * `Subscription` instances that are added to this instance will automatically remove themselves\n * if they are unsubscribed. Functions and {@link Unsubscribable} objects that you wish to remove\n * will need to be removed manually with {@link #remove}\n *\n * @param teardown The finalization logic to add to this subscription.\n */\n add(teardown: TeardownLogic): void {\n // Only add the finalizer if it's not undefined\n // and don't add a subscription to itself.\n if (teardown && teardown !== this) {\n if (this.closed) {\n // If this subscription is already closed,\n // execute whatever finalizer is handed to it automatically.\n execFinalizer(teardown);\n } else {\n if (teardown instanceof Subscription) {\n // We don't add closed subscriptions, and we don't add the same subscription\n // twice. Subscription unsubscribe is idempotent.\n if (teardown.closed || teardown._hasParent(this)) {\n return;\n }\n teardown._addParent(this);\n }\n (this._finalizers = this._finalizers ?? []).push(teardown);\n }\n }\n }\n\n /**\n * Checks to see if a this subscription already has a particular parent.\n * This will signal that this subscription has already been added to the parent in question.\n * @param parent the parent to check for\n */\n private _hasParent(parent: Subscription) {\n const { _parentage } = this;\n return _parentage === parent || (Array.isArray(_parentage) && _parentage.includes(parent));\n }\n\n /**\n * Adds a parent to this subscription so it can be removed from the parent if it\n * unsubscribes on it's own.\n *\n * NOTE: THIS ASSUMES THAT {@link _hasParent} HAS ALREADY BEEN CHECKED.\n * @param parent The parent subscription to add\n */\n private _addParent(parent: Subscription) {\n const { _parentage } = this;\n this._parentage = Array.isArray(_parentage) ? (_parentage.push(parent), _parentage) : _parentage ? [_parentage, parent] : parent;\n }\n\n /**\n * Called on a child when it is removed via {@link #remove}.\n * @param parent The parent to remove\n */\n private _removeParent(parent: Subscription) {\n const { _parentage } = this;\n if (_parentage === parent) {\n this._parentage = null;\n } else if (Array.isArray(_parentage)) {\n arrRemove(_parentage, parent);\n }\n }\n\n /**\n * Removes a finalizer from this subscription that was previously added with the {@link #add} method.\n *\n * Note that `Subscription` instances, when unsubscribed, will automatically remove themselves\n * from every other `Subscription` they have been added to. This means that using the `remove` method\n * is not a common thing and should be used thoughtfully.\n *\n * If you add the same finalizer instance of a function or an unsubscribable object to a `Subscription` instance\n * more than once, you will need to call `remove` the same number of times to remove all instances.\n *\n * All finalizer instances are removed to free up memory upon unsubscription.\n *\n * @param teardown The finalizer to remove from this subscription\n */\n remove(teardown: Exclude): void {\n const { _finalizers } = this;\n _finalizers && arrRemove(_finalizers, teardown);\n\n if (teardown instanceof Subscription) {\n teardown._removeParent(this);\n }\n }\n}\n\nexport const EMPTY_SUBSCRIPTION = Subscription.EMPTY;\n\nexport function isSubscription(value: any): value is Subscription {\n return (\n value instanceof Subscription ||\n (value && 'closed' in value && isFunction(value.remove) && isFunction(value.add) && isFunction(value.unsubscribe))\n );\n}\n\nfunction execFinalizer(finalizer: Unsubscribable | (() => void)) {\n if (isFunction(finalizer)) {\n finalizer();\n } else {\n finalizer.unsubscribe();\n }\n}\n", "import { Subscriber } from './Subscriber';\nimport { ObservableNotification } from './types';\n\n/**\n * The {@link GlobalConfig} object for RxJS. It is used to configure things\n * like how to react on unhandled errors.\n */\nexport const config: GlobalConfig = {\n onUnhandledError: null,\n onStoppedNotification: null,\n Promise: undefined,\n useDeprecatedSynchronousErrorHandling: false,\n useDeprecatedNextContext: false,\n};\n\n/**\n * The global configuration object for RxJS, used to configure things\n * like how to react on unhandled errors. Accessible via {@link config}\n * object.\n */\nexport interface GlobalConfig {\n /**\n * A registration point for unhandled errors from RxJS. These are errors that\n * cannot were not handled by consuming code in the usual subscription path. For\n * example, if you have this configured, and you subscribe to an observable without\n * providing an error handler, errors from that subscription will end up here. This\n * will _always_ be called asynchronously on another job in the runtime. This is because\n * we do not want errors thrown in this user-configured handler to interfere with the\n * behavior of the library.\n */\n onUnhandledError: ((err: any) => void) | null;\n\n /**\n * A registration point for notifications that cannot be sent to subscribers because they\n * have completed, errored or have been explicitly unsubscribed. By default, next, complete\n * and error notifications sent to stopped subscribers are noops. However, sometimes callers\n * might want a different behavior. For example, with sources that attempt to report errors\n * to stopped subscribers, a caller can configure RxJS to throw an unhandled error instead.\n * This will _always_ be called asynchronously on another job in the runtime. This is because\n * we do not want errors thrown in this user-configured handler to interfere with the\n * behavior of the library.\n */\n onStoppedNotification: ((notification: ObservableNotification, subscriber: Subscriber) => void) | null;\n\n /**\n * The promise constructor used by default for {@link Observable#toPromise toPromise} and {@link Observable#forEach forEach}\n * methods.\n *\n * @deprecated As of version 8, RxJS will no longer support this sort of injection of a\n * Promise constructor. If you need a Promise implementation other than native promises,\n * please polyfill/patch Promise as you see appropriate. Will be removed in v8.\n */\n Promise?: PromiseConstructorLike;\n\n /**\n * If true, turns on synchronous error rethrowing, which is a deprecated behavior\n * in v6 and higher. This behavior enables bad patterns like wrapping a subscribe\n * call in a try/catch block. It also enables producer interference, a nasty bug\n * where a multicast can be broken for all observers by a downstream consumer with\n * an unhandled error. DO NOT USE THIS FLAG UNLESS IT'S NEEDED TO BUY TIME\n * FOR MIGRATION REASONS.\n *\n * @deprecated As of version 8, RxJS will no longer support synchronous throwing\n * of unhandled errors. All errors will be thrown on a separate call stack to prevent bad\n * behaviors described above. Will be removed in v8.\n */\n useDeprecatedSynchronousErrorHandling: boolean;\n\n /**\n * If true, enables an as-of-yet undocumented feature from v5: The ability to access\n * `unsubscribe()` via `this` context in `next` functions created in observers passed\n * to `subscribe`.\n *\n * This is being removed because the performance was severely problematic, and it could also cause\n * issues when types other than POJOs are passed to subscribe as subscribers, as they will likely have\n * their `this` context overwritten.\n *\n * @deprecated As of version 8, RxJS will no longer support altering the\n * context of next functions provided as part of an observer to Subscribe. Instead,\n * you will have access to a subscription or a signal or token that will allow you to do things like\n * unsubscribe and test closed status. Will be removed in v8.\n */\n useDeprecatedNextContext: boolean;\n}\n", "import type { TimerHandle } from './timerHandle';\ntype SetTimeoutFunction = (handler: () => void, timeout?: number, ...args: any[]) => TimerHandle;\ntype ClearTimeoutFunction = (handle: TimerHandle) => void;\n\ninterface TimeoutProvider {\n setTimeout: SetTimeoutFunction;\n clearTimeout: ClearTimeoutFunction;\n delegate:\n | {\n setTimeout: SetTimeoutFunction;\n clearTimeout: ClearTimeoutFunction;\n }\n | undefined;\n}\n\nexport const timeoutProvider: TimeoutProvider = {\n // When accessing the delegate, use the variable rather than `this` so that\n // the functions can be called without being bound to the provider.\n setTimeout(handler: () => void, timeout?: number, ...args) {\n const { delegate } = timeoutProvider;\n if (delegate?.setTimeout) {\n return delegate.setTimeout(handler, timeout, ...args);\n }\n return setTimeout(handler, timeout, ...args);\n },\n clearTimeout(handle) {\n const { delegate } = timeoutProvider;\n return (delegate?.clearTimeout || clearTimeout)(handle as any);\n },\n delegate: undefined,\n};\n", "import { config } from '../config';\nimport { timeoutProvider } from '../scheduler/timeoutProvider';\n\n/**\n * Handles an error on another job either with the user-configured {@link onUnhandledError},\n * or by throwing it on that new job so it can be picked up by `window.onerror`, `process.on('error')`, etc.\n *\n * This should be called whenever there is an error that is out-of-band with the subscription\n * or when an error hits a terminal boundary of the subscription and no error handler was provided.\n *\n * @param err the error to report\n */\nexport function reportUnhandledError(err: any) {\n timeoutProvider.setTimeout(() => {\n const { onUnhandledError } = config;\n if (onUnhandledError) {\n // Execute the user-configured error handler.\n onUnhandledError(err);\n } else {\n // Throw so it is picked up by the runtime's uncaught error mechanism.\n throw err;\n }\n });\n}\n", "/* tslint:disable:no-empty */\nexport function noop() { }\n", "import { CompleteNotification, NextNotification, ErrorNotification } from './types';\n\n/**\n * A completion object optimized for memory use and created to be the\n * same \"shape\" as other notifications in v8.\n * @internal\n */\nexport const COMPLETE_NOTIFICATION = (() => createNotification('C', undefined, undefined) as CompleteNotification)();\n\n/**\n * Internal use only. Creates an optimized error notification that is the same \"shape\"\n * as other notifications.\n * @internal\n */\nexport function errorNotification(error: any): ErrorNotification {\n return createNotification('E', undefined, error) as any;\n}\n\n/**\n * Internal use only. Creates an optimized next notification that is the same \"shape\"\n * as other notifications.\n * @internal\n */\nexport function nextNotification(value: T) {\n return createNotification('N', value, undefined) as NextNotification;\n}\n\n/**\n * Ensures that all notifications created internally have the same \"shape\" in v8.\n *\n * TODO: This is only exported to support a crazy legacy test in `groupBy`.\n * @internal\n */\nexport function createNotification(kind: 'N' | 'E' | 'C', value: any, error: any) {\n return {\n kind,\n value,\n error,\n };\n}\n", "import { config } from '../config';\n\nlet context: { errorThrown: boolean; error: any } | null = null;\n\n/**\n * Handles dealing with errors for super-gross mode. Creates a context, in which\n * any synchronously thrown errors will be passed to {@link captureError}. Which\n * will record the error such that it will be rethrown after the call back is complete.\n * TODO: Remove in v8\n * @param cb An immediately executed function.\n */\nexport function errorContext(cb: () => void) {\n if (config.useDeprecatedSynchronousErrorHandling) {\n const isRoot = !context;\n if (isRoot) {\n context = { errorThrown: false, error: null };\n }\n cb();\n if (isRoot) {\n const { errorThrown, error } = context!;\n context = null;\n if (errorThrown) {\n throw error;\n }\n }\n } else {\n // This is the general non-deprecated path for everyone that\n // isn't crazy enough to use super-gross mode (useDeprecatedSynchronousErrorHandling)\n cb();\n }\n}\n\n/**\n * Captures errors only in super-gross mode.\n * @param err the error to capture\n */\nexport function captureError(err: any) {\n if (config.useDeprecatedSynchronousErrorHandling && context) {\n context.errorThrown = true;\n context.error = err;\n }\n}\n", "import { isFunction } from './util/isFunction';\nimport { Observer, ObservableNotification } from './types';\nimport { isSubscription, Subscription } from './Subscription';\nimport { config } from './config';\nimport { reportUnhandledError } from './util/reportUnhandledError';\nimport { noop } from './util/noop';\nimport { nextNotification, errorNotification, COMPLETE_NOTIFICATION } from './NotificationFactories';\nimport { timeoutProvider } from './scheduler/timeoutProvider';\nimport { captureError } from './util/errorContext';\n\n/**\n * Implements the {@link Observer} interface and extends the\n * {@link Subscription} class. While the {@link Observer} is the public API for\n * consuming the values of an {@link Observable}, all Observers get converted to\n * a Subscriber, in order to provide Subscription-like capabilities such as\n * `unsubscribe`. Subscriber is a common type in RxJS, and crucial for\n * implementing operators, but it is rarely used as a public API.\n *\n * @class Subscriber\n */\nexport class Subscriber extends Subscription implements Observer {\n /**\n * A static factory for a Subscriber, given a (potentially partial) definition\n * of an Observer.\n * @param next The `next` callback of an Observer.\n * @param error The `error` callback of an\n * Observer.\n * @param complete The `complete` callback of an\n * Observer.\n * @return A Subscriber wrapping the (partially defined)\n * Observer represented by the given arguments.\n * @nocollapse\n * @deprecated Do not use. Will be removed in v8. There is no replacement for this\n * method, and there is no reason to be creating instances of `Subscriber` directly.\n * If you have a specific use case, please file an issue.\n */\n static create(next?: (x?: T) => void, error?: (e?: any) => void, complete?: () => void): Subscriber {\n return new SafeSubscriber(next, error, complete);\n }\n\n /** @deprecated Internal implementation detail, do not use directly. Will be made internal in v8. */\n protected isStopped: boolean = false;\n /** @deprecated Internal implementation detail, do not use directly. Will be made internal in v8. */\n protected destination: Subscriber | Observer; // this `any` is the escape hatch to erase extra type param (e.g. R)\n\n /**\n * @deprecated Internal implementation detail, do not use directly. Will be made internal in v8.\n * There is no reason to directly create an instance of Subscriber. This type is exported for typings reasons.\n */\n constructor(destination?: Subscriber | Observer) {\n super();\n if (destination) {\n this.destination = destination;\n // Automatically chain subscriptions together here.\n // if destination is a Subscription, then it is a Subscriber.\n if (isSubscription(destination)) {\n destination.add(this);\n }\n } else {\n this.destination = EMPTY_OBSERVER;\n }\n }\n\n /**\n * The {@link Observer} callback to receive notifications of type `next` from\n * the Observable, with a value. The Observable may call this method 0 or more\n * times.\n * @param {T} [value] The `next` value.\n * @return {void}\n */\n next(value?: T): void {\n if (this.isStopped) {\n handleStoppedNotification(nextNotification(value), this);\n } else {\n this._next(value!);\n }\n }\n\n /**\n * The {@link Observer} callback to receive notifications of type `error` from\n * the Observable, with an attached `Error`. Notifies the Observer that\n * the Observable has experienced an error condition.\n * @param {any} [err] The `error` exception.\n * @return {void}\n */\n error(err?: any): void {\n if (this.isStopped) {\n handleStoppedNotification(errorNotification(err), this);\n } else {\n this.isStopped = true;\n this._error(err);\n }\n }\n\n /**\n * The {@link Observer} callback to receive a valueless notification of type\n * `complete` from the Observable. Notifies the Observer that the Observable\n * has finished sending push-based notifications.\n * @return {void}\n */\n complete(): void {\n if (this.isStopped) {\n handleStoppedNotification(COMPLETE_NOTIFICATION, this);\n } else {\n this.isStopped = true;\n this._complete();\n }\n }\n\n unsubscribe(): void {\n if (!this.closed) {\n this.isStopped = true;\n super.unsubscribe();\n this.destination = null!;\n }\n }\n\n protected _next(value: T): void {\n this.destination.next(value);\n }\n\n protected _error(err: any): void {\n try {\n this.destination.error(err);\n } finally {\n this.unsubscribe();\n }\n }\n\n protected _complete(): void {\n try {\n this.destination.complete();\n } finally {\n this.unsubscribe();\n }\n }\n}\n\n/**\n * This bind is captured here because we want to be able to have\n * compatibility with monoid libraries that tend to use a method named\n * `bind`. In particular, a library called Monio requires this.\n */\nconst _bind = Function.prototype.bind;\n\nfunction bind any>(fn: Fn, thisArg: any): Fn {\n return _bind.call(fn, thisArg);\n}\n\n/**\n * Internal optimization only, DO NOT EXPOSE.\n * @internal\n */\nclass ConsumerObserver implements Observer {\n constructor(private partialObserver: Partial>) {}\n\n next(value: T): void {\n const { partialObserver } = this;\n if (partialObserver.next) {\n try {\n partialObserver.next(value);\n } catch (error) {\n handleUnhandledError(error);\n }\n }\n }\n\n error(err: any): void {\n const { partialObserver } = this;\n if (partialObserver.error) {\n try {\n partialObserver.error(err);\n } catch (error) {\n handleUnhandledError(error);\n }\n } else {\n handleUnhandledError(err);\n }\n }\n\n complete(): void {\n const { partialObserver } = this;\n if (partialObserver.complete) {\n try {\n partialObserver.complete();\n } catch (error) {\n handleUnhandledError(error);\n }\n }\n }\n}\n\nexport class SafeSubscriber extends Subscriber {\n constructor(\n observerOrNext?: Partial> | ((value: T) => void) | null,\n error?: ((e?: any) => void) | null,\n complete?: (() => void) | null\n ) {\n super();\n\n let partialObserver: Partial>;\n if (isFunction(observerOrNext) || !observerOrNext) {\n // The first argument is a function, not an observer. The next\n // two arguments *could* be observers, or they could be empty.\n partialObserver = {\n next: (observerOrNext ?? undefined) as (((value: T) => void) | undefined),\n error: error ?? undefined,\n complete: complete ?? undefined,\n };\n } else {\n // The first argument is a partial observer.\n let context: any;\n if (this && config.useDeprecatedNextContext) {\n // This is a deprecated path that made `this.unsubscribe()` available in\n // next handler functions passed to subscribe. This only exists behind a flag\n // now, as it is *very* slow.\n context = Object.create(observerOrNext);\n context.unsubscribe = () => this.unsubscribe();\n partialObserver = {\n next: observerOrNext.next && bind(observerOrNext.next, context),\n error: observerOrNext.error && bind(observerOrNext.error, context),\n complete: observerOrNext.complete && bind(observerOrNext.complete, context),\n };\n } else {\n // The \"normal\" path. Just use the partial observer directly.\n partialObserver = observerOrNext;\n }\n }\n\n // Wrap the partial observer to ensure it's a full observer, and\n // make sure proper error handling is accounted for.\n this.destination = new ConsumerObserver(partialObserver);\n }\n}\n\nfunction handleUnhandledError(error: any) {\n if (config.useDeprecatedSynchronousErrorHandling) {\n captureError(error);\n } else {\n // Ideal path, we report this as an unhandled error,\n // which is thrown on a new call stack.\n reportUnhandledError(error);\n }\n}\n\n/**\n * An error handler used when no error handler was supplied\n * to the SafeSubscriber -- meaning no error handler was supplied\n * do the `subscribe` call on our observable.\n * @param err The error to handle\n */\nfunction defaultErrorHandler(err: any) {\n throw err;\n}\n\n/**\n * A handler for notifications that cannot be sent to a stopped subscriber.\n * @param notification The notification being sent\n * @param subscriber The stopped subscriber\n */\nfunction handleStoppedNotification(notification: ObservableNotification, subscriber: Subscriber) {\n const { onStoppedNotification } = config;\n onStoppedNotification && timeoutProvider.setTimeout(() => onStoppedNotification(notification, subscriber));\n}\n\n/**\n * The observer used as a stub for subscriptions where the user did not\n * pass any arguments to `subscribe`. Comes with the default error handling\n * behavior.\n */\nexport const EMPTY_OBSERVER: Readonly> & { closed: true } = {\n closed: true,\n next: noop,\n error: defaultErrorHandler,\n complete: noop,\n};\n", "/**\n * Symbol.observable or a string \"@@observable\". Used for interop\n *\n * @deprecated We will no longer be exporting this symbol in upcoming versions of RxJS.\n * Instead polyfill and use Symbol.observable directly *or* use https://www.npmjs.com/package/symbol-observable\n */\nexport const observable: string | symbol = (() => (typeof Symbol === 'function' && Symbol.observable) || '@@observable')();\n", "/**\n * This function takes one parameter and just returns it. Simply put,\n * this is like `(x: T): T => x`.\n *\n * ## Examples\n *\n * This is useful in some cases when using things like `mergeMap`\n *\n * ```ts\n * import { interval, take, map, range, mergeMap, identity } from 'rxjs';\n *\n * const source$ = interval(1000).pipe(take(5));\n *\n * const result$ = source$.pipe(\n * map(i => range(i)),\n * mergeMap(identity) // same as mergeMap(x => x)\n * );\n *\n * result$.subscribe({\n * next: console.log\n * });\n * ```\n *\n * Or when you want to selectively apply an operator\n *\n * ```ts\n * import { interval, take, identity } from 'rxjs';\n *\n * const shouldLimit = () => Math.random() < 0.5;\n *\n * const source$ = interval(1000);\n *\n * const result$ = source$.pipe(shouldLimit() ? take(5) : identity);\n *\n * result$.subscribe({\n * next: console.log\n * });\n * ```\n *\n * @param x Any value that is returned by this function\n * @returns The value passed as the first parameter to this function\n */\nexport function identity(x: T): T {\n return x;\n}\n", "import { identity } from './identity';\nimport { UnaryFunction } from '../types';\n\nexport function pipe(): typeof identity;\nexport function pipe(fn1: UnaryFunction): UnaryFunction;\nexport function pipe(fn1: UnaryFunction, fn2: UnaryFunction): UnaryFunction;\nexport function pipe(fn1: UnaryFunction, fn2: UnaryFunction, fn3: UnaryFunction): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction,\n fn6: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction,\n fn6: UnaryFunction,\n fn7: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction,\n fn6: UnaryFunction,\n fn7: UnaryFunction,\n fn8: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction,\n fn6: UnaryFunction,\n fn7: UnaryFunction,\n fn8: UnaryFunction,\n fn9: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction,\n fn6: UnaryFunction,\n fn7: UnaryFunction,\n fn8: UnaryFunction,\n fn9: UnaryFunction,\n ...fns: UnaryFunction[]\n): UnaryFunction;\n\n/**\n * pipe() can be called on one or more functions, each of which can take one argument (\"UnaryFunction\")\n * and uses it to return a value.\n * It returns a function that takes one argument, passes it to the first UnaryFunction, and then\n * passes the result to the next one, passes that result to the next one, and so on. \n */\nexport function pipe(...fns: Array>): UnaryFunction {\n return pipeFromArray(fns);\n}\n\n/** @internal */\nexport function pipeFromArray(fns: Array>): UnaryFunction {\n if (fns.length === 0) {\n return identity as UnaryFunction;\n }\n\n if (fns.length === 1) {\n return fns[0];\n }\n\n return function piped(input: T): R {\n return fns.reduce((prev: any, fn: UnaryFunction) => fn(prev), input as any);\n };\n}\n", "import { Operator } from './Operator';\nimport { SafeSubscriber, Subscriber } from './Subscriber';\nimport { isSubscription, Subscription } from './Subscription';\nimport { TeardownLogic, OperatorFunction, Subscribable, Observer } from './types';\nimport { observable as Symbol_observable } from './symbol/observable';\nimport { pipeFromArray } from './util/pipe';\nimport { config } from './config';\nimport { isFunction } from './util/isFunction';\nimport { errorContext } from './util/errorContext';\n\n/**\n * A representation of any set of values over any amount of time. This is the most basic building block\n * of RxJS.\n *\n * @class Observable\n */\nexport class Observable implements Subscribable {\n /**\n * @deprecated Internal implementation detail, do not use directly. Will be made internal in v8.\n */\n source: Observable | undefined;\n\n /**\n * @deprecated Internal implementation detail, do not use directly. Will be made internal in v8.\n */\n operator: Operator | undefined;\n\n /**\n * @constructor\n * @param {Function} subscribe the function that is called when the Observable is\n * initially subscribed to. This function is given a Subscriber, to which new values\n * can be `next`ed, or an `error` method can be called to raise an error, or\n * `complete` can be called to notify of a successful completion.\n */\n constructor(subscribe?: (this: Observable, subscriber: Subscriber) => TeardownLogic) {\n if (subscribe) {\n this._subscribe = subscribe;\n }\n }\n\n // HACK: Since TypeScript inherits static properties too, we have to\n // fight against TypeScript here so Subject can have a different static create signature\n /**\n * Creates a new Observable by calling the Observable constructor\n * @owner Observable\n * @method create\n * @param {Function} subscribe? the subscriber function to be passed to the Observable constructor\n * @return {Observable} a new observable\n * @nocollapse\n * @deprecated Use `new Observable()` instead. Will be removed in v8.\n */\n static create: (...args: any[]) => any = (subscribe?: (subscriber: Subscriber) => TeardownLogic) => {\n return new Observable(subscribe);\n };\n\n /**\n * Creates a new Observable, with this Observable instance as the source, and the passed\n * operator defined as the new observable's operator.\n * @method lift\n * @param operator the operator defining the operation to take on the observable\n * @return a new observable with the Operator applied\n * @deprecated Internal implementation detail, do not use directly. Will be made internal in v8.\n * If you have implemented an operator using `lift`, it is recommended that you create an\n * operator by simply returning `new Observable()` directly. See \"Creating new operators from\n * scratch\" section here: https://rxjs.dev/guide/operators\n */\n lift(operator?: Operator): Observable {\n const observable = new Observable();\n observable.source = this;\n observable.operator = operator;\n return observable;\n }\n\n subscribe(observerOrNext?: Partial> | ((value: T) => void)): Subscription;\n /** @deprecated Instead of passing separate callback arguments, use an observer argument. Signatures taking separate callback arguments will be removed in v8. Details: https://rxjs.dev/deprecations/subscribe-arguments */\n subscribe(next?: ((value: T) => void) | null, error?: ((error: any) => void) | null, complete?: (() => void) | null): Subscription;\n /**\n * Invokes an execution of an Observable and registers Observer handlers for notifications it will emit.\n *\n * Use it when you have all these Observables, but still nothing is happening.\n *\n * `subscribe` is not a regular operator, but a method that calls Observable's internal `subscribe` function. It\n * might be for example a function that you passed to Observable's constructor, but most of the time it is\n * a library implementation, which defines what will be emitted by an Observable, and when it be will emitted. This means\n * that calling `subscribe` is actually the moment when Observable starts its work, not when it is created, as it is often\n * the thought.\n *\n * Apart from starting the execution of an Observable, this method allows you to listen for values\n * that an Observable emits, as well as for when it completes or errors. You can achieve this in two\n * of the following ways.\n *\n * The first way is creating an object that implements {@link Observer} interface. It should have methods\n * defined by that interface, but note that it should be just a regular JavaScript object, which you can create\n * yourself in any way you want (ES6 class, classic function constructor, object literal etc.). In particular, do\n * not attempt to use any RxJS implementation details to create Observers - you don't need them. Remember also\n * that your object does not have to implement all methods. If you find yourself creating a method that doesn't\n * do anything, you can simply omit it. Note however, if the `error` method is not provided and an error happens,\n * it will be thrown asynchronously. Errors thrown asynchronously cannot be caught using `try`/`catch`. Instead,\n * use the {@link onUnhandledError} configuration option or use a runtime handler (like `window.onerror` or\n * `process.on('error)`) to be notified of unhandled errors. Because of this, it's recommended that you provide\n * an `error` method to avoid missing thrown errors.\n *\n * The second way is to give up on Observer object altogether and simply provide callback functions in place of its methods.\n * This means you can provide three functions as arguments to `subscribe`, where the first function is equivalent\n * of a `next` method, the second of an `error` method and the third of a `complete` method. Just as in case of an Observer,\n * if you do not need to listen for something, you can omit a function by passing `undefined` or `null`,\n * since `subscribe` recognizes these functions by where they were placed in function call. When it comes\n * to the `error` function, as with an Observer, if not provided, errors emitted by an Observable will be thrown asynchronously.\n *\n * You can, however, subscribe with no parameters at all. This may be the case where you're not interested in terminal events\n * and you also handled emissions internally by using operators (e.g. using `tap`).\n *\n * Whichever style of calling `subscribe` you use, in both cases it returns a Subscription object.\n * This object allows you to call `unsubscribe` on it, which in turn will stop the work that an Observable does and will clean\n * up all resources that an Observable used. Note that cancelling a subscription will not call `complete` callback\n * provided to `subscribe` function, which is reserved for a regular completion signal that comes from an Observable.\n *\n * Remember that callbacks provided to `subscribe` are not guaranteed to be called asynchronously.\n * It is an Observable itself that decides when these functions will be called. For example {@link of}\n * by default emits all its values synchronously. Always check documentation for how given Observable\n * will behave when subscribed and if its default behavior can be modified with a `scheduler`.\n *\n * #### Examples\n *\n * Subscribe with an {@link guide/observer Observer}\n *\n * ```ts\n * import { of } from 'rxjs';\n *\n * const sumObserver = {\n * sum: 0,\n * next(value) {\n * console.log('Adding: ' + value);\n * this.sum = this.sum + value;\n * },\n * error() {\n * // We actually could just remove this method,\n * // since we do not really care about errors right now.\n * },\n * complete() {\n * console.log('Sum equals: ' + this.sum);\n * }\n * };\n *\n * of(1, 2, 3) // Synchronously emits 1, 2, 3 and then completes.\n * .subscribe(sumObserver);\n *\n * // Logs:\n * // 'Adding: 1'\n * // 'Adding: 2'\n * // 'Adding: 3'\n * // 'Sum equals: 6'\n * ```\n *\n * Subscribe with functions ({@link deprecations/subscribe-arguments deprecated})\n *\n * ```ts\n * import { of } from 'rxjs'\n *\n * let sum = 0;\n *\n * of(1, 2, 3).subscribe(\n * value => {\n * console.log('Adding: ' + value);\n * sum = sum + value;\n * },\n * undefined,\n * () => console.log('Sum equals: ' + sum)\n * );\n *\n * // Logs:\n * // 'Adding: 1'\n * // 'Adding: 2'\n * // 'Adding: 3'\n * // 'Sum equals: 6'\n * ```\n *\n * Cancel a subscription\n *\n * ```ts\n * import { interval } from 'rxjs';\n *\n * const subscription = interval(1000).subscribe({\n * next(num) {\n * console.log(num)\n * },\n * complete() {\n * // Will not be called, even when cancelling subscription.\n * console.log('completed!');\n * }\n * });\n *\n * setTimeout(() => {\n * subscription.unsubscribe();\n * console.log('unsubscribed!');\n * }, 2500);\n *\n * // Logs:\n * // 0 after 1s\n * // 1 after 2s\n * // 'unsubscribed!' after 2.5s\n * ```\n *\n * @param {Observer|Function} observerOrNext (optional) Either an observer with methods to be called,\n * or the first of three possible handlers, which is the handler for each value emitted from the subscribed\n * Observable.\n * @param {Function} error (optional) A handler for a terminal event resulting from an error. If no error handler is provided,\n * the error will be thrown asynchronously as unhandled.\n * @param {Function} complete (optional) A handler for a terminal event resulting from successful completion.\n * @return {Subscription} a subscription reference to the registered handlers\n * @method subscribe\n */\n subscribe(\n observerOrNext?: Partial> | ((value: T) => void) | null,\n error?: ((error: any) => void) | null,\n complete?: (() => void) | null\n ): Subscription {\n const subscriber = isSubscriber(observerOrNext) ? observerOrNext : new SafeSubscriber(observerOrNext, error, complete);\n\n errorContext(() => {\n const { operator, source } = this;\n subscriber.add(\n operator\n ? // We're dealing with a subscription in the\n // operator chain to one of our lifted operators.\n operator.call(subscriber, source)\n : source\n ? // If `source` has a value, but `operator` does not, something that\n // had intimate knowledge of our API, like our `Subject`, must have\n // set it. We're going to just call `_subscribe` directly.\n this._subscribe(subscriber)\n : // In all other cases, we're likely wrapping a user-provided initializer\n // function, so we need to catch errors and handle them appropriately.\n this._trySubscribe(subscriber)\n );\n });\n\n return subscriber;\n }\n\n /** @internal */\n protected _trySubscribe(sink: Subscriber): TeardownLogic {\n try {\n return this._subscribe(sink);\n } catch (err) {\n // We don't need to return anything in this case,\n // because it's just going to try to `add()` to a subscription\n // above.\n sink.error(err);\n }\n }\n\n /**\n * Used as a NON-CANCELLABLE means of subscribing to an observable, for use with\n * APIs that expect promises, like `async/await`. You cannot unsubscribe from this.\n *\n * **WARNING**: Only use this with observables you *know* will complete. If the source\n * observable does not complete, you will end up with a promise that is hung up, and\n * potentially all of the state of an async function hanging out in memory. To avoid\n * this situation, look into adding something like {@link timeout}, {@link take},\n * {@link takeWhile}, or {@link takeUntil} amongst others.\n *\n * #### Example\n *\n * ```ts\n * import { interval, take } from 'rxjs';\n *\n * const source$ = interval(1000).pipe(take(4));\n *\n * async function getTotal() {\n * let total = 0;\n *\n * await source$.forEach(value => {\n * total += value;\n * console.log('observable -> ' + value);\n * });\n *\n * return total;\n * }\n *\n * getTotal().then(\n * total => console.log('Total: ' + total)\n * );\n *\n * // Expected:\n * // 'observable -> 0'\n * // 'observable -> 1'\n * // 'observable -> 2'\n * // 'observable -> 3'\n * // 'Total: 6'\n * ```\n *\n * @param next a handler for each value emitted by the observable\n * @return a promise that either resolves on observable completion or\n * rejects with the handled error\n */\n forEach(next: (value: T) => void): Promise;\n\n /**\n * @param next a handler for each value emitted by the observable\n * @param promiseCtor a constructor function used to instantiate the Promise\n * @return a promise that either resolves on observable completion or\n * rejects with the handled error\n * @deprecated Passing a Promise constructor will no longer be available\n * in upcoming versions of RxJS. This is because it adds weight to the library, for very\n * little benefit. If you need this functionality, it is recommended that you either\n * polyfill Promise, or you create an adapter to convert the returned native promise\n * to whatever promise implementation you wanted. Will be removed in v8.\n */\n forEach(next: (value: T) => void, promiseCtor: PromiseConstructorLike): Promise;\n\n forEach(next: (value: T) => void, promiseCtor?: PromiseConstructorLike): Promise {\n promiseCtor = getPromiseCtor(promiseCtor);\n\n return new promiseCtor((resolve, reject) => {\n const subscriber = new SafeSubscriber({\n next: (value) => {\n try {\n next(value);\n } catch (err) {\n reject(err);\n subscriber.unsubscribe();\n }\n },\n error: reject,\n complete: resolve,\n });\n this.subscribe(subscriber);\n }) as Promise;\n }\n\n /** @internal */\n protected _subscribe(subscriber: Subscriber): TeardownLogic {\n return this.source?.subscribe(subscriber);\n }\n\n /**\n * An interop point defined by the es7-observable spec https://github.com/zenparsing/es-observable\n * @method Symbol.observable\n * @return {Observable} this instance of the observable\n */\n [Symbol_observable]() {\n return this;\n }\n\n /* tslint:disable:max-line-length */\n pipe(): Observable;\n pipe(op1: OperatorFunction): Observable;\n pipe(op1: OperatorFunction, op2: OperatorFunction): Observable;\n pipe(op1: OperatorFunction, op2: OperatorFunction, op3: OperatorFunction): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction,\n op6: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction,\n op6: OperatorFunction,\n op7: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction,\n op6: OperatorFunction,\n op7: OperatorFunction,\n op8: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction,\n op6: OperatorFunction,\n op7: OperatorFunction,\n op8: OperatorFunction,\n op9: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction,\n op6: OperatorFunction,\n op7: OperatorFunction,\n op8: OperatorFunction,\n op9: OperatorFunction,\n ...operations: OperatorFunction[]\n ): Observable;\n /* tslint:enable:max-line-length */\n\n /**\n * Used to stitch together functional operators into a chain.\n * @method pipe\n * @return {Observable} the Observable result of all of the operators having\n * been called in the order they were passed in.\n *\n * ## Example\n *\n * ```ts\n * import { interval, filter, map, scan } from 'rxjs';\n *\n * interval(1000)\n * .pipe(\n * filter(x => x % 2 === 0),\n * map(x => x + x),\n * scan((acc, x) => acc + x)\n * )\n * .subscribe(x => console.log(x));\n * ```\n */\n pipe(...operations: OperatorFunction[]): Observable {\n return pipeFromArray(operations)(this);\n }\n\n /* tslint:disable:max-line-length */\n /** @deprecated Replaced with {@link firstValueFrom} and {@link lastValueFrom}. Will be removed in v8. Details: https://rxjs.dev/deprecations/to-promise */\n toPromise(): Promise;\n /** @deprecated Replaced with {@link firstValueFrom} and {@link lastValueFrom}. Will be removed in v8. Details: https://rxjs.dev/deprecations/to-promise */\n toPromise(PromiseCtor: typeof Promise): Promise;\n /** @deprecated Replaced with {@link firstValueFrom} and {@link lastValueFrom}. Will be removed in v8. Details: https://rxjs.dev/deprecations/to-promise */\n toPromise(PromiseCtor: PromiseConstructorLike): Promise;\n /* tslint:enable:max-line-length */\n\n /**\n * Subscribe to this Observable and get a Promise resolving on\n * `complete` with the last emission (if any).\n *\n * **WARNING**: Only use this with observables you *know* will complete. If the source\n * observable does not complete, you will end up with a promise that is hung up, and\n * potentially all of the state of an async function hanging out in memory. To avoid\n * this situation, look into adding something like {@link timeout}, {@link take},\n * {@link takeWhile}, or {@link takeUntil} amongst others.\n *\n * @method toPromise\n * @param [promiseCtor] a constructor function used to instantiate\n * the Promise\n * @return A Promise that resolves with the last value emit, or\n * rejects on an error. If there were no emissions, Promise\n * resolves with undefined.\n * @deprecated Replaced with {@link firstValueFrom} and {@link lastValueFrom}. Will be removed in v8. Details: https://rxjs.dev/deprecations/to-promise\n */\n toPromise(promiseCtor?: PromiseConstructorLike): Promise {\n promiseCtor = getPromiseCtor(promiseCtor);\n\n return new promiseCtor((resolve, reject) => {\n let value: T | undefined;\n this.subscribe(\n (x: T) => (value = x),\n (err: any) => reject(err),\n () => resolve(value)\n );\n }) as Promise;\n }\n}\n\n/**\n * Decides between a passed promise constructor from consuming code,\n * A default configured promise constructor, and the native promise\n * constructor and returns it. If nothing can be found, it will throw\n * an error.\n * @param promiseCtor The optional promise constructor to passed by consuming code\n */\nfunction getPromiseCtor(promiseCtor: PromiseConstructorLike | undefined) {\n return promiseCtor ?? config.Promise ?? Promise;\n}\n\nfunction isObserver(value: any): value is Observer {\n return value && isFunction(value.next) && isFunction(value.error) && isFunction(value.complete);\n}\n\nfunction isSubscriber(value: any): value is Subscriber {\n return (value && value instanceof Subscriber) || (isObserver(value) && isSubscription(value));\n}\n", "import { Observable } from '../Observable';\nimport { Subscriber } from '../Subscriber';\nimport { OperatorFunction } from '../types';\nimport { isFunction } from './isFunction';\n\n/**\n * Used to determine if an object is an Observable with a lift function.\n */\nexport function hasLift(source: any): source is { lift: InstanceType['lift'] } {\n return isFunction(source?.lift);\n}\n\n/**\n * Creates an `OperatorFunction`. Used to define operators throughout the library in a concise way.\n * @param init The logic to connect the liftedSource to the subscriber at the moment of subscription.\n */\nexport function operate(\n init: (liftedSource: Observable, subscriber: Subscriber) => (() => void) | void\n): OperatorFunction {\n return (source: Observable) => {\n if (hasLift(source)) {\n return source.lift(function (this: Subscriber, liftedSource: Observable) {\n try {\n return init(liftedSource, this);\n } catch (err) {\n this.error(err);\n }\n });\n }\n throw new TypeError('Unable to lift unknown Observable type');\n };\n}\n", "import { Subscriber } from '../Subscriber';\n\n/**\n * Creates an instance of an `OperatorSubscriber`.\n * @param destination The downstream subscriber.\n * @param onNext Handles next values, only called if this subscriber is not stopped or closed. Any\n * error that occurs in this function is caught and sent to the `error` method of this subscriber.\n * @param onError Handles errors from the subscription, any errors that occur in this handler are caught\n * and send to the `destination` error handler.\n * @param onComplete Handles completion notification from the subscription. Any errors that occur in\n * this handler are sent to the `destination` error handler.\n * @param onFinalize Additional teardown logic here. This will only be called on teardown if the\n * subscriber itself is not already closed. This is called after all other teardown logic is executed.\n */\nexport function createOperatorSubscriber(\n destination: Subscriber