Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Edit mixed precision pages in feature guide. #3755

Open
wants to merge 1 commit into
base: develop
Choose a base branch
from

Conversation

dwelsch-esi
Copy link
Contributor

No description provided.

Prerequisites
-------------

Manual mixed precision is supported only on PyTorch models.
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is this true? I assumed it is because I didn't see APIs for TF or ONNX.

* Change the precision of all the layers in the model of a certain type
* Change the precision of model input tensors (or only a subset of input tensors)
* Change the precision of model output tensors (or only a subset of output tensors)
* A leaf layer
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

How is a leaf layer defined?

Not supported.


The ``apply`` call generates a report detailing how the request was inferred, propagated to other layers, and eventually realized.
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Where is the report saved?

Layer Groups are defined as a group of layers grouped together based on certain rules.
This helps in reducing search space over which the mixed precision algorithm operates.
It also ensures that we search only over the valid bit-width settings for parameters and activations.
Layer Groups are defined based on certain rules.
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What are the rules and how are they defined?


.. image:: ../../images/pareto.png
:width: 900px
An example of a Pareto list:
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

How is this list different from the accuracy list in the previous phase?


Use Cases
=========
Conversion operations (convert ops) are introduced in the mixed-precision model for transition between ops with different activation bit widths or data types (float vs int). Convert ops contribute to the inference time along with bit-operations of ops.
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I assume that 'convert op' is jargon for 'conversion operation'. Is that the case?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant