|
| 1 | +strict digraph { |
| 2 | +x [id=0, type="nncf_model_input", metatype=PTInputNoopMetatype]; |
| 3 | +"/unbind/0" [id=1, type=unbind, metatype=PTSplitMetatype]; |
| 4 | +"__nncf_hooks.pre_hooks./scaled_dot_product_attention/0__0.0._scale_param_storage" [id=2, type="nncf_model_const", metatype=PTConstNoopMetatype]; |
| 5 | +"pre_hook__-scaled_dot_product_attention-0__0[0]/symmetric_quantize/0" [id=3, type="symmetric_quantize", metatype=UnknownMetatype]; |
| 6 | +"__nncf_hooks.pre_hooks./scaled_dot_product_attention/0__1.0._scale_param_storage" [id=4, type="nncf_model_const", metatype=PTConstNoopMetatype]; |
| 7 | +"pre_hook__-scaled_dot_product_attention-0__1[0]/symmetric_quantize/0" [id=5, type="symmetric_quantize", metatype=UnknownMetatype]; |
| 8 | +"/scaled_dot_product_attention/0" [id=6, type="scaled_dot_product_attention", metatype=PTScaledDotProductAttentionMetatype]; |
| 9 | +output [id=7, type="nncf_model_output", metatype=PTOutputNoopMetatype]; |
| 10 | +x -> "/unbind/0" [dtype=float, shape="(3, 1, 8, 16)", out_port_id=0, in_port_id=0]; |
| 11 | +"/unbind/0" -> "pre_hook__-scaled_dot_product_attention-0__0[0]/symmetric_quantize/0" [dtype=float, shape="(1, 8, 16)", out_port_id=0, in_port_id=0]; |
| 12 | +"/unbind/0" -> "pre_hook__-scaled_dot_product_attention-0__1[0]/symmetric_quantize/0" [dtype=float, shape="(1, 8, 16)", out_port_id=1, in_port_id=0]; |
| 13 | +"/unbind/0" -> "/scaled_dot_product_attention/0" [dtype=float, shape="(1, 8, 16)", out_port_id=2, in_port_id=2]; |
| 14 | +"__nncf_hooks.pre_hooks./scaled_dot_product_attention/0__0.0._scale_param_storage" -> "pre_hook__-scaled_dot_product_attention-0__0[0]/symmetric_quantize/0" [dtype=float, shape="(1,)", out_port_id=0, in_port_id=4]; |
| 15 | +"pre_hook__-scaled_dot_product_attention-0__0[0]/symmetric_quantize/0" -> "/scaled_dot_product_attention/0" [dtype=float, shape="(1, 8, 16)", out_port_id=0, in_port_id=0]; |
| 16 | +"__nncf_hooks.pre_hooks./scaled_dot_product_attention/0__1.0._scale_param_storage" -> "pre_hook__-scaled_dot_product_attention-0__1[0]/symmetric_quantize/0" [dtype=float, shape="(1,)", out_port_id=0, in_port_id=4]; |
| 17 | +"pre_hook__-scaled_dot_product_attention-0__1[0]/symmetric_quantize/0" -> "/scaled_dot_product_attention/0" [dtype=float, shape="(1, 8, 16)", out_port_id=0, in_port_id=1]; |
| 18 | +"/scaled_dot_product_attention/0" -> output [dtype=float, shape="(1, 8, 16)", out_port_id=0, in_port_id=0]; |
| 19 | +} |
0 commit comments