Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[xdoctest] reformat example code with google style in No.386-389 #58206

Merged
merged 5 commits into from
Oct 25, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
25 changes: 19 additions & 6 deletions python/paddle/nn/functional/flash_attention.py
Original file line number Diff line number Diff line change
Expand Up @@ -183,10 +183,22 @@ def flash_attention(

>>> import paddle

>>> paddle.seed(1)
>>> paddle.seed(2023)
>>> q = paddle.rand((1, 128, 2, 16))

>>> output = paddle.nn.functional.flash_attention.flash_attention(q, q, q, 0.9, False, False)
>>> print(output)
(Tensor(shape=[1, 128, 2, 16], dtype=float32, place=Place(cpu), stop_gradient=True,
[[[[0.34992966, 0.34456208, 0.45826620, ..., 0.39883569,
0.42132431, 0.39157745],
[0.76687670, 0.65837246, 0.69117945, ..., 0.82817286,
0.76690865, 0.71485823]],
...,
[[0.71662450, 0.57275224, 0.57053083, ..., 0.48108247,
0.53336465, 0.54540104],
[0.59137970, 0.51350880, 0.50449550, ..., 0.38860250,
0.40526697, 0.60541755]]]]), None)

"""
head_dim = query.shape[3]
sdp_func_name = _select_sdp(head_dim)
Expand Down Expand Up @@ -340,11 +352,12 @@ def flash_attn_unpadded(
.. code-block:: python

>>> import paddle
>>> paddle.seed(1)
>>> q = paddle.rand((1, 128, 2, 16))
>>> paddle.seed(2023)
>>> q = paddle.rand((2, 128, 8, 16), dtype='float16')
>>> cu = paddle.arange(0, 384, 128, dtype='int32')
>>> qq = paddle.reshape(q, [256, 8, 16])
>>> output = paddle.nn.functional.flash_attention.flash_attn_unpadded(qq, qq, qq, cu, cu, 128, 128, 0.25, 0.0, False, False)

>>> output = paddle.nn.functional.flash_attention.flash_attn_unpadded(q, q, q, 0.9, False, False)
>>> print(output)
"""
if in_dynamic_mode():
(
Expand Down Expand Up @@ -461,7 +474,7 @@ def scaled_dot_product_attention(
Examples:
.. code-block:: python

>>> # doctest: +SKIP()
>>> # doctest: +SKIP('bfloat need V100 compile')
>>> import paddle
>>> q = paddle.rand((1, 128, 2, 16), dtype=paddle.bfloat16)
>>> output = paddle.nn.functional.scaled_dot_product_attention(q, q, q, None, 0.9, False)
Expand Down
6 changes: 3 additions & 3 deletions python/paddle/quantization/config.py
Original file line number Diff line number Diff line change
Expand Up @@ -127,7 +127,7 @@ def add_layer_config(
>>> quanter = FakeQuanterWithAbsMaxObserver(moving_rate=0.9)
>>> q_config = QuantConfig(activation=None, weight=None)
>>> q_config.add_layer_config([model.fc], activation=quanter, weight=quanter)
>>> # doctest: +SKIP
>>> # doctest: +SKIP('random memory address')
>>> print(q_config)
Global config:
None
Expand Down Expand Up @@ -176,7 +176,7 @@ def add_name_config(
>>> quanter = FakeQuanterWithAbsMaxObserver(moving_rate=0.9)
>>> q_config = QuantConfig(activation=None, weight=None)
>>> q_config.add_name_config([model.fc.full_name()], activation=quanter, weight=quanter)
>>> # doctest: +SKIP
>>> # doctest: +SKIP('random memory address')
>>> print(q_config)
Global config:
None
Expand Down Expand Up @@ -226,7 +226,7 @@ def add_type_config(
>>> quanter = FakeQuanterWithAbsMaxObserver(moving_rate=0.9)
>>> q_config = QuantConfig(activation=None, weight=None)
>>> q_config.add_type_config([Linear], activation=quanter, weight=quanter)
>>> # doctest: +SKIP
>>> # doctest: +SKIP('random memory address')
>>> print(q_config)
Global config:
None
Expand Down
2 changes: 1 addition & 1 deletion python/paddle/quantization/factory.py
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ def quanter(class_name):
Examples:
.. code-block:: python

>>> # doctest: +SKIP
>>> # doctest: +SKIP('need 2 file to run example')
>>> # Given codes in ./customized_quanter.py
>>> from paddle.quantization import quanter
>>> from paddle.quantization import BaseQuanter
Expand Down
14 changes: 7 additions & 7 deletions python/paddle/tensor/linalg.py
Original file line number Diff line number Diff line change
Expand Up @@ -1958,14 +1958,12 @@ def slogdet(x, name=None):

>>> import paddle
>>> paddle.seed(2023)
>>> x = paddle.randn([3,3,3])
>>> x = paddle.randn([3, 3, 3])
>>> A = paddle.linalg.slogdet(x)
>>> print(A)
>>> # doctest: +SKIP
Tensor(shape=[2, 3], dtype=float32, place=Place(cpu), stop_gradient=True,
[[-1. , 1. , 1. ],
[ 0.25681755, -0.25061053, -0.10809582]])
>>> # doctest: -SKIP

"""
if in_dynamic_mode():
Expand Down Expand Up @@ -2801,10 +2799,12 @@ def eigh(x, UPLO='L', name=None):
property. For more information, please refer to :ref:`api_guide_Name`.

Returns:
- out_value(Tensor): A Tensor with shape [*, N] and data type of float32 and float64.
The eigenvalues of eigh op.
- out_vector(Tensor): A Tensor with shape [*, N, N] and data type of float32,float64,
complex64 and complex128. The eigenvectors of eigh op.
2-element tuple containing

- out_value(Tensor): A Tensor with shape :math:`[*, N]` and data type of float32 and float64.
The eigenvalues of eigh op.
- out_vector(Tensor): A Tensor with shape :math:`[*, N, N]` and data type of float32, float64,
complex64 and complex128. The eigenvectors of eigh op.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

预览好像还是不行..两个段落之间空行试一下?@enkilee


Examples:
.. code-block:: python
Expand Down