Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: support group_norm, batch_norm, and layer_norm #2330

Merged
merged 7 commits into from
Oct 10, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
151 changes: 113 additions & 38 deletions py/torch_tensorrt/dynamo/conversion/aten_ops_converters.py
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,23 @@ def get_ir(target: Target) -> SourceIR:
return SourceIR.UNKNOWN


def one_user_validator(node: Node) -> bool:
# Validate only one user, which is a getitem node that accesses the first element in the list
return (
len(node.users) == 1
and list(node.users)[0].target == operator.getitem
and list(node.users)[0].args[1] == 0
)


@dynamo_tensorrt_converter(torch.ops.aten.native_batch_norm.default, capability_validator=one_user_validator) # type: ignore[misc]
@dynamo_tensorrt_converter(torch.ops.aten.batch_norm.default) # type: ignore[misc]
@dynamo_tensorrt_converter(torch.ops.aten.batch_norm) # type: ignore[misc]
@enforce_tensor_types(
{
0: (TRTTensor,),
}
) # type: ignore[misc]
def aten_ops_batch_norm(
ctx: ConversionContext,
target: Target,
Expand All @@ -59,14 +75,103 @@ def aten_ops_batch_norm(
target,
SourceIR.ATEN,
name,
args[0],
args[1],
args[2],
args[3],
args[4],
args[5],
args[6],
args[7],
input=args[0],
zewenli98 marked this conversation as resolved.
Show resolved Hide resolved
weight=args[1],
bias=args[2],
running_mean=args[3],
running_var=args[4],
training=args[5],
momentum=args[6],
eps=args[7],
cudnn_enabled=args_bounds_check(args, 8, True),
return_mean_rstd=(target == torch.ops.aten.native_batch_norm.default),
)


@dynamo_tensorrt_converter(torch.ops.aten.native_layer_norm.default, capability_validator=one_user_validator) # type: ignore[misc]
@dynamo_tensorrt_converter(torch.ops.aten.layer_norm.default) # type: ignore[misc]
zewenli98 marked this conversation as resolved.
Show resolved Hide resolved
@dynamo_tensorrt_converter(torch.ops.aten.layer_norm) # type: ignore[misc]
@enforce_tensor_types(
{
0: (TRTTensor,),
}
) # type: ignore[misc]
def aten_ops_layer_norm(
ctx: ConversionContext,
target: Target,
args: Tuple[Argument, ...],
kwargs: Dict[str, Argument],
name: str,
) -> Union[TRTTensor, Sequence[TRTTensor]]:
return impl.normalization.layer_norm(
ctx,
target,
SourceIR.ATEN,
name,
input=args[0],
normalized_shape=args[1],
weight=args_bounds_check(args, 2),
bias=args_bounds_check(args, 3),
eps=args_bounds_check(args, 4, 1e-05),
cudnn_enable=args_bounds_check(args, 5, True),
return_mean_rstd=(target == torch.ops.aten.native_layer_norm.default),
)


@dynamo_tensorrt_converter(torch.ops.aten.native_group_norm.default, capability_validator=one_user_validator) # type: ignore[misc]
@enforce_tensor_types(
{
0: (TRTTensor,),
}
) # type: ignore[misc]
def aten_ops_native_group_norm(
ctx: ConversionContext,
target: Target,
args: Tuple[Argument, ...],
kwargs: Dict[str, Argument],
name: str,
) -> Union[TRTTensor, Sequence[TRTTensor]]:
return impl.normalization.native_group_norm(
ctx,
target,
SourceIR.ATEN,
name,
input=args[0],
weight=args[1],
bias=args[2],
N=args[3],
C=args[4],
HxW=args[5],
group=args[6],
eps=args[7],
)


@dynamo_tensorrt_converter(torch.ops.aten.group_norm.default) # type: ignore[misc]
@dynamo_tensorrt_converter(torch.ops.aten.group_norm) # type: ignore[misc]
@enforce_tensor_types(
{
0: (TRTTensor,),
}
) # type: ignore[misc]
def aten_ops_group_norm(
ctx: ConversionContext,
target: Target,
args: Tuple[Argument, ...],
kwargs: Dict[str, Argument],
name: str,
) -> Union[TRTTensor, Sequence[TRTTensor]]:
return impl.normalization.group_norm(
ctx,
target,
SourceIR.ATEN,
name,
input=args[0],
num_groups=args[1],
weight=args_bounds_check(args, 2, None),
bias=args_bounds_check(args, 3, None),
eps=args_bounds_check(args, 4, 1e-05),
cudnn_enabled=args_bounds_check(args, 5, True),
)


Expand Down Expand Up @@ -328,27 +433,6 @@ def aten_ops_matmul(
)


@dynamo_tensorrt_converter(torch.ops.aten.layer_norm.default) # type: ignore[misc]
def aten_ops_layernorm(
ctx: ConversionContext,
target: Target,
args: Tuple[Argument, ...],
kwargs: Dict[str, Argument],
name: str,
) -> Union[TRTTensor, Sequence[TRTTensor]]:
return impl.normalization.layer_norm(
ctx,
target,
SourceIR.ATEN,
name,
args[0],
args[1],
args[2],
args[3],
args[4],
)


@dynamo_tensorrt_converter(torch.ops.aten.rsqrt.default) # type: ignore[misc]
def aten_ops_rsqrt(
ctx: ConversionContext,
Expand Down Expand Up @@ -763,15 +847,6 @@ def aten_ops_prod(
)


def one_user_validator(node: Node) -> bool:
# Validate only one user, which is a getitem node that accesses the first element in the list
return (
len(node.users) == 1
and list(node.users)[0].target == operator.getitem
and list(node.users)[0].args[1] == 0
)


@dynamo_tensorrt_converter(torch.ops.aten.max.default) # type: ignore[misc]
@dynamo_tensorrt_converter(torch.ops.aten.max.dim, capability_validator=one_user_validator) # type: ignore[misc]
def aten_ops_max(
Expand Down
Loading
Loading