Skip to content

Instantly share code, notes, and snippets.

// -----// IR Dump After mlir::iree_compiler::IREE::ABI::WrapEntryPointsPass //----- //
#loc1 = loc("<stdin>":2:3)
#loc2 = loc("<stdin>":2:17)
#loc17 = loc("<stdin>":17:10)
#loc18 = loc("<stdin>":17:25)
#loc19 = loc("<stdin>":17:40)
#loc20 = loc("<stdin>":17:55)
module {
func @forward(%arg0: !hal.buffer_view loc("<stdin>":2:3)) -> !hal.buffer_view attributes {iree.abi.stub} {
%0 = hal.buffer_view.dim %arg0, 0 : index loc(#loc1)
// -----// IR Dump After mlir::iree_compiler::IREE::ABI::WrapEntryPointsPass //----- //
#loc1 = loc("<stdin>":2:3)
#loc2 = loc("<stdin>":2:17)
module {
func @forward(%arg0: !hal.buffer_view loc("<stdin>":2:3)) -> !hal.buffer_view attributes {iree.abi.stub} {
%0 = hal.buffer_view.dim %arg0, 0 : index loc(#loc1)
%1 = hal.buffer_view.dim %arg0, 1 : index loc(#loc1)
%2 = hal.buffer_view.dim %arg0, 2 : index loc(#loc1)
%3 = hal.tensor.cast %arg0 : !hal.buffer_view -> tensor<?x?x?xf32>{%0, %1, %2} loc(#loc1)
%4 = call @_forward(%3) : (tensor<?x?x?xf32>) -> tensor<?x?xf32> loc(#loc1)
builtin.func@test(%intx :!torch.int, %145 : !torch.bool, %136 : !torch.optional<!torch.tensor>, %137 : !torch.optional<!torch.tensor>) -> !torch.optional<!torch.tensor> {
%146 = torch.prim.If %145 -> (!torch.optional<!torch.tensor>) {
%int0 = torch.constant.int 0
%int1 = torch.constant.int 1
%int-1 = torch.constant.int -1
%304 = torch.prim.unchecked_cast %137 : !torch.optional<!torch.tensor> -> !torch.tensor
%305 = torch.operator "aten.contiguous"(%304, %int0) : (!torch.tensor, !torch.int) -> !torch.tensor
%int8_72 = torch.constant.int 8
%306 = torch.aten.mul.int %intx, %int8_72 : !torch.int, !torch.int -> !torch.int
%int64_73 = torch.constant.int 64
builtin.func@test(%intx :!torch.int, %145 : !torch.bool, %136 : !torch.optional<!torch.tensor>, %137 : !torch.optional<!torch.tensor>) -> !torch.optional<!torch.tensor> {
%146 = torch.prim.If %145 -> (!torch.optional<!torch.tensor>) {
%int0 = torch.constant.int 0
%int1 = torch.constant.int 1
%int-1 = torch.constant.int -1
%304 = torch.prim.unchecked_cast %137 : !torch.optional<!torch.tensor> -> !torch.tensor
%305 = torch.operator "aten.contiguous"(%304, %int0) : (!torch.tensor, !torch.int) -> !torch.tensor
%int8_72 = torch.constant.int 8
%306 = torch.aten.mul.int %intx, %int8_72 : !torch.int, !torch.int -> !torch.int
%int64_73 = torch.constant.int 64
func @forward(%arg0: !torch.vtensor<[?,?],si64>) -> !torch.vtensor<*,f32> {
%false = torch.constant.bool false
%float1.000000e-01 = torch.constant.float 1.000000e-01
%float1.000000e00 = torch.constant.float 1.000000e+00
%int6 = torch.constant.int 6
%int-2 = torch.constant.int -2
%int-1 = torch.constant.int -1
%int3 = torch.constant.int 3
%int2 = torch.constant.int 2
%int64 = torch.constant.int 64
func @forward(%arg0: !torch.vtensor<[?,?],si64>) -> !torch.vtensor<*,f32> {
%false = torch.constant.bool false
%float1.000000e-01 = torch.constant.float 1.000000e-01
%float1.000000e00 = torch.constant.float 1.000000e+00
%int6 = torch.constant.int 6
%int-2 = torch.constant.int -2
%int-1 = torch.constant.int -1
%int3 = torch.constant.int 3
%int2 = torch.constant.int 2
%int32 = torch.constant.int 32
This file has been truncated, but you can view the full file.
ninja: Entering directory `/usr/local/google/home/cathyzhyi/npcomp/mlir-npcomp/build'
ninja: no work to do.
ninja: Entering directory `/usr/local/google/home/cathyzhyi/npcomp/mlir-npcomp/build'
ninja: no work to do.
// -----// IR Dump After SymbolDCE //----- //
module {
func private @__torch__.torch.nn.modules.sparse.___torch_mangle_227.Embedding.forward(%arg0: !torch.nn.Module<"__torch__.torch.nn.modules.sparse.___torch_mangle_227.Embedding">, %arg1: !torch.tensor) -> !torch.tensor {
%false_0 = torch.constant.bool false
%int0 = torch.constant.int 0
%225 = torch.prim.GetAttr %arg0["weight"] : !torch.nn.Module<"__torch__.torch.nn.modules.sparse.___torch_mangle_227.Embedding"> -> !torch.tensor
This file has been truncated, but you can view the full file.
ninja: Entering directory `/usr/local/google/home/cathyzhyi/npcomp/mlir-npcomp/build'
ninja: no work to do.
ninja: Entering directory `/usr/local/google/home/cathyzhyi/npcomp/mlir-npcomp/build'
ninja: no work to do.
// -----// IR Dump After SymbolDCE //----- //
module {
func private @__torch__.torch.nn.modules.sparse.___torch_mangle_227.Embedding.forward(%arg0: !torch.nn.Module<"__torch__.torch.nn.modules.sparse.___torch_mangle_227.Embedding">, %arg1: !torch.tensor) -> !torch.tensor {
%false_0 = torch.constant.bool false
%int0 = torch.constant.int 0
%225 = torch.prim.GetAttr %arg0["weight"] : !torch.nn.Module<"__torch__.torch.nn.modules.sparse.___torch_mangle_227.Embedding"> -> !torch.tensor
" File: ~/.init.vim
" Discription: NEOVIM Configuration
" Author: Yi Zhang
" Time: 2019/09/21
"
"
"
" --- if want to use the VIM configuration ---
" set runtimepath+=~/.vim,~/.vim/after
" set packpath+=~/.vim
//===-------------------------------------------===//
Legalizing operation : 'builtin.module'(0x5a3c2a0) {
* Fold {
} -> FAILURE : unable to fold
} -> FAILURE : no matched legalization pattern
//===-------------------------------------------===//
//===-------------------------------------------===//
Legalizing operation : 'builtin.func'(0x5a449d0) {
* Fold {