forked from apache/tvm
-
Notifications
You must be signed in to change notification settings - Fork 5
kurisu add assume attr patch 1 #8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Closed
kurisu6912
wants to merge
801
commits into
tile-ai:tilelang_codebase
from
kurisu6912:kurisu-add-assume-attr-patch-1
Closed
kurisu add assume attr patch 1 #8
kurisu6912
wants to merge
801
commits into
tile-ai:tilelang_codebase
from
kurisu6912:kurisu-add-assume-attr-patch-1
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
* fix reduce buffer allocation position * fix test_tir_analysis_detect_buffer_access_lca.py::test_buffer_load_store
* fix test_flatten * re-enable test_split * fix test_to_copy * re-enable test_batchnorm2d
* Update fx_translator.py * Update base_fx_graph_translator.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * fix lint
…er (apache#17812) * Update exported_program_translator.py * Update test_frontend_from_exported_program.py * Update test_frontend_from_exported_program.py * Update test_frontend_from_exported_program.py
In this PR I have added jit support for sampling flashinfer kernel. I have also added a unit test to test the jit compiled flashinfer kernel.
…ls (apache#17796) In this PR I have made changes so that we can support CUBLAS dispatch operations for bfloat16 data type.
…atten.int`, `hardtanh_.default`, `dropout_.default`, `silu_.default`, `add_.Tensor` and `relu_.default` (apache#17813) * support `relu_.default` * support `add_.Tensor` * support `silu_.default` * support `dropout_.default` * support `hardswish_.default` * support `hardtanh_.default` * support `unflatten.int` * fix lint error
…pache#17817) support dynamic shape
…graph (apache#17806) * add softplus op into exported program and fx graph frontend * fixing trailing whitespace issue * fixing lint issues * fix lint issue on docs * modify description to avoid cpplints issue * update softplus function with threshold attr * remove trailing spaces in softplus func * fix lint issues in legalize func * fixing cpp lints issue * test script for both exported and fx graph * trim trailing spaces iin test script * fix lint issues in test script * unit test script is added in test frontend op files * fixing lint issues in test_op_nn file * fixing attribute error in test script * fixing lint issues in test script functions * adding softplus wrapper function in op file --------- Co-authored-by: deivanayakisankaralingam <deiva@Deivanayaki>
…7822) * move gelu, relu, selu, sigmoid, silu tests to test_basic_unary_ops * remove unused torchversion * we don't need to manually call `test_*` functions * remove unused variable
* Update fx_translator.py * Update base_fx_graph_translator.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py
…ms translator (apache#17814) * stack correct * sum correct in side script * all pass
This PR fixes a bug where running "pip install -e /path-to-tvm/python" fails if installation files remain in python/tvm. The fix includes: - Preventing libraries from `python/tvm` from being appended to the library list, resolving the shutil.SameFileError exception raised by shutil.copy() - Adding cleanup logic earlier in case it was not executed due to a previous pip installation failure, resolving the FileExistsError exception raised by shutil.copytree()
* enhance missing func types finding in exported program and fx graph frontend * fix trailing space issue * fix lint issues by formatting the code * fix name error in fx frontend --------- Co-authored-by: deivanayakisankaralingam <deiva@Deivanayaki>
…ram importer (apache#17830) * Update exported_program_translator.py * Update test_frontend_from_exported_program.py * Update test_frontend_from_exported_program.py * Update test_frontend_from_exported_program.py
…e#17832) * unit test * full.default * linting * ones ok * tests for ones, full, and full like work
…pache#17838) This PR fixes a bug in reverse-compute-inline of tir Schedule, which generates incorrect TIR after inlining a transpose block into a reduction block.
…ends (apache#17840) * combine missing op logic of export and fx graph into common utilities * move func call above builder and fix lint issue * add type hint for nodes in helper function --------- Co-authored-by: deivanayakisankaralingam <deiva@Deivanayaki>
* Update fx_translator.py * Update base_fx_graph_translator.py * Update test_frontend_from_fx.py * Update base_fx_graph_translator.py * Update test_frontend_from_fx.py * Update base_fx_graph_translator.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py * Update test_frontend_from_fx.py
…e#17803) * Add support for logaddexp core operator * Add test script for logaddexp * Add fix for lint issues * Adjust trailing spaces * Adjust leading whitespace * Add fix for lint inssues * Add fix for logaddexp test script * Fix lint issues * decomposition at op level * unity check --------- Co-authored-by: Pratheesh <[email protected]>
Update upsampling.py fix the incorrect docstring
* use `ubuntu-latest` for github ci * use `ubuntu-22.04` for android build
…ph (apache#17816) * prelu op support and test script added * end-of-file issue fixed * trailing whitespace issue fixed * fixing lint issues * fix assertion error in test_op_nn.py file * add test script in test_frontend_nn_op.py * include wrapper function for prelu in op.py * fixing unity check issue by modifying test func * conflicts resolved * add doc for prelu op axis arg * fixed failing checks issue --------- Co-authored-by: deivanayakisankaralingam <deiva@Deivanayaki>
…e#17850) * Add masked_fill support in exportedProgram * Fix lint issues
apache#17849) * add rsub op support into exported and fx graph frontend * fix trailing whitespace issue * fix lint issues in test scripts --------- Co-authored-by: deivanayakisankaralingam <deiva@Deivanayaki>
* Update batch_to_space_nd.py * Update bitserial_util.py
This PR modules reflection of ffi into registry.h and accessor.h the dependent items are updated accordingly
…ef (apache#18148) This PR migrates the remaining global def reg to use the new mechanism. It also phases out the TVM_FFI_REGISTER_GLOBAL macro in favor of the GlobalDef mechanism.
This PR changes the function global dup registration to log and throw so we have clear error message about the function duplication.
…command for tvm_cython target
* [COMMUNITY] Add new key for release signing * [Misc] Update test_release_package.sh Fix release script according tianqi advice (apache#17861 (comment)).
This PR decouples deep equal from structural equal implementation by providing a more direct implementatio through functor. DeepEqual is being used at heart of arith simplification as subroutine and it performs more direct nested checking without doing var remapping as structural equal for efficiency reasons. It also do not need to trace the wrong comparison since the failed path is also expected to happen often. This step likely will improve the deep equal efficiency because of the more direct approach and gives us opportunity to run simplify future refactor of structural equal to focus on struct path tracing.
This PR add initial support for structural equal and hash via the new reflection mechanism. It will helps us to streamline the structural equality/hash with broader support and clean error reports via AccessPath. It also gives us ability to unify all struct equal/hash registration into the extra meta-data in reflection registration.
apache#18120) * Root cause * Update
This PR fixes a build failure in nccl.cc due to the recent switch of global function registration.
This PR introduces TypeAttr to reflection to bring extra optional attribute registration that can be used to extend behaviors such as structural equality. Also renames TypeExtraInfo to TypeMetadata for better clarity.
…of BoolOp nodes, improving code clarity.
… details in the error message for better debugging context.
…ity and maintainability.
…ctRef> to Map<String, Any> for improved flexibility.
… compatibility." This reverts commit 9574805.
…ing variable naming consistency.
… Map<String, Any> for enhanced flexibility in handling annotations.
…ith various attributes for enhanced GPU compatibility (tile-ai#7) Co-authored-by: xinyxiao <[email protected]>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
ci_cpu
([CI] Use LLVM17 for tests onci_cpu
apache/tvm#16931)arm_cpu
targets ([TOPI] Revert unification of conv2d NHWC hybrid scheduling forarm_cpu
targets apache/tvm#16951)vscale_range()
function attribute ([SVE] Add codegen support forvscale_range()
function attribute apache/tvm#16962)blockIdx.z
in topi sort ([TOPI] RemoveblockIdx.z
in topi sort apache/tvm#16977)num_workers
property fordisco.Session
([Disco] Implementnum_workers
property fordisco.Session
apache/tvm#16978)arm_cpu
([TOPI][Testing] Enable conv2d NHWC fp16 topi testing forarm_cpu
apache/tvm#17007)int8x4
intou32
([WebGPU] Translateint8x4
intou32
apache/tvm#17071)input_axis_separator
to allow 2D to 1D conversion ([Relax] Supportinput_axis_separator
to allow 2D to 1D conversion apache/tvm#17115)tir.dp4a
([WebGPU] Addtir.dp4a
apache/tvm#17124)tir.dp4a
with WGSL built-in functiondot4I8Packed
([WebGPU] Implementtir.dp4a
with WGSL built-in functiondot4I8Packed
apache/tvm#16976)@type_check
forset_axis_separator
([TIR][Schedule] Remove@type_check
forset_axis_separator
apache/tvm#17134)unity/pr-head
step ([CI] Remove lint step fromunity/pr-head
step apache/tvm#17155)packaging.version.parse
instead ofdistutils.version.LooseVersion
(Usepackaging.version.parse
instead ofdistutils.version.LooseVersion
apache/tvm#17173)packaging
topython/gen_requirements.py
(Addpackaging
topython/gen_requirements.py
apache/tvm#17188)xgboost.rabit
withxgboost.collective
because it's deprecated ([MetaSchedule] Replacexgboost.rabit
withxgboost.collective
because it's deprecated apache/tvm#17166)distutils.util.strtobool()
(Remove and replace deprecateddistutils.util.strtobool()
apache/tvm#17185)x==x
expressions for all dtypes ([TIR][Analyzer] Simplifyx==x
expressions for all dtypes apache/tvm#17158)torch.nn.functional.max_pool2d
([Relax][PyTorch] Add support fortorch.nn.functional.max_pool2d
apache/tvm#17189)is_group
argument in IPC AllReduce ([Transform][Relax] Handleis_group
argument in IPC AllReduce apache/tvm#17201)aten::tile
([Relay][Pytorch] Add support foraten::tile
apache/tvm#17277)tir::Let
as var assignment explicitly ([Codegen] Emittir::Let
as var assignment explicitly apache/tvm#17278)using namespace tvm::runtime
from headers ([Cleanup] Removeusing namespace tvm::runtime
from headers apache/tvm#17246)tvm.
prefix from image name when./docker/build.sh
([Fix] Removetvm.
prefix from image name when./docker/build.sh
apache/tvm#17324)torch.nn.functional.conv*
([Relax][PyTorch] Add support fortorch.nn.functional.conv*
apache/tvm#17325)layer_norm
converter to supportimmutable_list
fornormalized_shape
([Relax][PyTorch][Bugfix] Updatelayer_norm
converter to supportimmutable_list
fornormalized_shape
apache/tvm#17330)_convert_torch_tensor_to_relax()
where possible ([Relax][PyTorch][Fix] use_convert_torch_tensor_to_relax()
where possible apache/tvm#17335)torch.ops.aten.sym_size.int
([Relax][PyTorch] Add support fortorch.ops.aten.sym_size.int
apache/tvm#17342)R.call_tir
([Relax] Require correct input/output shapesR.call_tir
apache/tvm#17285)_attention_sequence_prefill
function to … ([Relax][KV Cache] Refactor_attention_sequence_prefill
function to … apache/tvm#17362)torch.nn.functional.scaled_dot_product_attention
([Relax][PyTorch] Fix output shape oftorch.nn.functional.scaled_dot_product_attention
apache/tvm#17379)20240917-153130-9f281758
([CI] Upgrade unity image tag to20240917-153130-9f281758
apache/tvm#17410)torch.export.ExportedProgram
in Relax PyTorch Frontend ([Relax][PyTorch] Add support fortorch.export.ExportedProgram
in Relax PyTorch Frontend apache/tvm#17396)is_vector
Method to DataType class and update usages across Codebase ([TIR] Addis_vector
Method to DataType class and update usages across Codebase apache/tvm#17443)torch.export
insteamd offx.symbolic_trace
for tutorial ([Relax][PyTorch][Docs] Usetorch.export
insteamd offx.symbolic_trace
for tutorial apache/tvm#17436)np.int
withnp.int32
(Replacenp.int
withnp.int32
apache/tvm#17484)oneflow==0.9.0
([CI] Upgradeoneflow==0.9.0
apache/tvm#17503)operator[]
function ([TOPI] Move ewise.h -> elemwise.h apache/tvm#327) ([3rdparty] Update Picojson with constoperator[]
function (#327) apache/tvm#17532)20241105-030952-3e386fd3
([CI] Upgrade CI image to20241105-030952-3e386fd3
apache/tvm#17451)decorator
([Refactor] Phrase out python dependencydecorator
apache/tvm#17661)<chrono>
forstd::chrono
([Fix] Include<chrono>
forstd::chrono
apache/tvm#17697)tvm.compile
interface ([Refactor] Introduce base Executable class andtvm.compile
interface apache/tvm#17710)tvm.compile
([Refactor] Migrate build API totvm.compile
apache/tvm#17718)_to
(#17809)explict
typo (#17811)unflatten.int
,hardtanh_.default
,dropout_.default
,silu_.default
,add_.Tensor
andrelu_.default
(#17813)leaky_relu_.default
andreshape_as.default
in ExportedProgram frontend (#17851)triton.compile
interface (#17913)20250513-063354-70aa3797
(#17981)mapping.TENSOR_TYPE_TO_NP_TYPE
usage (#18016)GetFunction
of metal runtime (#18034)relax_vm
tovm
(#18049)__attribute__
with C++ standard attributes (#18114)Arg2Str
with a more powerfulfor_each
(#18117)Arg2Str
with a more powerfulfor_each
" (#18121)insertDeclare
(#18123)std::move()
calls (#18130)T.thread_return()
for early thread exit in CUDA kernels (#18134)TupleRewriterNode
(#18120)