diff --git a/README.md b/README.md index b080111b33cf7021e000c6cdfc9186d7fb54455b..bf9f2ddd7ef78ebf080e4e4cefae031874c39667 100644 --- a/README.md +++ b/README.md @@ -4,6 +4,3 @@ tags: - kernel --- -![Status](https://hubwebhook.dholtz.com/shield?repo=kernels-community/deformable-detr) - -## deformable-detr diff --git a/build.toml b/build.toml index f39de7e929172dc4ddd49c9e50efd99a432a1cfe..09496f4d32b405396fab6b6aba45e3a1e51402b5 100644 --- a/build.toml +++ b/build.toml @@ -1,20 +1,19 @@ [general] name = "deformable_detr" -universal = false [torch] src = [ - "torch-ext/torch_binding.cpp", - "torch-ext/torch_binding.h", + "torch-ext/torch_binding.cpp", + "torch-ext/torch_binding.h" ] [kernel.activation] -backend = "cuda" -depends = ["torch"] -include = ["."] +cuda-capabilities = [ "7.0", "7.2", "7.5", "8.0", "8.6", "8.7", "8.9", "9.0" ] src = [ - "deformable_detr/ms_deform_attn_cuda.cu", - "deformable_detr/ms_deform_im2col_cuda.cuh", - "deformable_detr/ms_deform_attn_cuda.cuh", - "deformable_detr/ms_deform_attn_cuda.h", + "deformable_detr/ms_deform_attn_cuda.cu", + "deformable_detr/ms_deform_im2col_cuda.cuh", + "deformable_detr/ms_deform_attn_cuda.cuh", + "deformable_detr/ms_deform_attn_cuda.h", ] +include = ["."] +depends = [ "torch" ] diff --git a/build/torch25-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch25-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index f622604e89058689647ba41b3c71ccbd3aa68ae7..0000000000000000000000000000000000000000 --- a/build/torch25-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:ae00c12295a458e2534149aea16da0289541447123c19fae59baaf6d6d2752f1 -size 6693656 diff --git a/build/torch25-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_cxy6p3o2latjs.abi3.so b/build/torch25-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_cxy6p3o2latjs.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..4e4da3ee7fdd606fe0498f09c1eb4d758861af5e --- /dev/null +++ b/build/torch25-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_cxy6p3o2latjs.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1cf71a0243675c22ba3207a6f895a907b0699f964575088e054220cea5e2fb2e +size 5870376 diff --git a/build/torch25-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py b/build/torch25-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..ffb73bc1bbd96dfb75a830b83b58676cd1377989 100644 --- a/build/torch25-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch25-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_cxy6p3o2latjs +ops = torch.ops._deformable_detr_cxy6p3o2latjs def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_cxy6p3o2latjs::{op_name}" \ No newline at end of file diff --git a/build/torch25-cxx11-cu121-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch25-cxx11-cu121-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 25768c52fa2f7041778baad9b1beaaee6772e8d3..0000000000000000000000000000000000000000 --- a/build/torch25-cxx11-cu121-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4ab8cf59779b768359df0fa268b6cd52be2f518dd4fafdd61baec31c64f44813 -size 6679440 diff --git a/build/torch25-cxx11-cu121-x86_64-linux/deformable_detr/_deformable_detr_esifsbuexbtbw.abi3.so b/build/torch25-cxx11-cu121-x86_64-linux/deformable_detr/_deformable_detr_esifsbuexbtbw.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..dd7fd6ca351b521f64a078d315551edad1f02f8b --- /dev/null +++ b/build/torch25-cxx11-cu121-x86_64-linux/deformable_detr/_deformable_detr_esifsbuexbtbw.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79dce2e84e09fb2a5bf1b47441b226343494807687d8829f141682af9b78e361 +size 5856160 diff --git a/build/torch25-cxx11-cu121-x86_64-linux/deformable_detr/_ops.py b/build/torch25-cxx11-cu121-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..7272205607f6b6db5e3f6aa7673bc77798647317 100644 --- a/build/torch25-cxx11-cu121-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch25-cxx11-cu121-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_esifsbuexbtbw +ops = torch.ops._deformable_detr_esifsbuexbtbw def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_esifsbuexbtbw::{op_name}" \ No newline at end of file diff --git a/build/torch25-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch25-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index a574c7a01d734f9f4fb7210058d3064c280c5564..0000000000000000000000000000000000000000 --- a/build/torch25-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:5cdcd6902a03140074cff4cd44bf6b47dc27a32e13e0515a93929c66be186cab -size 6652680 diff --git a/build/torch25-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_cuzn3o54ku5iq.abi3.so b/build/torch25-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_cuzn3o54ku5iq.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..1792ad2bb4c1dc835a4fdf2e54a4d9b0ad354ec9 --- /dev/null +++ b/build/torch25-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_cuzn3o54ku5iq.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:825c7cb6f9a4350bdcdffa4383d7a527d5fa7b0d9d83222f5d1e72f1c6087841 +size 5841688 diff --git a/build/torch25-cxx11-cu124-x86_64-linux/deformable_detr/_ops.py b/build/torch25-cxx11-cu124-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..3534aa4eb5695150d1dd7d8c0f439fb6452edca9 100644 --- a/build/torch25-cxx11-cu124-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch25-cxx11-cu124-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_cuzn3o54ku5iq +ops = torch.ops._deformable_detr_cuzn3o54ku5iq def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_cuzn3o54ku5iq::{op_name}" \ No newline at end of file diff --git a/build/torch25-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch25-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 6360ed5a03fd9fbe757e8b7c1816f9935b13764f..0000000000000000000000000000000000000000 --- a/build/torch25-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:82174ec2812ee672a447b94fb5ec907e348eb3d0be338daddf145a1d74969a6f -size 6686592 diff --git a/build/torch25-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_gom2c5vfrl2ic.abi3.so b/build/torch25-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_gom2c5vfrl2ic.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..95c51f0c1f8a7e90e248bc945259220e468dfd91 --- /dev/null +++ b/build/torch25-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_gom2c5vfrl2ic.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbe4c67fc885df711581660f72d86dbd0a237c7f106308e55a484725c88e9927 +size 5863312 diff --git a/build/torch25-cxx98-cu118-x86_64-linux/deformable_detr/_ops.py b/build/torch25-cxx98-cu118-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..734f23a0e93ba1c1e7e8bf8501df925bebdee046 100644 --- a/build/torch25-cxx98-cu118-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch25-cxx98-cu118-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_gom2c5vfrl2ic +ops = torch.ops._deformable_detr_gom2c5vfrl2ic def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_gom2c5vfrl2ic::{op_name}" \ No newline at end of file diff --git a/build/torch25-cxx98-cu121-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch25-cxx98-cu121-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index d23c2f012773987542cbcab34a84e3eb240bdff9..0000000000000000000000000000000000000000 --- a/build/torch25-cxx98-cu121-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c9d8540a4ffa00d331f60204fe6baf543a45667d6bba2c0a0b23aca9202b6233 -size 6672464 diff --git a/build/torch25-cxx98-cu121-x86_64-linux/deformable_detr/_deformable_detr_a7sajsuqrick6.abi3.so b/build/torch25-cxx98-cu121-x86_64-linux/deformable_detr/_deformable_detr_a7sajsuqrick6.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..0bd1ef4078bb421da435e98a7bb2f9311eda0739 --- /dev/null +++ b/build/torch25-cxx98-cu121-x86_64-linux/deformable_detr/_deformable_detr_a7sajsuqrick6.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56b4c64eb7931a6f580bd5b806eae1aea43b3bb8c0f115d5d202f151974a5e7b +size 5853280 diff --git a/build/torch25-cxx98-cu121-x86_64-linux/deformable_detr/_ops.py b/build/torch25-cxx98-cu121-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..4531116c7e3fbc3258dcd23ebdbe22768a8d566a 100644 --- a/build/torch25-cxx98-cu121-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch25-cxx98-cu121-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_a7sajsuqrick6 +ops = torch.ops._deformable_detr_a7sajsuqrick6 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_a7sajsuqrick6::{op_name}" \ No newline at end of file diff --git a/build/torch25-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch25-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index aaa5548734ba34766edf646277e25b48752d6b7f..0000000000000000000000000000000000000000 --- a/build/torch25-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fa748a4de72c06de09f46b4af4fec7f23cb2c76eb8683c117fefd20833cd3fd8 -size 6649800 diff --git a/build/torch25-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_tyogxwmtolvok.abi3.so b/build/torch25-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_tyogxwmtolvok.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..b53558717a3608677dd75c82cdc027ab1374179d --- /dev/null +++ b/build/torch25-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_tyogxwmtolvok.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7eef07a96ddf574e5b1e07476089a62659a70faa33c82fc79987c54fecb2711f +size 5834712 diff --git a/build/torch25-cxx98-cu124-x86_64-linux/deformable_detr/_ops.py b/build/torch25-cxx98-cu124-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..b3c0318eaacce12db912a3c5e91ad578adbe74ca 100644 --- a/build/torch25-cxx98-cu124-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch25-cxx98-cu124-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_tyogxwmtolvok +ops = torch.ops._deformable_detr_tyogxwmtolvok def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_tyogxwmtolvok::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_5kxpyt5yogkv2.abi3.so b/build/torch26-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_5kxpyt5yogkv2.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..11b825669658ee3d1a209111ce5c954f47627b64 --- /dev/null +++ b/build/torch26-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_5kxpyt5yogkv2.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1c5bb5376002363e2008eb6db64ebe0c9f6c31f9a635b7420ddfb46dce16b02 +size 5870352 diff --git a/build/torch26-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch26-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 64ea4bcc2dfb73951fa3315247c594ea5b6aee1c..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:57e38bac3087c1446307e504b1e22e61ae584d1de7f5b3d15bd7a60780c3431c -size 6693632 diff --git a/build/torch26-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py b/build/torch26-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..68d3b35da5c2af46145714ea08cdb287d2a50d5d 100644 --- a/build/torch26-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch26-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_5kxpyt5yogkv2 +ops = torch.ops._deformable_detr_5kxpyt5yogkv2 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_5kxpyt5yogkv2::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch26-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 658ac25c98e67be8affb211b20fc2af46efcbca2..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1d7c09d3bedd89d7119e7023a07784724d3a3f79664b75fce37b778ef3bcfe52 -size 6648656 diff --git a/build/torch26-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_titoehueyfqjg.abi3.so b/build/torch26-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_titoehueyfqjg.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..b64a727062504abecbb7bd162d6f8c6848de3b53 --- /dev/null +++ b/build/torch26-cxx11-cu124-x86_64-linux/deformable_detr/_deformable_detr_titoehueyfqjg.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76b74d4bdbb1f562474b987fd23430d12b9f033183198f35a7dfd21fcc8ce4e1 +size 5837664 diff --git a/build/torch26-cxx11-cu124-x86_64-linux/deformable_detr/_ops.py b/build/torch26-cxx11-cu124-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..0df7bff1cf00c94ce4da6eee12f6e49f513bc3b8 100644 --- a/build/torch26-cxx11-cu124-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch26-cxx11-cu124-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_titoehueyfqjg +ops = torch.ops._deformable_detr_titoehueyfqjg def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_titoehueyfqjg::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/__init__.py b/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index d1b1fe211439d00ca4681305f56e91b4027fe45e..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6856b6efe5130f019f6cb7f964d7a2073f1ecc5cd7afc850334e64798f871dae -size 6833224 diff --git a/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/_ops.py b/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/_ops.py deleted file mode 100644 index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/layers.py b/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu126-aarch64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch26-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch26-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 10bcc5c4bad459abe354c507166121a5a47d5d74..0000000000000000000000000000000000000000 --- a/build/torch26-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c76ad874b78882d3108a7fdaf49f8c00b6a6a7dceec63912118f8fa7d07e5f30 -size 6800656 diff --git a/build/torch26-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_imqt5tuqtmyt4.abi3.so b/build/torch26-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_imqt5tuqtmyt4.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..61011f23d656b543339a1eb2e1481684e3d66e10 --- /dev/null +++ b/build/torch26-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_imqt5tuqtmyt4.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1acd032c2f3bc530872e0839d8bec8950b01668c913539a2e14008a1e652560f +size 5944608 diff --git a/build/torch26-cxx11-cu126-x86_64-linux/deformable_detr/_ops.py b/build/torch26-cxx11-cu126-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..4ea7469cf391720f056311bda128fbec93749437 100644 --- a/build/torch26-cxx11-cu126-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch26-cxx11-cu126-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_imqt5tuqtmyt4 +ops = torch.ops._deformable_detr_imqt5tuqtmyt4 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_imqt5tuqtmyt4::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch26-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index a15ec9d8735e3de07ebd3bfaae1b2ded8de87447..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d1fb3a24fd95c1cc3cba080ae1c9d4217f377435770c7e423de53b11ecc437dc -size 6686600 diff --git a/build/torch26-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_qbnaho3zp2d3o.abi3.so b/build/torch26-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_qbnaho3zp2d3o.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..353504e4fa52b281748f32dfcee02781e450794a --- /dev/null +++ b/build/torch26-cxx98-cu118-x86_64-linux/deformable_detr/_deformable_detr_qbnaho3zp2d3o.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9e5074a5afdb137688e20182cf4c9f7cbb1e8a69651c08a570076aeedc8c76b +size 5863320 diff --git a/build/torch26-cxx98-cu118-x86_64-linux/deformable_detr/_ops.py b/build/torch26-cxx98-cu118-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..ed42e13a067a186a1f701cec376afa3194956f6b 100644 --- a/build/torch26-cxx98-cu118-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch26-cxx98-cu118-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_qbnaho3zp2d3o +ops = torch.ops._deformable_detr_qbnaho3zp2d3o def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_qbnaho3zp2d3o::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_5oxft6tr6jbvu.abi3.so b/build/torch26-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_5oxft6tr6jbvu.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..e799c57967e1132680a7e360d5cb6b197932b787 --- /dev/null +++ b/build/torch26-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_5oxft6tr6jbvu.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd4d0f47c165b9ce95c0328cb7a52e331e4c698746ea8e4d43c7d09c193e34bd +size 5834720 diff --git a/build/torch26-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch26-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 63c34c123c4b78349bc1a2bb92b0614c4b689a13..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu124-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d5a455975be5790964cc95c6813d293b1aba581f5c2dc132c9a08690bf6e5cad -size 6649808 diff --git a/build/torch26-cxx98-cu124-x86_64-linux/deformable_detr/_ops.py b/build/torch26-cxx98-cu124-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..bb7524858e03d0f913bd747f4c79db60d2737aa2 100644 --- a/build/torch26-cxx98-cu124-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch26-cxx98-cu124-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_5oxft6tr6jbvu +ops = torch.ops._deformable_detr_5oxft6tr6jbvu def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_5oxft6tr6jbvu::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/__init__.py b/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 6740aef68abf0d71d0756d6b2f78fc2da67c6752..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2c13dce2b080676eb192d87ba83df6ef1f6d0f1101727f4b29185d48dec7281d -size 6829872 diff --git a/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/_ops.py b/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/_ops.py deleted file mode 100644 index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file diff --git a/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/layers.py b/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-aarch64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch26-cxx98-cu126-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch26-cxx98-cu126-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index f3ce58cab856be1f382aa9f3cf1bb0b1ec06d41d..0000000000000000000000000000000000000000 --- a/build/torch26-cxx98-cu126-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c6ad344319579f0abef7fe1a9d3f479f1c8737994f563a540815a1445020959e -size 6797712 diff --git a/build/torch26-cxx98-cu126-x86_64-linux/deformable_detr/_deformable_detr_po264mz2i2ffg.abi3.so b/build/torch26-cxx98-cu126-x86_64-linux/deformable_detr/_deformable_detr_po264mz2i2ffg.abi3.so new file mode 100755 index 0000000000000000000000000000000000000000..952c1115177dbfead31b4cfe79df0fe6175b0fe1 --- /dev/null +++ b/build/torch26-cxx98-cu126-x86_64-linux/deformable_detr/_deformable_detr_po264mz2i2ffg.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:129844ba533ee201cd3f2bb0e17a354ee8aa35176c10896454926485acdacdac +size 5945760 diff --git a/build/torch26-cxx98-cu126-x86_64-linux/deformable_detr/_ops.py b/build/torch26-cxx98-cu126-x86_64-linux/deformable_detr/_ops.py index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..c411feebca4fae25684ed9b07abfdc21f1c55488 100644 --- a/build/torch26-cxx98-cu126-x86_64-linux/deformable_detr/_ops.py +++ b/build/torch26-cxx98-cu126-x86_64-linux/deformable_detr/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe +from . import _deformable_detr_po264mz2i2ffg +ops = torch.ops._deformable_detr_po264mz2i2ffg def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file + return f"_deformable_detr_po264mz2i2ffg::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/__init__.py b/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index ba3b9aa83048be82868858b43e50e1fcdea6aae6..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:20ee462e214e11abd65b44259847f7dbf535b94a3003357251c53a72f4ac4392 -size 6693728 diff --git a/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py b/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py deleted file mode 100644 index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/layers.py b/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu118-x86_64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/__init__.py b/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 63b666838685ae29d11438e79273b09b38f9df39..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:af2831b68229a910e8703cae2c9e720ded825e401745d38923548c444e56c37b -size 6833456 diff --git a/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/_ops.py b/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/_ops.py deleted file mode 100644 index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/layers.py b/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-aarch64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/__init__.py b/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 282b10d9f49f52aea681ca798589a4779343bb03..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c2a2064d5d7b5f62704d4cad6a530dd812804cfcc426889030fefdcc553828d3 -size 6800688 diff --git a/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/_ops.py b/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/_ops.py deleted file mode 100644 index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/layers.py b/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu126-x86_64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/__init__.py b/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 806a6cf1a90e99117a7f7e4f47a088345c2987f1..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:cf9d0207063dca2d03f2200aa78249daf14551288db1d247d56278bf06449def -size 9848208 diff --git a/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/_ops.py b/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/_ops.py deleted file mode 100644 index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/layers.py b/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu128-aarch64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/__init__.py b/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so b/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so deleted file mode 100755 index 53d8d25f1ede4d8b40761eebef271c19c8ebf201..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/_deformable_detr_7c33cbe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1f819d4d59e9ea2ce0afd8a0198784cb9f1c3100a0d352261bf84d5b8570cda7 -size 9796776 diff --git a/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/_ops.py b/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/_ops.py deleted file mode 100644 index 676ee6fea64b714dedb7ccd1d54148dcf75575a6..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_7c33cbe -ops = torch.ops._deformable_detr_7c33cbe - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_7c33cbe::{op_name}" \ No newline at end of file diff --git a/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/layers.py b/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch27-cxx11-cu128-x86_64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__init__.py b/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__pycache__/__init__.cpython-313.pyc b/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index 664a914e405aa872821d228b552f22a99cb39d97..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__pycache__/_ops.cpython-313.pyc b/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__pycache__/_ops.cpython-313.pyc deleted file mode 100644 index 273cdfd52347ebcf0e0c050945d37c5dc5094a9a..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__pycache__/_ops.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__pycache__/layers.cpython-313.pyc b/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__pycache__/layers.cpython-313.pyc deleted file mode 100644 index 082d05eb36d7ef831ca076e3e7d2eb070130c2ec..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/__pycache__/layers.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/_deformable_detr_a92c8ea_dirty.abi3.so b/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/_deformable_detr_a92c8ea_dirty.abi3.so deleted file mode 100755 index 3d206228684c315eb04160dc62c8a7cb6811156a..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/_deformable_detr_a92c8ea_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b4a5665b05309312200ca97a80cc61340c0f5de123ab33254e5307a5ec4ed2a0 -size 6901024 diff --git a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/_ops.py b/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/_ops.py deleted file mode 100644 index 019cf7ce30dc11d9b791075404417b1ac47500e7..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_a92c8ea_dirty -ops = torch.ops._deformable_detr_a92c8ea_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_a92c8ea_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/layers.py b/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-aarch64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/__init__.py b/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_e82640e_dirty.abi3.so b/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_e82640e_dirty.abi3.so deleted file mode 100755 index 2d6c73a98077e6cd6ed5738989d824fbd5196708..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/_deformable_detr_e82640e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:8877a2159e6f4f04905120af8d10ab9b8294aff80b97a584d6a6716922141b46 -size 6802152 diff --git a/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/_ops.py b/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/_ops.py deleted file mode 100644 index 9de8ab380209628599d241c68159af782399e651..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_e82640e_dirty -ops = torch.ops._deformable_detr_e82640e_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_e82640e_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/layers.py b/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu126-x86_64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__init__.py b/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__pycache__/__init__.cpython-313.pyc b/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index 28a5933156ae99fa286e692f8e98dd04e50f5174..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__pycache__/_ops.cpython-313.pyc b/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__pycache__/_ops.cpython-313.pyc deleted file mode 100644 index 0009fad4a606c6aacd228a2d7af5e6263d47d861..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__pycache__/_ops.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__pycache__/layers.cpython-313.pyc b/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__pycache__/layers.cpython-313.pyc deleted file mode 100644 index 18befc2ceb219257a717fe3c9800618101a3904c..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/__pycache__/layers.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/_deformable_detr_a92c8ea_dirty.abi3.so b/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/_deformable_detr_a92c8ea_dirty.abi3.so deleted file mode 100755 index 7a0b563d8285a538c60f993ba90ffee3d5b49ec6..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/_deformable_detr_a92c8ea_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:3f4a0a7850968822e26e3a59c801fd711231d5294193155efbf9583761e114ef -size 9849688 diff --git a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/_ops.py b/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/_ops.py deleted file mode 100644 index 019cf7ce30dc11d9b791075404417b1ac47500e7..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_a92c8ea_dirty -ops = torch.ops._deformable_detr_a92c8ea_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_a92c8ea_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/layers.py b/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-aarch64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/__init__.py b/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/_deformable_detr_e82640e_dirty.abi3.so b/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/_deformable_detr_e82640e_dirty.abi3.so deleted file mode 100755 index 035ad37af4ff31fd302bbc78d1ca6fe908af33e0..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/_deformable_detr_e82640e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:f6bfdf67f0c09d9f194c5c18e7090b0656303627b0725545a6d9bc1070e23773 -size 9793704 diff --git a/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/_ops.py b/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/_ops.py deleted file mode 100644 index 9de8ab380209628599d241c68159af782399e651..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_e82640e_dirty -ops = torch.ops._deformable_detr_e82640e_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_e82640e_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/layers.py b/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu128-x86_64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__init__.py b/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__pycache__/__init__.cpython-313.pyc b/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index 407fbdba08c1b094bc33f155d295878890022c6f..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__pycache__/_ops.cpython-313.pyc b/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__pycache__/_ops.cpython-313.pyc deleted file mode 100644 index dea1f36299952301b445e5092f16e9ec8c3d2630..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__pycache__/_ops.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__pycache__/layers.cpython-313.pyc b/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__pycache__/layers.cpython-313.pyc deleted file mode 100644 index 6cf1dd945d846d90e97fe0bf00f731e780d88011..0000000000000000000000000000000000000000 Binary files a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/__pycache__/layers.cpython-313.pyc and /dev/null differ diff --git a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/_deformable_detr_a92c8ea_dirty.abi3.so b/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/_deformable_detr_a92c8ea_dirty.abi3.so deleted file mode 100755 index 8648d72185c14428facc6db0fdcffc068f6a7f53..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/_deformable_detr_a92c8ea_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:bc660849ac99525cfc072e758b85d07ed8b3040e4e529efeb42dcd0d286690bd -size 9916656 diff --git a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/_ops.py b/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/_ops.py deleted file mode 100644 index 019cf7ce30dc11d9b791075404417b1ac47500e7..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_a92c8ea_dirty -ops = torch.ops._deformable_detr_a92c8ea_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_a92c8ea_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/layers.py b/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-aarch64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/__init__.py b/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/__init__.py deleted file mode 100644 index 33db73ca6e361af4707ba5bb5f55bf0e7c3005a4..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/__init__.py +++ /dev/null @@ -1,46 +0,0 @@ -from typing import List -import torch - -from ._ops import ops -from . import layers - - -def ms_deform_attn_backward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - grad_output: torch.Tensor, - im2col_step: int, -) -> List[torch.Tensor]: - return ops.ms_deform_attn_backward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - grad_output, - im2col_step, - ) - - -def ms_deform_attn_forward( - value: torch.Tensor, - spatial_shapes: torch.Tensor, - level_start_index: torch.Tensor, - sampling_loc: torch.Tensor, - attn_weight: torch.Tensor, - im2col_step: int, -) -> torch.Tensor: - return ops.ms_deform_attn_forward( - value, - spatial_shapes, - level_start_index, - sampling_loc, - attn_weight, - im2col_step, - ) - - -__all__ = ["layers", "ms_deform_attn_forward", "ms_deform_attn_backward"] diff --git a/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/_deformable_detr_e82640e_dirty.abi3.so b/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/_deformable_detr_e82640e_dirty.abi3.so deleted file mode 100755 index 47740ce20ca9edfe98552118acb7f99a81c2dd56..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/_deformable_detr_e82640e_dirty.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:90e39e7e29a04b5c6696b2c5dc12af375b217223366b0c7b627c1e9150f5fa9c -size 9811104 diff --git a/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/_ops.py b/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/_ops.py deleted file mode 100644 index 9de8ab380209628599d241c68159af782399e651..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/_ops.py +++ /dev/null @@ -1,9 +0,0 @@ -import torch -from . import _deformable_detr_e82640e_dirty -ops = torch.ops._deformable_detr_e82640e_dirty - -def add_op_namespace_prefix(op_name: str): - """ - Prefix op by namespace. - """ - return f"_deformable_detr_e82640e_dirty::{op_name}" \ No newline at end of file diff --git a/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/layers.py b/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/layers.py deleted file mode 100644 index db94032dea3d445f27017f923ae80468e18d2d77..0000000000000000000000000000000000000000 --- a/build/torch28-cxx11-cu129-x86_64-linux/deformable_detr/layers.py +++ /dev/null @@ -1,84 +0,0 @@ -from typing import List, Union, Tuple - -from torch import Tensor -from torch.autograd import Function -from torch.autograd.function import once_differentiable -import torch.nn as nn - -from ._ops import ops - - -class MultiScaleDeformableAttentionFunction(Function): - @staticmethod - def forward( - context, - value: Tensor, - value_spatial_shapes: Tensor, - value_level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - context.im2col_step = im2col_step - output = ops.ms_deform_attn_forward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - context.im2col_step, - ) - context.save_for_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) - return output - - @staticmethod - @once_differentiable - def backward(context, grad_output): - ( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - ) = context.saved_tensors - grad_value, grad_sampling_loc, grad_attn_weight = ops.ms_deform_attn_backward( - value, - value_spatial_shapes, - value_level_start_index, - sampling_locations, - attention_weights, - grad_output, - context.im2col_step, - ) - - return grad_value, None, None, grad_sampling_loc, grad_attn_weight, None - - -class MultiScaleDeformableAttention(nn.Module): - def forward( - self, - value: Tensor, - value_spatial_shapes: Tensor, - value_spatial_shapes_list: List[Tuple], - level_start_index: Tensor, - sampling_locations: Tensor, - attention_weights: Tensor, - im2col_step: int, - ): - return MultiScaleDeformableAttentionFunction.apply( - value, - value_spatial_shapes, - level_start_index, - sampling_locations, - attention_weights, - im2col_step, - ) - - -__all__ = ["MultiScaleDeformableAttention"] diff --git a/flake.lock b/flake.lock deleted file mode 100644 index 26319b67eda604001ec12fc6bdd52ee9bfe65846..0000000000000000000000000000000000000000 --- a/flake.lock +++ /dev/null @@ -1,169 +0,0 @@ -{ - "nodes": { - "flake-compat": { - "locked": { - "lastModified": 1747046372, - "narHash": "sha256-CIVLLkVgvHYbgI2UpXvIIBJ12HWgX+fjA8Xf8PUmqCY=", - "owner": "edolstra", - "repo": "flake-compat", - "rev": "9100a0f413b0c601e0533d1d94ffd501ce2e7885", - "type": "github" - }, - "original": { - "owner": "edolstra", - "repo": "flake-compat", - "type": "github" - } - }, - "flake-compat_2": { - "locked": { - "lastModified": 1733328505, - "narHash": "sha256-NeCCThCEP3eCl2l/+27kNNK7QrwZB1IJCrXfrbv5oqU=", - "owner": "edolstra", - "repo": "flake-compat", - "rev": "ff81ac966bb2cae68946d5ed5fc4994f96d0ffec", - "type": "github" - }, - "original": { - "owner": "edolstra", - "repo": "flake-compat", - "type": "github" - } - }, - "flake-utils": { - "inputs": { - "systems": "systems" - }, - "locked": { - "lastModified": 1731533236, - "narHash": "sha256-l0KFg5HjrsfsO/JpG+r7fRrqm12kzFHyUHqHCVpMMbI=", - "owner": "numtide", - "repo": "flake-utils", - "rev": "11707dc2f618dd54ca8739b309ec4fc024de578b", - "type": "github" - }, - "original": { - "owner": "numtide", - "repo": "flake-utils", - "type": "github" - } - }, - "flake-utils_2": { - "inputs": { - "systems": "systems_2" - }, - "locked": { - "lastModified": 1731533236, - "narHash": "sha256-l0KFg5HjrsfsO/JpG+r7fRrqm12kzFHyUHqHCVpMMbI=", - "owner": "numtide", - "repo": "flake-utils", - "rev": "11707dc2f618dd54ca8739b309ec4fc024de578b", - "type": "github" - }, - "original": { - "owner": "numtide", - "repo": "flake-utils", - "type": "github" - } - }, - "hf-nix": { - "inputs": { - "flake-compat": "flake-compat_2", - "flake-utils": "flake-utils_2", - "nixpkgs": "nixpkgs" - }, - "locked": { - "lastModified": 1753354560, - "narHash": "sha256-vmOfRmr0Qm/IbZTWB2sBn+UFrABSTTA/cTg+m27Yt/E=", - "owner": "huggingface", - "repo": "hf-nix", - "rev": "7f2aceda2a2e72cd573bdb25e5c0667fd75f89d3", - "type": "github" - }, - "original": { - "owner": "huggingface", - "repo": "hf-nix", - "type": "github" - } - }, - "kernel-builder": { - "inputs": { - "flake-compat": "flake-compat", - "flake-utils": "flake-utils", - "hf-nix": "hf-nix", - "nixpkgs": [ - "kernel-builder", - "hf-nix", - "nixpkgs" - ] - }, - "locked": { - "lastModified": 1753354632, - "narHash": "sha256-31SX3Raiyx0qCuY9JSlx9ZZgxljeUxvW+JdujjxbofQ=", - "owner": "huggingface", - "repo": "kernel-builder", - "rev": "524b628fd8e58525dbd28455bffb0628092c5265", - "type": "github" - }, - "original": { - "owner": "huggingface", - "ref": "torch-2.8", - "repo": "kernel-builder", - "type": "github" - } - }, - "nixpkgs": { - "locked": { - "lastModified": 1752785354, - "narHash": "sha256-Y33ryUz7MPqKrZwlbQcsYCUz2jAJCacRf8jbs0tYUlA=", - "owner": "nixos", - "repo": "nixpkgs", - "rev": "d38025438a6ee456758dc03188ca6873a415463b", - "type": "github" - }, - "original": { - "owner": "nixos", - "repo": "nixpkgs", - "rev": "d38025438a6ee456758dc03188ca6873a415463b", - "type": "github" - } - }, - "root": { - "inputs": { - "kernel-builder": "kernel-builder" - } - }, - "systems": { - "locked": { - "lastModified": 1681028828, - "narHash": "sha256-Vy1rq5AaRuLzOxct8nz4T6wlgyUR7zLU309k9mBC768=", - "owner": "nix-systems", - "repo": "default", - "rev": "da67096a3b9bf56a91d16901293e51ba5b49a27e", - "type": "github" - }, - "original": { - "owner": "nix-systems", - "repo": "default", - "type": "github" - } - }, - "systems_2": { - "locked": { - "lastModified": 1681028828, - "narHash": "sha256-Vy1rq5AaRuLzOxct8nz4T6wlgyUR7zLU309k9mBC768=", - "owner": "nix-systems", - "repo": "default", - "rev": "da67096a3b9bf56a91d16901293e51ba5b49a27e", - "type": "github" - }, - "original": { - "owner": "nix-systems", - "repo": "default", - "type": "github" - } - } - }, - "root": "root", - "version": 7 -} diff --git a/flake.nix b/flake.nix index be5c5cec2c87d7cc948aa5160ce53b14b9fb0823..6fe55afb5a309a71720974748f2796b5c26a2f05 100644 --- a/flake.nix +++ b/flake.nix @@ -2,7 +2,7 @@ description = "Flake for deformable_detr kernels"; inputs = { - kernel-builder.url = "github:huggingface/kernel-builder/torch-2.8"; + kernel-builder.url = "git+ssh://git@github.com/huggingface/kernel-builder"; }; outputs = @@ -10,8 +10,5 @@ self, kernel-builder, }: - kernel-builder.lib.genFlakeOutputs { - path = ./.; - rev = self.shortRev or self.dirtyShortRev or self.lastModifiedDate; - }; + kernel-builder.lib.genFlakeOutputs ./.; }