diff --git a/FateZero-main/data/attribute/bus_gpu/00001.png b/FateZero-main/data/attribute/bus_gpu/00001.png new file mode 100644 index 0000000000000000000000000000000000000000..f9acf901dd4e33f8c5e56cc081a3e8046c3047ea --- /dev/null +++ b/FateZero-main/data/attribute/bus_gpu/00001.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:524baafde24d513dd5705a3d89b4af2fe422dfca50ebbca3e09b566a0b3e2e7c +size 392503 diff --git a/FateZero-main/data/attribute/bus_gpu/00003.png b/FateZero-main/data/attribute/bus_gpu/00003.png new file mode 100644 index 0000000000000000000000000000000000000000..fc7d192676e419a7fb5b3e5477f268e46041cbf6 --- /dev/null +++ b/FateZero-main/data/attribute/bus_gpu/00003.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3c4d79136cc6bc1946bc522913614bcdcdbc96931b7e5a1cfab936b592b968e +size 440345 diff --git a/FateZero-main/data/attribute/bus_gpu/00005.png b/FateZero-main/data/attribute/bus_gpu/00005.png new file mode 100644 index 0000000000000000000000000000000000000000..f8227c55a686438ae93874d8372203608ae68e38 --- /dev/null +++ b/FateZero-main/data/attribute/bus_gpu/00005.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e29ce6aaea50eb621a93ee48bf0ce825c2bce7b82ceb15f9614078fa1debb15b +size 421307 diff --git a/FateZero-main/data/attribute/rabbit_strawberry/00000.png b/FateZero-main/data/attribute/rabbit_strawberry/00000.png new file mode 100644 index 0000000000000000000000000000000000000000..599fd2e1437957444895dc5fac6805e4eb089c59 --- /dev/null +++ b/FateZero-main/data/attribute/rabbit_strawberry/00000.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19266fcbda4e8e5e8807e40b1b56bb97b34f0c8a7186ca2b044789dc550adfbf +size 331622 diff --git a/FateZero-main/data/attribute/rabbit_strawberry/00001.png b/FateZero-main/data/attribute/rabbit_strawberry/00001.png new file mode 100644 index 0000000000000000000000000000000000000000..0f393a011513ce414420f248dd89e48bd42a5144 --- /dev/null +++ b/FateZero-main/data/attribute/rabbit_strawberry/00001.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c02f75063ba744983834e73fc7d40d5e9cd99d41ead280afbdcc35c181cdb6d7 +size 337483 diff --git a/FateZero-main/data/attribute/rabbit_strawberry/00002.png b/FateZero-main/data/attribute/rabbit_strawberry/00002.png new file mode 100644 index 0000000000000000000000000000000000000000..84d8244bd3476b055dfb64326193ba9a5ccc7ab8 --- /dev/null +++ b/FateZero-main/data/attribute/rabbit_strawberry/00002.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26c5749b8b8276dcef03f44066f6e63382f150a77fec7454c787f03ac89d7cca +size 339248 diff --git a/FateZero-main/data/attribute/rabbit_strawberry/00003.png b/FateZero-main/data/attribute/rabbit_strawberry/00003.png new file mode 100644 index 0000000000000000000000000000000000000000..be18c35c07db3d666a2e22be42bd21e6231d7fad --- /dev/null +++ b/FateZero-main/data/attribute/rabbit_strawberry/00003.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fba39d5fbfb63f66c761398704e8ea17340ea27629a4205ceb15b93a100e7258 +size 338491 diff --git a/FateZero-main/data/attribute/rabbit_strawberry/00004.png b/FateZero-main/data/attribute/rabbit_strawberry/00004.png new file mode 100644 index 0000000000000000000000000000000000000000..1feca13a269199036270816af9181bb508408079 --- /dev/null +++ b/FateZero-main/data/attribute/rabbit_strawberry/00004.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4256217865ebfa8645422e4d16452d6caec77ea7c135dbe4e0fd9ef5b8a8eb42 +size 344289 diff --git a/FateZero-main/data/attribute/rabbit_strawberry/00006.png b/FateZero-main/data/attribute/rabbit_strawberry/00006.png new file mode 100644 index 0000000000000000000000000000000000000000..45638121b4b569540437daf17797c2b483ff2a4d --- /dev/null +++ b/FateZero-main/data/attribute/rabbit_strawberry/00006.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f0916a3356732060fe59421d2be148ba977fec42d29b693b341abb21242f889 +size 351183 diff --git a/FateZero-main/data/attribute/rabbit_strawberry/00007.png b/FateZero-main/data/attribute/rabbit_strawberry/00007.png new file mode 100644 index 0000000000000000000000000000000000000000..14e4198979ed1b36074c3e54ac7e60807b122bfc --- /dev/null +++ b/FateZero-main/data/attribute/rabbit_strawberry/00007.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d05e1ae1625d0f5043587b7c9c0006e6228d6e989b5363fececadc92a216caf +size 348576 diff --git a/FateZero-main/data/shape/swan_swarov/00000.png b/FateZero-main/data/shape/swan_swarov/00000.png new file mode 100644 index 0000000000000000000000000000000000000000..477b1622704a978375f2253bec1147d58ef6801d --- /dev/null +++ b/FateZero-main/data/shape/swan_swarov/00000.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f17f6634276283a259a89aad1ce81b75e2bf4e867afe010b696019d229c56bd +size 426513 diff --git a/FateZero-main/data/shape/swan_swarov/00001.png b/FateZero-main/data/shape/swan_swarov/00001.png new file mode 100644 index 0000000000000000000000000000000000000000..be7f653006a24432a28ce0c3b3b98c1557a6f455 --- /dev/null +++ b/FateZero-main/data/shape/swan_swarov/00001.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5084b5a7c053b278ff726e0317c9a6b03e4442f3760b03b05253a30b2bebec08 +size 433256 diff --git a/FateZero-main/data/shape/swan_swarov/00003.png b/FateZero-main/data/shape/swan_swarov/00003.png new file mode 100644 index 0000000000000000000000000000000000000000..378d656179bf7172fa2702f10624d29a0fd2488f --- /dev/null +++ b/FateZero-main/data/shape/swan_swarov/00003.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c3e972e26e35713865891318fdd8cf1c58d82fecbe21ef5493343838a629cc2 +size 439649 diff --git a/FateZero-main/data/shape/swan_swarov/00005.png b/FateZero-main/data/shape/swan_swarov/00005.png new file mode 100644 index 0000000000000000000000000000000000000000..8574ef175a61cf3935af994325a9841b86008d24 --- /dev/null +++ b/FateZero-main/data/shape/swan_swarov/00005.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e48a1ec062e69faf7069e36d164a6e0372dab7c43f9b2a20c731f8db2b35146 +size 436798 diff --git a/FateZero-main/data/shape/teaser_car-turn/00000.png b/FateZero-main/data/shape/teaser_car-turn/00000.png new file mode 100644 index 0000000000000000000000000000000000000000..43bdbca268fc1bcd4a4cd18d7107529c98f05369 --- /dev/null +++ b/FateZero-main/data/shape/teaser_car-turn/00000.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a44ff824b031bcd6c6547f95eb7ff8532bdfbe9efbfa34506fc0cb3ed1f9923 +size 420934 diff --git a/FateZero-main/data/shape/teaser_car-turn/00001.png b/FateZero-main/data/shape/teaser_car-turn/00001.png new file mode 100644 index 0000000000000000000000000000000000000000..8cd34d9c6367e2dad0b55fb951c3b61c67ee27b8 --- /dev/null +++ b/FateZero-main/data/shape/teaser_car-turn/00001.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bd08cc5c2e5eb2078148e3f7d78e19922b66ec984ee2a3689a3197084ef7fa3 +size 425987 diff --git a/FateZero-main/data/shape/teaser_car-turn/00002.png b/FateZero-main/data/shape/teaser_car-turn/00002.png new file mode 100644 index 0000000000000000000000000000000000000000..1dc13a001485c6b6ba113ee479c59a771f957ba1 --- /dev/null +++ b/FateZero-main/data/shape/teaser_car-turn/00002.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e241abc23a5bbc63afd5d8607a7b4efd566635921562a7b1be157480fd09292 +size 432155 diff --git a/FateZero-main/data/shape/teaser_car-turn/00003.png b/FateZero-main/data/shape/teaser_car-turn/00003.png new file mode 100644 index 0000000000000000000000000000000000000000..d552439ac2bbbdb77b7750ce20889cf89a6e7b2f --- /dev/null +++ b/FateZero-main/data/shape/teaser_car-turn/00003.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c124f133164223458b136c538b67811282c5d862fb34f470ed55d9788436886d +size 425496 diff --git a/FateZero-main/data/shape/teaser_car-turn/00004.png b/FateZero-main/data/shape/teaser_car-turn/00004.png new file mode 100644 index 0000000000000000000000000000000000000000..17544235c28027ca6ca6deafe745f9b123a0175f --- /dev/null +++ b/FateZero-main/data/shape/teaser_car-turn/00004.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c57427aa5a33d0552be2dba4e34dcbbc324bccad7a28f21af2185776e5aac54 +size 414891 diff --git a/FateZero-main/data/shape/teaser_car-turn/00005.png b/FateZero-main/data/shape/teaser_car-turn/00005.png new file mode 100644 index 0000000000000000000000000000000000000000..73da096db80de920020e95a3aa2d33639501bf37 --- /dev/null +++ b/FateZero-main/data/shape/teaser_car-turn/00005.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1edef4dfee18a654d9828c24a69ddb8a73fe6c7c7c8589530fc8f0e240219a5d +size 431210 diff --git a/FateZero-main/data/shape/teaser_car-turn/00006.png b/FateZero-main/data/shape/teaser_car-turn/00006.png new file mode 100644 index 0000000000000000000000000000000000000000..a0bb1e5b108ae151ca367d0701ac9d8ee035d4ed --- /dev/null +++ b/FateZero-main/data/shape/teaser_car-turn/00006.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32a9e63c65c8797e6d8d95b8a278089ed48fc7510b2e1a444ce915e6bc7a992c +size 428020 diff --git a/FateZero-main/data/shape/teaser_car-turn/00007.png b/FateZero-main/data/shape/teaser_car-turn/00007.png new file mode 100644 index 0000000000000000000000000000000000000000..cd78a27963a442c15cf8f87926db6f46f60e7a86 --- /dev/null +++ b/FateZero-main/data/shape/teaser_car-turn/00007.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9dbe65f7358f2bbfae87958a9437f9715e31a738aef56d469508587f77533564 +size 415206 diff --git a/FateZero-main/data/style/blackswan.mp4 b/FateZero-main/data/style/blackswan.mp4 new file mode 100644 index 0000000000000000000000000000000000000000..d1e15b7e8e52a85aa46649b9b64a38fe9427b177 --- /dev/null +++ b/FateZero-main/data/style/blackswan.mp4 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7edf82f0c047b3c83af0d5c4adf7b5a799d04cd4b41225d0b7430f77e76ab097 +size 1010990 diff --git a/FateZero-main/data/style/blackswan/00000.jpg b/FateZero-main/data/style/blackswan/00000.jpg new file mode 100644 index 0000000000000000000000000000000000000000..d74432283dfb132046eda0c7232b440b99cf43d9 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00000.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f134391f3c757dd305e79f1fb3a3e6b0ec441013394f2e620d3cfe6b887a02c4 +size 129502 diff --git a/FateZero-main/data/style/blackswan/00001.jpg b/FateZero-main/data/style/blackswan/00001.jpg new file mode 100644 index 0000000000000000000000000000000000000000..44a561c8cd8b8c3ad01e9dcb8897494160332fb3 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00001.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cd884dd9a5bc083711fdc8deefbfa7e8f36cfb6832fbd70e41cbfadfb219c0d +size 128110 diff --git a/FateZero-main/data/style/blackswan/00002.jpg b/FateZero-main/data/style/blackswan/00002.jpg new file mode 100644 index 0000000000000000000000000000000000000000..7dc9fba2dd3b24a53908d3471d2a96a6ca7e711a --- /dev/null +++ b/FateZero-main/data/style/blackswan/00002.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:725a8a4175042dbc7d5423cf5fe67a001531548a44ccea5832f191c292ec7eff +size 127872 diff --git a/FateZero-main/data/style/blackswan/00003.jpg b/FateZero-main/data/style/blackswan/00003.jpg new file mode 100644 index 0000000000000000000000000000000000000000..6ba86eeec16a8bb144b030848aea97be236ecf7e --- /dev/null +++ b/FateZero-main/data/style/blackswan/00003.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0977844f5af8b00b60baa55e2ac814b517105f3870b39727fb0125fe2ba98dd9 +size 127545 diff --git a/FateZero-main/data/style/blackswan/00004.jpg b/FateZero-main/data/style/blackswan/00004.jpg new file mode 100644 index 0000000000000000000000000000000000000000..86028aa7e68ceefedc0ca767fa4bdda6f968654d --- /dev/null +++ b/FateZero-main/data/style/blackswan/00004.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebf44b1f86ef8208b2cbb6657bd3ccfab3d90f4403697b292c46aa5385c01076 +size 128063 diff --git a/FateZero-main/data/style/blackswan/00005.jpg b/FateZero-main/data/style/blackswan/00005.jpg new file mode 100644 index 0000000000000000000000000000000000000000..68d714be9df01887c7b09f7ce77a7409041c5b8e --- /dev/null +++ b/FateZero-main/data/style/blackswan/00005.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c444f058072ef8a698e46a7bda7968d48226325d5ad0157ca74551edb33fca0 +size 129303 diff --git a/FateZero-main/data/style/blackswan/00006.jpg b/FateZero-main/data/style/blackswan/00006.jpg new file mode 100644 index 0000000000000000000000000000000000000000..5a424d3012cba690d3c044753690a7d163a4af07 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00006.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5d6cdee16c894cdc0202bb783538efe3dbfe0ced1d5ad1d215ddf571abdd7ee +size 131153 diff --git a/FateZero-main/data/style/blackswan/00007.jpg b/FateZero-main/data/style/blackswan/00007.jpg new file mode 100644 index 0000000000000000000000000000000000000000..d98f17a75e14b2818ebf09e6324b523dbdf96a68 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00007.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:674fc72e09dd205ed295a19616c08f4c73f67aeaabd8b34f95114e014a4b99c1 +size 129290 diff --git a/FateZero-main/data/style/blackswan/00008.jpg b/FateZero-main/data/style/blackswan/00008.jpg new file mode 100644 index 0000000000000000000000000000000000000000..b651a631f423582d3a1a155aa9c9142d326241b1 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00008.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:316dc7179c8dc8c800c3945feee50191c1c440e6ffbf62029dab1ebeb09c7be6 +size 129509 diff --git a/FateZero-main/data/style/blackswan/00009.jpg b/FateZero-main/data/style/blackswan/00009.jpg new file mode 100644 index 0000000000000000000000000000000000000000..c54ae3b836694a0d8bc3b03865e0e76ef4356e60 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00009.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bef36ec6d7e8205c3e5c2ce30eb31231eeee0c653be8a27d73c195b320912219 +size 130575 diff --git a/FateZero-main/data/style/blackswan/00010.jpg b/FateZero-main/data/style/blackswan/00010.jpg new file mode 100644 index 0000000000000000000000000000000000000000..5e3583529bab93b6f86d334488a7bfff55fdb982 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00010.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9191bf19b1c803caf755387852cc3883f32495dab8f232a871019de2d9dfaa3 +size 128755 diff --git a/FateZero-main/data/style/blackswan/00011.jpg b/FateZero-main/data/style/blackswan/00011.jpg new file mode 100644 index 0000000000000000000000000000000000000000..f728a364aeb0a2aff55c6c9429d6f8cda39a6fae --- /dev/null +++ b/FateZero-main/data/style/blackswan/00011.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82bb0d8faeabb37e945ee0c4983215a55d1715a789b1b64b8fe34c4c72240d96 +size 128140 diff --git a/FateZero-main/data/style/blackswan/00012.jpg b/FateZero-main/data/style/blackswan/00012.jpg new file mode 100644 index 0000000000000000000000000000000000000000..2ab8c44438d173576012327685d647f73ee69313 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00012.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5519e6ae768671f9b9b99df8db0b6b989c1faf8a15170a4af7c0feb528c2b206 +size 129129 diff --git a/FateZero-main/data/style/blackswan/00013.jpg b/FateZero-main/data/style/blackswan/00013.jpg new file mode 100644 index 0000000000000000000000000000000000000000..9f5d2f2ddb166dfc45c988beba81bb9081f831c1 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00013.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c53d1452643de431652ba46c2b9bafe0e474fa7307875d9ff50f0f4bad4896f +size 122198 diff --git a/FateZero-main/data/style/blackswan/00014.jpg b/FateZero-main/data/style/blackswan/00014.jpg new file mode 100644 index 0000000000000000000000000000000000000000..8d91d2449c3285867a084d5e3fc77f404d8190c1 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00014.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4eda9417c2bd16a5519a03757687421682aceec93c7d0b1c4321856154075cb1 +size 122897 diff --git a/FateZero-main/data/style/blackswan/00015.jpg b/FateZero-main/data/style/blackswan/00015.jpg new file mode 100644 index 0000000000000000000000000000000000000000..79764f88f6cb0317cd3bf813da55cd631ec4f715 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00015.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f47904d5d43c1ccaa40e475f957a21bb369e0c53f265d8b2c81040b9ebcfba01 +size 124116 diff --git a/FateZero-main/data/style/blackswan/00016.jpg b/FateZero-main/data/style/blackswan/00016.jpg new file mode 100644 index 0000000000000000000000000000000000000000..99509d3791ee0f18d06540851794fe064079bdb0 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00016.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e8f22604d3fceb251f07ad5a7c43249eee3daeba4ecb1429fd9a2a07a2637ce +size 122740 diff --git a/FateZero-main/data/style/blackswan/00017.jpg b/FateZero-main/data/style/blackswan/00017.jpg new file mode 100644 index 0000000000000000000000000000000000000000..b7d12a9fac835c27caaf04ea7b47a6ef552e6b68 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00017.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3df441bcd420f3e3962289ed56d17d1e2eaf26f38652aa8ac9343463a068d1b2 +size 124000 diff --git a/FateZero-main/data/style/blackswan/00018.jpg b/FateZero-main/data/style/blackswan/00018.jpg new file mode 100644 index 0000000000000000000000000000000000000000..fcd003c0666bcbf2912510ac4e03d12e8357e243 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00018.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0c4a418423e00f222c81c1b89a3748d589c729c6838dc5d906acad057e2e595 +size 127320 diff --git a/FateZero-main/data/style/blackswan/00019.jpg b/FateZero-main/data/style/blackswan/00019.jpg new file mode 100644 index 0000000000000000000000000000000000000000..77d330a5db40558cd25e1072bf9e5334b6a10a5a --- /dev/null +++ b/FateZero-main/data/style/blackswan/00019.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f596d9a41662af01511b76818c89d5b8a9f10c33e65a8431df5eaf2937286dda +size 126609 diff --git a/FateZero-main/data/style/blackswan/00020.jpg b/FateZero-main/data/style/blackswan/00020.jpg new file mode 100644 index 0000000000000000000000000000000000000000..a6d4d2712907bf8abe03ad88ebb3b0499ac11aff --- /dev/null +++ b/FateZero-main/data/style/blackswan/00020.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58227d78be7958d8f41bbddf83cbe05efcc6287a7a07201916544c987bd49176 +size 126680 diff --git a/FateZero-main/data/style/blackswan/00021.jpg b/FateZero-main/data/style/blackswan/00021.jpg new file mode 100644 index 0000000000000000000000000000000000000000..ad94f5d0f028fe9a2dfee41cb358b445a0b7c230 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00021.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbc7914cfdf1d69e9d301a02c535ce3708ed027d5b152bebb428cdba9a9db88b +size 127989 diff --git a/FateZero-main/data/style/blackswan/00022.jpg b/FateZero-main/data/style/blackswan/00022.jpg new file mode 100644 index 0000000000000000000000000000000000000000..b99d3438f17c4891fe6eab9b317d16b6cccf4029 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00022.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c1b340446e23566c3eac711fd3e6b039ac45dbcaead4fedea4dfe3dfa6b66e6 +size 125181 diff --git a/FateZero-main/data/style/blackswan/00023.jpg b/FateZero-main/data/style/blackswan/00023.jpg new file mode 100644 index 0000000000000000000000000000000000000000..14569cfcbdff12b882be8244568429ced29bce3b --- /dev/null +++ b/FateZero-main/data/style/blackswan/00023.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc4c6dd389852985cdbca956ce36bd7d987ada18eb9bd2ed0551a875b852f7e8 +size 124167 diff --git a/FateZero-main/data/style/blackswan/00024.jpg b/FateZero-main/data/style/blackswan/00024.jpg new file mode 100644 index 0000000000000000000000000000000000000000..00bc58535496433c78f1cce036c5d2c1d4edf09a --- /dev/null +++ b/FateZero-main/data/style/blackswan/00024.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8bc6ee28025a1657039afcee74bb4f28ac2a1cfdae30bd206de09f09d68c865e +size 124410 diff --git a/FateZero-main/data/style/blackswan/00025.jpg b/FateZero-main/data/style/blackswan/00025.jpg new file mode 100644 index 0000000000000000000000000000000000000000..0a0c3c1a27ea5ca5d0a34082224a675c7a8f222f --- /dev/null +++ b/FateZero-main/data/style/blackswan/00025.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8b2cc99a0459a681f625542d2a2cd0f35bc00f886dcd167cb28445f13c13907 +size 122714 diff --git a/FateZero-main/data/style/blackswan/00026.jpg b/FateZero-main/data/style/blackswan/00026.jpg new file mode 100644 index 0000000000000000000000000000000000000000..dae112a2f49e1ed3a5c65736fa0a89113ab56179 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00026.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e36f55efac7c1bc3de0c92de00f76004c19d81f2b5ef4963db1741aa155f68fc +size 123548 diff --git a/FateZero-main/data/style/blackswan/00027.jpg b/FateZero-main/data/style/blackswan/00027.jpg new file mode 100644 index 0000000000000000000000000000000000000000..48595b867e0d636b21f263b84dad117b9d19b39b --- /dev/null +++ b/FateZero-main/data/style/blackswan/00027.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1033f60cdea0ff6ed196d09e6a9ebd40d0ca9986739695928397b05c61895eee +size 125724 diff --git a/FateZero-main/data/style/blackswan/00028.jpg b/FateZero-main/data/style/blackswan/00028.jpg new file mode 100644 index 0000000000000000000000000000000000000000..a41f85836eb0f4280fbf3d4d75770951e60afb05 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00028.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a19e5ef1e51be7b1328d03a54c50e77c3def925250e057df1f8d42b41bb7853 +size 124422 diff --git a/FateZero-main/data/style/blackswan/00029.jpg b/FateZero-main/data/style/blackswan/00029.jpg new file mode 100644 index 0000000000000000000000000000000000000000..53e957da09a267dec0514c95115604219dc36c23 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00029.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f1b1d6bb4547c78c2c4d372e150239e8f3f50488da8fe0a1392c52acec82e58 +size 125538 diff --git a/FateZero-main/data/style/blackswan/00030.jpg b/FateZero-main/data/style/blackswan/00030.jpg new file mode 100644 index 0000000000000000000000000000000000000000..664032ad5b3c953e18243f5d89894c940317de60 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00030.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f07e5726382cf1d44573212be93292650e89f3365a7c49267b3e3864b79a686 +size 127319 diff --git a/FateZero-main/data/style/blackswan/00031.jpg b/FateZero-main/data/style/blackswan/00031.jpg new file mode 100644 index 0000000000000000000000000000000000000000..ee15f9aef4c5d1c04369900553b90e0d27ef6fbb --- /dev/null +++ b/FateZero-main/data/style/blackswan/00031.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17cd1865395422d54b57cf52163a1ff33f3f6168a4233350627e48796871b14d +size 126217 diff --git a/FateZero-main/data/style/blackswan/00032.jpg b/FateZero-main/data/style/blackswan/00032.jpg new file mode 100644 index 0000000000000000000000000000000000000000..fb5ef5814e4fb4e022b0000ac6b9baf964ba6639 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00032.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a4c5ddaf6cecd18ede05c8d2c3b945b5843ac21124f7955c866c4ac017e37fd +size 126320 diff --git a/FateZero-main/data/style/blackswan/00033.jpg b/FateZero-main/data/style/blackswan/00033.jpg new file mode 100644 index 0000000000000000000000000000000000000000..d8459c8ca4c978f8320bf06d1bf002ce3e9706b8 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00033.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:856a351109d16fbb38eec6b6ef9dfd37cd7c42d7bc5c818b3953d69be3cda41d +size 127394 diff --git a/FateZero-main/data/style/blackswan/00034.jpg b/FateZero-main/data/style/blackswan/00034.jpg new file mode 100644 index 0000000000000000000000000000000000000000..9499fae14dda9e34fdaba27bd195d52d58f78b19 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00034.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:854db0109f5c56a497ebc843cd817cafeb83596a34c3e615a6698cc628f6ae45 +size 127221 diff --git a/FateZero-main/data/style/blackswan/00035.jpg b/FateZero-main/data/style/blackswan/00035.jpg new file mode 100644 index 0000000000000000000000000000000000000000..82b4a83e6ca70d34335ab87e4feff9e00c437b45 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00035.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01c802933151f6dced55e1a5638b8a9b852bd8b71ec8f0564145181ceb025c4b +size 127784 diff --git a/FateZero-main/data/style/blackswan/00036.jpg b/FateZero-main/data/style/blackswan/00036.jpg new file mode 100644 index 0000000000000000000000000000000000000000..e50ab12046ab5370122b20c65f3d2546024d22a5 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00036.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e415e05f5d9af62346d766a36ddefe53a138495803d0628a71aef0810a8df8ee +size 129678 diff --git a/FateZero-main/data/style/blackswan/00037.jpg b/FateZero-main/data/style/blackswan/00037.jpg new file mode 100644 index 0000000000000000000000000000000000000000..366df50e5f376effc9b5620c1b3a03d3afafde82 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00037.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e2c3e42a6e6a5673f3720fb788f16e91b606c5705e8a271dbc974d794e322a8 +size 128982 diff --git a/FateZero-main/data/style/blackswan/00038.jpg b/FateZero-main/data/style/blackswan/00038.jpg new file mode 100644 index 0000000000000000000000000000000000000000..54ffac270fa6477cdfc85e7a02fe6427f7024947 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00038.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afc902dc59ff933afc8b8344a800c23ddfa5139b9740b1aaa8e39d15f653c900 +size 130042 diff --git a/FateZero-main/data/style/blackswan/00039.jpg b/FateZero-main/data/style/blackswan/00039.jpg new file mode 100644 index 0000000000000000000000000000000000000000..f403c9d5944ebf4526a0256c60e375c494f7d928 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00039.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a06477c166a66d887161e13216403ad78f122851bc9e240d4d193d708abb4e60 +size 131695 diff --git a/FateZero-main/data/style/blackswan/00040.jpg b/FateZero-main/data/style/blackswan/00040.jpg new file mode 100644 index 0000000000000000000000000000000000000000..9f5b1cd32d2cc6d18eeb26d35711a655f6270ff5 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00040.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca3fb8647375b6aed42823d5aad00930ea4cff396c512b3987bb9201c8d4cc21 +size 129893 diff --git a/FateZero-main/data/style/blackswan/00041.jpg b/FateZero-main/data/style/blackswan/00041.jpg new file mode 100644 index 0000000000000000000000000000000000000000..d95d33690a0b40a4e85975fc1dc2aefca3e1bd9d --- /dev/null +++ b/FateZero-main/data/style/blackswan/00041.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa2ed808073717f5a35c19315ec826d179feab93abe0db2be8dde28f519fb974 +size 129910 diff --git a/FateZero-main/data/style/blackswan/00042.jpg b/FateZero-main/data/style/blackswan/00042.jpg new file mode 100644 index 0000000000000000000000000000000000000000..0bc4e76b50aec3238f709ec40698ef51c5e357f3 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00042.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32ea488909a8e7c3b305a046c15a60d40f471ddf8ed87fbcb004fdfbf34879bc +size 130682 diff --git a/FateZero-main/data/style/blackswan/00043.jpg b/FateZero-main/data/style/blackswan/00043.jpg new file mode 100644 index 0000000000000000000000000000000000000000..c49f34edf8ca0f8f467d7dba1ae2c44e4308daaf --- /dev/null +++ b/FateZero-main/data/style/blackswan/00043.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb80c5f998198326ea79b5e3e5137e794d55ac88a1f179c33a73756917601d27 +size 129024 diff --git a/FateZero-main/data/style/blackswan/00044.jpg b/FateZero-main/data/style/blackswan/00044.jpg new file mode 100644 index 0000000000000000000000000000000000000000..fe30ceb0ff5f99d5ed890323ec11e87211852878 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00044.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8a269e6fb2aa38569c19fda60f6df531bff06ad70e177b373dff11087a12d7f +size 129770 diff --git a/FateZero-main/data/style/blackswan/00045.jpg b/FateZero-main/data/style/blackswan/00045.jpg new file mode 100644 index 0000000000000000000000000000000000000000..191b29de5447d1638b9815b68ae02c8808ecc97f --- /dev/null +++ b/FateZero-main/data/style/blackswan/00045.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9d71274e94e1a587402bd5b0e676896f52db53c78920b1ecbca68736bdec021 +size 131099 diff --git a/FateZero-main/data/style/blackswan/00046.jpg b/FateZero-main/data/style/blackswan/00046.jpg new file mode 100644 index 0000000000000000000000000000000000000000..b55d16200328425a16c44c66eec9a05f60b8038f --- /dev/null +++ b/FateZero-main/data/style/blackswan/00046.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35dbd0249b01d37fb2590eee6e143ce5c8b96402b776903989fe462fe49a00e1 +size 128072 diff --git a/FateZero-main/data/style/blackswan/00047.jpg b/FateZero-main/data/style/blackswan/00047.jpg new file mode 100644 index 0000000000000000000000000000000000000000..4280d437ff5d46b4ba8f2d1101f32b50b7ba2fb6 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00047.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4d5580efaefa782290b1c59b8679c0aa800b536a1516a9899daef6da25235be +size 126866 diff --git a/FateZero-main/data/style/blackswan/00048.jpg b/FateZero-main/data/style/blackswan/00048.jpg new file mode 100644 index 0000000000000000000000000000000000000000..d089f62b85becb853e5487c6bc2f7aadafb855d0 --- /dev/null +++ b/FateZero-main/data/style/blackswan/00048.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe9929815c6470efb58751f1e478a3065c619b8985c054118467868860a02a0e +size 127388 diff --git a/FateZero-main/data/style/blackswan/00049.jpg b/FateZero-main/data/style/blackswan/00049.jpg new file mode 100644 index 0000000000000000000000000000000000000000..4ed1b7b511fe0669f9d007021ec9dbe990507adb --- /dev/null +++ b/FateZero-main/data/style/blackswan/00049.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b4242587d32afa96a41cea04fe9d9682d584fb1bc1f912b9898551a0a459829 +size 125243 diff --git a/FateZero-main/data/style/rabit.mp4 b/FateZero-main/data/style/rabit.mp4 new file mode 100644 index 0000000000000000000000000000000000000000..2ec84016d46584562ab4d1b0aa7e312790351fe9 --- /dev/null +++ b/FateZero-main/data/style/rabit.mp4 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c8e1da7ce60d04108624d59316e718e84f77c493723d55cce81880194c5beba +size 98675 diff --git a/FateZero-main/data/style/rabit/rabit-00.png b/FateZero-main/data/style/rabit/rabit-00.png new file mode 100644 index 0000000000000000000000000000000000000000..312b5dee45141a2e1e1fcc9fa743d3eeaabce10c --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-00.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdb0ebce819c5a67dc95ee02daf2000f9ebc346b195cd8319b726175ce96fc3d +size 70626 diff --git a/FateZero-main/data/style/rabit/rabit-01.png b/FateZero-main/data/style/rabit/rabit-01.png new file mode 100644 index 0000000000000000000000000000000000000000..7d56de63eabdba025859c2f9709850b829487268 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-01.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84c14e2a6f2b7597f5be0c8e638290d827d61490992f5c22bc887b2cd078afc8 +size 151940 diff --git a/FateZero-main/data/style/rabit/rabit-02.png b/FateZero-main/data/style/rabit/rabit-02.png new file mode 100644 index 0000000000000000000000000000000000000000..5ea7d9fee885c354e557e58515defe589d99d3c2 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-02.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:768b419d189ce544c83bf8f90b16f4adbdc42ab4eee29961f2538d151d88728e +size 154714 diff --git a/FateZero-main/data/style/rabit/rabit-03.png b/FateZero-main/data/style/rabit/rabit-03.png new file mode 100644 index 0000000000000000000000000000000000000000..973cc39cc0fa22db6dfe054454f80356fd0f6ae3 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-03.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d23334ee00e6a28e89b636ffffc05b659b3438bab67b80285305fae8f8648d6 +size 73023 diff --git a/FateZero-main/data/style/rabit/rabit-04.png b/FateZero-main/data/style/rabit/rabit-04.png new file mode 100644 index 0000000000000000000000000000000000000000..5e971c7600cffee5a5a7781d66455b18f3d8b95e --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-04.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0afcfb02189069ea6b200ac3f9d8497fd304865b86a01530e62f2ed4fa2f3465 +size 73028 diff --git a/FateZero-main/data/style/rabit/rabit-05.png b/FateZero-main/data/style/rabit/rabit-05.png new file mode 100644 index 0000000000000000000000000000000000000000..513202d98369c94eb6c37b48802d19bb67d39bc5 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-05.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf857e13cb80aa24e48c6c2e702b1f67ef1fee51675b473f200f21dc0bf96ab2 +size 72704 diff --git a/FateZero-main/data/style/rabit/rabit-06.png b/FateZero-main/data/style/rabit/rabit-06.png new file mode 100644 index 0000000000000000000000000000000000000000..0dbf031a82679553a8794656211aba2e900a681c --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-06.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26e82d7a145a89e1c326d65eb2c86057a1c2d3c5bdbe3c82765b65722eff57dc +size 158727 diff --git a/FateZero-main/data/style/rabit/rabit-07.png b/FateZero-main/data/style/rabit/rabit-07.png new file mode 100644 index 0000000000000000000000000000000000000000..5c3a6c54a21bd46ffae5dbdff0fb7089b3afe53d --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-07.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63186599c2bb4a8b6bff375244d70aef29b1fe10c51b5ce33bcd13c6cd86c8c8 +size 74603 diff --git a/FateZero-main/data/style/rabit/rabit-08.png b/FateZero-main/data/style/rabit/rabit-08.png new file mode 100644 index 0000000000000000000000000000000000000000..b4736b996ac495ee6e551613ce7d7e01106f8f34 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-08.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3888adac1507e71e773e763f39f8ba92dfd641624bdddec7d347877bc2c4e2db +size 74144 diff --git a/FateZero-main/data/style/rabit/rabit-09.png b/FateZero-main/data/style/rabit/rabit-09.png new file mode 100644 index 0000000000000000000000000000000000000000..f0beaaecd73b0f37d3a588e8be5747e857925a7d --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-09.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2217158be58c93359c6ef80493e9c6d7a1fe1448f519c8a3ef63bfd91033089c +size 75171 diff --git a/FateZero-main/data/style/rabit/rabit-10.png b/FateZero-main/data/style/rabit/rabit-10.png new file mode 100644 index 0000000000000000000000000000000000000000..7929273259b4a1c8f6f82e2f119cb214af3183de --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-10.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ddcc86a05edc568a942e3f3cb2f8e2764bda61c65771e3cfd32c5611b6482cd +size 74481 diff --git a/FateZero-main/data/style/rabit/rabit-11.png b/FateZero-main/data/style/rabit/rabit-11.png new file mode 100644 index 0000000000000000000000000000000000000000..580cc30198ca9f5d49864b4dd273d52668bc601e --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-11.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c905901e8e00c0b42e0dab509b8d69d1064f2cad593bacc44c11d107c4f3ca7 +size 74916 diff --git a/FateZero-main/data/style/rabit/rabit-12.png b/FateZero-main/data/style/rabit/rabit-12.png new file mode 100644 index 0000000000000000000000000000000000000000..4e2a51d58af30b87613cc19f4883c42cab8a52c1 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-12.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c96ff13647275ec30f24906009dd50e8047f4a8f47f876d709df4a6b69acd22 +size 161085 diff --git a/FateZero-main/data/style/rabit/rabit-13.png b/FateZero-main/data/style/rabit/rabit-13.png new file mode 100644 index 0000000000000000000000000000000000000000..e46b5b0455b07f8c459c0e62b5c7ea9531cdd45f --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-13.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:889b4a2914e5c63be4f528977ec2287cdae08b2bdd1d0a7939e932c410b1a5b1 +size 157069 diff --git a/FateZero-main/data/style/rabit/rabit-14.png b/FateZero-main/data/style/rabit/rabit-14.png new file mode 100644 index 0000000000000000000000000000000000000000..e5a6aa8bcaec567ef9b886528c127bdf4afb6ecc --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-14.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:479a2858b839af18ee1cf61687fc688835eb470a98333a1c8bfe6c0d5fdea051 +size 158813 diff --git a/FateZero-main/data/style/rabit/rabit-15.png b/FateZero-main/data/style/rabit/rabit-15.png new file mode 100644 index 0000000000000000000000000000000000000000..161c1d7c9511c723e32b4676e533ce47145e25a3 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-15.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04277bc236a223f612ed6a3bfdcfd946dedb07056eb1cde11bf3967a4026e7f8 +size 158099 diff --git a/FateZero-main/data/style/rabit/rabit-16.png b/FateZero-main/data/style/rabit/rabit-16.png new file mode 100644 index 0000000000000000000000000000000000000000..b5a07751251bc9211ebe131a533f9a6decd18531 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-16.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:523fa5c5da1b6b1a12df8532f6c182a3500caaa30b282343ace0f5e67f9103d5 +size 158290 diff --git a/FateZero-main/data/style/rabit/rabit-17.png b/FateZero-main/data/style/rabit/rabit-17.png new file mode 100644 index 0000000000000000000000000000000000000000..9d6158cd60669795c62cbd810a46e335bdc1ac1d --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-17.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5615c006df57205f8a6e82ccd44789ddd0cd50e738aa94a144efa60f2aab171d +size 158377 diff --git a/FateZero-main/data/style/rabit/rabit-18.png b/FateZero-main/data/style/rabit/rabit-18.png new file mode 100644 index 0000000000000000000000000000000000000000..8a60692b3900d3c64ebea402a4ac1f0c1e610a7b --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-18.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b901f97f1abd4d209a66c814eafeb1bdf511bbb24b87a0375689c8a2847294a +size 156353 diff --git a/FateZero-main/data/style/rabit/rabit-19.png b/FateZero-main/data/style/rabit/rabit-19.png new file mode 100644 index 0000000000000000000000000000000000000000..2620b8a3ba2641b4be4d9b584730596547f95da1 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-19.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bd3a758f704975a93a0c583b6b727ae0280170006ee004b424f03f182532c66 +size 153193 diff --git a/FateZero-main/data/style/rabit/rabit-20.png b/FateZero-main/data/style/rabit/rabit-20.png new file mode 100644 index 0000000000000000000000000000000000000000..b1e4568da22d356104828731bc6dd42482c910b1 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-20.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c76d54d05efc3c647a0c2eb76f88238c1f707659c06fb1f6d148d9d583aea10d +size 155612 diff --git a/FateZero-main/data/style/rabit/rabit-21.png b/FateZero-main/data/style/rabit/rabit-21.png new file mode 100644 index 0000000000000000000000000000000000000000..9f457f5052ab853daa012fbfcc90f7b9cdfbf369 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-21.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8597ad2095ebe2298cb3831f5189a9dba1aeee19441396647d433ccb862c76a6 +size 154157 diff --git a/FateZero-main/data/style/rabit/rabit-22.png b/FateZero-main/data/style/rabit/rabit-22.png new file mode 100644 index 0000000000000000000000000000000000000000..03b9ee9d1e5acf9034a4ddda904178524374d6bb --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-22.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b4f5f51b51403dd5ee39318200386544a1cd945ef59441c25342246ca1144aa +size 155132 diff --git a/FateZero-main/data/style/rabit/rabit-23.png b/FateZero-main/data/style/rabit/rabit-23.png new file mode 100644 index 0000000000000000000000000000000000000000..a14915a4eacc8bdeeadc6c0b887b83712be28f89 --- /dev/null +++ b/FateZero-main/data/style/rabit/rabit-23.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21d51dc3ff09e9f06e22923f41b0a56402798c36a624f8c2cb2398f97194350e +size 154874 diff --git a/FateZero-main/data/style/red_water_lily_opening.mp4 b/FateZero-main/data/style/red_water_lily_opening.mp4 new file mode 100644 index 0000000000000000000000000000000000000000..7db01b4647efb996b0eaad6e2903e84d9a654202 --- /dev/null +++ b/FateZero-main/data/style/red_water_lily_opening.mp4 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31a503dc2fcfad3cf81e9a3deb89429925b5658d364ced7cc028927f417ca431 +size 1660321 diff --git a/FateZero-main/data/style/sunflower.mp4 b/FateZero-main/data/style/sunflower.mp4 new file mode 100644 index 0000000000000000000000000000000000000000..b4e445e53baa9a4a1fac6979da335832005967d2 --- /dev/null +++ b/FateZero-main/data/style/sunflower.mp4 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9c8c9ed5cd4af51d87547813be4a94c3d2eca2c4e427521f880a00c70c6cddd +size 129363 diff --git a/FateZero-main/data/style/sunflower/00000.png b/FateZero-main/data/style/sunflower/00000.png new file mode 100644 index 0000000000000000000000000000000000000000..b12a609baf697563d8cc40a311f736ad0fc659ed --- /dev/null +++ b/FateZero-main/data/style/sunflower/00000.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c116e8789e7b0f97e9b325198b1eccd6949b0bbc0cff575caaba191eab68ca22 +size 302261 diff --git a/FateZero-main/data/style/sunflower/00001.png b/FateZero-main/data/style/sunflower/00001.png new file mode 100644 index 0000000000000000000000000000000000000000..3c96adeaf1ad666dcfb815fb1664ada6c1385299 --- /dev/null +++ b/FateZero-main/data/style/sunflower/00001.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e50caa22525d1dccb97d1a913692a5a7720fdd5f5e3c4412e1e3e54e44b389f +size 299605 diff --git a/FateZero-main/data/style/sunflower/00002.png b/FateZero-main/data/style/sunflower/00002.png new file mode 100644 index 0000000000000000000000000000000000000000..6c1ec784f792e47af4f27fa8b50ccedac8a182aa --- /dev/null +++ b/FateZero-main/data/style/sunflower/00002.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cebe1c2bda9208920b7bd7f27c06907eebde8a36b32234b9162f2eb9c8be9cf +size 314949 diff --git a/FateZero-main/data/style/sunflower/00003.png b/FateZero-main/data/style/sunflower/00003.png new file mode 100644 index 0000000000000000000000000000000000000000..1bcd58b65fd1b7104549cc0605aaf4cae0e40ddd --- /dev/null +++ b/FateZero-main/data/style/sunflower/00003.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0df0594504a93a8b39aa6f4aca916a3bf4619e19fdc72ddaa87d6fb437ba18c7 +size 342092 diff --git a/FateZero-main/data/style/sunflower/00004.png b/FateZero-main/data/style/sunflower/00004.png new file mode 100644 index 0000000000000000000000000000000000000000..af37b0f4a9bd4970b040f0ad23a6a43015c4f09d --- /dev/null +++ b/FateZero-main/data/style/sunflower/00004.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:640ab0eaf5bd9a9f33c3a82558ec4f6d5930c97e25efc2b09df1d3d07c1c6b01 +size 343244 diff --git a/FateZero-main/data/style/sunflower/00005.png b/FateZero-main/data/style/sunflower/00005.png new file mode 100644 index 0000000000000000000000000000000000000000..27a2c24993ccc43d373d71520b94e389f44ead7c --- /dev/null +++ b/FateZero-main/data/style/sunflower/00005.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5b4f1e5427302ca49fefa80315b0d2a21f29a72b681cd599fa7f3d2386dbe4d +size 349710 diff --git a/FateZero-main/data/style/sunflower/00006.png b/FateZero-main/data/style/sunflower/00006.png new file mode 100644 index 0000000000000000000000000000000000000000..5f1d7d5850ae6b9798a110d0c62f4a345e2cec6a --- /dev/null +++ b/FateZero-main/data/style/sunflower/00006.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:514edaa3e8fa1a1f1ae220135afdcfc26a1f453a60519299d2beb52495408418 +size 349896 diff --git a/FateZero-main/data/style/sunflower/00007.png b/FateZero-main/data/style/sunflower/00007.png new file mode 100644 index 0000000000000000000000000000000000000000..abf33e32ba19239bfff06dbeedb95961849be187 --- /dev/null +++ b/FateZero-main/data/style/sunflower/00007.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d98d9cecd5f5313d3eb444b8ab505951c83c5a8cd9fb430cde0e7f6eab4aace +size 350385 diff --git a/FateZero-main/data/style/surf.mp4 b/FateZero-main/data/style/surf.mp4 new file mode 100644 index 0000000000000000000000000000000000000000..b6f1d0dbab5ccaec6d6b615f25930c8e0b19051f --- /dev/null +++ b/FateZero-main/data/style/surf.mp4 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c32af6211ba863597f442c79cc157532cb231a7a74a6978bb8a22738d7940506 +size 156069 diff --git a/FateZero-main/data/style/train.mp4 b/FateZero-main/data/style/train.mp4 new file mode 100644 index 0000000000000000000000000000000000000000..2ba58c87724a73ea0a9b27fb0cdcf823ce824868 --- /dev/null +++ b/FateZero-main/data/style/train.mp4 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:439e609a404deff78a9dbf960dafd86b667157d327c5b12ad3c08bdaedd4c2b2 +size 856799 diff --git a/FateZero-main/data/style/train/00003.jpg b/FateZero-main/data/style/train/00003.jpg new file mode 100644 index 0000000000000000000000000000000000000000..a305a9ec221459b4bdc20f69c329e139da49d6f2 --- /dev/null +++ b/FateZero-main/data/style/train/00003.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5ec273610e08975405563adbfb5bcd22069446a6b7c4b75aae75fe3bef6fa2d +size 102269 diff --git a/FateZero-main/data/style/train/00011.jpg b/FateZero-main/data/style/train/00011.jpg new file mode 100644 index 0000000000000000000000000000000000000000..bc7afcdce59ea9125223916577af883e2cc70a67 --- /dev/null +++ b/FateZero-main/data/style/train/00011.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:acdf7250d84d3d83eaab74934fea75a68b258916541a2edb8e48fc6ce2eac7c3 +size 104160 diff --git a/FateZero-main/data/style/train/00020.jpg b/FateZero-main/data/style/train/00020.jpg new file mode 100644 index 0000000000000000000000000000000000000000..4663956bad762c1799517c68eb83678446896164 --- /dev/null +++ b/FateZero-main/data/style/train/00020.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71201f537a298d699c0714ae9100f5f8d47d34d9741fcb08d9bd5296b7d762e8 +size 104243 diff --git a/FateZero-main/data/style/train/00036.jpg b/FateZero-main/data/style/train/00036.jpg new file mode 100644 index 0000000000000000000000000000000000000000..1c5e8b7141f73762311831976be1ee0b597adb06 --- /dev/null +++ b/FateZero-main/data/style/train/00036.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d31fc2cb362a68fa0832a0ba7d5f201e7f1d629d1e7df24f605029e0bac5199d +size 103010 diff --git a/FateZero-main/data/style/train/00037.jpg b/FateZero-main/data/style/train/00037.jpg new file mode 100644 index 0000000000000000000000000000000000000000..5f679b99675b571ffb0d8e5b964513cea31185e2 --- /dev/null +++ b/FateZero-main/data/style/train/00037.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbf63ec4b785e4682f8999856f37a765d3492d9301f2c153ac5584eb0bb980cc +size 103128 diff --git a/FateZero-main/data/style/train/00040.jpg b/FateZero-main/data/style/train/00040.jpg new file mode 100644 index 0000000000000000000000000000000000000000..8487ba63b10dd0dffa081f6b152132b35fcbd731 --- /dev/null +++ b/FateZero-main/data/style/train/00040.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:000f4cc6da9b7c58d5f08b6ae84fd4b9dec72f5bc217688a9872b92eb7ec8a3f +size 102967 diff --git a/FateZero-main/data/style/train/00043.jpg b/FateZero-main/data/style/train/00043.jpg new file mode 100644 index 0000000000000000000000000000000000000000..9834dadf71b43f20e975f6a204667ddc2b27dd4a --- /dev/null +++ b/FateZero-main/data/style/train/00043.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:228328177c97b9d7e46da3f2fe30080a576d5ec607b6e9a013fd87d763079146 +size 103098 diff --git a/FateZero-main/data/style/train/00055.jpg b/FateZero-main/data/style/train/00055.jpg new file mode 100644 index 0000000000000000000000000000000000000000..c58b0fc9f365d5d614ff3b241520edcc850bef90 --- /dev/null +++ b/FateZero-main/data/style/train/00055.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15bcad24fde436fb99e999361ff6e92e0efbf1dc6773bf15576f0bc1622daf5d +size 103547 diff --git a/FateZero-main/data/style/train/00060.jpg b/FateZero-main/data/style/train/00060.jpg new file mode 100644 index 0000000000000000000000000000000000000000..7f9809190620d190c43a9f8c560f894f83b002d7 --- /dev/null +++ b/FateZero-main/data/style/train/00060.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d25449a5cd7aab1f947cac4e04ee6226b297c53109c0df76bbed5d527238da54 +size 102836 diff --git a/FateZero-main/data/style/train/00067.jpg b/FateZero-main/data/style/train/00067.jpg new file mode 100644 index 0000000000000000000000000000000000000000..147e7df39c6595abed1b8fd245fc544bb54050e8 --- /dev/null +++ b/FateZero-main/data/style/train/00067.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a38912cf51ed4b74e52ea65a2b6862fcaf697f4565d12a846e02cab7ce2dae7 +size 102968 diff --git a/FateZero-main/data/style/train/00068.jpg b/FateZero-main/data/style/train/00068.jpg new file mode 100644 index 0000000000000000000000000000000000000000..67add01dbefdd767b4fbb345568f255440bd2fd3 --- /dev/null +++ b/FateZero-main/data/style/train/00068.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eec71485a7fde5c787d03dd4a9ff127e638130bfeb1749f40406684169f18404 +size 102985 diff --git a/FateZero-main/data/style/train/00069.jpg b/FateZero-main/data/style/train/00069.jpg new file mode 100644 index 0000000000000000000000000000000000000000..b2d181731b98f3a8d12b502c5a502464840195ba --- /dev/null +++ b/FateZero-main/data/style/train/00069.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9737fa7c7496df9d75df9a80da9cc36ca0bf5be460ece923e8ed597c2ad02f7 +size 103020 diff --git a/FateZero-main/data/style/train/00070.jpg b/FateZero-main/data/style/train/00070.jpg new file mode 100644 index 0000000000000000000000000000000000000000..caaeec1fdcac5edcb56c99a3f0602e6a7a46de1a --- /dev/null +++ b/FateZero-main/data/style/train/00070.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3ec3b36001c884840042668f69346949be61c1c7ac0fb9ba24f1f91d971b801 +size 103248 diff --git a/FateZero-main/data/style/train/00083.jpg b/FateZero-main/data/style/train/00083.jpg new file mode 100644 index 0000000000000000000000000000000000000000..d8e56b471571cfb9e4f01ec1de6fefc2099f05d1 --- /dev/null +++ b/FateZero-main/data/style/train/00083.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f87ee348994a88bd72a191f69665775c4faa9748e4a795def98c0c4db2e9d22 +size 103515 diff --git a/FateZero-main/data/style/train/00085.jpg b/FateZero-main/data/style/train/00085.jpg new file mode 100644 index 0000000000000000000000000000000000000000..02fc69f62168800f56392fc67b5134180231cd08 --- /dev/null +++ b/FateZero-main/data/style/train/00085.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:165355d179ef48b788d5fa95aa98d939f7f5b4e16a705dd34747e7074b49a749 +size 103619 diff --git a/FateZero-main/data/style/train/00092.jpg b/FateZero-main/data/style/train/00092.jpg new file mode 100644 index 0000000000000000000000000000000000000000..b85f29c819cda3a5ec2b307f6b52cef3162d0cae --- /dev/null +++ b/FateZero-main/data/style/train/00092.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d31f85d49c573f1f6a4a13312088546737f1a3deef56e9ca6918d63a6cc5ec5 +size 102894 diff --git a/FateZero-main/data/style/train/00100.jpg b/FateZero-main/data/style/train/00100.jpg new file mode 100644 index 0000000000000000000000000000000000000000..e75373622095c737138cac0b318530abe7629a1e --- /dev/null +++ b/FateZero-main/data/style/train/00100.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed89ad44a600dd12f18c14c266575c9ead4d22007443360e8de0a45887d5ab75 +size 102796 diff --git a/FateZero-main/data/style/train/00110.jpg b/FateZero-main/data/style/train/00110.jpg new file mode 100644 index 0000000000000000000000000000000000000000..64a9557c0c570123b1bba9bd3518a1cc87075fb8 --- /dev/null +++ b/FateZero-main/data/style/train/00110.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85ddc77320139e5e39134839e8bf9f733aee4254adf507f9fe076fbe229de865 +size 102728 diff --git a/FateZero-main/data/style/train/00115.jpg b/FateZero-main/data/style/train/00115.jpg new file mode 100644 index 0000000000000000000000000000000000000000..6be9cc34bb3950eaa009dd986d86003fe0372a04 --- /dev/null +++ b/FateZero-main/data/style/train/00115.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f42b643f5b237d9d1170c582233765097966ff93ce1bac39b05fac2b7ec672c +size 102588 diff --git a/FateZero-main/data/style/train/00118.jpg b/FateZero-main/data/style/train/00118.jpg new file mode 100644 index 0000000000000000000000000000000000000000..381f08b80d36e2d886214804b0a515751bf2ae00 --- /dev/null +++ b/FateZero-main/data/style/train/00118.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e41ac663a7ed5ca65053bf9fc14088afe89326d1dbfbf95768f216093a9e7b16 +size 102648 diff --git a/FateZero-main/data/style/train/00121.jpg b/FateZero-main/data/style/train/00121.jpg new file mode 100644 index 0000000000000000000000000000000000000000..91d4fe4beb7ead7c4f194a0b7b4a2f566b646996 --- /dev/null +++ b/FateZero-main/data/style/train/00121.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76fd340d51ccaaf8e4061f5a4e7525be4cd0631a3f264f3f463db6040a95f6dd +size 102527 diff --git a/FateZero-main/data/style/train/00129.jpg b/FateZero-main/data/style/train/00129.jpg new file mode 100644 index 0000000000000000000000000000000000000000..482dd6cb44713d658cf8a2013242dd6ea5d1e588 --- /dev/null +++ b/FateZero-main/data/style/train/00129.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78bd2f6808fd24c1ace1f57dbb49c1385e5955cb7d9e53b700b3196b5995c768 +size 102277 diff --git a/FateZero-main/data/style/train/00140.jpg b/FateZero-main/data/style/train/00140.jpg new file mode 100644 index 0000000000000000000000000000000000000000..36966856db02d134fad7cf371b385663979a9cf9 --- /dev/null +++ b/FateZero-main/data/style/train/00140.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:387c28c207a57b5d1405e679e67e8fe0c9375e1eb308fe15f3063376187847ce +size 102326 diff --git a/FateZero-main/data/style/train/00144.jpg b/FateZero-main/data/style/train/00144.jpg new file mode 100644 index 0000000000000000000000000000000000000000..f6f913903a6bd656ab6f3b01793992320d55572c --- /dev/null +++ b/FateZero-main/data/style/train/00144.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:211fc255e41f375e5e88002f22497952936a76ea7eb07f8dd0967c4e26b49786 +size 102244 diff --git a/FateZero-main/data/style/train/00151.jpg b/FateZero-main/data/style/train/00151.jpg new file mode 100644 index 0000000000000000000000000000000000000000..60fb8302275d19818c4e85e8cb8a569eea48e6a4 --- /dev/null +++ b/FateZero-main/data/style/train/00151.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7ab0c845e1adf37c4966eab216ef458af68f2984c835a7a8a469082156d9973 +size 101496 diff --git a/FateZero-main/data/style/train/00158.jpg b/FateZero-main/data/style/train/00158.jpg new file mode 100644 index 0000000000000000000000000000000000000000..18b4fd32824c17d68ad78b65f134875c755c9c5a --- /dev/null +++ b/FateZero-main/data/style/train/00158.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71db8685638e03690818187daeea5243457646fc0ea18a4527b3c0f0b5478cdd +size 101537 diff --git a/FateZero-main/data/style/train/00167.jpg b/FateZero-main/data/style/train/00167.jpg new file mode 100644 index 0000000000000000000000000000000000000000..2c241f08a51afe330093887eb03ac7694d078307 --- /dev/null +++ b/FateZero-main/data/style/train/00167.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e87bb78010804c5726998cadb2db432bf167915198635c5e7ae5e083caa7323 +size 101765 diff --git a/FateZero-main/data/style/train/00177.jpg b/FateZero-main/data/style/train/00177.jpg new file mode 100644 index 0000000000000000000000000000000000000000..7b75faa1a79578bf09af9eba0a41c76cfd0adb70 --- /dev/null +++ b/FateZero-main/data/style/train/00177.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49ab91c105b6aa35b986ccd1a108d84e5e389a2e16c87cf169c6f4f0a239fd08 +size 102418 diff --git a/FateZero-main/data/style/train/00180.jpg b/FateZero-main/data/style/train/00180.jpg new file mode 100644 index 0000000000000000000000000000000000000000..0a8664dabcc995683b4850d06e08bae44d6af883 --- /dev/null +++ b/FateZero-main/data/style/train/00180.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d47165c1b8bed835fcafbe4a33aec763856283914575d1eef3601e4cd8880394 +size 101876 diff --git a/FateZero-main/data/style/train/00181.jpg b/FateZero-main/data/style/train/00181.jpg new file mode 100644 index 0000000000000000000000000000000000000000..f12568bf266d56f13c7baa70a2082c7df2a2ce65 --- /dev/null +++ b/FateZero-main/data/style/train/00181.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec67c939d528a5f96502d7bcc4797260cd38aa9912c1b65834c51062664d3c15 +size 101906 diff --git a/FateZero-main/data/style/train/00184.jpg b/FateZero-main/data/style/train/00184.jpg new file mode 100644 index 0000000000000000000000000000000000000000..b6298efc92010a4fa8bc2f2962f75b2b1dad132c --- /dev/null +++ b/FateZero-main/data/style/train/00184.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5291417483344cb554df46300b444a51e2e38351326e853a2e61d32316d81a05 +size 101586 diff --git a/FateZero-main/data/style/train/00185.jpg b/FateZero-main/data/style/train/00185.jpg new file mode 100644 index 0000000000000000000000000000000000000000..09711f4c589e021f9b7582f6a63bc86bd72460e5 --- /dev/null +++ b/FateZero-main/data/style/train/00185.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dac17d3c083e4bb2837117985514d6abb78524d9c57763f97c632c3db0ac9a59 +size 101600 diff --git a/FateZero-main/data/style/train/00190.jpg b/FateZero-main/data/style/train/00190.jpg new file mode 100644 index 0000000000000000000000000000000000000000..7449dc665beb604de7680da45c3b3d6c2e5dc62c --- /dev/null +++ b/FateZero-main/data/style/train/00190.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:995d644aa491b68061dad7f68151638e496c5d85b1738aa7880ec24c46d7a8b5 +size 101626 diff --git a/FateZero-main/data/style/train/00197.jpg b/FateZero-main/data/style/train/00197.jpg new file mode 100644 index 0000000000000000000000000000000000000000..f79868cc25f48c638da1e4792e2d1b5f20fb3422 --- /dev/null +++ b/FateZero-main/data/style/train/00197.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6bfc52d02574046c0a5f8b18397b13bdfb7b50fb5cc6124c98ff837eb94f668 +size 101703 diff --git a/FateZero-main/data/style/train/00202.jpg b/FateZero-main/data/style/train/00202.jpg new file mode 100644 index 0000000000000000000000000000000000000000..1e0fcb140937cdd1b4fb571e35feef78dc5e9482 --- /dev/null +++ b/FateZero-main/data/style/train/00202.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a96d33b74fb42ad4396b953eac9ba4551f8475a3d8d35e26e52b40b8f5bb8867 +size 101765 diff --git a/FateZero-main/data/style/train/00203.jpg b/FateZero-main/data/style/train/00203.jpg new file mode 100644 index 0000000000000000000000000000000000000000..162758ccc616f4fccc31c870bc451f5e898b8ce1 --- /dev/null +++ b/FateZero-main/data/style/train/00203.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f2ff696b057ad6bdc305e648855e64f542e369c755bd808d7d38b12d2df81cc +size 101677 diff --git a/FateZero-main/data/style/train/00206.jpg b/FateZero-main/data/style/train/00206.jpg new file mode 100644 index 0000000000000000000000000000000000000000..e1ec09579605d2f554b0dced0256323747213ee7 --- /dev/null +++ b/FateZero-main/data/style/train/00206.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cac8a66b4f7f4526243b972a61d443da52b1628a2536c4a1ecc3db4aba1fcc79 +size 101559 diff --git a/FateZero-main/data/style/train/00212.jpg b/FateZero-main/data/style/train/00212.jpg new file mode 100644 index 0000000000000000000000000000000000000000..88d5355d3351317cf100d69e5fee3f6a47d4d7f7 --- /dev/null +++ b/FateZero-main/data/style/train/00212.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:578098e117a0915bdae138f6b98d8982e1bc6dd166e21415bdefc9e1c5170aeb +size 101821 diff --git a/FateZero-main/data/style/train/00220.jpg b/FateZero-main/data/style/train/00220.jpg new file mode 100644 index 0000000000000000000000000000000000000000..be7fee191e4d882fd2c3a1c6348e6b2a782f2e11 --- /dev/null +++ b/FateZero-main/data/style/train/00220.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8c23abc459cabb34623338a13aca343a2d59785dba1492a641bf2953870d389 +size 101466 diff --git a/FateZero-main/output.mp4 b/FateZero-main/output.mp4 new file mode 100644 index 0000000000000000000000000000000000000000..821a4fce3bb0ca7773c68e725933a93dc73b6b43 --- /dev/null +++ b/FateZero-main/output.mp4 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bfd86ba01fabfacc61308c11d975e7cac406b9df3285f8c9ac42c7d4c9228fc7 +size 5164110 diff --git a/RAVE-main/annotator/binary/__init__.py b/RAVE-main/annotator/binary/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..2d13ad692ffc109ad95789334bb5524d52794acc --- /dev/null +++ b/RAVE-main/annotator/binary/__init__.py @@ -0,0 +1,14 @@ +import cv2 + + +def apply_binary(img, bin_threshold): + img_gray = cv2.cvtColor(img, cv2.COLOR_RGB2GRAY) + + if bin_threshold == 0 or bin_threshold == 255: + # Otsu's threshold + otsu_threshold, img_bin = cv2.threshold(img_gray, 0, 255, cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU) + print("Otsu threshold:", otsu_threshold) + else: + _, img_bin = cv2.threshold(img_gray, bin_threshold, 255, cv2.THRESH_BINARY_INV) + + return cv2.cvtColor(img_bin, cv2.COLOR_GRAY2RGB) diff --git a/RAVE-main/annotator/color/__init__.py b/RAVE-main/annotator/color/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..65799a2a83efd18dc556600c99d43292845aa6f2 --- /dev/null +++ b/RAVE-main/annotator/color/__init__.py @@ -0,0 +1,20 @@ +import cv2 + +def cv2_resize_shortest_edge(image, size): + h, w = image.shape[:2] + if h < w: + new_h = size + new_w = int(round(w / h * size)) + else: + new_w = size + new_h = int(round(h / w * size)) + resized_image = cv2.resize(image, (new_w, new_h), interpolation=cv2.INTER_AREA) + return resized_image + +def apply_color(img, res=512): + img = cv2_resize_shortest_edge(img, res) + h, w = img.shape[:2] + + input_img_color = cv2.resize(img, (w//64, h//64), interpolation=cv2.INTER_CUBIC) + input_img_color = cv2.resize(input_img_color, (w, h), interpolation=cv2.INTER_NEAREST) + return input_img_color \ No newline at end of file diff --git a/RAVE-main/annotator/lineart_anime/LICENSE b/RAVE-main/annotator/lineart_anime/LICENSE new file mode 100644 index 0000000000000000000000000000000000000000..16a9d56a3d4c15e4f34ac5426459c58487b01520 --- /dev/null +++ b/RAVE-main/annotator/lineart_anime/LICENSE @@ -0,0 +1,21 @@ +MIT License + +Copyright (c) 2022 Caroline Chan + +Permission is hereby granted, free of charge, to any person obtaining a copy +of this software and associated documentation files (the "Software"), to deal +in the Software without restriction, including without limitation the rights +to use, copy, modify, merge, publish, distribute, sublicense, and/or sell +copies of the Software, and to permit persons to whom the Software is +furnished to do so, subject to the following conditions: + +The above copyright notice and this permission notice shall be included in all +copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +SOFTWARE. \ No newline at end of file diff --git a/RAVE-main/annotator/lineart_anime/__init__.py b/RAVE-main/annotator/lineart_anime/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..bcc3cf94b2dbe56019a339c6d62b806ea7502b60 --- /dev/null +++ b/RAVE-main/annotator/lineart_anime/__init__.py @@ -0,0 +1,160 @@ +import numpy as np +import torch +import torch.nn as nn +import functools + +import os +import cv2 +from einops import rearrange +from annotator.annotator_path import models_path, DEVICE + + +class UnetGenerator(nn.Module): + """Create a Unet-based generator""" + + def __init__(self, input_nc, output_nc, num_downs, ngf=64, norm_layer=nn.BatchNorm2d, use_dropout=False): + """Construct a Unet generator + Parameters: + input_nc (int) -- the number of channels in input images + output_nc (int) -- the number of channels in output images + num_downs (int) -- the number of downsamplings in UNet. For example, # if |num_downs| == 7, + image of size 128x128 will become of size 1x1 # at the bottleneck + ngf (int) -- the number of filters in the last conv layer + norm_layer -- normalization layer + We construct the U-Net from the innermost layer to the outermost layer. + It is a recursive process. + """ + super(UnetGenerator, self).__init__() + # construct unet structure + unet_block = UnetSkipConnectionBlock(ngf * 8, ngf * 8, input_nc=None, submodule=None, norm_layer=norm_layer, innermost=True) # add the innermost layer + for _ in range(num_downs - 5): # add intermediate layers with ngf * 8 filters + unet_block = UnetSkipConnectionBlock(ngf * 8, ngf * 8, input_nc=None, submodule=unet_block, norm_layer=norm_layer, use_dropout=use_dropout) + # gradually reduce the number of filters from ngf * 8 to ngf + unet_block = UnetSkipConnectionBlock(ngf * 4, ngf * 8, input_nc=None, submodule=unet_block, norm_layer=norm_layer) + unet_block = UnetSkipConnectionBlock(ngf * 2, ngf * 4, input_nc=None, submodule=unet_block, norm_layer=norm_layer) + unet_block = UnetSkipConnectionBlock(ngf, ngf * 2, input_nc=None, submodule=unet_block, norm_layer=norm_layer) + self.model = UnetSkipConnectionBlock(output_nc, ngf, input_nc=input_nc, submodule=unet_block, outermost=True, norm_layer=norm_layer) # add the outermost layer + + def forward(self, input): + """Standard forward""" + return self.model(input) + + +class UnetSkipConnectionBlock(nn.Module): + """Defines the Unet submodule with skip connection. + X -------------------identity---------------------- + |-- downsampling -- |submodule| -- upsampling --| + """ + + def __init__(self, outer_nc, inner_nc, input_nc=None, + submodule=None, outermost=False, innermost=False, norm_layer=nn.BatchNorm2d, use_dropout=False): + """Construct a Unet submodule with skip connections. + Parameters: + outer_nc (int) -- the number of filters in the outer conv layer + inner_nc (int) -- the number of filters in the inner conv layer + input_nc (int) -- the number of channels in input images/features + submodule (UnetSkipConnectionBlock) -- previously defined submodules + outermost (bool) -- if this module is the outermost module + innermost (bool) -- if this module is the innermost module + norm_layer -- normalization layer + use_dropout (bool) -- if use dropout layers. + """ + super(UnetSkipConnectionBlock, self).__init__() + self.outermost = outermost + if type(norm_layer) == functools.partial: + use_bias = norm_layer.func == nn.InstanceNorm2d + else: + use_bias = norm_layer == nn.InstanceNorm2d + if input_nc is None: + input_nc = outer_nc + downconv = nn.Conv2d(input_nc, inner_nc, kernel_size=4, + stride=2, padding=1, bias=use_bias) + downrelu = nn.LeakyReLU(0.2, True) + downnorm = norm_layer(inner_nc) + uprelu = nn.ReLU(True) + upnorm = norm_layer(outer_nc) + + if outermost: + upconv = nn.ConvTranspose2d(inner_nc * 2, outer_nc, + kernel_size=4, stride=2, + padding=1) + down = [downconv] + up = [uprelu, upconv, nn.Tanh()] + model = down + [submodule] + up + elif innermost: + upconv = nn.ConvTranspose2d(inner_nc, outer_nc, + kernel_size=4, stride=2, + padding=1, bias=use_bias) + down = [downrelu, downconv] + up = [uprelu, upconv, upnorm] + model = down + up + else: + upconv = nn.ConvTranspose2d(inner_nc * 2, outer_nc, + kernel_size=4, stride=2, + padding=1, bias=use_bias) + down = [downrelu, downconv, downnorm] + up = [uprelu, upconv, upnorm] + + if use_dropout: + model = down + [submodule] + up + [nn.Dropout(0.5)] + else: + model = down + [submodule] + up + + self.model = nn.Sequential(*model) + + def forward(self, x): + if self.outermost: + return self.model(x) + else: # add skip connections + return torch.cat([x, self.model(x)], 1) + + +class LineartAnimeDetector: + model_dir = os.path.join(models_path, "lineart_anime") + + def __init__(self): + self.model = None + self.device = DEVICE + + def load_model(self): + remote_model_path = "https://huggingface.co/lllyasviel/Annotators/resolve/main/netG.pth" + modelpath = os.path.join(self.model_dir, "netG.pth") + if not os.path.exists(modelpath): + from basicsr.utils.download_util import load_file_from_url + load_file_from_url(remote_model_path, model_dir=self.model_dir) + norm_layer = functools.partial(nn.InstanceNorm2d, affine=False, track_running_stats=False) + net = UnetGenerator(3, 1, 8, 64, norm_layer=norm_layer, use_dropout=False) + ckpt = torch.load(modelpath) + for key in list(ckpt.keys()): + if 'module.' in key: + ckpt[key.replace('module.', '')] = ckpt[key] + del ckpt[key] + net.load_state_dict(ckpt) + net.eval() + self.model = net.to(self.device) + + def unload_model(self): + if self.model is not None: + self.model.cpu() + + def __call__(self, input_image): + if self.model is None: + self.load_model() + self.model.to(self.device) + + H, W, C = input_image.shape + Hn = 256 * int(np.ceil(float(H) / 256.0)) + Wn = 256 * int(np.ceil(float(W) / 256.0)) + img = cv2.resize(input_image, (Wn, Hn), interpolation=cv2.INTER_CUBIC) + with torch.no_grad(): + image_feed = torch.from_numpy(img).float().to(self.device) + image_feed = image_feed / 127.5 - 1.0 + image_feed = rearrange(image_feed, 'h w c -> 1 c h w') + + line = self.model(image_feed)[0, 0] * 127.5 + 127.5 + line = line.cpu().numpy() + + line = cv2.resize(line, (W, H), interpolation=cv2.INTER_CUBIC) + line = line.clip(0, 255).astype(np.uint8) + return line + diff --git a/RAVE-main/annotator/lineart_anime/__pycache__/__init__.cpython-38.pyc b/RAVE-main/annotator/lineart_anime/__pycache__/__init__.cpython-38.pyc new file mode 100644 index 0000000000000000000000000000000000000000..1eaed7554572467ff0af91badf954dabc7d1c984 Binary files /dev/null and b/RAVE-main/annotator/lineart_anime/__pycache__/__init__.cpython-38.pyc differ diff --git a/RAVE-main/annotator/pidinet/LICENSE b/RAVE-main/annotator/pidinet/LICENSE new file mode 100644 index 0000000000000000000000000000000000000000..913b6cf92c19d37b6ee4f7bc99c65f655e7f840c --- /dev/null +++ b/RAVE-main/annotator/pidinet/LICENSE @@ -0,0 +1,21 @@ +It is just for research purpose, and commercial use should be contacted with authors first. + +Copyright (c) 2021 Zhuo Su + +Permission is hereby granted, free of charge, to any person obtaining a copy +of this software and associated documentation files (the "Software"), to deal +in the Software without restriction, including without limitation the rights +to use, copy, modify, merge, publish, distribute, sublicense, and/or sell +copies of the Software, and to permit persons to whom the Software is +furnished to do so, subject to the following conditions: + +The above copyright notice and this permission notice shall be included in all +copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +SOFTWARE. \ No newline at end of file diff --git a/RAVE-main/annotator/pidinet/__init__.py b/RAVE-main/annotator/pidinet/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..e2e69e7caca9940a8e96d7d89f98a72ef227148a --- /dev/null +++ b/RAVE-main/annotator/pidinet/__init__.py @@ -0,0 +1,63 @@ +import os +import torch +import numpy as np +from einops import rearrange +from annotator.pidinet.model import pidinet +from annotator.util import safe_step +from annotator.annotator_path import models_path, DEVICE +import safetensors.torch +# from modules.safe import unsafe_torch_load + +def get_state_dict(d): + return d.get("state_dict", d) + +def load_state_dict(ckpt_path, location="cpu"): + _, extension = os.path.splitext(ckpt_path) + if extension.lower() == ".safetensors": + state_dict = safetensors.torch.load_file(ckpt_path, device=location) + else: + state_dict = torch.load(ckpt_path, map_location=torch.device(location)) + state_dict = get_state_dict(state_dict) + return state_dict + +netNetwork = None +remote_model_path = "https://huggingface.co/lllyasviel/Annotators/resolve/main/table5_pidinet.pth" +modeldir = os.path.join(models_path, "pidinet") +old_modeldir = os.path.dirname(os.path.realpath(__file__)) + +def apply_pidinet(input_image, is_safe=False, apply_fliter=False): + global netNetwork + if netNetwork is None: + modelpath = os.path.join(modeldir, "table5_pidinet.pth") + old_modelpath = os.path.join(old_modeldir, "table5_pidinet.pth") + if os.path.exists(old_modelpath): + modelpath = old_modelpath + elif not os.path.exists(modelpath): + from basicsr.utils.download_util import load_file_from_url + load_file_from_url(remote_model_path, model_dir=modeldir) + netNetwork = pidinet() + ckp = load_state_dict(modelpath) + netNetwork.load_state_dict({k.replace('module.',''):v for k, v in ckp.items()}) + + netNetwork = netNetwork.to(DEVICE) + netNetwork.eval() + assert input_image.ndim == 3 + input_image = input_image[:, :, ::-1].copy() + with torch.no_grad(): + image_pidi = torch.from_numpy(input_image).float().to(DEVICE) + image_pidi = image_pidi / 255.0 + image_pidi = rearrange(image_pidi, 'h w c -> 1 c h w') + edge = netNetwork(image_pidi)[-1] + edge = edge.cpu().numpy() + if apply_fliter: + edge = edge > 0.5 + if is_safe: + edge = safe_step(edge) + edge = (edge * 255.0).clip(0, 255).astype(np.uint8) + + return edge[0][0] + +def unload_pid_model(): + global netNetwork + if netNetwork is not None: + netNetwork.cpu() \ No newline at end of file diff --git a/RAVE-main/annotator/pidinet/__pycache__/__init__.cpython-38.pyc b/RAVE-main/annotator/pidinet/__pycache__/__init__.cpython-38.pyc new file mode 100644 index 0000000000000000000000000000000000000000..3d8ede0b8e783f3d6368bc718ff8e39659391ee1 Binary files /dev/null and b/RAVE-main/annotator/pidinet/__pycache__/__init__.cpython-38.pyc differ diff --git a/RAVE-main/annotator/pidinet/__pycache__/model.cpython-38.pyc b/RAVE-main/annotator/pidinet/__pycache__/model.cpython-38.pyc new file mode 100644 index 0000000000000000000000000000000000000000..880e4503147770431b5dd1229ac930365525f886 Binary files /dev/null and b/RAVE-main/annotator/pidinet/__pycache__/model.cpython-38.pyc differ diff --git a/RAVE-main/annotator/pidinet/model.py b/RAVE-main/annotator/pidinet/model.py new file mode 100644 index 0000000000000000000000000000000000000000..82780faa650a389d681f3b01eb3f4f9bef994795 --- /dev/null +++ b/RAVE-main/annotator/pidinet/model.py @@ -0,0 +1,654 @@ +""" +Author: Zhuo Su, Wenzhe Liu +Date: Feb 18, 2021 +""" + +import math + +import cv2 +import numpy as np +import torch +import torch.nn as nn +import torch.nn.functional as F +from basicsr.utils import img2tensor +from annotator.annotator_path import DEVICE + +nets = { + 'baseline': { + 'layer0': 'cv', + 'layer1': 'cv', + 'layer2': 'cv', + 'layer3': 'cv', + 'layer4': 'cv', + 'layer5': 'cv', + 'layer6': 'cv', + 'layer7': 'cv', + 'layer8': 'cv', + 'layer9': 'cv', + 'layer10': 'cv', + 'layer11': 'cv', + 'layer12': 'cv', + 'layer13': 'cv', + 'layer14': 'cv', + 'layer15': 'cv', + }, + 'c-v15': { + 'layer0': 'cd', + 'layer1': 'cv', + 'layer2': 'cv', + 'layer3': 'cv', + 'layer4': 'cv', + 'layer5': 'cv', + 'layer6': 'cv', + 'layer7': 'cv', + 'layer8': 'cv', + 'layer9': 'cv', + 'layer10': 'cv', + 'layer11': 'cv', + 'layer12': 'cv', + 'layer13': 'cv', + 'layer14': 'cv', + 'layer15': 'cv', + }, + 'a-v15': { + 'layer0': 'ad', + 'layer1': 'cv', + 'layer2': 'cv', + 'layer3': 'cv', + 'layer4': 'cv', + 'layer5': 'cv', + 'layer6': 'cv', + 'layer7': 'cv', + 'layer8': 'cv', + 'layer9': 'cv', + 'layer10': 'cv', + 'layer11': 'cv', + 'layer12': 'cv', + 'layer13': 'cv', + 'layer14': 'cv', + 'layer15': 'cv', + }, + 'r-v15': { + 'layer0': 'rd', + 'layer1': 'cv', + 'layer2': 'cv', + 'layer3': 'cv', + 'layer4': 'cv', + 'layer5': 'cv', + 'layer6': 'cv', + 'layer7': 'cv', + 'layer8': 'cv', + 'layer9': 'cv', + 'layer10': 'cv', + 'layer11': 'cv', + 'layer12': 'cv', + 'layer13': 'cv', + 'layer14': 'cv', + 'layer15': 'cv', + }, + 'cvvv4': { + 'layer0': 'cd', + 'layer1': 'cv', + 'layer2': 'cv', + 'layer3': 'cv', + 'layer4': 'cd', + 'layer5': 'cv', + 'layer6': 'cv', + 'layer7': 'cv', + 'layer8': 'cd', + 'layer9': 'cv', + 'layer10': 'cv', + 'layer11': 'cv', + 'layer12': 'cd', + 'layer13': 'cv', + 'layer14': 'cv', + 'layer15': 'cv', + }, + 'avvv4': { + 'layer0': 'ad', + 'layer1': 'cv', + 'layer2': 'cv', + 'layer3': 'cv', + 'layer4': 'ad', + 'layer5': 'cv', + 'layer6': 'cv', + 'layer7': 'cv', + 'layer8': 'ad', + 'layer9': 'cv', + 'layer10': 'cv', + 'layer11': 'cv', + 'layer12': 'ad', + 'layer13': 'cv', + 'layer14': 'cv', + 'layer15': 'cv', + }, + 'rvvv4': { + 'layer0': 'rd', + 'layer1': 'cv', + 'layer2': 'cv', + 'layer3': 'cv', + 'layer4': 'rd', + 'layer5': 'cv', + 'layer6': 'cv', + 'layer7': 'cv', + 'layer8': 'rd', + 'layer9': 'cv', + 'layer10': 'cv', + 'layer11': 'cv', + 'layer12': 'rd', + 'layer13': 'cv', + 'layer14': 'cv', + 'layer15': 'cv', + }, + 'cccv4': { + 'layer0': 'cd', + 'layer1': 'cd', + 'layer2': 'cd', + 'layer3': 'cv', + 'layer4': 'cd', + 'layer5': 'cd', + 'layer6': 'cd', + 'layer7': 'cv', + 'layer8': 'cd', + 'layer9': 'cd', + 'layer10': 'cd', + 'layer11': 'cv', + 'layer12': 'cd', + 'layer13': 'cd', + 'layer14': 'cd', + 'layer15': 'cv', + }, + 'aaav4': { + 'layer0': 'ad', + 'layer1': 'ad', + 'layer2': 'ad', + 'layer3': 'cv', + 'layer4': 'ad', + 'layer5': 'ad', + 'layer6': 'ad', + 'layer7': 'cv', + 'layer8': 'ad', + 'layer9': 'ad', + 'layer10': 'ad', + 'layer11': 'cv', + 'layer12': 'ad', + 'layer13': 'ad', + 'layer14': 'ad', + 'layer15': 'cv', + }, + 'rrrv4': { + 'layer0': 'rd', + 'layer1': 'rd', + 'layer2': 'rd', + 'layer3': 'cv', + 'layer4': 'rd', + 'layer5': 'rd', + 'layer6': 'rd', + 'layer7': 'cv', + 'layer8': 'rd', + 'layer9': 'rd', + 'layer10': 'rd', + 'layer11': 'cv', + 'layer12': 'rd', + 'layer13': 'rd', + 'layer14': 'rd', + 'layer15': 'cv', + }, + 'c16': { + 'layer0': 'cd', + 'layer1': 'cd', + 'layer2': 'cd', + 'layer3': 'cd', + 'layer4': 'cd', + 'layer5': 'cd', + 'layer6': 'cd', + 'layer7': 'cd', + 'layer8': 'cd', + 'layer9': 'cd', + 'layer10': 'cd', + 'layer11': 'cd', + 'layer12': 'cd', + 'layer13': 'cd', + 'layer14': 'cd', + 'layer15': 'cd', + }, + 'a16': { + 'layer0': 'ad', + 'layer1': 'ad', + 'layer2': 'ad', + 'layer3': 'ad', + 'layer4': 'ad', + 'layer5': 'ad', + 'layer6': 'ad', + 'layer7': 'ad', + 'layer8': 'ad', + 'layer9': 'ad', + 'layer10': 'ad', + 'layer11': 'ad', + 'layer12': 'ad', + 'layer13': 'ad', + 'layer14': 'ad', + 'layer15': 'ad', + }, + 'r16': { + 'layer0': 'rd', + 'layer1': 'rd', + 'layer2': 'rd', + 'layer3': 'rd', + 'layer4': 'rd', + 'layer5': 'rd', + 'layer6': 'rd', + 'layer7': 'rd', + 'layer8': 'rd', + 'layer9': 'rd', + 'layer10': 'rd', + 'layer11': 'rd', + 'layer12': 'rd', + 'layer13': 'rd', + 'layer14': 'rd', + 'layer15': 'rd', + }, + 'carv4': { + 'layer0': 'cd', + 'layer1': 'ad', + 'layer2': 'rd', + 'layer3': 'cv', + 'layer4': 'cd', + 'layer5': 'ad', + 'layer6': 'rd', + 'layer7': 'cv', + 'layer8': 'cd', + 'layer9': 'ad', + 'layer10': 'rd', + 'layer11': 'cv', + 'layer12': 'cd', + 'layer13': 'ad', + 'layer14': 'rd', + 'layer15': 'cv', + }, + } + +def createConvFunc(op_type): + assert op_type in ['cv', 'cd', 'ad', 'rd'], 'unknown op type: %s' % str(op_type) + if op_type == 'cv': + return F.conv2d + + if op_type == 'cd': + def func(x, weights, bias=None, stride=1, padding=0, dilation=1, groups=1): + assert dilation in [1, 2], 'dilation for cd_conv should be in 1 or 2' + assert weights.size(2) == 3 and weights.size(3) == 3, 'kernel size for cd_conv should be 3x3' + assert padding == dilation, 'padding for cd_conv set wrong' + + weights_c = weights.sum(dim=[2, 3], keepdim=True) + yc = F.conv2d(x, weights_c, stride=stride, padding=0, groups=groups) + y = F.conv2d(x, weights, bias, stride=stride, padding=padding, dilation=dilation, groups=groups) + return y - yc + return func + elif op_type == 'ad': + def func(x, weights, bias=None, stride=1, padding=0, dilation=1, groups=1): + assert dilation in [1, 2], 'dilation for ad_conv should be in 1 or 2' + assert weights.size(2) == 3 and weights.size(3) == 3, 'kernel size for ad_conv should be 3x3' + assert padding == dilation, 'padding for ad_conv set wrong' + + shape = weights.shape + weights = weights.view(shape[0], shape[1], -1) + weights_conv = (weights - weights[:, :, [3, 0, 1, 6, 4, 2, 7, 8, 5]]).view(shape) # clock-wise + y = F.conv2d(x, weights_conv, bias, stride=stride, padding=padding, dilation=dilation, groups=groups) + return y + return func + elif op_type == 'rd': + def func(x, weights, bias=None, stride=1, padding=0, dilation=1, groups=1): + assert dilation in [1, 2], 'dilation for rd_conv should be in 1 or 2' + assert weights.size(2) == 3 and weights.size(3) == 3, 'kernel size for rd_conv should be 3x3' + padding = 2 * dilation + + shape = weights.shape + if weights.is_cuda: + buffer = torch.cuda.FloatTensor(shape[0], shape[1], 5 * 5).fill_(0).to(DEVICE) + else: + buffer = torch.zeros(shape[0], shape[1], 5 * 5).to(DEVICE) + weights = weights.view(shape[0], shape[1], -1) + buffer[:, :, [0, 2, 4, 10, 14, 20, 22, 24]] = weights[:, :, 1:] + buffer[:, :, [6, 7, 8, 11, 13, 16, 17, 18]] = -weights[:, :, 1:] + buffer[:, :, 12] = 0 + buffer = buffer.view(shape[0], shape[1], 5, 5) + y = F.conv2d(x, buffer, bias, stride=stride, padding=padding, dilation=dilation, groups=groups) + return y + return func + else: + print('impossible to be here unless you force that') + return None + +class Conv2d(nn.Module): + def __init__(self, pdc, in_channels, out_channels, kernel_size, stride=1, padding=0, dilation=1, groups=1, bias=False): + super(Conv2d, self).__init__() + if in_channels % groups != 0: + raise ValueError('in_channels must be divisible by groups') + if out_channels % groups != 0: + raise ValueError('out_channels must be divisible by groups') + self.in_channels = in_channels + self.out_channels = out_channels + self.kernel_size = kernel_size + self.stride = stride + self.padding = padding + self.dilation = dilation + self.groups = groups + self.weight = nn.Parameter(torch.Tensor(out_channels, in_channels // groups, kernel_size, kernel_size)) + if bias: + self.bias = nn.Parameter(torch.Tensor(out_channels)) + else: + self.register_parameter('bias', None) + self.reset_parameters() + self.pdc = pdc + + def reset_parameters(self): + nn.init.kaiming_uniform_(self.weight, a=math.sqrt(5)) + if self.bias is not None: + fan_in, _ = nn.init._calculate_fan_in_and_fan_out(self.weight) + bound = 1 / math.sqrt(fan_in) + nn.init.uniform_(self.bias, -bound, bound) + + def forward(self, input): + + return self.pdc(input, self.weight, self.bias, self.stride, self.padding, self.dilation, self.groups) + +class CSAM(nn.Module): + """ + Compact Spatial Attention Module + """ + def __init__(self, channels): + super(CSAM, self).__init__() + + mid_channels = 4 + self.relu1 = nn.ReLU() + self.conv1 = nn.Conv2d(channels, mid_channels, kernel_size=1, padding=0) + self.conv2 = nn.Conv2d(mid_channels, 1, kernel_size=3, padding=1, bias=False) + self.sigmoid = nn.Sigmoid() + nn.init.constant_(self.conv1.bias, 0) + + def forward(self, x): + y = self.relu1(x) + y = self.conv1(y) + y = self.conv2(y) + y = self.sigmoid(y) + + return x * y + +class CDCM(nn.Module): + """ + Compact Dilation Convolution based Module + """ + def __init__(self, in_channels, out_channels): + super(CDCM, self).__init__() + + self.relu1 = nn.ReLU() + self.conv1 = nn.Conv2d(in_channels, out_channels, kernel_size=1, padding=0) + self.conv2_1 = nn.Conv2d(out_channels, out_channels, kernel_size=3, dilation=5, padding=5, bias=False) + self.conv2_2 = nn.Conv2d(out_channels, out_channels, kernel_size=3, dilation=7, padding=7, bias=False) + self.conv2_3 = nn.Conv2d(out_channels, out_channels, kernel_size=3, dilation=9, padding=9, bias=False) + self.conv2_4 = nn.Conv2d(out_channels, out_channels, kernel_size=3, dilation=11, padding=11, bias=False) + nn.init.constant_(self.conv1.bias, 0) + + def forward(self, x): + x = self.relu1(x) + x = self.conv1(x) + x1 = self.conv2_1(x) + x2 = self.conv2_2(x) + x3 = self.conv2_3(x) + x4 = self.conv2_4(x) + return x1 + x2 + x3 + x4 + + +class MapReduce(nn.Module): + """ + Reduce feature maps into a single edge map + """ + def __init__(self, channels): + super(MapReduce, self).__init__() + self.conv = nn.Conv2d(channels, 1, kernel_size=1, padding=0) + nn.init.constant_(self.conv.bias, 0) + + def forward(self, x): + return self.conv(x) + + +class PDCBlock(nn.Module): + def __init__(self, pdc, inplane, ouplane, stride=1): + super(PDCBlock, self).__init__() + self.stride=stride + + self.stride=stride + if self.stride > 1: + self.pool = nn.MaxPool2d(kernel_size=2, stride=2) + self.shortcut = nn.Conv2d(inplane, ouplane, kernel_size=1, padding=0) + self.conv1 = Conv2d(pdc, inplane, inplane, kernel_size=3, padding=1, groups=inplane, bias=False) + self.relu2 = nn.ReLU() + self.conv2 = nn.Conv2d(inplane, ouplane, kernel_size=1, padding=0, bias=False) + + def forward(self, x): + if self.stride > 1: + x = self.pool(x) + y = self.conv1(x) + y = self.relu2(y) + y = self.conv2(y) + if self.stride > 1: + x = self.shortcut(x) + y = y + x + return y + +class PDCBlock_converted(nn.Module): + """ + CPDC, APDC can be converted to vanilla 3x3 convolution + RPDC can be converted to vanilla 5x5 convolution + """ + def __init__(self, pdc, inplane, ouplane, stride=1): + super(PDCBlock_converted, self).__init__() + self.stride=stride + + if self.stride > 1: + self.pool = nn.MaxPool2d(kernel_size=2, stride=2) + self.shortcut = nn.Conv2d(inplane, ouplane, kernel_size=1, padding=0) + if pdc == 'rd': + self.conv1 = nn.Conv2d(inplane, inplane, kernel_size=5, padding=2, groups=inplane, bias=False) + else: + self.conv1 = nn.Conv2d(inplane, inplane, kernel_size=3, padding=1, groups=inplane, bias=False) + self.relu2 = nn.ReLU() + self.conv2 = nn.Conv2d(inplane, ouplane, kernel_size=1, padding=0, bias=False) + + def forward(self, x): + if self.stride > 1: + x = self.pool(x) + y = self.conv1(x) + y = self.relu2(y) + y = self.conv2(y) + if self.stride > 1: + x = self.shortcut(x) + y = y + x + return y + +class PiDiNet(nn.Module): + def __init__(self, inplane, pdcs, dil=None, sa=False, convert=False): + super(PiDiNet, self).__init__() + self.sa = sa + if dil is not None: + assert isinstance(dil, int), 'dil should be an int' + self.dil = dil + + self.fuseplanes = [] + + self.inplane = inplane + if convert: + if pdcs[0] == 'rd': + init_kernel_size = 5 + init_padding = 2 + else: + init_kernel_size = 3 + init_padding = 1 + self.init_block = nn.Conv2d(3, self.inplane, + kernel_size=init_kernel_size, padding=init_padding, bias=False) + block_class = PDCBlock_converted + else: + self.init_block = Conv2d(pdcs[0], 3, self.inplane, kernel_size=3, padding=1) + block_class = PDCBlock + + self.block1_1 = block_class(pdcs[1], self.inplane, self.inplane) + self.block1_2 = block_class(pdcs[2], self.inplane, self.inplane) + self.block1_3 = block_class(pdcs[3], self.inplane, self.inplane) + self.fuseplanes.append(self.inplane) # C + + inplane = self.inplane + self.inplane = self.inplane * 2 + self.block2_1 = block_class(pdcs[4], inplane, self.inplane, stride=2) + self.block2_2 = block_class(pdcs[5], self.inplane, self.inplane) + self.block2_3 = block_class(pdcs[6], self.inplane, self.inplane) + self.block2_4 = block_class(pdcs[7], self.inplane, self.inplane) + self.fuseplanes.append(self.inplane) # 2C + + inplane = self.inplane + self.inplane = self.inplane * 2 + self.block3_1 = block_class(pdcs[8], inplane, self.inplane, stride=2) + self.block3_2 = block_class(pdcs[9], self.inplane, self.inplane) + self.block3_3 = block_class(pdcs[10], self.inplane, self.inplane) + self.block3_4 = block_class(pdcs[11], self.inplane, self.inplane) + self.fuseplanes.append(self.inplane) # 4C + + self.block4_1 = block_class(pdcs[12], self.inplane, self.inplane, stride=2) + self.block4_2 = block_class(pdcs[13], self.inplane, self.inplane) + self.block4_3 = block_class(pdcs[14], self.inplane, self.inplane) + self.block4_4 = block_class(pdcs[15], self.inplane, self.inplane) + self.fuseplanes.append(self.inplane) # 4C + + self.conv_reduces = nn.ModuleList() + if self.sa and self.dil is not None: + self.attentions = nn.ModuleList() + self.dilations = nn.ModuleList() + for i in range(4): + self.dilations.append(CDCM(self.fuseplanes[i], self.dil)) + self.attentions.append(CSAM(self.dil)) + self.conv_reduces.append(MapReduce(self.dil)) + elif self.sa: + self.attentions = nn.ModuleList() + for i in range(4): + self.attentions.append(CSAM(self.fuseplanes[i])) + self.conv_reduces.append(MapReduce(self.fuseplanes[i])) + elif self.dil is not None: + self.dilations = nn.ModuleList() + for i in range(4): + self.dilations.append(CDCM(self.fuseplanes[i], self.dil)) + self.conv_reduces.append(MapReduce(self.dil)) + else: + for i in range(4): + self.conv_reduces.append(MapReduce(self.fuseplanes[i])) + + self.classifier = nn.Conv2d(4, 1, kernel_size=1) # has bias + nn.init.constant_(self.classifier.weight, 0.25) + nn.init.constant_(self.classifier.bias, 0) + + # print('initialization done') + + def get_weights(self): + conv_weights = [] + bn_weights = [] + relu_weights = [] + for pname, p in self.named_parameters(): + if 'bn' in pname: + bn_weights.append(p) + elif 'relu' in pname: + relu_weights.append(p) + else: + conv_weights.append(p) + + return conv_weights, bn_weights, relu_weights + + def forward(self, x): + H, W = x.size()[2:] + + x = self.init_block(x) + + x1 = self.block1_1(x) + x1 = self.block1_2(x1) + x1 = self.block1_3(x1) + + x2 = self.block2_1(x1) + x2 = self.block2_2(x2) + x2 = self.block2_3(x2) + x2 = self.block2_4(x2) + + x3 = self.block3_1(x2) + x3 = self.block3_2(x3) + x3 = self.block3_3(x3) + x3 = self.block3_4(x3) + + x4 = self.block4_1(x3) + x4 = self.block4_2(x4) + x4 = self.block4_3(x4) + x4 = self.block4_4(x4) + + x_fuses = [] + if self.sa and self.dil is not None: + for i, xi in enumerate([x1, x2, x3, x4]): + x_fuses.append(self.attentions[i](self.dilations[i](xi))) + elif self.sa: + for i, xi in enumerate([x1, x2, x3, x4]): + x_fuses.append(self.attentions[i](xi)) + elif self.dil is not None: + for i, xi in enumerate([x1, x2, x3, x4]): + x_fuses.append(self.dilations[i](xi)) + else: + x_fuses = [x1, x2, x3, x4] + + e1 = self.conv_reduces[0](x_fuses[0]) + e1 = F.interpolate(e1, (H, W), mode="bilinear", align_corners=False) + + e2 = self.conv_reduces[1](x_fuses[1]) + e2 = F.interpolate(e2, (H, W), mode="bilinear", align_corners=False) + + e3 = self.conv_reduces[2](x_fuses[2]) + e3 = F.interpolate(e3, (H, W), mode="bilinear", align_corners=False) + + e4 = self.conv_reduces[3](x_fuses[3]) + e4 = F.interpolate(e4, (H, W), mode="bilinear", align_corners=False) + + outputs = [e1, e2, e3, e4] + + output = self.classifier(torch.cat(outputs, dim=1)) + #if not self.training: + # return torch.sigmoid(output) + + outputs.append(output) + outputs = [torch.sigmoid(r) for r in outputs] + return outputs + +def config_model(model): + model_options = list(nets.keys()) + assert model in model_options, \ + 'unrecognized model, please choose from %s' % str(model_options) + + # print(str(nets[model])) + + pdcs = [] + for i in range(16): + layer_name = 'layer%d' % i + op = nets[model][layer_name] + pdcs.append(createConvFunc(op)) + + return pdcs + +def pidinet(): + pdcs = config_model('carv4') + dil = 24 #if args.dil else None + return PiDiNet(60, pdcs, dil=dil, sa=True) + + +if __name__ == '__main__': + model = pidinet() + ckp = torch.load('table5_pidinet.pth')['state_dict'] + model.load_state_dict({k.replace('module.',''):v for k, v in ckp.items()}) + im = cv2.imread('examples/test_my/cat_v4.png') + im = img2tensor(im).unsqueeze(0)/255. + res = model(im)[-1] + res = res>0.5 + res = res.float() + res = (res[0,0].cpu().data.numpy()*255.).astype(np.uint8) + print(res.shape) + cv2.imwrite('edge.png', res) \ No newline at end of file diff --git a/RAVE-main/annotator/uniformer/configs/_base_/models/emanet_r50-d8.py b/RAVE-main/annotator/uniformer/configs/_base_/models/emanet_r50-d8.py new file mode 100644 index 0000000000000000000000000000000000000000..26adcd430926de0862204a71d345f2543167f27b --- /dev/null +++ b/RAVE-main/annotator/uniformer/configs/_base_/models/emanet_r50-d8.py @@ -0,0 +1,47 @@ +# model settings +norm_cfg = dict(type='SyncBN', requires_grad=True) +model = dict( + type='EncoderDecoder', + pretrained='open-mmlab://resnet50_v1c', + backbone=dict( + type='ResNetV1c', + depth=50, + num_stages=4, + out_indices=(0, 1, 2, 3), + dilations=(1, 1, 2, 4), + strides=(1, 2, 1, 1), + norm_cfg=norm_cfg, + norm_eval=False, + style='pytorch', + contract_dilation=True), + decode_head=dict( + type='EMAHead', + in_channels=2048, + in_index=3, + channels=256, + ema_channels=512, + num_bases=64, + num_stages=3, + momentum=0.1, + dropout_ratio=0.1, + num_classes=19, + norm_cfg=norm_cfg, + align_corners=False, + loss_decode=dict( + type='CrossEntropyLoss', use_sigmoid=False, loss_weight=1.0)), + auxiliary_head=dict( + type='FCNHead', + in_channels=1024, + in_index=2, + channels=256, + num_convs=1, + concat_input=False, + dropout_ratio=0.1, + num_classes=19, + norm_cfg=norm_cfg, + align_corners=False, + loss_decode=dict( + type='CrossEntropyLoss', use_sigmoid=False, loss_weight=0.4)), + # model training and testing settings + train_cfg=dict(), + test_cfg=dict(mode='whole')) diff --git a/RAVE-main/annotator/uniformer/configs/_base_/models/fast_scnn.py b/RAVE-main/annotator/uniformer/configs/_base_/models/fast_scnn.py new file mode 100644 index 0000000000000000000000000000000000000000..32fdeb659355a5ce5ef2cc7c2f30742703811cdf --- /dev/null +++ b/RAVE-main/annotator/uniformer/configs/_base_/models/fast_scnn.py @@ -0,0 +1,57 @@ +# model settings +norm_cfg = dict(type='SyncBN', requires_grad=True, momentum=0.01) +model = dict( + type='EncoderDecoder', + backbone=dict( + type='FastSCNN', + downsample_dw_channels=(32, 48), + global_in_channels=64, + global_block_channels=(64, 96, 128), + global_block_strides=(2, 2, 1), + global_out_channels=128, + higher_in_channels=64, + lower_in_channels=128, + fusion_out_channels=128, + out_indices=(0, 1, 2), + norm_cfg=norm_cfg, + align_corners=False), + decode_head=dict( + type='DepthwiseSeparableFCNHead', + in_channels=128, + channels=128, + concat_input=False, + num_classes=19, + in_index=-1, + norm_cfg=norm_cfg, + align_corners=False, + loss_decode=dict( + type='CrossEntropyLoss', use_sigmoid=True, loss_weight=0.4)), + auxiliary_head=[ + dict( + type='FCNHead', + in_channels=128, + channels=32, + num_convs=1, + num_classes=19, + in_index=-2, + norm_cfg=norm_cfg, + concat_input=False, + align_corners=False, + loss_decode=dict( + type='CrossEntropyLoss', use_sigmoid=True, loss_weight=0.4)), + dict( + type='FCNHead', + in_channels=64, + channels=32, + num_convs=1, + num_classes=19, + in_index=-3, + norm_cfg=norm_cfg, + concat_input=False, + align_corners=False, + loss_decode=dict( + type='CrossEntropyLoss', use_sigmoid=True, loss_weight=0.4)), + ], + # model training and testing settings + train_cfg=dict(), + test_cfg=dict(mode='whole')) diff --git a/RAVE-main/annotator/uniformer/configs/_base_/models/fcn_unet_s5-d16.py b/RAVE-main/annotator/uniformer/configs/_base_/models/fcn_unet_s5-d16.py new file mode 100644 index 0000000000000000000000000000000000000000..a33e7972877f902d0e7d18401ca675e3e4e60a18 --- /dev/null +++ b/RAVE-main/annotator/uniformer/configs/_base_/models/fcn_unet_s5-d16.py @@ -0,0 +1,51 @@ +# model settings +norm_cfg = dict(type='SyncBN', requires_grad=True) +model = dict( + type='EncoderDecoder', + pretrained=None, + backbone=dict( + type='UNet', + in_channels=3, + base_channels=64, + num_stages=5, + strides=(1, 1, 1, 1, 1), + enc_num_convs=(2, 2, 2, 2, 2), + dec_num_convs=(2, 2, 2, 2), + downsamples=(True, True, True, True), + enc_dilations=(1, 1, 1, 1, 1), + dec_dilations=(1, 1, 1, 1), + with_cp=False, + conv_cfg=None, + norm_cfg=norm_cfg, + act_cfg=dict(type='ReLU'), + upsample_cfg=dict(type='InterpConv'), + norm_eval=False), + decode_head=dict( + type='FCNHead', + in_channels=64, + in_index=4, + channels=64, + num_convs=1, + concat_input=False, + dropout_ratio=0.1, + num_classes=2, + norm_cfg=norm_cfg, + align_corners=False, + loss_decode=dict( + type='CrossEntropyLoss', use_sigmoid=False, loss_weight=1.0)), + auxiliary_head=dict( + type='FCNHead', + in_channels=128, + in_index=3, + channels=64, + num_convs=1, + concat_input=False, + dropout_ratio=0.1, + num_classes=2, + norm_cfg=norm_cfg, + align_corners=False, + loss_decode=dict( + type='CrossEntropyLoss', use_sigmoid=False, loss_weight=0.4)), + # model training and testing settings + train_cfg=dict(), + test_cfg=dict(mode='slide', crop_size=256, stride=170)) diff --git a/RAVE-main/annotator/uniformer/mmcv_custom/__init__.py b/RAVE-main/annotator/uniformer/mmcv_custom/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..4b958738b9fd93bfcec239c550df1d9a44b8c536 --- /dev/null +++ b/RAVE-main/annotator/uniformer/mmcv_custom/__init__.py @@ -0,0 +1,5 @@ +# -*- coding: utf-8 -*- + +from .checkpoint import load_checkpoint + +__all__ = ['load_checkpoint'] \ No newline at end of file diff --git a/RAVE-main/annotator/uniformer/mmcv_custom/checkpoint.py b/RAVE-main/annotator/uniformer/mmcv_custom/checkpoint.py new file mode 100644 index 0000000000000000000000000000000000000000..48c1b16b53107cb1301edf6cc07ccfe6f7010da6 --- /dev/null +++ b/RAVE-main/annotator/uniformer/mmcv_custom/checkpoint.py @@ -0,0 +1,508 @@ +# Copyright (c) Open-MMLab. All rights reserved. +import io +import os +import os.path as osp +import pkgutil +import time +import warnings +from collections import OrderedDict +from importlib import import_module +from tempfile import TemporaryDirectory + +import torch +import torchvision +from torch.optim import Optimizer +from torch.utils import model_zoo +from torch.nn import functional as F + +try: + import mmcv as mmcv + from mmcv.fileio import FileClient + from mmcv.fileio import load as load_file + from mmcv.parallel import is_module_wrapper + from mmcv.utils import mkdir_or_exist + from mmcv.runner import get_dist_info +except ImportError: + import annotator.mmpkg.mmcv as mmcv + from annotator.mmpkg.mmcv.fileio import FileClient + from annotator.mmpkg.mmcv.fileio import load as load_file + from annotator.mmpkg.mmcv.parallel import is_module_wrapper + from annotator.mmpkg.mmcv.utils import mkdir_or_exist + from annotator.mmpkg.mmcv.runner import get_dist_info + +ENV_MMCV_HOME = 'MMCV_HOME' +ENV_XDG_CACHE_HOME = 'XDG_CACHE_HOME' +DEFAULT_CACHE_DIR = '~/.cache' + + +def _get_mmcv_home(): + mmcv_home = os.path.expanduser( + os.getenv( + ENV_MMCV_HOME, + os.path.join( + os.getenv(ENV_XDG_CACHE_HOME, DEFAULT_CACHE_DIR), 'mmcv'))) + + mkdir_or_exist(mmcv_home) + return mmcv_home + + +def load_state_dict(module, state_dict, strict=False, logger=None): + """Load state_dict to a module. + + This method is modified from :meth:`torch.nn.Module.load_state_dict`. + Default value for ``strict`` is set to ``False`` and the message for + param mismatch will be shown even if strict is False. + + Args: + module (Module): Module that receives the state_dict. + state_dict (OrderedDict): Weights. + strict (bool): whether to strictly enforce that the keys + in :attr:`state_dict` match the keys returned by this module's + :meth:`~torch.nn.Module.state_dict` function. Default: ``False``. + logger (:obj:`logging.Logger`, optional): Logger to log the error + message. If not specified, print function will be used. + """ + unexpected_keys = [] + all_missing_keys = [] + err_msg = [] + + metadata = getattr(state_dict, '_metadata', None) + state_dict = state_dict.copy() + if metadata is not None: + state_dict._metadata = metadata + + # use _load_from_state_dict to enable checkpoint version control + def load(module, prefix=''): + # recursively check parallel module in case that the model has a + # complicated structure, e.g., nn.Module(nn.Module(DDP)) + if is_module_wrapper(module): + module = module.module + local_metadata = {} if metadata is None else metadata.get( + prefix[:-1], {}) + module._load_from_state_dict(state_dict, prefix, local_metadata, True, + all_missing_keys, unexpected_keys, + err_msg) + for name, child in module._modules.items(): + if child is not None: + load(child, prefix + name + '.') + + load(module) + load = None # break load->load reference cycle + + # ignore "num_batches_tracked" of BN layers + missing_keys = [ + key for key in all_missing_keys if 'num_batches_tracked' not in key + ] + + if unexpected_keys: + err_msg.append('unexpected key in source ' + f'state_dict: {", ".join(unexpected_keys)}\n') + if missing_keys: + err_msg.append( + f'missing keys in source state_dict: {", ".join(missing_keys)}\n') + + rank, _ = get_dist_info() + if len(err_msg) > 0 and rank == 0: + err_msg.insert( + 0, 'The model and loaded state dict do not match exactly\n') + err_msg = '\n'.join(err_msg) + if strict: + raise RuntimeError(err_msg) + elif logger is not None: + logger.warning(err_msg) + else: + print(err_msg) + + +def load_url_dist(url, model_dir=None): + """In distributed setting, this function only download checkpoint at local + rank 0.""" + rank, world_size = get_dist_info() + rank = int(os.environ.get('LOCAL_RANK', rank)) + if rank == 0: + checkpoint = model_zoo.load_url(url, model_dir=model_dir) + if world_size > 1: + torch.distributed.barrier() + if rank > 0: + checkpoint = model_zoo.load_url(url, model_dir=model_dir) + return checkpoint + + +def load_pavimodel_dist(model_path, map_location=None): + """In distributed setting, this function only download checkpoint at local + rank 0.""" + try: + from pavi import modelcloud + except ImportError: + raise ImportError( + 'Please install pavi to load checkpoint from modelcloud.') + rank, world_size = get_dist_info() + rank = int(os.environ.get('LOCAL_RANK', rank)) + if rank == 0: + model = modelcloud.get(model_path) + with TemporaryDirectory() as tmp_dir: + downloaded_file = osp.join(tmp_dir, model.name) + model.download(downloaded_file) + checkpoint = torch.load(downloaded_file, map_location=map_location) + if world_size > 1: + torch.distributed.barrier() + if rank > 0: + model = modelcloud.get(model_path) + with TemporaryDirectory() as tmp_dir: + downloaded_file = osp.join(tmp_dir, model.name) + model.download(downloaded_file) + checkpoint = torch.load( + downloaded_file, map_location=map_location) + return checkpoint + + +def load_fileclient_dist(filename, backend, map_location): + """In distributed setting, this function only download checkpoint at local + rank 0.""" + rank, world_size = get_dist_info() + rank = int(os.environ.get('LOCAL_RANK', rank)) + allowed_backends = ['ceph'] + if backend not in allowed_backends: + raise ValueError(f'Load from Backend {backend} is not supported.') + if rank == 0: + fileclient = FileClient(backend=backend) + buffer = io.BytesIO(fileclient.get(filename)) + checkpoint = torch.load(buffer, map_location=map_location) + if world_size > 1: + torch.distributed.barrier() + if rank > 0: + fileclient = FileClient(backend=backend) + buffer = io.BytesIO(fileclient.get(filename)) + checkpoint = torch.load(buffer, map_location=map_location) + return checkpoint + + +def get_torchvision_models(): + model_urls = dict() + for _, name, ispkg in pkgutil.walk_packages(torchvision.models.__path__): + if ispkg: + continue + _zoo = import_module(f'torchvision.models.{name}') + if hasattr(_zoo, 'model_urls'): + _urls = getattr(_zoo, 'model_urls') + model_urls.update(_urls) + return model_urls + + +def get_external_models(): + mmcv_home = _get_mmcv_home() + default_json_path = osp.join(mmcv.__path__[0], 'model_zoo/open_mmlab.json') + default_urls = load_file(default_json_path) + assert isinstance(default_urls, dict) + external_json_path = osp.join(mmcv_home, 'open_mmlab.json') + if osp.exists(external_json_path): + external_urls = load_file(external_json_path) + assert isinstance(external_urls, dict) + default_urls.update(external_urls) + + return default_urls + + +def get_mmcls_models(): + mmcls_json_path = osp.join(mmcv.__path__[0], 'model_zoo/mmcls.json') + mmcls_urls = load_file(mmcls_json_path) + + return mmcls_urls + + +def get_deprecated_model_names(): + deprecate_json_path = osp.join(mmcv.__path__[0], + 'model_zoo/deprecated.json') + deprecate_urls = load_file(deprecate_json_path) + assert isinstance(deprecate_urls, dict) + + return deprecate_urls + + +def _process_mmcls_checkpoint(checkpoint): + state_dict = checkpoint['state_dict'] + new_state_dict = OrderedDict() + for k, v in state_dict.items(): + if k.startswith('backbone.'): + new_state_dict[k[9:]] = v + new_checkpoint = dict(state_dict=new_state_dict) + + return new_checkpoint + + +def _load_checkpoint(filename, map_location=None): + """Load checkpoint from somewhere (modelzoo, file, url). + + Args: + filename (str): Accept local filepath, URL, ``torchvision://xxx``, + ``open-mmlab://xxx``. Please refer to ``docs/model_zoo.md`` for + details. + map_location (str | None): Same as :func:`torch.load`. Default: None. + + Returns: + dict | OrderedDict: The loaded checkpoint. It can be either an + OrderedDict storing model weights or a dict containing other + information, which depends on the checkpoint. + """ + if filename.startswith('modelzoo://'): + warnings.warn('The URL scheme of "modelzoo://" is deprecated, please ' + 'use "torchvision://" instead') + model_urls = get_torchvision_models() + model_name = filename[11:] + checkpoint = load_url_dist(model_urls[model_name]) + elif filename.startswith('torchvision://'): + model_urls = get_torchvision_models() + model_name = filename[14:] + checkpoint = load_url_dist(model_urls[model_name]) + elif filename.startswith('open-mmlab://'): + model_urls = get_external_models() + model_name = filename[13:] + deprecated_urls = get_deprecated_model_names() + if model_name in deprecated_urls: + warnings.warn(f'open-mmlab://{model_name} is deprecated in favor ' + f'of open-mmlab://{deprecated_urls[model_name]}') + model_name = deprecated_urls[model_name] + model_url = model_urls[model_name] + # check if is url + if model_url.startswith(('http://', 'https://')): + checkpoint = load_url_dist(model_url) + else: + filename = osp.join(_get_mmcv_home(), model_url) + if not osp.isfile(filename): + raise IOError(f'{filename} is not a checkpoint file') + checkpoint = torch.load(filename, map_location=map_location) + elif filename.startswith('mmcls://'): + model_urls = get_mmcls_models() + model_name = filename[8:] + checkpoint = load_url_dist(model_urls[model_name]) + checkpoint = _process_mmcls_checkpoint(checkpoint) + elif filename.startswith(('http://', 'https://')): + checkpoint = load_url_dist(filename) + elif filename.startswith('pavi://'): + model_path = filename[7:] + checkpoint = load_pavimodel_dist(model_path, map_location=map_location) + elif filename.startswith('s3://'): + checkpoint = load_fileclient_dist( + filename, backend='ceph', map_location=map_location) + else: + if not osp.isfile(filename): + raise IOError(f'{filename} is not a checkpoint file') + checkpoint = torch.load(filename, map_location=map_location) + return checkpoint + + +def load_checkpoint(model, + filename, + map_location='cpu', + strict=False, + logger=None): + """Load checkpoint from a file or URI. + + Args: + model (Module): Module to load checkpoint. + filename (str): Accept local filepath, URL, ``torchvision://xxx``, + ``open-mmlab://xxx``. Please refer to ``docs/model_zoo.md`` for + details. + map_location (str): Same as :func:`torch.load`. + strict (bool): Whether to allow different params for the model and + checkpoint. + logger (:mod:`logging.Logger` or None): The logger for error message. + + Returns: + dict or OrderedDict: The loaded checkpoint. + """ + checkpoint = _load_checkpoint(filename, map_location) + # OrderedDict is a subclass of dict + if not isinstance(checkpoint, dict): + raise RuntimeError( + f'No state_dict found in checkpoint file {filename}') + # get state_dict from checkpoint + if 'state_dict' in checkpoint: + state_dict = checkpoint['state_dict'] + elif 'model' in checkpoint: + state_dict = checkpoint['model'] + else: + state_dict = checkpoint + # strip prefix of state_dict + if list(state_dict.keys())[0].startswith('module.'): + state_dict = {k[7:]: v for k, v in state_dict.items()} + + # for MoBY, load model of online branch + if sorted(list(state_dict.keys()))[0].startswith('encoder'): + state_dict = {k.replace('encoder.', ''): v for k, v in state_dict.items() if k.startswith('encoder.')} + + # reshape absolute position embedding + if state_dict.get('absolute_pos_embed') is not None: + absolute_pos_embed = state_dict['absolute_pos_embed'] + N1, L, C1 = absolute_pos_embed.size() + N2, C2, H, W = model.absolute_pos_embed.size() + if N1 != N2 or C1 != C2 or L != H*W: + logger.warning("Error in loading absolute_pos_embed, pass") + else: + state_dict['absolute_pos_embed'] = absolute_pos_embed.view(N2, H, W, C2).permute(0, 3, 1, 2) + + # interpolate position bias table if needed + relative_position_bias_table_keys = [k for k in state_dict.keys() if "relative_position_bias_table" in k] + for table_key in relative_position_bias_table_keys: + table_pretrained = state_dict[table_key] + table_current = model.state_dict()[table_key] + L1, nH1 = table_pretrained.size() + L2, nH2 = table_current.size() + if nH1 != nH2: + logger.warning(f"Error in loading {table_key}, pass") + else: + if L1 != L2: + S1 = int(L1 ** 0.5) + S2 = int(L2 ** 0.5) + table_pretrained_resized = F.interpolate( + table_pretrained.permute(1, 0).view(1, nH1, S1, S1), + size=(S2, S2), mode='bicubic') + state_dict[table_key] = table_pretrained_resized.view(nH2, L2).permute(1, 0) + + # load state_dict + load_state_dict(model, state_dict, strict, logger) + return checkpoint + + +def weights_to_cpu(state_dict): + """Copy a model state_dict to cpu. + + Args: + state_dict (OrderedDict): Model weights on GPU. + + Returns: + OrderedDict: Model weights on GPU. + """ + state_dict_cpu = OrderedDict() + for key, val in state_dict.items(): + state_dict_cpu[key] = val.cpu() + return state_dict_cpu + + +def _save_to_state_dict(module, destination, prefix, keep_vars): + """Saves module state to `destination` dictionary. + + This method is modified from :meth:`torch.nn.Module._save_to_state_dict`. + + Args: + module (nn.Module): The module to generate state_dict. + destination (dict): A dict where state will be stored. + prefix (str): The prefix for parameters and buffers used in this + module. + """ + for name, param in module._parameters.items(): + if param is not None: + destination[prefix + name] = param if keep_vars else param.detach() + for name, buf in module._buffers.items(): + # remove check of _non_persistent_buffers_set to allow nn.BatchNorm2d + if buf is not None: + destination[prefix + name] = buf if keep_vars else buf.detach() + + +def get_state_dict(module, destination=None, prefix='', keep_vars=False): + """Returns a dictionary containing a whole state of the module. + + Both parameters and persistent buffers (e.g. running averages) are + included. Keys are corresponding parameter and buffer names. + + This method is modified from :meth:`torch.nn.Module.state_dict` to + recursively check parallel module in case that the model has a complicated + structure, e.g., nn.Module(nn.Module(DDP)). + + Args: + module (nn.Module): The module to generate state_dict. + destination (OrderedDict): Returned dict for the state of the + module. + prefix (str): Prefix of the key. + keep_vars (bool): Whether to keep the variable property of the + parameters. Default: False. + + Returns: + dict: A dictionary containing a whole state of the module. + """ + # recursively check parallel module in case that the model has a + # complicated structure, e.g., nn.Module(nn.Module(DDP)) + if is_module_wrapper(module): + module = module.module + + # below is the same as torch.nn.Module.state_dict() + if destination is None: + destination = OrderedDict() + destination._metadata = OrderedDict() + destination._metadata[prefix[:-1]] = local_metadata = dict( + version=module._version) + _save_to_state_dict(module, destination, prefix, keep_vars) + for name, child in module._modules.items(): + if child is not None: + get_state_dict( + child, destination, prefix + name + '.', keep_vars=keep_vars) + for hook in module._state_dict_hooks.values(): + hook_result = hook(module, destination, prefix, local_metadata) + if hook_result is not None: + destination = hook_result + return destination + + +def save_checkpoint(model, filename, optimizer=None, meta=None): + """Save checkpoint to file. + + The checkpoint will have 3 fields: ``meta``, ``state_dict`` and + ``optimizer``. By default ``meta`` will contain version and time info. + + Args: + model (Module): Module whose params are to be saved. + filename (str): Checkpoint filename. + optimizer (:obj:`Optimizer`, optional): Optimizer to be saved. + meta (dict, optional): Metadata to be saved in checkpoint. + """ + if meta is None: + meta = {} + elif not isinstance(meta, dict): + raise TypeError(f'meta must be a dict or None, but got {type(meta)}') + meta.update(mmcv_version=mmcv.__version__, time=time.asctime()) + + if is_module_wrapper(model): + model = model.module + + if hasattr(model, 'CLASSES') and model.CLASSES is not None: + # save class name to the meta + meta.update(CLASSES=model.CLASSES) + + checkpoint = { + 'meta': meta, + 'state_dict': weights_to_cpu(get_state_dict(model)) + } + # save optimizer state dict in the checkpoint + if isinstance(optimizer, Optimizer): + checkpoint['optimizer'] = optimizer.state_dict() + elif isinstance(optimizer, dict): + checkpoint['optimizer'] = {} + for name, optim in optimizer.items(): + checkpoint['optimizer'][name] = optim.state_dict() + + if filename.startswith('pavi://'): + try: + from pavi import modelcloud + from pavi.exception import NodeNotFoundError + except ImportError: + raise ImportError( + 'Please install pavi to load checkpoint from modelcloud.') + model_path = filename[7:] + root = modelcloud.Folder() + model_dir, model_name = osp.split(model_path) + try: + model = modelcloud.get(model_dir) + except NodeNotFoundError: + model = root.create_training_model(model_dir) + with TemporaryDirectory() as tmp_dir: + checkpoint_file = osp.join(tmp_dir, model_name) + with open(checkpoint_file, 'wb') as f: + torch.save(checkpoint, f) + f.flush() + model.create_file(checkpoint_file, name=model_name) + else: + mmcv.mkdir_or_exist(osp.dirname(filename)) + # immediately flush buffer + with open(filename, 'wb') as f: + torch.save(checkpoint, f) + f.flush() \ No newline at end of file diff --git a/RAVE-main/pipelines/__pycache__/sd_controlnet_rave.cpython-38.pyc b/RAVE-main/pipelines/__pycache__/sd_controlnet_rave.cpython-38.pyc new file mode 100644 index 0000000000000000000000000000000000000000..56898e71cbd81fd545548a3743612e100df6f327 Binary files /dev/null and b/RAVE-main/pipelines/__pycache__/sd_controlnet_rave.cpython-38.pyc differ diff --git a/RAVE-main/pipelines/__pycache__/sd_multicontrolnet_rave.cpython-38.pyc b/RAVE-main/pipelines/__pycache__/sd_multicontrolnet_rave.cpython-38.pyc new file mode 100644 index 0000000000000000000000000000000000000000..80efa2369366037e8cce00f48d58aff9b3ce410d Binary files /dev/null and b/RAVE-main/pipelines/__pycache__/sd_multicontrolnet_rave.cpython-38.pyc differ diff --git a/RAVE-main/pipelines/sd_controlnet_rave.py b/RAVE-main/pipelines/sd_controlnet_rave.py new file mode 100644 index 0000000000000000000000000000000000000000..b5ce2e2227d1d3834fea4c543c1d708a01681ece --- /dev/null +++ b/RAVE-main/pipelines/sd_controlnet_rave.py @@ -0,0 +1,421 @@ +import random +import os +import PIL +import torch +import warnings +warnings.filterwarnings("ignore") + +from transformers import set_seed +from tqdm import tqdm +from transformers import logging +from diffusers import ControlNetModel, StableDiffusionControlNetImg2ImgPipeline, DDIMScheduler + +import torch.nn as nn +import numpy as np +import utils.feature_utils as fu +import utils.preprocesser_utils as pu +import utils.image_process_utils as ipu + + +logging.set_verbosity_error() + +def set_seed_lib(seed): + np.random.seed(seed) + torch.manual_seed(seed) + torch.cuda.manual_seed(seed) + random.seed(seed) + set_seed(seed) + +@torch.no_grad() +class RAVE(nn.Module): + def __init__(self, device): + super().__init__() + + self.device = device + self.dtype = torch.float + + @torch.no_grad() + def __init_pipe(self, hf_cn_path, hf_path): + controlnet = ControlNetModel.from_pretrained(hf_cn_path, torch_dtype=self.dtype).to(self.device, self.dtype) + + pipe = StableDiffusionControlNetImg2ImgPipeline.from_pretrained(hf_path, controlnet=controlnet, torch_dtype=self.dtype).to(self.device, self.dtype) + pipe.enable_model_cpu_offload() + pipe.enable_xformers_memory_efficient_attention() + return pipe + + @torch.no_grad() + def init_models(self, hf_cn_path, hf_path, preprocess_name, model_id=None): + if model_id is None or model_id == "None": + pipe = self.__init_pipe(hf_cn_path, hf_path) + else: + pipe = self.__init_pipe(hf_cn_path, model_id) + self.preprocess_name = preprocess_name + + + self._prepare_control_image = pipe.prepare_control_image + self.run_safety_checker = pipe.run_safety_checker + self.tokenizer = pipe.tokenizer + self.text_encoder = pipe.text_encoder + + self.vae = pipe.vae + self.unet = pipe.unet + + self.controlnet = pipe.controlnet + self.scheduler_config = pipe.scheduler.config + + del pipe + + @torch.no_grad() + def get_text_embeds(self, prompt, negative_prompt): + + cond_input = self.tokenizer(prompt, padding='max_length', max_length=self.tokenizer.model_max_length, truncation=True, return_tensors='pt') + cond_embeddings = self.text_encoder(cond_input.input_ids.to(self.device))[0] + + + uncond_input = self.tokenizer(negative_prompt, padding='max_length', max_length=self.tokenizer.model_max_length, return_tensors='pt') + + uncond_embeddings = self.text_encoder(uncond_input.input_ids.to(self.device))[0] + + + return cond_embeddings, uncond_embeddings + + @torch.no_grad() + def prepare_control_image(self, control_pil, width, height): + + control_image = self._prepare_control_image( + image=control_pil, + width=width, + height=height, + device=self.device, + dtype=self.controlnet.dtype, + batch_size=1, + num_images_per_prompt=1 + ) + + return control_image + + @torch.no_grad() + def pred_controlnet_sampling(self, current_sampling_percent, latent_model_input, t, text_embeddings, control_image): + if (current_sampling_percent < self.controlnet_guidance_start or current_sampling_percent > self.controlnet_guidance_end): + down_block_res_samples = None + mid_block_res_sample = None + else: + + down_block_res_samples, mid_block_res_sample = self.controlnet( + latent_model_input, + t, + conditioning_scale=self.controlnet_conditioning_scale, + encoder_hidden_states=text_embeddings, + controlnet_cond=control_image, + return_dict=False, + ) + noise_pred = self.unet(latent_model_input, t, encoder_hidden_states=text_embeddings, + down_block_additional_residuals=down_block_res_samples, + mid_block_additional_residual=mid_block_res_sample)['sample'] + return noise_pred + + + @torch.no_grad() + def denoising_step(self, latents, control_image, text_embeddings, t, guidance_scale, current_sampling_percent): + + latent_model_input = torch.cat([latents] * 2) + control_image = torch.cat([control_image] * 2) + latent_model_input = self.scheduler.scale_model_input(latent_model_input, t) + + + noise_pred = self.pred_controlnet_sampling(current_sampling_percent, latent_model_input, t, text_embeddings, control_image) + + noise_pred_uncond, noise_pred_text = noise_pred.chunk(2) + noise_pred = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) + + + latents = self.scheduler.step(noise_pred, t, latents)['prev_sample'] + return latents + + + @torch.no_grad() + def preprocess_control_grid(self, image_pil): + + list_of_image_pils = fu.pil_grid_to_frames(image_pil, grid_size=self.grid) # List[C, W, H] -> len = num_frames + list_of_pils = [pu.pixel_perfect_process(np.array(frame_pil, dtype='uint8'), self.preprocess_name) for frame_pil in list_of_image_pils] + control_images = np.array(list_of_pils) + control_img = ipu.create_grid_from_numpy(control_images, grid_size=self.grid) + control_img = PIL.Image.fromarray(control_img).convert("L") + + return control_img + + @torch.no_grad() + def shuffle_latents(self, latents, control_image, indices): + rand_i = torch.randperm(self.total_frame_number).tolist() + + latents_l, controls_l, randx = [], [], [] + for j in range(self.sample_size): + rand_indices = rand_i[j*self.grid_frame_number:(j+1)*self.grid_frame_number] + + latents_keyframe, _ = fu.prepare_key_grid_latents(latents, self.grid, self.grid, rand_indices) + control_keyframe, _ = fu.prepare_key_grid_latents(control_image, self.grid, self.grid, rand_indices) + latents_l.append(latents_keyframe) + controls_l.append(control_keyframe) + randx.extend(rand_indices) + rand_i = randx.copy() + latents = torch.cat(latents_l, dim=0) + control_image = torch.cat(controls_l, dim=0) + indices = [indices[i] for i in rand_i] + return latents, indices, control_image + + @torch.no_grad() + def batch_denoise(self, latents, control_image, indices, t, guidance_scale, current_sampling_percent): + latents_l, controls_l = [], [] + control_split = control_image.split(self.batch_size, dim=0) + latents_split = latents.split(self.batch_size, dim=0) + for idx in range(len(control_split)): + txt_embed = torch.cat([self.uncond_embeddings] * len(latents_split[idx]) + [self.cond_embeddings] * len(latents_split[idx])) + + + latents = self.denoising_step(latents_split[idx], control_split[idx], txt_embed, t, guidance_scale, current_sampling_percent) + + latents_l.append(latents) + controls_l.append(control_split[idx]) + latents = torch.cat(latents_l, dim=0) + controls = torch.cat(controls_l, dim=0) + return latents, indices, controls + + @torch.no_grad() + def reverse_diffusion(self, latents=None, control_image=None, guidance_scale=7.5, indices=None): + self.scheduler.set_timesteps(self.num_inference_steps, device=self.device) + with torch.autocast('cuda'): + + for i, t in tqdm(enumerate(self.scheduler.timesteps), desc='reverse_diffusion'): + indices = list(indices) + current_sampling_percent = i / len(self.scheduler.timesteps) + + if self.is_shuffle: + latents, indices, control_image = self.shuffle_latents(latents, control_image, indices) + + if self.cond_step_start < current_sampling_percent: + latents, indices, controls = self.batch_denoise(latents, control_image, indices, t, guidance_scale, current_sampling_percent) + else: + latents, indices, controls = self.batch_denoise(latents, control_image, indices, t, 0.0, current_sampling_percent) + + return latents, indices, controls + + @torch.no_grad() + def encode_imgs(self, img_torch): + latents_l = [] + splits = img_torch.split(self.batch_size_vae, dim=0) + for split in splits: + image = 2 * split - 1 + posterior = self.vae.encode(image).latent_dist + latents = posterior.mean * self.vae.config.scaling_factor + latents_l.append(latents) + + + return torch.cat(latents_l, dim=0) + + @torch.no_grad() + def decode_latents(self, latents: torch.Tensor): + image_l = [] + splits = latents.split(self.batch_size_vae, dim=0) + for split in splits: + image = self.vae.decode(split / self.vae.config.scaling_factor, return_dict=False)[0] + image = (image / 2 + 0.5).clamp(0, 1) + image_l.append(image) + return torch.cat(image_l, dim=0) + + + + + @torch.no_grad() + def controlnet_pred(self, latent_model_input, t, text_embed_input, controlnet_cond): + down_block_res_samples, mid_block_res_sample = self.controlnet( + latent_model_input, + t, + encoder_hidden_states=text_embed_input, + controlnet_cond=controlnet_cond, + conditioning_scale=1, + return_dict=False, + ) + noise_pred = self.unet( + latent_model_input, + t, + encoder_hidden_states=text_embed_input, + cross_attention_kwargs={}, + down_block_additional_residuals=down_block_res_samples, + mid_block_additional_residual=mid_block_res_sample, + return_dict=False, + )[0] + return noise_pred + + @torch.no_grad() + def ddim_inversion(self, latents, control_batch, indices): + k = None + els = os.listdir(self.inverse_path) + els = [el for el in els if el.endswith('.pt')] + + for k,inv_path in enumerate(sorted(els, key=lambda x: int(x.split('.')[0]))): + latents[k] = torch.load(os.path.join(self.inverse_path, inv_path)).to(device=self.device) + + self.inverse_scheduler = DDIMScheduler.from_config(self.scheduler_config) + self.inverse_scheduler.set_timesteps(self.num_inversion_step, device=self.device) + self.timesteps = reversed(self.inverse_scheduler.timesteps) + + if k == (latents.shape[0]-1): + return latents, indices, control_batch + inv_cond = torch.cat([self.inv_uncond_embeddings] * 1 + [self.inv_cond_embeddings] * 1)[1].unsqueeze(0) + for i, t in enumerate(tqdm(self.timesteps)): + + alpha_prod_t = self.inverse_scheduler.alphas_cumprod[t] + alpha_prod_t_prev = (self.inverse_scheduler.alphas_cumprod[self.timesteps[i - 1]] if i > 0 else self.inverse_scheduler.final_alpha_cumprod) + + if k is not None: + if len(latents[:k+1].shape) == 3: + latents[:k+1] = latents[:k+1].unsqueeze(0) + latents_l = [] if k is None else [latents[:k+1]] + latents_split = latents.split(self.inv_batch_size, dim=0) if k is None else latents[k+1:].split(self.inv_batch_size, dim=0) + control_batch_split = control_batch.split(self.inv_batch_size, dim=0) if k is None else control_batch[k+1:].split(self.inv_batch_size, dim=0) + for idx in range(len(latents_split)): + cond_batch = inv_cond.repeat(latents_split[idx].shape[0], 1, 1) + latents = self.ddim_step(latents_split[idx], t, cond_batch, alpha_prod_t, alpha_prod_t_prev, control_batch_split[idx]) + latents_l.append(latents) + latents = torch.cat(latents_l, dim=0) + + for k,i in enumerate(latents): + torch.save(i.detach().cpu(), f'{self.inverse_path}/{str(k).zfill(5)}.pt') + + return latents, indices, control_batch + + + def ddim_step(self, latent_frames, t, cond_batch, alpha_prod_t, alpha_prod_t_prev, control_batch): + mu = alpha_prod_t ** 0.5 + mu_prev = alpha_prod_t_prev ** 0.5 + sigma = (1 - alpha_prod_t) ** 0.5 + sigma_prev = (1 - alpha_prod_t_prev) ** 0.5 + if self.give_control_inversion: + eps = self.controlnet_pred(latent_frames, t, text_embed_input=cond_batch, controlnet_cond=control_batch) + else: + eps = self.unet(latent_frames, t, encoder_hidden_states=cond_batch, return_dict=False)[0] + pred_x0 = (latent_frames - sigma_prev * eps) / mu_prev + latent_frames = mu * pred_x0 + sigma * eps + return latent_frames + + + def process_image_batch(self, image_pil_list): + if len(os.listdir(self.controls_path)) > 0: + control_torch = torch.load(os.path.join(self.controls_path, 'control.pt')).to(self.device) + img_torch = torch.load(os.path.join(self.controls_path, 'img.pt')).to(self.device) + else: + image_torch_list = [] + control_torch_list = [] + for image_pil in image_pil_list: + width, height = image_pil.size + control_pil = self.preprocess_control_grid(image_pil) + control_image = self.prepare_control_image(control_pil, width, height) + control_torch_list.append(control_image) + image_torch_list.append(ipu.pil_img_to_torch_tensor(image_pil)) + control_torch = torch.cat(control_torch_list, dim=0).to(self.device) + img_torch = torch.cat(image_torch_list, dim=0).to(self.device) + torch.save(control_torch, os.path.join(self.controls_path, 'control.pt')) + torch.save(img_torch, os.path.join(self.controls_path, 'img.pt')) + + return img_torch, control_torch + + def order_grids(self, list_of_pils, indices): + k = [] + for i in range(len(list_of_pils)): + k.extend(fu.pil_grid_to_frames(list_of_pils[i], self.grid)) + + frames = [k[indices.index(i)] for i in np.arange(len(indices))] + return frames + + + @torch.autocast(dtype=torch.float16, device_type='cuda') + def batched_denoise_step(self, x, t, indices): + batch_size = self.config["batch_size"] + denoised_latents = [] + pivotal_idx = torch.randint(batch_size, (len(x)//batch_size,)) + torch.arange(0,len(x),batch_size) + + self.denoise_step(x[pivotal_idx], t, indices[pivotal_idx]) + for i, b in enumerate(range(0, len(x), batch_size)): + denoised_latents.append(self.denoise_step(x[b:b + batch_size], t, indices[b:b + batch_size])) + denoised_latents = torch.cat(denoised_latents) + return denoised_latents + + @torch.no_grad() + def __preprocess_inversion_input(self, init_latents, control_batch): + list_of_flattens = [fu.flatten_grid(el.unsqueeze(0), self.grid) for el in init_latents] + init_latents = torch.cat(list_of_flattens, dim=-1) + init_latents = torch.cat(torch.chunk(init_latents, self.total_frame_number, dim=-1), dim=0) + control_batch_flattens = [fu.flatten_grid(el.unsqueeze(0), self.grid) for el in control_batch] + control_batch = torch.cat(control_batch_flattens, dim=-1) + control_batch = torch.cat(torch.chunk(control_batch, self.total_frame_number, dim=-1), dim=0) + return init_latents, control_batch + + @torch.no_grad() + def __postprocess_inversion_input(self, latents_inverted, control_batch): + latents_inverted = torch.cat([fu.unflatten_grid(torch.cat([a for a in latents_inverted[i*self.grid_frame_number:(i+1)*self.grid_frame_number]], dim=-1).unsqueeze(0), self.grid) for i in range(self.sample_size)] , dim=0) + control_batch = torch.cat([fu.unflatten_grid(torch.cat([a for a in control_batch[i*self.grid_frame_number:(i+1)*self.grid_frame_number]], dim=-1).unsqueeze(0), self.grid) for i in range(self.sample_size)] , dim=0) + return latents_inverted, control_batch + + + @torch.no_grad() + def __call__(self, input_dict): + set_seed_lib(input_dict['seed']) + + self.grid_size = input_dict['grid_size'] + self.sample_size = input_dict['sample_size'] + + self.grid_frame_number = self.grid_size * self.grid_size + self.total_frame_number = (self.grid_frame_number) * self.sample_size + self.grid = [self.grid_size, self.grid_size] + + self.cond_step_start = input_dict['cond_step_start'] + + self.controlnet_guidance_start = input_dict['controlnet_guidance_start'] + self.controlnet_guidance_end = input_dict['controlnet_guidance_end'] + self.controlnet_conditioning_scale = input_dict['controlnet_conditioning_scale'] + + self.positive_prompts = input_dict['positive_prompts'] + self.negative_prompts = input_dict['negative_prompts'] + self.inversion_prompt = input_dict['inversion_prompt'] + + self.batch_size = input_dict['batch_size'] + self.inv_batch_size = self.batch_size * self.grid_size * self.grid_size + self.batch_size_vae = input_dict['batch_size_vae'] + + self.num_inference_steps = input_dict['num_inference_steps'] + self.num_inversion_step = input_dict['num_inversion_step'] + self.inverse_path = input_dict['inverse_path'] + self.controls_path = input_dict['control_path'] + + self.is_ddim_inversion = input_dict['is_ddim_inversion'] + self.is_shuffle = input_dict['is_shuffle'] + self.give_control_inversion = input_dict['give_control_inversion'] + + self.guidance_scale = input_dict['guidance_scale'] + + indices = list(np.arange(self.total_frame_number)) + + + img_batch, control_batch = self.process_image_batch(input_dict['image_pil_list']) + init_latents_pre = self.encode_imgs(img_batch) + + self.scheduler = DDIMScheduler.from_config(self.scheduler_config) + self.scheduler.set_timesteps(self.num_inference_steps, device=self.device) + self.inv_cond_embeddings, self.inv_uncond_embeddings = self.get_text_embeds(self.inversion_prompt, "") + if self.is_ddim_inversion: + init_latents, control_batch = self.__preprocess_inversion_input(init_latents_pre, control_batch) + latents_inverted, indices, control_batch = self.ddim_inversion(init_latents, control_batch, indices) + latents_inverted, control_batch = self.__postprocess_inversion_input(latents_inverted, control_batch) + else: + init_latents_pre = torch.cat([init_latents_pre], dim=0) + noise = torch.randn_like(init_latents_pre) + latents_inverted = self.scheduler.add_noise(init_latents_pre, noise, self.scheduler.timesteps[:1]) + + self.cond_embeddings, self.uncond_embeddings = self.get_text_embeds(self.positive_prompts, self.negative_prompts) + latents_denoised, indices, controls = self.reverse_diffusion(latents_inverted, control_batch, self.guidance_scale, indices=indices) + + image_torch = self.decode_latents(latents_denoised) + ordered_img_frames = self.order_grids(ipu.torch_to_pil_img_batch(image_torch), indices) + ordered_control_frames = self.order_grids(ipu.torch_to_pil_img_batch(controls), indices) + return ordered_img_frames, ordered_control_frames + diff --git a/RAVE-main/pipelines/sd_multicontrolnet_rave.py b/RAVE-main/pipelines/sd_multicontrolnet_rave.py new file mode 100644 index 0000000000000000000000000000000000000000..95532ede14087e64c9264ffbdceb6a3cba3960d0 --- /dev/null +++ b/RAVE-main/pipelines/sd_multicontrolnet_rave.py @@ -0,0 +1,466 @@ +import random +import os +import PIL +import torch +import warnings + +warnings.filterwarnings("ignore") + +from transformers import set_seed +from tqdm import tqdm +from transformers import logging +from diffusers import ControlNetModel, StableDiffusionControlNetImg2ImgPipeline, DDIMScheduler + +import torch.nn as nn +import numpy as np + +import utils.constants as const +import utils.feature_utils as fu +import utils.preprocesser_utils as pu +import utils.image_process_utils as ipu + + +logging.set_verbosity_error() + +def set_seed_lib(seed): + np.random.seed(seed) + torch.manual_seed(seed) + torch.cuda.manual_seed(seed) + random.seed(seed) + set_seed(seed) + +@torch.no_grad() +class RAVE_MultiControlNet(nn.Module): + def __init__(self, device): + super().__init__() + + self.device = device + self.dtype = torch.float + + @torch.no_grad() + def __init_pipe(self, hf_cn_path, hf_path): + controlnet_1 = ControlNetModel.from_pretrained(hf_cn_path[0], torch_dtype=self.dtype).to(self.device, self.dtype) + controlnet_2 = ControlNetModel.from_pretrained(hf_cn_path[1], torch_dtype=self.dtype).to(self.device, self.dtype) + pipe = StableDiffusionControlNetImg2ImgPipeline.from_pretrained(hf_path, controlnet=[controlnet_1, controlnet_2], torch_dtype=self.dtype).to(self.device, self.dtype) + pipe.enable_model_cpu_offload() + pipe.enable_xformers_memory_efficient_attention() + return pipe + + @torch.no_grad() + def init_models(self, hf_cn_path, hf_path, preprocess_name, model_id=None): + if model_id is None or model_id == "None": + pipe = self.__init_pipe(hf_cn_path, hf_path) + else: + pipe = self.__init_pipe(hf_cn_path, model_id) + self.preprocess_name_1, self.preprocess_name_2 = preprocess_name.split('-') + + + self._prepare_control_image = pipe.prepare_control_image + self.run_safety_checker = pipe.run_safety_checker + self.tokenizer = pipe.tokenizer + self.text_encoder = pipe.text_encoder + + self.vae = pipe.vae + self.unet = pipe.unet + + self.controlnet = pipe.controlnet + self.scheduler_config = pipe.scheduler.config + + del pipe + + + + @torch.no_grad() + def get_text_embeds(self, prompt, negative_prompt): + # prompt, negative_prompt: [str] + + # Tokenize text and get embeddings + cond_input = self.tokenizer(prompt, padding='max_length', max_length=self.tokenizer.model_max_length, truncation=True, return_tensors='pt') + cond_embeddings = self.text_encoder(cond_input.input_ids.to(self.device))[0] + + # Do the same for unconditional embeddings + uncond_input = self.tokenizer(negative_prompt, padding='max_length', max_length=self.tokenizer.model_max_length, return_tensors='pt') + + uncond_embeddings = self.text_encoder(uncond_input.input_ids.to(self.device))[0] + + # Cat for final embeddings + + return cond_embeddings, uncond_embeddings + + @torch.no_grad() + def prepare_control_image(self, control_pil, width, height): + + control_image = self._prepare_control_image( + image=control_pil, + width=width, + height=height, + device=self.device, + dtype=self.controlnet.dtype, + batch_size=1, + num_images_per_prompt=1 + ) + + return control_image + + @torch.no_grad() + def pred_controlnet_sampling(self, current_sampling_percent, latent_model_input, t, text_embeddings, control_image): + if (current_sampling_percent < self.controlnet_guidance_start or current_sampling_percent > self.controlnet_guidance_end): + down_block_res_samples = None + mid_block_res_sample = None + else: + + down_block_res_samples, mid_block_res_sample = self.controlnet( + latent_model_input, + t, + conditioning_scale=self.controlnet_conditioning_scale, + encoder_hidden_states=text_embeddings, + controlnet_cond=control_image, + return_dict=False, + ) + noise_pred = self.unet(latent_model_input, t, encoder_hidden_states=text_embeddings, + down_block_additional_residuals=down_block_res_samples, + mid_block_additional_residual=mid_block_res_sample)['sample'] + return noise_pred + + + @torch.no_grad() + def denoising_step(self, latents, control_image_1, control_image_2, text_embeddings, t, guidance_scale, current_sampling_percent): + # expand the latents if we are doing classifier-free guidance to avoid doing two forward passes. + + latent_model_input = torch.cat([latents] * 2) + control_image_1 = torch.cat([control_image_1] * 2) + control_image_2 = torch.cat([control_image_2] * 2) + latent_model_input = self.scheduler.scale_model_input(latent_model_input, t) + # compute the percentage of total steps we are at + + + noise_pred = self.pred_controlnet_sampling(current_sampling_percent, latent_model_input, t, text_embeddings, [control_image_1, control_image_2]) + + # perform guidance + noise_pred_uncond, noise_pred_text = noise_pred.chunk(2) + noise_pred = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) + + # compute the previous noisy sample x_t -> x_t-1 + + latents = self.scheduler.step(noise_pred, t, latents)['prev_sample'] + return latents + + + @torch.no_grad() + def preprocess_control_grid(self, image_pil): + + list_of_image_pils = fu.pil_grid_to_frames(image_pil, grid_size=self.grid) # List[C, W, H] -> len = num_frames + list_of_pils_1, list_of_pils_2 = [], [] + for frame_pil in list_of_image_pils: + frame_pil_1 = pu.pixel_perfect_process(np.array(frame_pil, dtype='uint8'), self.preprocess_name_1) + frame_pil_2 = pu.pixel_perfect_process(np.array(frame_pil, dtype='uint8'), self.preprocess_name_2) + list_of_pils_1.append(frame_pil_1) + list_of_pils_2.append(frame_pil_2) + control_images_1 = np.array(list_of_pils_1) + control_images_2 = np.array(list_of_pils_2) + + control_img_1 = ipu.create_grid_from_numpy(control_images_1, grid_size=self.grid) + control_img_1 = PIL.Image.fromarray(control_img_1).convert("L") + + control_img_2 = ipu.create_grid_from_numpy(control_images_2, grid_size=self.grid) + control_img_2 = PIL.Image.fromarray(control_img_2).convert("L") + + return control_img_1, control_img_2 + + @torch.no_grad() + def shuffle_latents(self, latents, control_image_1, control_image_2, indices): + rand_i = torch.randperm(self.total_frame_number).tolist() + # latents, _ = fu.prepare_key_grid_latents(latents, self.grid, self.grid, rand_i) + # control_image, _ = fu.prepare_key_grid_latents(control_image, self.grid, self.grid, rand_i) + + latents_l, controls_l_1, controls_l_2, randx = [], [], [], [] + for j in range(self.sample_size): + rand_indices = rand_i[j*self.grid_frame_number:(j+1)*self.grid_frame_number] + + latents_keyframe, _ = fu.prepare_key_grid_latents(latents, self.grid, self.grid, rand_indices) + control_keyframe_1, _ = fu.prepare_key_grid_latents(control_image_1, self.grid, self.grid, rand_indices) + control_keyframe_2, _ = fu.prepare_key_grid_latents(control_image_2, self.grid, self.grid, rand_indices) + latents_l.append(latents_keyframe) + controls_l_1.append(control_keyframe_1) + controls_l_2.append(control_keyframe_2) + randx.extend(rand_indices) + rand_i = randx.copy() + latents = torch.cat(latents_l, dim=0) + control_image_1 = torch.cat(controls_l_1, dim=0) + control_image_2 = torch.cat(controls_l_2, dim=0) + indices = [indices[i] for i in rand_i] + return latents, indices, control_image_1, control_image_2 + + @torch.no_grad() + def batch_denoise(self, latents, control_image_1, control_image_2, indices, t, guidance_scale, current_sampling_percent): + + + latents_l, controls_l_1, controls_l_2 = [], [], [] + control_split_1 = control_image_1.split(self.batch_size, dim=0) + control_split_2 = control_image_2.split(self.batch_size, dim=0) + latents_split = latents.split(self.batch_size, dim=0) + + + for idx in range(len(control_split_1)): + txt_embed = torch.cat([self.uncond_embeddings] * len(latents_split[idx]) + [self.cond_embeddings] * len(latents_split[idx])) + + + latents = self.denoising_step(latents_split[idx], control_split_1[idx], control_split_2[idx], txt_embed, t, guidance_scale, current_sampling_percent) + + latents_l.append(latents) + controls_l_1.append(control_split_1[idx]) + controls_l_2.append(control_split_2[idx]) + + latents = torch.cat(latents_l, dim=0) + controls_1 = torch.cat(controls_l_1, dim=0) + controls_2 = torch.cat(controls_l_2, dim=0) + return latents, indices, controls_1, controls_2 + + @torch.no_grad() + def reverse_diffusion(self, latents=None, control_image_1=None, control_image_2=None, guidance_scale=7.5, indices=None): + self.scheduler.set_timesteps(self.num_inference_steps, device=self.device) + with torch.autocast('cuda'): + + for i, t in tqdm(enumerate(self.scheduler.timesteps), desc='reverse_diffusion'): + indices = list(indices) + current_sampling_percent = i / len(self.scheduler.timesteps) + + if self.is_shuffle: + latents, indices, control_image_1, control_image_2 = self.shuffle_latents(latents, control_image_1, control_image_2, indices) + + if self.cond_step_start < current_sampling_percent: + latents, indices, control_image_1, control_image_2 = self.batch_denoise(latents, control_image_1, control_image_2, indices, t, guidance_scale, current_sampling_percent) + else: + latents, indices, control_image_1, control_image_2 = self.batch_denoise(latents, control_image_1, control_image_2, indices, t, 0.0, current_sampling_percent) + + return latents, indices, control_image_1, control_image_2 + + @torch.no_grad() + def encode_imgs(self, img_torch): + latents_l = [] + splits = img_torch.split(self.batch_size_vae, dim=0) + for split in splits: + image = 2 * split - 1 + posterior = self.vae.encode(image).latent_dist + latents = posterior.mean * self.vae.config.scaling_factor + latents_l.append(latents) + + + return torch.cat(latents_l, dim=0) + + @torch.no_grad() + def decode_latents(self, latents): + image_l = [] + splits = latents.split(self.batch_size_vae, dim=0) + for split in splits: + image = self.vae.decode(split / self.vae.config.scaling_factor, return_dict=False)[0] + image = (image / 2 + 0.5).clamp(0, 1) + image_l.append(image) + return torch.cat(image_l, dim=0) + + + @torch.no_grad() + def controlnet_pred(self, latent_model_input, t, text_embed_input, controlnet_cond): + down_block_res_samples, mid_block_res_sample = self.controlnet( + latent_model_input, + t, + encoder_hidden_states=text_embed_input, + controlnet_cond=controlnet_cond, + conditioning_scale=self.controlnet_conditioning_scale, + return_dict=False, + ) + + # apply the denoising network + noise_pred = self.unet( + latent_model_input, + t, + encoder_hidden_states=text_embed_input, + cross_attention_kwargs={}, + down_block_additional_residuals=down_block_res_samples, + mid_block_additional_residual=mid_block_res_sample, + return_dict=False, + )[0] + return noise_pred + + @torch.no_grad() + def ddim_inversion(self, latents, control_batch_1, control_batch_2, indices): + k = None + els = os.listdir(self.inverse_path) + els = [el for el in els if el.endswith('.pt')] + for k,inv_path in enumerate(sorted(els, key=lambda x: int(x.split('.')[0]))): + latents[k] = torch.load(os.path.join(self.inverse_path, inv_path)).to(device=self.device) + + self.inverse_scheduler = DDIMScheduler.from_config(self.scheduler_config) + self.inverse_scheduler.set_timesteps(self.num_inversion_step, device=self.device) + self.timesteps = reversed(self.inverse_scheduler.timesteps) + + if k == (latents.shape[0]-1): + return latents, indices, control_batch_1, control_batch_2 + inv_cond = torch.cat([self.inv_uncond_embeddings] * 1 + [self.inv_cond_embeddings] * 1)[1].unsqueeze(0) + for i, t in enumerate(tqdm(self.timesteps)): + + alpha_prod_t = self.inverse_scheduler.alphas_cumprod[t] + alpha_prod_t_prev = (self.inverse_scheduler.alphas_cumprod[self.timesteps[i - 1]] if i > 0 else self.inverse_scheduler.final_alpha_cumprod) + + # latent, indices, control_batch = self.shuffle_latents(latent, control_batch, indices) + + latents_l = [] + latents_split = latents.split(self.batch_size, dim=0) + control_batch_split_1 = control_batch_1.split(self.batch_size, dim=0) + control_batch_split_2 = control_batch_2.split(self.batch_size, dim=0) + for idx in range(len(latents_split)): + cond_batch = inv_cond.repeat(latents_split[idx].shape[0], 1, 1) + # print(cond_batch.shape, latents_split[idx].shape, control_batch_split_1[idx].shape, control_batch_split_2[idx].shape) + # input() + latents = self.ddim_step(latents_split[idx], t, cond_batch, alpha_prod_t, alpha_prod_t_prev, control_batch_split_1[idx], control_batch_split_2[idx]) + latents_l.append(latents) + latents = torch.cat(latents_l, dim=0) + for k,i in enumerate(latents): + torch.save(i.detach().cpu(), f'{self.inverse_path}/{str(k).zfill(5)}.pt') + return latents, indices, control_batch_1, control_batch_2 + + + def ddim_step(self, latent_frames, t, cond_batch, alpha_prod_t, alpha_prod_t_prev, control_batch_1, control_batch_2): + mu = alpha_prod_t ** 0.5 + mu_prev = alpha_prod_t_prev ** 0.5 + sigma = (1 - alpha_prod_t) ** 0.5 + sigma_prev = (1 - alpha_prod_t_prev) ** 0.5 + if self.give_control_inversion: + eps = self.controlnet_pred(latent_frames, t, text_embed_input=cond_batch, controlnet_cond=[control_batch_1, control_batch_2]) + else: + eps = self.unet(latent_frames, t, encoder_hidden_states=cond_batch, return_dict=False)[0] + pred_x0 = (latent_frames - sigma_prev * eps) / mu_prev + latent_frames = mu * pred_x0 + sigma * eps + return latent_frames + + + def process_image_batch(self, image_pil_list): + if len(os.listdir(self.controls_path)) > 0: + control_torch_1 = torch.load(os.path.join(self.controls_path, 'control_1.pt')).to(self.device) + control_torch_2 = torch.load(os.path.join(self.controls_path, 'control_2.pt')).to(self.device) + img_torch = torch.load(os.path.join(self.controls_path, 'img.pt')).to(self.device) + else: + image_torch_list = [] + control_torch_list_1, control_torch_list_2 = [], [] + for image_pil in image_pil_list: + width, height = image_pil.size + # control_pil = PIL.Image.fromarray(pu.pixel_perfect_process(np.array(image_pil, dtype='uint8'), self.preprocess_name)) + control_pil_1, control_pil_2 = self.preprocess_control_grid(image_pil) + control_image_1 = self.prepare_control_image(control_pil_1, width, height) + control_image_2 = self.prepare_control_image(control_pil_2, width, height) + + control_torch_list_1.append(control_image_1) + control_torch_list_2.append(control_image_2) + image_torch_list.append(ipu.pil_img_to_torch_tensor(image_pil)) + control_torch_1 = torch.cat(control_torch_list_1, dim=0).to(self.device) + control_torch_2 = torch.cat(control_torch_list_2, dim=0).to(self.device) + img_torch = torch.cat(image_torch_list, dim=0).to(self.device) + torch.save(control_torch_1, os.path.join(self.controls_path, 'control_1.pt')) + torch.save(control_torch_2, os.path.join(self.controls_path, 'control_2.pt')) + torch.save(img_torch, os.path.join(self.controls_path, 'img.pt')) + + return img_torch, control_torch_1, control_torch_2 + + def order_grids(self, list_of_pils, indices): + k = [] + for i in range(len(list_of_pils)): + k.extend(fu.pil_grid_to_frames(list_of_pils[i], self.grid)) + + frames = [k[indices.index(i)] for i in np.arange(len(indices))] + return frames + + + @torch.no_grad() + def __preprocess_inversion_input(self, init_latents, control_batch_1, control_batch_2): + list_of_flattens = [fu.flatten_grid(el.unsqueeze(0), self.grid) for el in init_latents] + init_latents = torch.cat(list_of_flattens, dim=-1) + init_latents = torch.cat(torch.chunk(init_latents, self.total_frame_number, dim=-1), dim=0) + + control_batch_flattens_1 = [fu.flatten_grid(el.unsqueeze(0), self.grid) for el in control_batch_1] + control_batch_1 = torch.cat(control_batch_flattens_1, dim=-1) + control_batch_1 = torch.cat(torch.chunk(control_batch_1, self.total_frame_number, dim=-1), dim=0) + + control_batch_flattens_2 = [fu.flatten_grid(el.unsqueeze(0), self.grid) for el in control_batch_2] + control_batch_2 = torch.cat(control_batch_flattens_2, dim=-1) + control_batch_2 = torch.cat(torch.chunk(control_batch_2, self.total_frame_number, dim=-1), dim=0) + + return init_latents, control_batch_1, control_batch_2 + + + @torch.no_grad() + def __postprocess_inversion_input(self, latents_inverted, control_batch_1, control_batch_2): + latents_inverted = torch.cat([fu.unflatten_grid(torch.cat([a for a in latents_inverted[i*self.grid_frame_number:(i+1)*self.grid_frame_number]], dim=-1).unsqueeze(0), self.grid) for i in range(self.sample_size)] , dim=0) + control_batch_1 = torch.cat([fu.unflatten_grid(torch.cat([a for a in control_batch_1[i*self.grid_frame_number:(i+1)*self.grid_frame_number]], dim=-1).unsqueeze(0), self.grid) for i in range(self.sample_size)] , dim=0) + control_batch_2 = torch.cat([fu.unflatten_grid(torch.cat([a for a in control_batch_2[i*self.grid_frame_number:(i+1)*self.grid_frame_number]], dim=-1).unsqueeze(0), self.grid) for i in range(self.sample_size)] , dim=0) + return latents_inverted, control_batch_1, control_batch_2 + + + + + @torch.no_grad() + def __call__(self, input_dict): + set_seed_lib(input_dict['seed']) + + self.grid_size = input_dict['grid_size'] + self.sample_size = input_dict['sample_size'] + + self.grid_frame_number = self.grid_size * self.grid_size + self.total_frame_number = (self.grid_frame_number) * self.sample_size + self.grid = [self.grid_size, self.grid_size] + + self.cond_step_start = input_dict['cond_step_start'] + + self.controlnet_guidance_start = input_dict['controlnet_guidance_start'] + self.controlnet_guidance_end = input_dict['controlnet_guidance_end'] + self.controlnet_conditioning_scale = [float(x) for x in input_dict['controlnet_conditioning_scale'].split('-')] + + self.positive_prompts = input_dict['positive_prompts'] + self.negative_prompts = input_dict['negative_prompts'] + self.inversion_prompt = input_dict['inversion_prompt'] + + self.batch_size = input_dict['batch_size'] + self.batch_size_vae = input_dict['batch_size_vae'] + + self.num_inference_steps = input_dict['num_inference_steps'] + self.num_inversion_step = input_dict['num_inversion_step'] + self.inverse_path = input_dict['inverse_path'] + self.controls_path = input_dict['control_path'] + + self.is_ddim_inversion = input_dict['is_ddim_inversion'] + + self.is_shuffle = input_dict['is_shuffle'] + self.give_control_inversion = input_dict['give_control_inversion'] + + self.guidance_scale = input_dict['guidance_scale'] + + + + indices = list(np.arange(self.total_frame_number)) + + + img_batch, control_batch_1, control_batch_2 = self.process_image_batch(input_dict['image_pil_list']) + init_latents_pre = self.encode_imgs(img_batch) + + self.scheduler = DDIMScheduler.from_config(self.scheduler_config) + self.scheduler.set_timesteps(self.num_inference_steps, device=self.device) + self.inv_cond_embeddings, self.inv_uncond_embeddings = self.get_text_embeds(self.inversion_prompt, "") + if self.is_ddim_inversion: + init_latents, control_batch_1, control_batch_2 = self.__preprocess_inversion_input(init_latents_pre, control_batch_1, control_batch_2) + latents_inverted, indices, control_batch_1, control_batch_2 = self.ddim_inversion(init_latents, control_batch_1, control_batch_2, indices) + latents_inverted, control_batch_1, control_batch_2 = self.__postprocess_inversion_input(latents_inverted, control_batch_1, control_batch_2) + else: + + init_latents_pre = torch.cat([init_latents_pre], dim=0) + noise = torch.randn_like(init_latents_pre) + latents_inverted = self.scheduler.add_noise(init_latents_pre, noise, self.scheduler.timesteps[:1]) + + self.cond_embeddings, self.uncond_embeddings = self.get_text_embeds(self.positive_prompts, self.negative_prompts) + + latents_denoised, indices, controls_1, controls_2 = self.reverse_diffusion(latents_inverted, control_batch_1, control_batch_2, self.guidance_scale, indices=indices) + + image_torch = self.decode_latents(latents_denoised) + ordered_img_frames = self.order_grids(ipu.torch_to_pil_img_batch(image_torch), indices) + ordered_control_frames_1 = self.order_grids(ipu.torch_to_pil_img_batch(controls_1), indices) + ordered_control_frames_2 = self.order_grids(ipu.torch_to_pil_img_batch(controls_2), indices) + return ordered_img_frames, ordered_control_frames_1, ordered_control_frames_2 + diff --git a/RAVE-main/results/02-27-2025/truck/truck/Wooden trucks drive on a racetrack-00001/config.yaml b/RAVE-main/results/02-27-2025/truck/truck/Wooden trucks drive on a racetrack-00001/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..85d276921789c3c5bc99b775b125799ab1ce1c89 --- /dev/null +++ b/RAVE-main/results/02-27-2025/truck/truck/Wooden trucks drive on a racetrack-00001/config.yaml @@ -0,0 +1,33 @@ +batch_size: 4 +batch_size_vae: 1 +cond_step_start: 0.0 +control_path: /home/wangjuntong/RAVE-main/generated/data/controls/truck/depth_zoe_3x3_1 +controlnet_conditioning_scale: 1.0 +controlnet_guidance_end: 1.0 +controlnet_guidance_start: 0.0 +give_control_inversion: true +grid_size: 3 +guidance_scale: 7.5 +hf_cn_path: lllyasviel/control_v11f1p_sd15_depth +hf_path: runwayml/stable-diffusion-v1-5 +inverse_path: /home/wangjuntong/RAVE-main/generated/data/inverses/truck/depth_zoe_None_3x3_1 +inversion_prompt: '' +is_ddim_inversion: true +is_shuffle: true +model_id: None +negative_prompts: '' +num_inference_steps: 50 +num_inversion_step: 50 +pad: 1 +positive_prompts: Wooden trucks drive on a racetrack +preprocess_name: depth_zoe +sample_size: 2 +save_folder: truck/truck +save_path: /home/wangjuntong/RAVE-main/results/02-27-2025/truck/truck/Wooden trucks + drive on a racetrack-00001 +sec_per_frame: 12.963384277777777 +seed: 0 +total_number_of_frames: 18 +total_time: 233.340917 +video_name: truck +video_path: /home/wangjuntong/RAVE-main/data/mp4_videos/truck.mp4