{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":733566298,"defaultBranch":"main","name":"Awesome-Text2X-Resources","ownerLogin":"ALEEEHU","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2023-12-19T16:04:53.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/57009216?v=4","public":true,"private":false,"isOrgOwned":false},"refInfo":{"name":"","listCacheKey":"v0:1703001894.597337","currentOid":""},"activityList":{"items":[{"before":"2d64828055317714206f0c612d16de7a7d2b2572","after":"4ebd8ecd3c6d15c584eacc3250beaa68c6aec854","ref":"refs/heads/main","pushedAt":"2024-06-02T11:53:12.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text24d]: update one paper\"PLA4D: Pixel-Level Alignments for Text-to-4D Gaussian Splatting\", arXiv on 30 May 2024.","shortMessageHtmlLink":"Update README.md"}},{"before":"dc8bdbaf2a4bc6942fab987b8df799acbc9788b3","after":"2d64828055317714206f0c612d16de7a7d2b2572","ref":"refs/heads/main","pushedAt":"2024-06-01T01:42:13.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text24d]: update one new paper\"EG4D: Explicit Generation of 4D Object without Score Distillation\", arXiv on 28 May 2024.","shortMessageHtmlLink":"Update README.md"}},{"before":"47b728f54b00983d08ca1a1fc64eb9dabee5562c","after":"dc8bdbaf2a4bc6942fab987b8df799acbc9788b3","ref":"refs/heads/main","pushedAt":"2024-05-28T06:39:24.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text24d]: update one paper \"Diffusion4D: Fast Spatial-temporal Consistent 4D Generation via Video Diffusion Models\", arXiv on 26 May,2024.","shortMessageHtmlLink":"Update README.md"}},{"before":"067c51765482ad0de16511f159f33f459d1cd4db","after":"47b728f54b00983d08ca1a1fc64eb9dabee5562c","ref":"refs/heads/main","pushedAt":"2024-05-27T11:34:40.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[4d section]: add one new paper \"Vidu4D: Single Generated Video to High-Fidelity 4D Reconstruction with Dynamic Gaussian Surfels\".","shortMessageHtmlLink":"Update README.md"}},{"before":"4e1d4993f17393f545d3aab128d76e071e7d750e","after":"067c51765482ad0de16511f159f33f459d1cd4db","ref":"refs/heads/main","pushedAt":"2024-05-24T05:41:43.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]:the paper\"Multi-Track Timeline Control for Text-Driven 3D Human Motion Generation\", has released its code recently!","shortMessageHtmlLink":"Update README.md"}},{"before":"9eeb08431237fbebf2f44d75ea3d6be98e30d694","after":"4e1d4993f17393f545d3aab128d76e071e7d750e","ref":"refs/heads/main","pushedAt":"2024-05-23T03:42:12.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text24d]:\"Dream-in-4D: A Unified Approach for Text- and Image-guided 4D Scene Generation\" code has released.","shortMessageHtmlLink":"Update README.md"}},{"before":"e4f0adda27e92b04ac530d02bac5d9a32bb359f4","after":"9eeb08431237fbebf2f44d75ea3d6be98e30d694","ref":"refs/heads/main","pushedAt":"2024-05-21T14:46:27.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]: update one paper\"Flexible Motion In-betweening with Diffusion Models\" got accepted by SIGGRAPH 2024 and arXiv on 17 May 2024.","shortMessageHtmlLink":"Update README.md"}},{"before":"8ba9f9a063baf7a0f8977f25ff01903f2c2fffa5","after":"e4f0adda27e92b04ac530d02bac5d9a32bb359f4","ref":"refs/heads/main","pushedAt":"2024-05-20T13:22:14.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2scene]: update one new paper \"ART3D\", got accepted by CVPR 2024 Workshop on AI3DG! Congrats!","shortMessageHtmlLink":"Update README.md"}},{"before":"7dc85003df4626b2ab7233d6e93b7a4cd91a9d3f","after":"8ba9f9a063baf7a0f8977f25ff01903f2c2fffa5","ref":"refs/heads/main","pushedAt":"2024-05-19T13:30:39.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2human]:add one new paper \"MagicMirror\".","shortMessageHtmlLink":"Update README.md"}},{"before":"464519eee30176ad822c038c96da27f05f2885a4","after":"7dc85003df4626b2ab7233d6e93b7a4cd91a9d3f","ref":"refs/heads/main","pushedAt":"2024-05-18T05:36:05.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]:update one paper\"StableMoFusion: Towards Robust and Efficient Diffusion-based Motion Generation Framework\",arXiv on 9 May 2024.","shortMessageHtmlLink":"Update README.md"}},{"before":"0de42b4f86625b841bd047cdbe5649e556427d3b","after":"464519eee30176ad822c038c96da27f05f2885a4","ref":"refs/heads/main","pushedAt":"2024-05-13T09:20:48.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]: congrats to \"HumanTOMATO\" accepted by ICML-2024!","shortMessageHtmlLink":"Update README.md"}},{"before":"7b4549f7c786d29f35e3bbf486d8ec6b3f8fb803","after":"0de42b4f86625b841bd047cdbe5649e556427d3b","ref":"refs/heads/main","pushedAt":"2024-05-13T09:12:06.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\nCongrats to \"Disentangled 3D Scene Generation with Layout Learning\" accepted by ICML 2024!","shortMessageHtmlLink":"Update README.md"}},{"before":"f90e18f10ba25905018bee018c78468c8cc14afa","after":"7b4549f7c786d29f35e3bbf486d8ec6b3f8fb803","ref":"refs/heads/main","pushedAt":"2024-05-08T13:25:34.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\nUpdate one new survey paper on 3DGS-\"Gaussian Splatting: 3D Reconstruction and Novel View Synthesis, a Review\" into the section [Related Resources--Survey and Awesome Repos--Survey]","shortMessageHtmlLink":"Update README.md"}},{"before":"67a1b91069407f51d762a377c03450a83494df8b","after":"f90e18f10ba25905018bee018c78468c8cc14afa","ref":"refs/heads/main","pushedAt":"2024-05-06T11:17:12.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n1.[text24d]: update one new paper\"DreamScene4D\", arXiv on 3 May 2024.\r\n2.update the 4D Gaussian Splatting section, congrats to 3DGStream got CVPR24 Highlight!","shortMessageHtmlLink":"Update README.md"}},{"before":"ad067d8b404e6461eb1a03fc48d015a0bdb7b550","after":"67a1b91069407f51d762a377c03450a83494df8b","ref":"refs/heads/main","pushedAt":"2024-05-04T12:53:34.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]: add one new paper \"SATO: Stable Text-to-Motion Framework\", arXiv on 2 May 2024.","shortMessageHtmlLink":"Update README.md"}},{"before":"0eb267f2fd35f7a2f4c47a0005f1f74c8e80dbf7","after":"ad067d8b404e6461eb1a03fc48d015a0bdb7b550","ref":"refs/heads/main","pushedAt":"2024-05-02T15:15:23.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion updates]: one paper got accepted by CVPRW 24, and \"self-correcting-self-consuming\" got accepted by ICML 24!","shortMessageHtmlLink":"Update README.md"}},{"before":"a0b283bf4085ee4a2cca045ef17385396fb1e973","after":"0eb267f2fd35f7a2f4c47a0005f1f74c8e80dbf7","ref":"refs/heads/main","pushedAt":"2024-05-02T14:59:07.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2scene updates]: congrats to \"GALA3D\", got accepted by ICML 2024!","shortMessageHtmlLink":"Update README.md"}},{"before":"faa05b113341d024fe04f628beba9ae63d2061de","after":"a0b283bf4085ee4a2cca045ef17385396fb1e973","ref":"refs/heads/main","pushedAt":"2024-05-01T16:58:34.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]: update one new paper \"MotionLCM\", arXiv on 30 April.","shortMessageHtmlLink":"Update README.md"}},{"before":"65e0f00c1670340704e50e09f5b026d2bf219905","after":"faa05b113341d024fe04f628beba9ae63d2061de","ref":"refs/heads/main","pushedAt":"2024-04-29T04:08:54.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]:update one paper 'in2IN', accepted at CVPR 2024 Workshop HuMoGen!","shortMessageHtmlLink":"Update README.md"}},{"before":"e9bf2e1fa610f0ef45b9aafbd53758b48cd18c96","after":"65e0f00c1670340704e50e09f5b026d2bf219905","ref":"refs/heads/main","pushedAt":"2024-04-24T16:30:53.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]: update one paper\"Move as You Say, Interact as You Can: Language-guided Human Motion Generation with Scene Affordance\", got accepted by CVPR24!","shortMessageHtmlLink":"Update README.md"}},{"before":"69ded6326701b07c7978e3e082569fbb3bfc51ef","after":"e9bf2e1fa610f0ef45b9aafbd53758b48cd18c96","ref":"refs/heads/main","pushedAt":"2024-04-19T03:32:54.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2scene]: add one new paper:\"DreamScene: 3D Gaussian-based Text-to-3D Scene Generation via Formation Pattern Sampling\",on arXiv.","shortMessageHtmlLink":"Update README.md"}},{"before":"8bafcb65be21cf6ee1b8406f59f4f2484d4851c5","after":"69ded6326701b07c7978e3e082569fbb3bfc51ef","ref":"refs/heads/main","pushedAt":"2024-04-18T15:47:48.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]: add one new paper\"Generating Human Interaction Motions in Scenes with Text Control\", on arXiv 16 April,2024.","shortMessageHtmlLink":"Update README.md"}},{"before":"4953b865c33cf639234c0f762c74f46aab404a84","after":"8bafcb65be21cf6ee1b8406f59f4f2484d4851c5","ref":"refs/heads/main","pushedAt":"2024-04-14T06:16:12.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2human]: add one new paper \"InstructHumans: Editing Animated 3D Human Textures with Instructions\", arXiv on 5 Apr 2024 (This paper is more like text to human texture editing)","shortMessageHtmlLink":"Update README.md"}},{"before":"c5feca6811b8d906e9316483e3fe7764913829fb","after":"4953b865c33cf639234c0f762c74f46aab404a84","ref":"refs/heads/main","pushedAt":"2024-04-11T11:33:53.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2scene]: add one new paper \"DreamScene360: Unconstrained Text-to-3D Scene Generation with Panoramic Gaussian Splatting\", arXiv on April 10 2024","shortMessageHtmlLink":"Update README.md"}},{"before":"85e0792adf4ff148682da997b32cf73dc6066f6e","after":"c5feca6811b8d906e9316483e3fe7764913829fb","ref":"refs/heads/main","pushedAt":"2024-04-11T11:30:03.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2scene]: add one new paper \"RealmDreamer: Text-Driven 3D Scene Generation with Inpainting and Depth Diffusion\", arXiv on April 10 2024","shortMessageHtmlLink":"Update README.md"}},{"before":"908fbdcbdff0e4218fd95e29eb151ef31bd65be4","after":"85e0792adf4ff148682da997b32cf73dc6066f6e","ref":"refs/heads/main","pushedAt":"2024-04-09T03:20:57.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text24d]: add one new paper \"SC4D: Sparse-Controlled Video-to-4D Generation and Motion Transfer\", arXiv on 4 Apr 2024.","shortMessageHtmlLink":"Update README.md"}},{"before":"d95d79b45c9f9942a618de06366507cfd320aec1","after":"908fbdcbdff0e4218fd95e29eb151ef31bd65be4","ref":"refs/heads/main","pushedAt":"2024-04-08T02:06:38.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]: add one paper \"Large Motion Model for Unified Multi-Modal Motion Generation\", arXiv on 1 Apr 2024","shortMessageHtmlLink":"Update README.md"}},{"before":"fd70d1ec9d9b0e42d3965350d23351e3768ebb38","after":"d95d79b45c9f9942a618de06366507cfd320aec1","ref":"refs/heads/main","pushedAt":"2024-04-07T11:05:12.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2motion]: add one paper accepted by CVPR24,\"Space-Time Diffusion Features for Zero-Shot Text-Driven Motion Transfer\".","shortMessageHtmlLink":"Update README.md"}},{"before":"e3ba0a15593c2920e3e04cece38b59e683e50749","after":"fd70d1ec9d9b0e42d3965350d23351e3768ebb38","ref":"refs/heads/main","pushedAt":"2024-04-05T10:45:34.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\nupdate logs","shortMessageHtmlLink":"Update README.md"}},{"before":"c9d73f1d11a2afb700a064ae2f5e915c8bbd5032","after":"e3ba0a15593c2920e3e04cece38b59e683e50749","ref":"refs/heads/main","pushedAt":"2024-04-01T05:07:14.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"ALEEEHU","name":"Yuqi HU","path":"/ALEEEHU","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57009216?s=80&v=4"},"commit":{"message":"Update README.md\n\n[text2scene]: 3D-SceneDreamer: Text-Driven 3D-Consistent Scene Generation, on arXiv 14 Mar 2024","shortMessageHtmlLink":"Update README.md"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEWiYHugA","startCursor":null,"endCursor":null}},"title":"Activity · ALEEEHU/Awesome-Text2X-Resources"}