{"payload":{"pageCount":3,"repositories":[{"type":"Public","name":"Large-Time-Series-Model","owner":"thuml","isFork":false,"description":"Official code, datasets and checkpoints for \"Timer: Generative Pre-trained Transformers Are Large Time Series Models\" (ICML 2024) ","allTopics":["time-series","transformer","large-models","time-series-analysis"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":75,"forksCount":2,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,7,0,0,0,0,0,2,0,0,3,0,0,0,8,1,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-19T06:47:55.694Z"}},{"type":"Public","name":"depyf","owner":"thuml","isFork":false,"description":"depyf is a tool to help you understand and adapt to PyTorch compiler torch.compile.","allTopics":["deep-learning","compiler","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":370,"forksCount":8,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,73,35,0,2,0,12,50,26,40,9,0,16,29,79,49,30,41,4,0,2,3,1,4,0,0,1,0,1,13,0,6,0,0,0,0,0,0,8,14,0,7,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-18T01:34:49.643Z"}},{"type":"Public","name":"Transolver","owner":"thuml","isFork":false,"description":"About code release of \"Transolver: A Fast Transformer Solver for PDEs on General Geometries\", ICML 2024 Spotlight. https://arxiv.org/abs/2402.02366","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":24,"forksCount":1,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-17T14:30:47.976Z"}},{"type":"Public","name":"iTransformer","owner":"thuml","isFork":false,"description":"Official implementation for \"iTransformer: Inverted Transformers Are Effective for Time Series Forecasting\" (ICLR 2024 Spotlight), https://openreview.net/forum?id=JePfAI8fah","allTopics":["transformer","time-series-forecasting"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":7,"starsCount":876,"forksCount":164,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-16T09:08:07.815Z"}},{"type":"Public","name":"Koopa","owner":"thuml","isFork":false,"description":"Code release for \"Koopa: Learning Non-stationary Time Series Dynamics with Koopman Predictors\" (NeurIPS 2023), https://arxiv.org/abs/2305.18803","allTopics":["deep-learning","time-series","forecasting","non-stationary"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":158,"forksCount":20,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,2,0,1,0,0,0,0,0,7,4,3,0,1,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-14T12:01:13.534Z"}},{"type":"Public","name":"Nonstationary_Transformers","owner":"thuml","isFork":false,"description":"Code release for \"Non-stationary Transformers: Exploring the Stationarity in Time Series Forecasting\" (NeurIPS 2022), https://arxiv.org/abs/2205.14415","allTopics":["deep-learning","time-series","forecasting","non-stationary"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":444,"forksCount":73,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-13T12:44:36.913Z"}},{"type":"Public","name":"TimeSiam","owner":"thuml","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,2,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-11T10:10:07.307Z"}},{"type":"Public","name":"HarmonyDream","owner":"thuml","isFork":false,"description":"Code release for \"HarmonyDream: Task Harmonization Inside World Models\" (ICML 2024), https://arxiv.org/abs/2310.00344","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-10T16:37:14.451Z"}},{"type":"Public","name":"AutoTimes","owner":"thuml","isFork":false,"description":"Official implementation for \"AutoTimes: Autoregressive Time Series Forecasters via Large Language Models\"","allTopics":["time-series-forecasting","large-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":31,"forksCount":2,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T07:55:28.914Z"}},{"type":"Public","name":"timer","owner":"thuml","isFork":false,"description":"See the official code and checkpoints for \"Timer: Generative Pre-trained Transformers Are Large Time Series Models\"","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":1,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,0,0,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T16:59:36.544Z"}},{"type":"Public","name":"Time-Series-Library","owner":"thuml","isFork":false,"description":"A Library for Advanced Deep Time Series Models.","allTopics":["deep-learning","time-series"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":17,"starsCount":4909,"forksCount":833,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T13:23:58.141Z"}},{"type":"Public","name":"HelmFluid","owner":"thuml","isFork":false,"description":"About code release of \"HelmFluid: Learning Helmholtz Dynamics for Interpretable Fluid Prediction\", ICML 2024. https://arxiv.org/pdf/2310.10565","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T16:13:58.134Z"}},{"type":"Public","name":"Multi-Embedding","owner":"thuml","isFork":false,"description":"Multi-Embedding for Recommendation to mitigate embedding collapse","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T07:39:13.752Z"}},{"type":"Public","name":"iVideoGPT","owner":"thuml","isFork":false,"description":"Official repo for \"iVideoGPT: Interactive VideoGPTs are Scalable World Models\", https://arxiv.org/abs/2405.15223","allTopics":["model-based-reinforcement-learning","video-prediction","visual-planning","world-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":38,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T07:00:13.841Z"}},{"type":"Public","name":"Transfer-Learning-Library","owner":"thuml","isFork":false,"description":"Transfer Learning Library for Domain Adaptation, Task Adaptation, and Domain Generalization","allTopics":["semi-supervised-learning","self-training","transfer-learning","image-translation","adversarial-learning","finetune","unsupervised-domain-adaptation","dann","out-of-distribution-generalization","deep-learning","domain-adaptation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":6,"starsCount":3225,"forksCount":540,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-10T06:59:42.213Z"}},{"type":"Public","name":"SimMTM","owner":"thuml","isFork":false,"description":"About Code release for \"SimMTM: A Simple Pre-Training Framework for Masked Time-Series Modeling\" (NeurIPS 2023 Spotlight), https://arxiv.org/abs/2302.00861","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":93,"forksCount":13,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-06T11:40:14.396Z"}},{"type":"Public","name":"TimesNet","owner":"thuml","isFork":false,"description":"About Code release for \"TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis\" (ICLR 2023), https://openreview.net/pdf?id=ju_Uqw384Oq","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":617,"forksCount":54,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-02T17:50:56.374Z"}},{"type":"Public","name":"Latent-Spectral-Models","owner":"thuml","isFork":false,"description":"About Code Release for \"Solving High-Dimensional PDEs with Latent Spectral Models\" (ICML 2023), https://arxiv.org/abs/2301.12664","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":54,"forksCount":5,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-28T14:11:43.816Z"}},{"type":"Public","name":"ContextWM","owner":"thuml","isFork":false,"description":"Code release for \"Pre-training Contextualized World Models with In-the-wild Videos for Reinforcement Learning\" (NeurIPS 2023), https://arxiv.org/abs/2305.18499","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":49,"forksCount":2,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-22T13:01:59.386Z"}},{"type":"Public","name":"Autoformer","owner":"thuml","isFork":false,"description":"About Code release for \"Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting\" (NeurIPS 2021), https://arxiv.org/abs/2106.13008","allTopics":["deep-learning","time-series"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":3,"starsCount":1789,"forksCount":395,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-15T08:58:09.706Z"}},{"type":"Public","name":"ForkMerge","owner":"thuml","isFork":false,"description":"Code release of paper \"ForkMerge: Mitigating Negative Transfer in Auxiliary-Task Learning\" (NeurIPS 2023)","allTopics":["multi-task-learning","auxiliary-task-learning"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":1,"starsCount":13,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-30T07:05:40.858Z"}},{"type":"Public","name":"Anomaly-Transformer","owner":"thuml","isFork":false,"description":"About Code release for \"Anomaly Transformer: Time Series Anomaly Detection with Association Discrepancy\" (ICLR 2022 Spotlight), https://openreview.net/forum?id=LzQQ89U1qm_","allTopics":["deep-learning","time-series","anomaly-detection"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":23,"starsCount":653,"forksCount":169,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-29T11:36:58.993Z"}},{"type":"Public","name":"learn_torch.compile","owner":"thuml","isFork":false,"description":"torch.compile artifacts for common deep learning models, can be used as a learning resource for torch.compile ","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":11,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-22T07:19:38.668Z"}},{"type":"Public","name":"Corrformer","owner":"thuml","isFork":false,"description":"About code release of \"Interpretable Weather Forecasting for Worldwide Stations with a Unified Deep Model\", Nature Machine Intelligence, 2023. https://www.nature.com/articles/s42256-023-00667-9","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":149,"forksCount":23,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-08T08:32:28.271Z"}},{"type":"Public","name":"predrnn-pytorch","owner":"thuml","isFork":false,"description":"Official implementation for NIPS'17 paper: PredRNN: Recurrent Neural Networks for Predictive Learning Using Spatiotemporal LSTMs.","allTopics":["recurrent-neural-networks","predictive-learning","video-prediction"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":6,"starsCount":416,"forksCount":103,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-04T01:54:31.943Z"}},{"type":"Public","name":"Flowformer","owner":"thuml","isFork":false,"description":"About Code release for \"Flowformer: Linearizing Transformers with Conservation Flows\" (ICML 2022), https://arxiv.org/pdf/2202.06258.pdf","allTopics":["deep-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":297,"forksCount":28,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-03T06:56:57.147Z"}},{"type":"Public","name":"LogME","owner":"thuml","isFork":false,"description":"Code release for \"LogME: Practical Assessment of Pre-trained Models for Transfer Learning\" (ICML 2021) and Ranking and Tuning Pre-trained Models: A New Paradigm for Exploiting Model Hubs (JMLR 2022)","allTopics":["deep-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":7,"starsCount":199,"forksCount":18,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-06T00:57:50.236Z"}},{"type":"Public","name":"CLIPood","owner":"thuml","isFork":false,"description":"About Code Release for \"CLIPood: Generalizing CLIP to Out-of-Distributions\" (ICML 2023), https://arxiv.org/abs/2302.00864","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":56,"forksCount":4,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-17T06:53:51.752Z"}},{"type":"Public","name":"MotionRNN","owner":"thuml","isFork":false,"description":"About Code release for \"MotionRNN: A Flexible Model for Video Prediction with Spacetime-Varying Motions\" (CVPR 2021) https://arxiv.org/abs/2103.02243","allTopics":["deep-learning","video-prediction"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":49,"forksCount":7,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-07-10T04:07:58.009Z"}},{"type":"Public","name":"MitNet","owner":"thuml","isFork":false,"description":"About Code Release for \"Estimating Heterogeneous Treatment Effects: Mutual Information Bounds and Learning Algorithms\" (ICML 2023)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":3,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-07-06T11:07:43.049Z"}}],"repositoryCount":79,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}