{"payload":{"pageCount":4,"repositories":[{"type":"Public","name":"CogVLM2","owner":"THUDM","isFork":false,"description":"GPT4V-level open-source multi-modal model based on Llama3-8B","allTopics":["pretrained-models","language-model","multi-modal","cogvlm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":29,"starsCount":1478,"forksCount":80,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,29,15,2,6,2,1,0,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-09T12:50:58.848Z"}},{"type":"Public","name":"CodeGeeX4","owner":"THUDM","isFork":false,"description":"CodeGeeX4-ALL-9B, a versatile model for all AI software development scenarios, including code completion, code interpreter, web search, function calling, repository-level Q&A and much more.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":8,"starsCount":411,"forksCount":31,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,11],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-09T12:16:57.040Z"}},{"type":"Public","name":"GLM-4","owner":"THUDM","isFork":false,"description":"GLM-4 series: Open Multilingual Multimodal Chat LMs | 开源多语言多模态对话模型","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":3553,"forksCount":255,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,49,12,11,7,11],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-09T10:14:02.267Z"}},{"type":"Public","name":"LVBench","owner":"THUDM","isFork":false,"description":"LVBench: An Extreme Long Video Understanding Benchmark","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":33,"forksCount":1,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-09T08:29:00.384Z"}},{"type":"Public","name":"CodeGeeX","owner":"THUDM","isFork":false,"description":"CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)","allTopics":["tools","code-generation","pretrained-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":152,"starsCount":7965,"forksCount":573,"license":"Apache License 2.0","participation":[0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-07T10:49:15.638Z"}},{"type":"Public","name":"CodeGeeX2","owner":"THUDM","isFork":false,"description":"CodeGeeX2: A More Powerful Multilingual Code Generation Model","allTopics":["tool","code","code-generation","pretrained-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":207,"starsCount":7600,"forksCount":535,"license":"Apache License 2.0","participation":[0,2,37,10,6,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-07T10:48:01.073Z"}},{"type":"Public","name":"RecDCL","owner":"THUDM","isFork":false,"description":"RecDCL: Dual Contrastive Learning for Recommendation (WWW'24, Oral)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":14,"forksCount":2,"license":"MIT License","participation":[0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-06T02:27:39.096Z"}},{"type":"Public","name":"Inf-DiT","owner":"THUDM","isFork":false,"description":"Official implementation of Inf-DiT: Upsampling Any-Resolution Image with Memory-Efficient Diffusion Transformer","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":12,"starsCount":298,"forksCount":12,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,1,1,0,0,1,0,0,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T10:35:10.090Z"}},{"type":"Public","name":"CogCoM","owner":"THUDM","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":13,"starsCount":140,"forksCount":9,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,3,3,8,0,1,0,0,0,0,0,0,0,0,0,0,12,0,0,5,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T09:56:19.915Z"}},{"type":"Public","name":"ChatGLM3","owner":"THUDM","isFork":false,"description":"ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":12,"starsCount":13037,"forksCount":1504,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,29,45,24,22,26,5,12,5,13,14,19,8,10,5,7,7,2,15,15,7,0,5,12,8,11,1,7,0,7,3,7,3,4,4,7,4,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-04T06:00:28.575Z"}},{"type":"Public","name":"P2TAG","owner":"THUDM","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-02T06:10:53.482Z"}},{"type":"Public","name":"VisualGLM-6B","owner":"THUDM","isFork":false,"description":"Chinese and English multimodal conversational language model | 多模态中英双语对话语言模型","allTopics":["gpt","multi-modal","chatglm-6b"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":257,"starsCount":4036,"forksCount":416,"license":"Apache License 2.0","participation":[2,0,2,0,2,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-28T06:19:24.861Z"}},{"type":"Public","name":"ChatGLM-6B","owner":"THUDM","isFork":false,"description":"ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":41,"issueCount":545,"starsCount":39984,"forksCount":5152,"license":"Apache License 2.0","participation":[3,1,1,0,0,0,0,2,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-27T04:05:25.542Z"}},{"type":"Public","name":"ChatGLM2-6B","owner":"THUDM","isFork":false,"description":"ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型","allTopics":["large-language-models","llm","chatglm","chatglm-6b"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":26,"issueCount":425,"starsCount":15617,"forksCount":1852,"license":"Other","participation":[11,2,5,0,0,0,2,1,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,3,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-27T04:05:08.884Z"}},{"type":"Public","name":"paper-source-trace","owner":"THUDM","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":12,"forksCount":1,"license":null,"participation":[0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,3,0,2,0,5,3,2,3,2,1,1,1,0,0,0,1,0,0,11,0,0,0,0,0,0,0,0,0,3,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-21T05:02:05.625Z"}},{"type":"Public","name":"AutoRE","owner":"THUDM","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":7,"starsCount":36,"forksCount":4,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,9,0,1,51,2,0,0,3,0,3,0,8,0,12,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-20T15:44:12.539Z"}},{"type":"Public","name":"AgentBench","owner":"THUDM","isFork":false,"description":"A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)","allTopics":["gpt-4","llm","chatgpt","llm-agent"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":33,"starsCount":2003,"forksCount":136,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,13,17,3,1,0,0,0,0,0,1,5,0,0,0,0,0,2,0,0,3,3,0,0,0,0,0,0,0,1,0,0,0,2,1,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-19T13:48:34.946Z"}},{"type":"Public","name":"AlignBench","owner":"THUDM","isFork":false,"description":"大模型多维度中文对齐评测基准 (ACL 2024)","allTopics":["large-language-models","llm","chatgpt","chatglm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":12,"starsCount":253,"forksCount":20,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,2,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-17T12:36:31.491Z"}},{"type":"Public","name":"MSAGPT","owner":"THUDM","isFork":false,"description":"MSAGPT","allTopics":["protein-language-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":13,"forksCount":1,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,5,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-12T10:58:48.090Z"}},{"type":"Public","name":"SwissArmyTransformer","owner":"THUDM","isFork":false,"description":"SwissArmyTransformer is a flexible and powerful library to develop your own Transformer variants.","allTopics":["transformer","pretrained-models","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":35,"starsCount":867,"forksCount":83,"license":"Apache License 2.0","participation":[21,4,1,0,9,6,1,3,10,1,4,9,2,0,4,2,6,3,2,5,0,4,0,6,7,1,17,20,10,4,0,0,2,2,1,2,2,3,6,1,0,0,1,0,2,0,0,0,1,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-12T06:50:17.348Z"}},{"type":"Public","name":"ReST-MCTS","owner":"THUDM","isFork":false,"description":"ReST-MCTS*: LLM Self-Training via Process Reward Guided Tree Search","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":45,"forksCount":3,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T02:03:03.462Z"}},{"type":"Public","name":"CogVLM","owner":"THUDM","isFork":false,"description":"a state-of-the-art-level open visual language model | 多模态预训练模型","allTopics":["pretrained-models","language-model","multi-modal","cross-modality","visual-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":29,"starsCount":5617,"forksCount":391,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,1,26,14,3,2,1,4,6,5,5,12,51,19,7,11,0,0,2,2,3,0,0,0,0,2,0,0,3,0,1,0,0,0,2,0,2,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T10:01:33.231Z"}},{"type":"Public","name":"OAG-AQA","owner":"THUDM","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":0,"starsCount":6,"forksCount":5,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,9,5,0,0,0,0,0,0,0,0,0,0,0,0,11,0,1,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T11:54:55.207Z"}},{"type":"Public","name":"NaturalCodeBench","owner":"THUDM","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":40,"forksCount":1,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,9,0,0,6,0,0,0,0,7,0,5,1,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-20T03:18:18.520Z"}},{"type":"Public","name":"Megatron-LM","owner":"THUDM","isFork":true,"description":"Ongoing research training transformer models at scale","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":2105,"license":"Other","participation":[30,56,13,42,34,65,63,41,53,45,30,58,98,112,128,96,35,42,76,24,37,45,39,30,9,21,48,43,57,42,26,17,13,22,34,46,72,48,67,65,50,13,24,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-06T07:54:45.673Z"}},{"type":"Public","name":"AutoWebGLM","owner":"THUDM","isFork":false,"description":"An LLM-based Web Navigating Agent (KDD'24)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":9,"starsCount":525,"forksCount":41,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-05T13:21:33.386Z"}},{"type":"Public","name":"ScenarioMeta","owner":"THUDM","isFork":false,"description":"Source code and dataset for KDD 2019 paper \"Sequential Scenario-Specific Meta Learner for Online Recommendation\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":2,"starsCount":81,"forksCount":11,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-03T19:44:30.025Z"}},{"type":"Public","name":"kgTransformer","owner":"THUDM","isFork":false,"description":"kgTransformer: pre-training for reasoning over complex KG queries (KDD 22)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":83,"forksCount":12,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-29T18:43:45.602Z"}},{"type":"Public","name":"RelayDiffusion","owner":"THUDM","isFork":false,"description":"The official implementation of \"Relay Diffusion: Unifying diffusion process across resolutions for image synthesis\" [ICLR 2024 Spotlight]","allTopics":["machine-learning","generative-model","image-synthesis","diffusion-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":247,"forksCount":19,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,1,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-29T09:29:51.054Z"}},{"type":"Public","name":"tot-prediction","owner":"THUDM","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":1,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-24T16:23:20.121Z"}}],"repositoryCount":109,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"THUDM repositories"}