{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":336274588,"defaultBranch":"develop","name":"PaddleNLP","ownerLogin":"PaddlePaddle","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2021-02-05T13:07:42.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/23534030?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1715166012.0","currentOid":""},"activityList":{"items":[{"before":"85ba5735c38750fcb00847080a3eb2b23d04b825","after":"c1cfe631df59d6273f8736843c4142b1805a92d0","ref":"refs/heads/develop","pushedAt":"2024-05-28T10:52:56.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"modify block_multihead_attention api (#8456)\n\n* modify block_multihead_attention api\r\n\r\n* add param to blha\r\n\r\n* modify fused_transformer_layers\r\n\r\n* fix bug","shortMessageHtmlLink":"modify block_multihead_attention api (#8456)"}},{"before":"d4edd19194aef7ce8f9fee936395bc279b2fd606","after":"85ba5735c38750fcb00847080a3eb2b23d04b825","ref":"refs/heads/develop","pushedAt":"2024-05-28T02:12:40.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"[AutoParallel] Support qwen for auto_parallel (#8312)\n\n* run 4096\r\n\r\n* support 3d semi auto run of qwen model\r\n\r\n* [AutoParallel] Add qwen auto parallel.\r\n\r\n* Delete useless files.\r\n\r\n* Fix code style problem.\r\n\r\n* Fix problems.\r\n\r\n* Polish code.\r\n\r\n* Add CI-cases.\r\n\r\n* Add bf16 ci testcase.\r\n\r\n* Fix some problems.\r\n\r\n---------\r\n\r\nCo-authored-by: pangengzheng \r\nCo-authored-by: pangengzheng ","shortMessageHtmlLink":"[AutoParallel] Support qwen for auto_parallel (#8312)"}},{"before":"773497e2e63b7c434502bda582b4ba9479634e6e","after":"d4edd19194aef7ce8f9fee936395bc279b2fd606","ref":"refs/heads/develop","pushedAt":"2024-05-27T06:30:23.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"Fix rng_state in llm models (#8396)","shortMessageHtmlLink":"Fix rng_state in llm models (#8396)"}},{"before":"d8e1a6b2b6290a2f25b45ca830dcf0a642a69ad7","after":"773497e2e63b7c434502bda582b4ba9479634e6e","ref":"refs/heads/develop","pushedAt":"2024-05-27T04:51:20.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"quick fix from pretrained. (#8486)","shortMessageHtmlLink":"quick fix from pretrained. (#8486)"}},{"before":"a90f16353d32b9ed5f2a2418a449406ad7fc7612","after":"d8e1a6b2b6290a2f25b45ca830dcf0a642a69ad7","ref":"refs/heads/develop","pushedAt":"2024-05-27T02:54:28.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"【AutoParallel】Add split_backward for vpp (#8479)\n\n* add split_backward for vpp\r\n\r\n* polish\r\n\r\n* add test cast\r\n\r\n* polish\r\n\r\n* update test case\r\n\r\n* change the config\r\n\r\n* polish","shortMessageHtmlLink":"【AutoParallel】Add split_backward for vpp (#8479)"}},{"before":"0cd8fe7ee58180522168411606152dfe4ddda681","after":"a90f16353d32b9ed5f2a2418a449406ad7fc7612","ref":"refs/heads/develop","pushedAt":"2024-05-24T08:51:23.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"fix run_benchmark for llama2_70b in auto_parallel (#8484)\n\n* remove tsinghua pypi\r\n\r\n* modify gpt dateset addr for benchmark\r\n\r\n* fix run_benchmark for llama2_70b in auto_parallel","shortMessageHtmlLink":"fix run_benchmark for llama2_70b in auto_parallel (#8484)"}},{"before":"7aaa788069fc07ca109d5ac77e14603ed295f02d","after":"0cd8fe7ee58180522168411606152dfe4ddda681","ref":"refs/heads/develop","pushedAt":"2024-05-24T07:28:09.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"Revert \"Support Sharding Overlap (#8473)\" (#8491)\n\nThis reverts commit 7aaa788069fc07ca109d5ac77e14603ed295f02d.","shortMessageHtmlLink":"Revert \"Support Sharding Overlap (#8473)\" (#8491)"}},{"before":"8879f79f9857dc7831403064631ae32b0a0def23","after":"bbf945b64ab611e491c429ef86887cf84f43d3a5","ref":"refs/heads/release/2.8","pushedAt":"2024-05-24T05:43:46.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"Release/2.8 (#8437)\n\n* [XPU] llama add xpu support (#8282)\r\n\r\n* [XPU] llama add xpu support\r\n\r\n* fix\r\n\r\n* use try import\r\n\r\n* fix\r\n\r\n* refine\r\n\r\n* refine\r\n\r\n* refine\r\n\r\n* refine\r\n\r\n* update (#8399)\r\n\r\n* [LLM] Support fuse attention q, k, v weights (#8202)\r\n\r\n1. add use-interface & fuse action\r\n\r\n1.1. modify 1., code order\r\n\r\n2. switch to name_mapping\r\n\r\n3. solve tp branch\r\n\r\n3.2 follow hui, handel qkv separately\r\n\r\n3.3 handle pdparams\r\n\r\n3.4 from torch\r\n\r\n3.5 abandon low_cpu_mem_usage\r\n\r\n3.6 solve shard branch\r\n\r\n* 3.6.1 solve shard branch after rebase develop\r\n\r\n* code clean\r\n\r\n* remove debug comment\r\n\r\n* Redefine fuse and split functions\r\n\r\n* Redefine fuse and split functions\r\n\r\n* comment and fix\r\n\r\n* update method\r\n\r\n* update QKV fuse and split\r\n\r\n* support fuse weights in multi-files\r\n\r\n* add precision compare\r\n\r\n* simplify function call\r\n\r\n* support use_fast_ffn\r\n\r\n* clean modeling and configuration\r\n\r\n* add test for gpt and opt\r\n\r\n* fix tp_actions get\r\n\r\n* add fast_ffn test\r\n\r\n* add Qwen2Moe\r\n\r\n* Revert \"add Qwen2Moe\"\r\n\r\nThis reverts commit 113b8838a7c53f1d131928c30bf1071dfa583445.\r\n\r\n* add test for split\r\n\r\n* update doc\r\n\r\n* update filter_dict_keys\r\n\r\n---------\r\n\r\nCo-authored-by: Zii \r\n\r\n* [LLM] Fix fuse or split with same key (#8378)\r\n\r\n* fix fuse or split with same key\r\n\r\n* fix\r\n\r\n* fix eps\r\n\r\n* update format\r\n\r\n* [LLM] add decay steps option for finetuning (#8251)\r\n\r\n* [LLM] add memory stats to logger of trainer (#8269)\r\n\r\n* [Distributed] fix lora (#8325)\r\n\r\n* [LLM] fix lora target modules on llama (#8372)\r\n\r\n* [Distributed] metric calculation supports tp logits (#8370)\r\n\r\n* Update model_utils.py\r\n\r\n* Update model_utils.py\r\n\r\n* Update model_utils.py\r\n\r\n---------\r\n\r\nCo-authored-by: Jianbang Yang \r\nCo-authored-by: DrownFish19 \r\nCo-authored-by: Zii \r\nCo-authored-by: Tian <121000916+SylarTiaNII@users.noreply.github.com>","shortMessageHtmlLink":"Release/2.8 (#8437)"}},{"before":"c6f4159d9c972732e3798903a7f75b33eee7ac25","after":"7aaa788069fc07ca109d5ac77e14603ed295f02d","ref":"refs/heads/develop","pushedAt":"2024-05-24T02:17:21.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"Support Sharding Overlap (#8473)\n\n* update\r\n\r\n* update is_casual_mask to use_casual_mask\r\n\r\n* update by environment","shortMessageHtmlLink":"Support Sharding Overlap (#8473)"}},{"before":"7a24bccfd15348e818036c56335fccf984fd95d5","after":"8879f79f9857dc7831403064631ae32b0a0def23","ref":"refs/heads/release/2.8","pushedAt":"2024-05-23T13:09:14.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"ZHUI","name":"Zhong Hui","path":"/ZHUI","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16911935?s=80&v=4"},"commit":{"message":"quick fix from pretrained. (#8487)","shortMessageHtmlLink":"quick fix from pretrained. (#8487)"}},{"before":"a315046f4eecb0f5b8a776db1c73e0a9ff7cdc6a","after":"6b96825447ee5387d849812b13e8aa5ae23b5fa1","ref":"refs/heads/paddlenlp-dev-fleetv","pushedAt":"2024-05-23T09:09:36.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"sneaxiy","name":null,"path":"/sneaxiy","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/32832641?s=80&v=4"},"commit":{"message":"fix bug of sharding format (#8483)","shortMessageHtmlLink":"fix bug of sharding format (#8483)"}},{"before":"621118e4870bcce20b942329ba7a7ca7d3c94bc6","after":"c6f4159d9c972732e3798903a7f75b33eee7ac25","ref":"refs/heads/develop","pushedAt":"2024-05-23T03:26:21.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"[Ops] RoPE kernel support theta input (#8440)\n\n* [RoPE] feat: support theta\r\n\r\n* Update modeling.py\r\n\r\n* [RoPE] feat: support theta for qwen","shortMessageHtmlLink":"[Ops] RoPE kernel support theta input (#8440)"}},{"before":"70bffa8df3986d45bba0e4358adf4c4b92e04057","after":"621118e4870bcce20b942329ba7a7ca7d3c94bc6","ref":"refs/heads/develop","pushedAt":"2024-05-22T12:44:25.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"fix rotary_emb for llama (#8470)","shortMessageHtmlLink":"fix rotary_emb for llama (#8470)"}},{"before":"328d0a032714a02fa9bdd8709888f25be5e985f0","after":"a315046f4eecb0f5b8a776db1c73e0a9ff7cdc6a","ref":"refs/heads/paddlenlp-dev-fleetv","pushedAt":"2024-05-22T08:50:07.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"sneaxiy","name":null,"path":"/sneaxiy","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/32832641?s=80&v=4"},"commit":{"message":"Add release grad for SD (#8478)","shortMessageHtmlLink":"Add release grad for SD (#8478)"}},{"before":"87e4c4fa876a9eaa76d22860a6db3a7832d1322b","after":"70bffa8df3986d45bba0e4358adf4c4b92e04057","ref":"refs/heads/develop","pushedAt":"2024-05-22T05:58:25.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"[LLM] fix bug when loss is None in llama modeling.py (#8459)","shortMessageHtmlLink":"[LLM] fix bug when loss is None in llama modeling.py (#8459)"}},{"before":"08898bf1e0429db3da6d0b3e8a95e8b7d8c817d7","after":"7a24bccfd15348e818036c56335fccf984fd95d5","ref":"refs/heads/release/2.8","pushedAt":"2024-05-21T03:24:11.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"ZHUI","name":"Zhong Hui","path":"/ZHUI","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16911935?s=80&v=4"},"commit":{"message":"Cherry pick type promotion fix. (#8463)","shortMessageHtmlLink":"Cherry pick type promotion fix. (#8463)"}},{"before":"9064078e8d3c5fed6213a690a07e38c7a75d794f","after":"87e4c4fa876a9eaa76d22860a6db3a7832d1322b","ref":"refs/heads/develop","pushedAt":"2024-05-20T14:15:10.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"[NPU] support npu llama2-13B export & inference (#8442)\n\n* [NPU] support npu llama2-13B export & inference\r\n\r\n* move csrc_npu to csrc/npu","shortMessageHtmlLink":"[NPU] support npu llama2-13B export & inference (#8442)"}},{"before":"b36b6a01a6dd65ea45f4e8db0177a61c80ae410e","after":"9064078e8d3c5fed6213a690a07e38c7a75d794f","ref":"refs/heads/develop","pushedAt":"2024-05-20T11:29:07.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"ZHUI","name":"Zhong Hui","path":"/ZHUI","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16911935?s=80&v=4"},"commit":{"message":"Fix CI tests. (#8471)","shortMessageHtmlLink":"Fix CI tests. (#8471)"}},{"before":"fc860a3289804fbaf197d12c6d858d0d79e741af","after":"08898bf1e0429db3da6d0b3e8a95e8b7d8c817d7","ref":"refs/heads/release/2.8","pushedAt":"2024-05-20T05:05:30.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"ZHUI","name":"Zhong Hui","path":"/ZHUI","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16911935?s=80&v=4"},"commit":{"message":"Cherry-Pick fast_safe_open (#8458)\n\n* [Performance] Optimize unified checkpoint save/load speed. (#8204)\r\n\r\n* opt unified checkpoint save/load speed.","shortMessageHtmlLink":"Cherry-Pick fast_safe_open (#8458)"}},{"before":"d84cfe4e1515e790b67d04a5e01132e4679eac95","after":"328d0a032714a02fa9bdd8709888f25be5e985f0","ref":"refs/heads/paddlenlp-dev-fleetv","pushedAt":"2024-05-18T16:24:32.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"sneaxiy","name":null,"path":"/sneaxiy","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/32832641?s=80&v=4"},"commit":{"message":"fix bug in sync_param_name (#8465)","shortMessageHtmlLink":"fix bug in sync_param_name (#8465)"}},{"before":"14f66ddf60ec6fa5d9c59f71a24c8fdc72622aa6","after":"d84cfe4e1515e790b67d04a5e01132e4679eac95","ref":"refs/heads/paddlenlp-dev-fleetv","pushedAt":"2024-05-18T10:35:02.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"sneaxiy","name":null,"path":"/sneaxiy","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/32832641?s=80&v=4"},"commit":{"message":"sync param/grad/moment (#8464)","shortMessageHtmlLink":"sync param/grad/moment (#8464)"}},{"before":"5d4ce560ef3442b26ee544182d6a149aad351ba3","after":"b36b6a01a6dd65ea45f4e8db0177a61c80ae410e","ref":"refs/heads/develop","pushedAt":"2024-05-17T12:56:30.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"zhiqiu","name":"Leo Chen","path":"/zhiqiu","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/6888866?s=80&v=4"},"commit":{"message":"Add llama2-70b for test_tipc (#8455)","shortMessageHtmlLink":"Add llama2-70b for test_tipc (#8455)"}},{"before":"9381e9fe7821e49c148122d3addcf7ff33a0c5d8","after":"5d4ce560ef3442b26ee544182d6a149aad351ba3","ref":"refs/heads/develop","pushedAt":"2024-05-17T08:12:50.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"sijunhe","name":"Sijun He","path":"/sijunhe","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11987277?s=80&v=4"},"commit":{"message":"[benchmark]Add llama2 auto (#8424)\n\n* fix ci requirements\r\n\r\n* add llama2_auto benchmark\r\n\r\n* fix\r\n\r\n* update config\r\n\r\n* Add ips log for per card","shortMessageHtmlLink":"[benchmark]Add llama2 auto (#8424)"}},{"before":"d9dcd9a79725979668d5874e14420ee3cf75e2bf","after":"9381e9fe7821e49c148122d3addcf7ff33a0c5d8","ref":"refs/heads/develop","pushedAt":"2024-05-17T08:10:50.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"sijunhe","name":"Sijun He","path":"/sijunhe","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11987277?s=80&v=4"},"commit":{"message":"[bugfix] fix erniedoc (#8393)\n\n* fix erniedoc\r\n\r\n* revert","shortMessageHtmlLink":"[bugfix] fix erniedoc (#8393)"}},{"before":"7f13396332b080fbec21ad511ee22e60e4e38f56","after":"d9dcd9a79725979668d5874e14420ee3cf75e2bf","ref":"refs/heads/develop","pushedAt":"2024-05-17T07:26:21.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"[GCU] Support llama for GCU (#8445)","shortMessageHtmlLink":"[GCU] Support llama for GCU (#8445)"}},{"before":"daf2f3ae6b61264927e261e62cd6abb7555290c3","after":"7f13396332b080fbec21ad511ee22e60e4e38f56","ref":"refs/heads/develop","pushedAt":"2024-05-17T04:52:56.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"update (#8448)\n\n* update\r\n\r\n* add llama-npu-opt-script\r\n\r\n* Update dev_opt_lora.sh\r\n\r\n* Update dev_opt_ppt.sh\r\n\r\n* Update dev_opt_lora.sh\r\n\r\n* Update dev_opt_ppt.sh\r\n\r\n* Update dev_opt_sft.sh\r\n\r\n* Rename dev_opt_lora.sh to llama_npu_opt_lora.sh\r\n\r\n* Update dev_opt_ppt.sh\r\n\r\n* Rename dev_opt_ppt.sh to llama_npu_opt_ppt.sh\r\n\r\n* Update llama_npu_opt_lora.sh\r\n\r\n* Update and rename dev_opt_sft.sh to llama_npu_opt_sft.sh\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* add funsion ops\r\n\r\n* update\r\n\r\n* Update fusion_ops.py\r\n\r\n* update\r\n\r\n* update","shortMessageHtmlLink":"update (#8448)"}},{"before":"ebe397ea52c0f6b0d0241f32a2b244b6158eef17","after":"daf2f3ae6b61264927e261e62cd6abb7555290c3","ref":"refs/heads/develop","pushedAt":"2024-05-17T03:15:40.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"fix load rng compatiblity. (#8450)","shortMessageHtmlLink":"fix load rng compatiblity. (#8450)"}},{"before":"562229c1da5e853f898c85583f5092eea1df9ab5","after":"ebe397ea52c0f6b0d0241f32a2b244b6158eef17","ref":"refs/heads/develop","pushedAt":"2024-05-16T10:37:30.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"Support fused_attention_qkv for auto_parallel llama (#8432)\n\n* add\r\n\r\n* add\r\n\r\n* add\r\n\r\n* add\r\n\r\n* add","shortMessageHtmlLink":"Support fused_attention_qkv for auto_parallel llama (#8432)"}},{"before":"21eb6bdcac0d52c18fb8525dad6a2bb3680050d4","after":"562229c1da5e853f898c85583f5092eea1df9ab5","ref":"refs/heads/develop","pushedAt":"2024-05-16T09:15:33.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"wawltor","name":null,"path":"/wawltor","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16698950?s=80&v=4"},"commit":{"message":"Fix/test gpu (#8452)\n\n* fix safe open.\r\n\r\n* fix online ci.","shortMessageHtmlLink":"Fix/test gpu (#8452)"}},{"before":"debb2ad92d08825f553818e88b971245b50d2433","after":"fc860a3289804fbaf197d12c6d858d0d79e741af","ref":"refs/heads/release/2.8","pushedAt":"2024-05-16T09:09:22.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"ZHUI","name":"Zhong Hui","path":"/ZHUI","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/16911935?s=80&v=4"},"commit":{"message":"Fix load RNG compatibility. (#8451)","shortMessageHtmlLink":"Fix load RNG compatibility. (#8451)"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEVbyo2gA","startCursor":null,"endCursor":null}},"title":"Activity · PaddlePaddle/PaddleNLP"}