{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":797172087,"defaultBranch":"main","name":"VILA","ownerLogin":"michael-heinrich","currentUserCanPush":false,"isFork":true,"isEmpty":false,"createdAt":"2024-05-07T10:33:00.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/33396188?v=4","public":true,"private":false,"isOrgOwned":false},"refInfo":{"name":"","listCacheKey":"v0:1715079523.0","currentOid":""},"activityList":{"items":[{"before":null,"after":"b81dfe4ce14df72a605aea2e256f3e8dd7d18783","ref":"refs/heads/older_gpus","pushedAt":"2024-05-07T10:58:43.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"michael-heinrich","name":null,"path":"/michael-heinrich","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/33396188?s=80&v=4"},"commit":{"message":"Fix for backwards compatibility\nChose LlamaAttention instead of LlamaFlashAttention2, if flash attention is not supported by the GPU architecture.","shortMessageHtmlLink":"Fix for backwards compatibility"}},{"before":"f85297fa156e85d95bab4b1c7c2f01283866e759","after":"8be8f8c531e5f946928e70e871b644346bd704cc","ref":"refs/heads/main","pushedAt":"2024-05-07T10:53:42.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"michael-heinrich","name":null,"path":"/michael-heinrich","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/33396188?s=80&v=4"},"commit":{"message":"Merge branch 'main' of https://github.com/Efficient-Large-Model/VILA","shortMessageHtmlLink":"Merge branch 'main' of https://github.com/Efficient-Large-Model/VILA"}}],"hasNextPage":false,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"Y3Vyc29yOnYyOpK7MjAyNC0wNS0wN1QxMDo1ODo0My4wMDAwMDBazwAAAARDdzdb","startCursor":"Y3Vyc29yOnYyOpK7MjAyNC0wNS0wN1QxMDo1ODo0My4wMDAwMDBazwAAAARDdzdb","endCursor":"Y3Vyc29yOnYyOpK7MjAyNC0wNS0wN1QxMDo1Mzo0Mi4wMDAwMDBazwAAAARDdiEx"}},"title":"Activity ยท michael-heinrich/VILA"}