{"payload":{"header_redesign_enabled":false,"results":[{"id":"585707656","archived":false,"color":"#3572A5","followers":10,"has_funding_file":false,"hl_name":"robflynnyh/hydra-linear-attention","hl_trunc_description":"Implementation of: Hydra Attention: Efficient Attention with Many Heads (https://arxiv.org/abs/2209.07484)","language":"Python","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":585707656,"name":"hydra-linear-attention","owner_id":44401820,"owner_login":"robflynnyh","updated_at":"2023-01-08T11:39:40.303Z","has_issues":true}},"sponsorable":false,"topics":["machine-learning","transformers","attention","linear-attention","efficient-attention"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":73,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Arobflynnyh%252Fhydra-linear-attention%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/robflynnyh/hydra-linear-attention/star":{"post":"2dr6_7AA4eE8oFEq5eMOkBnFZu8yHC8MerzkpepWtmkPqYT8oTwLcU_i1ZVk5oAYsrlyAt_HChrLIzz66IaR7w"},"/robflynnyh/hydra-linear-attention/unstar":{"post":"B7NWsTvjXIVHGEe6wyAg60YCQ5T-6D7ih3Jn31x0qvWC89dWFTICBVMji2nxD2yGYr7I3TAdx5b9AHfAVYiUhw"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"1gooG0dcQizAYBiMPrg-F71OeatZnjACcqhPlp2e9kZZJQCNb4ZrDPs_zTRd5wK4QkC1Uj2fMDhmoHFlBgllmQ"}}},"title":"Repository search results"}