{"payload":{"header_redesign_enabled":false,"results":[{"id":"200340627","archived":false,"color":"#3572A5","followers":640,"has_funding_file":false,"hl_name":"daiquocnguyen/Graph-Transformer","hl_trunc_description":"Universal Graph Transformer Self-Attention Networks (TheWebConf WWW 2022) (Pytorch and Tensorflow)","language":"Python","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":200340627,"name":"Graph-Transformer","owner_id":28601433,"owner_login":"daiquocnguyen","updated_at":"2022-08-16T12:35:43.493Z","has_issues":true}},"sponsorable":false,"topics":["text-classification","transformer","graph-machine-learning","graph-embeddings","graph-classification","self-attention","graph-neural-networks","graph-representation-learning","transformer-models","node-embeddings","graph-deep-learning","graph-transformer"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":60,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Adaiquocnguyen%252FGraph-Transformer%2B%2Blanguage%253APython","metadata":null,"warn_limited_results":false,"csrf_tokens":{"/daiquocnguyen/Graph-Transformer/star":{"post":"XiayyxqXmrGfOs4bsVp7sTX4k3mUe2zYyh0Vgf9B_FYYXLMv7zH-Nfs4skMa6IBVNtaP0GjdB-Qr9TQg6IIjWw"},"/daiquocnguyen/Graph-Transformer/unstar":{"post":"KcKxnNZzlIi-51f6eRmgOfb6hmzAcD7hqtTISrXQib2kW_sqtQGsSQW58ggyHpaXf4ahNDEW7XDnawfWy3A35w"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"Qo7eU4XT3nh1eleqyTM_81V8agZCB7r0tfe2T0IztWXERkjJlLFyLpqoyVGrSqtnlYeaYUvuPs8fOSrwok5WRg"}}},"title":"Repository search results"}