{"payload":{"header_redesign_enabled":false,"results":[{"id":"48015681","archived":false,"color":"#000080","followers":376,"has_funding_file":false,"hl_name":"GT-Vision-Lab/VQA_LSTM_CNN","hl_trunc_description":"Train a deeper LSTM and normalized CNN Visual Question Answering model. This current code can get 58.16 on OpenEnded and 63.09 on Multipl…","language":"Lua","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":48015681,"name":"VQA_LSTM_CNN","owner_id":12704830,"owner_login":"GT-Vision-Lab","updated_at":"2019-03-22T19:37:13.052Z","has_issues":true}},"sponsorable":false,"topics":[],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":69,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253AGT-Vision-Lab%252FVQA_LSTM_CNN%2B%2Blanguage%253ALua","metadata":null,"warn_limited_results":false,"csrf_tokens":{"/GT-Vision-Lab/VQA_LSTM_CNN/star":{"post":"Ldyk6PWeMg6dsNUzTLwNYdBLBR3HWnwrGUiz9p0Zj-OItmN36ES8fRtlqiEwrnh-Si0vikbzaVJ4UuCaB_-hpQ"},"/GT-Vision-Lab/VQA_LSTM_CNN/unstar":{"post":"bxAMqI7v4-9amWptEgcyNtEC2ncYcy-E-nnAAf7CvUSxsoJ_7MKcg6wmXHijiHy5PVWgnZ2T37WlWjnGCZCtuQ"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"X10WHJzTl1ZK-ks3nqhMQr66ByhqBksBkGRT5hCK-EtOT0hEbSVopYOPvqGePFoIfjW1kbmejCQxgftDEDpsbA"}}},"title":"Repository search results"}