{"payload":{"header_redesign_enabled":false,"results":[{"id":"246357476","archived":true,"color":"#DA5B0B","followers":4037,"has_funding_file":false,"hl_name":"google-research/simclr","hl_trunc_description":"SimCLRv2 - Big Self-Supervised Models are Strong Semi-Supervised Learners","language":"Jupyter Notebook","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":246357476,"name":"simclr","owner_id":43830688,"owner_login":"google-research","updated_at":"2023-05-22T04:42:53.898Z","has_issues":true}},"sponsorable":false,"topics":["computer-vision","representation-learning","unsupervised-learning","self-supervised-learning","simclr","contrastive-learning","simclrv2"],"type":"Public archive","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false},{"id":"385275507","archived":false,"color":"#3572A5","followers":3206,"has_funding_file":false,"hl_name":"google-research/scenic","hl_trunc_description":"Scenic: A Jax Library for Computer Vision Research and Beyond","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":385275507,"name":"scenic","owner_id":43830688,"owner_login":"google-research","updated_at":"2024-08-13T23:07:16.940Z","has_issues":true}},"sponsorable":false,"topics":["research","computer-vision","deep-learning","transformers","attention","jax","vision-transformer"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false},{"id":"192635589","archived":false,"color":"#3572A5","followers":2170,"has_funding_file":false,"hl_name":"google-research/uda","hl_trunc_description":"Unsupervised Data Augmentation (UDA)","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":192635589,"name":"uda","owner_id":43830688,"owner_login":"google-research","updated_at":"2021-08-28T07:16:56.630Z","has_issues":true}},"sponsorable":false,"topics":["nlp","natural-language-processing","computer-vision","tensorflow","cv","semi-supervised-learning"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false},{"id":"470020023","archived":true,"color":"#3572A5","followers":987,"has_funding_file":false,"hl_name":"google-research/maxim","hl_trunc_description":"[CVPR 2022 Oral] Official repository for \"MAXIM: Multi-Axis MLP for Image Processing\". SOTA for denoising, deblurring, deraining, dehazin…","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":470020023,"name":"maxim","owner_id":43830688,"owner_login":"google-research","updated_at":"2023-06-09T18:15:50.194Z","has_issues":true}},"sponsorable":false,"topics":["image","computer-vision","architecture","image-processing","transformer","mlp","enhancement","image-restoration","restoration","deblurring","denoising","dehazing","image-enhancement","low-level-vision","deraining","retouching"],"type":"Public archive","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false},{"id":"467614064","archived":false,"color":"#DA5B0B","followers":851,"has_funding_file":false,"hl_name":"google-research/pix2seq","hl_trunc_description":"Pix2Seq codebase: multi-tasks with generative modeling (autoregressive and diffusion)","language":"Jupyter Notebook","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":467614064,"name":"pix2seq","owner_id":43830688,"owner_login":"google-research","updated_at":"2023-11-07T08:25:57.509Z","has_issues":true}},"sponsorable":false,"topics":["computer-vision","deep-learning","object-detection","tensorflow2","vision-language","pix2seq"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false},{"id":"321489256","archived":false,"color":"#3572A5","followers":549,"has_funding_file":false,"hl_name":"google-research/ravens","hl_trunc_description":"Train robotic agents to learn pick and place with deep learning for vision-based manipulation in PyBullet. Transporter Nets, CoRL 2020.","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":321489256,"name":"ravens","owner_id":43830688,"owner_login":"google-research","updated_at":"2024-07-30T21:24:12.465Z","has_issues":true}},"sponsorable":false,"topics":["reinforcement-learning","computer-vision","deep-learning","robotics","tensorflow","openai-gym","artificial-intelligence","vision","manipulation","imitation-learning","pybullet","rearrangement","pick-and-place","transporter-nets"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false},{"id":"511682460","archived":true,"color":"#DA5B0B","followers":434,"has_funding_file":false,"hl_name":"google-research/maxvit","hl_trunc_description":"[ECCV 2022] Official repository for \"MaxViT: Multi-Axis Vision Transformer\". SOTA foundation models for classification, detection, segmen…","language":"Jupyter Notebook","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":511682460,"name":"maxvit","owner_id":43830688,"owner_login":"google-research","updated_at":"2023-06-02T00:28:27.326Z","has_issues":true}},"sponsorable":false,"topics":["image","computer-vision","architecture","image-processing","cnn","transformer","classification","segmentation","resnet","object-detection","mlp","transformer-architecture","vision-transformer"],"type":"Public archive","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false},{"id":"224050000","archived":false,"color":"#3572A5","followers":313,"has_funding_file":false,"hl_name":"google-research/rigl","hl_trunc_description":"End-to-end training of sparse deep neural networks with little-to-no performance loss. ","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":224050000,"name":"rigl","owner_id":43830688,"owner_login":"google-research","updated_at":"2023-01-26T17:47:14.930Z","has_issues":true}},"sponsorable":false,"topics":["machine-learning","computer-vision","neural-networks","sparse-training"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false},{"id":"507114594","archived":true,"color":"#DA5B0B","followers":38,"has_funding_file":false,"hl_name":"google-research/se3ds","hl_trunc_description":"This repository hosts the code for our paper, \"Simple and Effective Synthesis of Indoor 3D Scenes\".","language":"Jupyter Notebook","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":507114594,"name":"se3ds","owner_id":43830688,"owner_login":"google-research","updated_at":"2022-07-01T15:19:24.728Z","has_issues":true}},"sponsorable":false,"topics":["machine-learning","computer-vision","tensorflow"],"type":"Public archive","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false},{"id":"331371706","archived":false,"color":"#f1e05a","followers":18,"has_funding_file":false,"hl_name":"google-research/pangea","hl_trunc_description":"Panoramic Graph Environment Annotation toolkit, for collecting audio and text annotations in panoramic graph environments such as Matterp…","language":"JavaScript","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":331371706,"name":"pangea","owner_id":43830688,"owner_login":"google-research","updated_at":"2021-03-05T20:35:57.531Z","has_issues":true}},"sponsorable":false,"topics":["nlp","computer-vision","crowdsourcing","annotation-tool"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":186,"errors":[],"result_count":10,"facets":[{"kind":"FACET_KIND_LANGUAGE","entries":[{"name":"Python","language_color":"#3572A5","query":"language:Python"},{"name":"Jupyter Notebook","language_color":"#DA5B0B","query":"language:Jupyter Notebook"},{"name":"JavaScript","language_color":"#f1e05a","query":"language:JavaScript"}]}],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Dtopic%253Acomputer-vision%2Borg%253Agoogle-research%2Bfork%253Atrue%26type%3Drepositories","metadata":null,"warn_limited_results":false,"csrf_tokens":{"/google-research/simclr/star":{"post":"GqC0c8yNxia_iCBlaDwPq1PeYYsDU8dzoBhHZkJBPMZf_RtFuz1-q2_QFQzvCXjP7VuR1AdFO43ivxkobQRc8g"},"/google-research/simclr/unstar":{"post":"9DScKOIz7wKzZrniuEGl6PiV8EnpN8OppVLeWI4i6O6n4XcsM5yBSOOjxEkApmje2qwCNeqIWInBdXjZY7PnnA"},"/google-research/scenic/star":{"post":"K5XtiKC5pLbEnLCZX3QONOwUGLNNKhvO23Zhpe9cFJv1NWPk3rHL2QQCB8syHFaMF4Q88ayCaVUpVmaR0VhUNQ"},"/google-research/scenic/unstar":{"post":"QYeU-xUBBLHrw2Xb66b--_lfZp9l3LsStpjoza085WLqpDMZDf_9gJfmC-YxXWRPu1PfQlTZdKQlP8nRIJyUqQ"},"/google-research/uda/star":{"post":"xQ-QPH-brovZYc5OEKJv8fIliJS0rJK9adri5moI3m1p7CUgkG1d79PBKW4E_GKYMoK6t9NE0Peq78PAuIq6lw"},"/google-research/uda/unstar":{"post":"IivcIEqbD30QYvdCJEeIh8Dr_w5RJIBbizme3WNp5Q4HUGIUKv3IBIrJ0F5OwcPXVQe11zo61YG3J5prnZrADw"},"/google-research/maxim/star":{"post":"mQyfhecKJIdQ83ek8K_HJ8UbEySwjTStpTVjrW49C5vEg9D5wP7szbTuolcKVD_iuBK_3AVdjnISpsPX6qMa3Q"},"/google-research/maxim/unstar":{"post":"b97VbCqnMcOOrc4kRTxiXb7VdE53Fqa7krC9ggbR_Ee1fBgH3tLSE6eKo7e9iVmELTZruGT9vOXtKapWMFz1wA"},"/google-research/pix2seq/star":{"post":"3FDXD99W69HE_QiJ0mPkOdJ4YrygfDejpBiqXGuaqtSvirFIIDUAXcVvtCuX0b7FR3y0sAyjIvhO0O3uzobZcA"},"/google-research/pix2seq/unstar":{"post":"L3AHR7odLQ0N9bn8VC4C5PIxz62-2MV0Z3PomVmgDNKrZacfdEMc-kFXPtYK8ykraGjmB413J7TkhgItruxu0g"},"/google-research/ravens/star":{"post":"UtfGtCPtXtNTMCEW9cVZ65zsRQ4uzGd2XXf9dLleofEn1VZkw3xarAvaa7PpYC-hbg_y4eH_rKuxO43VcKqepQ"},"/google-research/ravens/unstar":{"post":"5yKGW1IZj0pHWHA2HMkp-3dG46pUfz5b2A9Hmsj-jZLOFI2hMemBReS7WcXXj2RNOderkErNZ3E8_nwE2NMkPA"},"/google-research/maxvit/star":{"post":"jFIC4lycEok12OrRDItPU8Kw96E_Np8zt3-5o9Y-AfyQJDsZPZ7zaCXf8XOqV2X4ZrTEsutNtpZIk76E3zsbqw"},"/google-research/maxvit/unstar":{"post":"8AwXI9YfAvSNCgQK7_jzqMPI77oXNlBDRj7qlD7ui_HB_fGKLttiEQKse4bUbacQsj1Wer34rsk52O2QNCHEcw"},"/google-research/rigl/star":{"post":"boZ6hzoElLtE0x3783EoPwJfbm-LVyitYTGii89T74jaWmUjz_JHRrQ81CDu23Zwa0MX-GhFzqA5ky1Wm2KCfw"},"/google-research/rigl/unstar":{"post":"5mOpYuQVuwyrIOCtoMPWte6KNuRMIQ01Tl3PEMNnSrOsG77JCWpA0g93XjPqQuJ2HAV6BU7JEXkX27f3gbitYw"},"/google-research/se3ds/star":{"post":"dXLEtjUgD8ETZho7E1SWeRUvrxB7ZBlP9Uf2TyiGrvxs0yKpBN5bcobIAHdUaSlfL7auzVl2AYFERXjGuFLwxQ"},"/google-research/se3ds/unstar":{"post":"8x2DHHi36IcL0AFA-Ao5DwEVuPuqmFQ_t_8Dg7issyxxicoJje3GVb_Ik37UOYs94KYagcBbzFuoAfkpZcEE0Q"},"/google-research/pangea/star":{"post":"qCytk4vZ_bpnta5CFavN-nx4yPZXtESfD4U6Mc4bsM8bY3_a7n1B4bHCP7RN-c10iKZqQSV8RZKqGcD-pBHoFQ"},"/google-research/pangea/unstar":{"post":"O61KiFjjNlPiPHcsKeoMO-w29kqrJHbS7MM6dBwhqd_PqTKdVJRKfFuDna3PfU9-iszbPYJDYpkgUFSj9bE2og"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"QtMMElpboH52uOKugCMUyF96nOZ2q7ZfZe4z53jVMy4PnAZ4cnWTNkelV9_x5RitcxuKFwanELcE7-Q0hkWvGA"}}},"title":"Repository search results"}