{"payload":{"pageCount":4,"repositories":[{"type":"Public","name":"NeMo","owner":"NVIDIA","isFork":false,"description":"A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)","topicNames":["machine-translation","tts","speech-synthesis","neural-networks","deeplearning","speaker-recognition","asr","multimodal","speech-translation","large-language-models"],"topicsNotShown":2,"allTopics":["machine-translation","tts","speech-synthesis","neural-networks","deeplearning","speaker-recognition","asr","multimodal","speech-translation","large-language-models","speaker-diariazation","generative-ai"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":93,"issueCount":52,"starsCount":10290,"forksCount":2193,"license":"Apache License 2.0","participation":[26,19,9,18,19,8,22,18,10,17,25,17,19,18,28,9,19,34,21,23,21,13,9,5,13,5,7,10,20,9,4,13,17,23,21,10,15,30,29,30,20,37,26,10,26,36,28,22,40,31,32,17],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T10:18:12.283Z"}},{"type":"Public","name":"NeMo-Aligner","owner":"NVIDIA","isFork":false,"description":"Scalable toolkit for efficient model alignment","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":22,"issueCount":42,"starsCount":283,"forksCount":31,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,9,2,12,3,1,3,1,1,2,0,7,0,3,0,2,3,2,3,2,2,4,3,3,2,5,0,1,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T06:26:15.850Z"}},{"type":"Public","name":"cloudai","owner":"NVIDIA","isFork":false,"description":"CloudAI Benchmark Framework","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":10,"issueCount":0,"starsCount":11,"forksCount":7,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,48,24],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T11:12:26.591Z"}},{"type":"Public","name":"earth2studio","owner":"NVIDIA","isFork":false,"description":"Open-source deep-learning framework for exploring, building and deploying AI weather/climate workflows.","topicNames":["weather","ai","deep-learning","climate-science"],"topicsNotShown":0,"allTopics":["weather","ai","deep-learning","climate-science"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":30,"forksCount":4,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,21,8,5,2,10,2,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T06:10:10.159Z"}},{"type":"Public","name":"TransformerEngine","owner":"NVIDIA","isFork":false,"description":"A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.","topicNames":["python","machine-learning","deep-learning","gpu","cuda","pytorch","jax","fp8"],"topicsNotShown":0,"allTopics":["python","machine-learning","deep-learning","gpu","cuda","pytorch","jax","fp8"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":30,"issueCount":95,"starsCount":1498,"forksCount":230,"license":"Apache License 2.0","participation":[4,10,7,12,3,2,11,10,5,5,10,8,4,2,3,2,9,6,10,9,9,4,3,9,7,2,13,4,13,0,0,11,5,12,10,12,7,5,6,8,7,5,6,3,13,5,14,9,9,6,9,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T01:34:57.865Z"}},{"type":"Public","name":"warp","owner":"NVIDIA","isFork":false,"description":"A Python framework for high performance GPU simulation and graphics","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":8,"issueCount":46,"starsCount":1736,"forksCount":146,"license":"Other","participation":[29,24,19,51,33,34,7,46,11,8,10,39,17,32,27,1,15,31,11,15,39,55,38,32,23,12,16,16,30,11,0,9,36,31,22,9,7,18,39,46,47,44,25,4,15,8,41,40,38,27,25,30],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-27T22:11:08.198Z"}},{"type":"Public","name":"Megatron-LM","owner":"NVIDIA","isFork":false,"description":"Ongoing research training transformer models at scale","topicNames":["transformers","model-para","large-language-models"],"topicsNotShown":0,"allTopics":["transformers","model-para","large-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":127,"issueCount":308,"starsCount":8896,"forksCount":1995,"license":"Other","participation":[35,29,3,14,15,46,30,56,13,42,34,65,63,41,53,45,30,58,98,112,128,96,35,42,76,24,37,45,39,30,9,21,48,43,57,43,26,17,13,22,34,47,74,48,67,65,51,11,38,14,26,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-27T11:42:20.265Z"}},{"type":"Public","name":"NeMo-Curator","owner":"NVIDIA","isFork":false,"description":"Scalable toolkit for data curation","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":24,"starsCount":269,"forksCount":27,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,5,1,4,1,2,4,5,4,1,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-26T08:58:56.218Z"}},{"type":"Public","name":"NeMo-Guardrails","owner":"NVIDIA","isFork":false,"description":"NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":22,"issueCount":159,"starsCount":3538,"forksCount":303,"license":"Other","participation":[5,9,8,56,39,8,29,28,27,18,14,17,18,45,68,23,46,37,74,58,52,67,46,42,73,54,29,44,39,20,3,26,35,27,24,48,49,55,39,41,25,58,75,34,1,10,36,3,5,14,3,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-25T16:26:46.768Z"}},{"type":"Public","name":"NVFlare","owner":"NVIDIA","isFork":false,"description":"NVIDIA Federated Learning Application Runtime Environment","topicNames":["python"],"topicsNotShown":0,"allTopics":["python"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":12,"issueCount":25,"starsCount":546,"forksCount":147,"license":"Apache License 2.0","participation":[16,12,3,9,0,11,8,10,11,14,10,12,11,11,13,16,23,18,4,6,8,14,16,14,8,7,13,8,13,17,4,13,9,5,11,13,3,1,5,1,7,1,11,2,13,8,16,2,13,8,6,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T22:42:51.159Z"}},{"type":"Public","name":"Megatron-Energon","owner":"NVIDIA","isFork":false,"description":"Megatron's multi-modal data loader","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":0,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,13,0,2,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T19:42:03.102Z"}},{"type":"Public","name":"GenerativeAIExamples","owner":"NVIDIA","isFork":false,"description":"Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.","topicNames":["microservice","gpu-acceleration","nemo","tensorrt","rag","triton-inference-server","large-language-models","llm","llm-inference","retrieval-augmented-generation"],"topicsNotShown":0,"allTopics":["microservice","gpu-acceleration","nemo","tensorrt","rag","triton-inference-server","large-language-models","llm","llm-inference","retrieval-augmented-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":18,"starsCount":1623,"forksCount":256,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,2,1,1,1,2,2,0,3,1,1,0,4,0,2,3,1,0,0,1,3,1,2,1,0,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T19:21:56.508Z"}},{"type":"Public","name":"ChatRTX","owner":"NVIDIA","isFork":false,"description":"A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":23,"starsCount":2421,"forksCount":260,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,2,10,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,3,0,0,0,0,0,0,0,0,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T16:45:47.293Z"}},{"type":"Public","name":"modulus","owner":"NVIDIA","isFork":false,"description":"Open-source deep-learning framework for building, training, and fine-tuning deep learning models using state-of-the-art Physics-ML methods","topicNames":["machine-learning","deep-learning","physics","pytorch","nvidia-gpu"],"topicsNotShown":0,"allTopics":["machine-learning","deep-learning","physics","pytorch","nvidia-gpu"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":18,"issueCount":90,"starsCount":705,"forksCount":146,"license":"Apache License 2.0","participation":[2,6,3,4,5,1,2,7,11,7,10,0,3,1,3,0,11,8,4,4,2,4,5,8,9,5,11,4,6,1,0,6,3,7,6,4,8,4,8,7,5,1,4,2,4,14,5,3,10,4,1,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T03:46:03.251Z"}},{"type":"Public","name":"spark-rapids-benchmarks","owner":"NVIDIA","isFork":false,"description":"Spark RAPIDS Benchmarks – benchmark sets and utilities for the RAPIDS Accelerator for Apache Spark","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":21,"starsCount":31,"forksCount":25,"license":"Apache License 2.0","participation":[0,0,0,1,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,1,2,0,1,1,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T21:49:43.136Z"}},{"type":"Public","name":"swift","owner":"NVIDIA","isFork":true,"description":"OpenStack Storage (Swift). Mirror of code maintained at opendev.org.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":1059,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-23T22:18:42.172Z"}},{"type":"Public","name":"air_sdk","owner":"NVIDIA","isFork":false,"description":"A Python SDK library for interacting with NVIDIA Air","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":5,"forksCount":4,"license":"Other","participation":[2,0,0,0,0,0,0,11,0,6,2,1,1,0,0,1,1,1,0,0,0,1,4,3,2,0,1,1,1,0,0,3,1,2,0,0,1,1,0,0,1,0,0,1,0,1,0,0,1,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-23T17:24:35.791Z"}},{"type":"Public","name":"numbast","owner":"NVIDIA","isFork":false,"description":"Numbast is a tool to build an automated pipeline that converts CUDA APIs into Numba bindings.","topicNames":["cuda","numba"],"topicsNotShown":0,"allTopics":["cuda","numba"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":17,"starsCount":13,"forksCount":3,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,15,1,1,5,8,1,2,2,1,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-27T22:34:07.312Z"}},{"type":"Public","name":"NeMo-Framework-Launcher","owner":"NVIDIA","isFork":false,"description":"NeMo Megatron launcher and tools","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":10,"issueCount":23,"starsCount":399,"forksCount":116,"license":"Apache License 2.0","participation":[41,11,5,8,2,7,6,18,6,18,16,1,12,5,17,22,5,1,10,34,11,7,7,1,20,13,17,37,28,14,5,19,24,27,25,10,57,25,35,22,27,42,5,7,14,16,24,21,14,36,50,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T23:19:20.665Z"}},{"type":"Public","name":"NeMo-text-processing","owner":"NVIDIA","isFork":false,"description":"NeMo text processing for ASR and TTS","topicNames":["text-normalization","inverse-text-n"],"topicsNotShown":0,"allTopics":["text-normalization","inverse-text-n"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":5,"starsCount":220,"forksCount":73,"license":"Apache License 2.0","participation":[0,4,2,1,4,1,1,0,0,1,1,2,1,2,2,0,0,1,2,1,1,4,1,0,0,1,0,1,0,0,0,0,0,1,0,0,0,2,0,1,0,4,1,1,0,0,1,2,4,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-27T00:26:12.322Z"}},{"type":"Public","name":"TensorRT-Model-Optimizer","owner":"NVIDIA","isFork":false,"description":"TensorRT Model Optimizer is a unified library of state-of-the-art model optimization techniques such as quantization and sparsity. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed on NVIDIA GPUs.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":7,"starsCount":219,"forksCount":13,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-21T18:02:35.173Z"}},{"type":"Public","name":"cuda-python","owner":"NVIDIA","isFork":false,"description":"CUDA Python Low-level Bindings","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":10,"starsCount":783,"forksCount":60,"license":"Other","participation":[0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,1,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-21T16:20:17.052Z"}},{"type":"Public","name":"workbench-example-hybrid-rag","owner":"NVIDIA","isFork":false,"description":"An NVIDIA AI Workbench example project for Retrieval Augmented Generation (RAG)","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":3,"starsCount":43,"forksCount":117,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,0,0,0,0,0,0,0,1,2,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-20T20:54:47.726Z"}},{"type":"Public","name":"audio-flamingo","owner":"NVIDIA","isFork":false,"description":"PyTorch implementation of Audio Flamingo: A Novel Audio Language Model with Few-Shot Learning and Dialogue Abilities.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":4,"forksCount":0,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-20T18:40:22.567Z"}},{"type":"Public","name":"NeMo-speech-data-processor","owner":"NVIDIA","isFork":false,"description":"A toolkit for processing speech data and creating speech datasets","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":1,"starsCount":63,"forksCount":20,"license":"Apache License 2.0","participation":[1,3,1,1,1,0,11,7,20,0,0,0,0,0,0,0,0,0,0,0,0,1,1,20,1,2,1,0,0,0,0,1,0,0,0,0,0,0,0,1,0,0,0,5,0,0,1,1,1,2,2,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-27T12:12:43.675Z"}},{"type":"Public","name":"hpc-container-maker","owner":"NVIDIA","isFork":false,"description":"HPC Container Maker","topicNames":["docker","containers","hpc","singularity"],"topicsNotShown":0,"allTopics":["docker","containers","hpc","singularity"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":11,"starsCount":439,"forksCount":86,"license":"Apache License 2.0","participation":[0,3,0,0,0,0,0,0,0,0,0,0,0,4,0,0,1,0,0,3,0,1,0,0,0,3,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-17T19:54:46.029Z"}},{"type":"Public","name":"nv-cloud-function-helpers","owner":"NVIDIA","isFork":false,"description":"Functions that simplify common tasks with NVIDIA Cloud Functions","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":12,"issueCount":0,"starsCount":7,"forksCount":2,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,6,2,9,5,0,1,0,0,0,0,1,0,0,0,6,0,3,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-17T16:11:27.744Z"}},{"type":"Public","name":"modulus-sym","owner":"NVIDIA","isFork":false,"description":"Framework providing pythonic APIs, algorithms and utilities to be used with Modulus core to physics inform model training as well as higher level abstraction for domain experts","topicNames":["machine-learning","deep-learning","physics","pytorch","nvidia-gpu"],"topicsNotShown":0,"allTopics":["machine-learning","deep-learning","physics","pytorch","nvidia-gpu"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":47,"starsCount":123,"forksCount":52,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-16T16:07:46.025Z"}},{"type":"Public","name":"framework-reproducibility","owner":"NVIDIA","isFork":false,"description":"Providing reproducibility in deep learning frameworks","topicNames":["seed","noise","reproducibility","seeder","noise-reduction","ngc","frameworks","variance-reduction","atomics","gpu-support"],"topicsNotShown":9,"allTopics":["seed","noise","reproducibility","seeder","noise-reduction","ngc","frameworks","variance-reduction","atomics","gpu-support","determinism","gpu-determinism","deterministic-ops","d9m","r13y","fwr13y","deep-learning","tensorflow","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":10,"starsCount":419,"forksCount":39,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-13T18:51:02.209Z"}},{"type":"Public","name":"Stable-Diffusion-WebUI-TensorRT","owner":"NVIDIA","isFork":false,"description":"TensorRT Extension for Stable Diffusion Web UI","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":15,"issueCount":139,"starsCount":1795,"forksCount":136,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,8,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-13T12:08:51.212Z"}}],"repositoryCount":108,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}