{"id":36435,"name":"flash-linear-attention","description":"🚀 Efficient implementations of state-of-the-art linear attention models","url":"https://github.com/fla-org/flash-linear-attention","last_synced_at":"2025-09-05T07:08:07.025Z","repository":{"id":214928648,"uuid":"733802106","full_name":"fla-org/flash-linear-attention","owner":"fla-org","description":"🚀 Efficient implementations of state-of-the-art linear attention models","archived":false,"fork":false,"pushed_at":"2025-08-28T21:12:36.000Z","size":4962,"stargazers_count":3075,"open_issues_count":47,"forks_count":239,"subscribers_count":27,"default_branch":"main","last_synced_at":"2025-08-28T21:49:01.774Z","etag":null,"topics":["large-language-models","machine-learning-systems","natural-language-processing"],"latest_commit_sha":null,"homepage":"https://github.com/fla-org/flash-linear-attention","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/fla-org.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":"CITATION.cff","codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null,"zenodo":null}},"created_at":"2023-12-20T06:50:18.000Z","updated_at":"2025-08-28T19:26:29.000Z","dependencies_parsed_at":"2024-03-22T14:45:26.380Z","dependency_job_id":"64e99b08-87a2-4813-9f2d-df563a014eab","html_url":"https://github.com/fla-org/flash-linear-attention","commit_stats":null,"previous_names":["sustcsonglin/flash-linear-attention","fla-org/flash-linear-attention"],"tags_count":8,"template":false,"template_full_name":null,"purl":"pkg:github/fla-org/flash-linear-attention","repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/fla-org","download_url":"https://codeload.github.com/fla-org/flash-linear-attention/tar.gz/refs/heads/main","sbom_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/sbom","scorecard":null,"host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":273723194,"owners_count":25156304,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","status":"online","status_checked_at":"2025-09-05T02:00:09.113Z","response_time":402,"last_error":null,"robots_txt_status":"success","robots_txt_updated_at":"2025-07-24T06:49:26.215Z","robots_txt_url":"https://github.com/robots.txt","online":true,"can_crawl_api":true,"host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"owner":{"login":"fla-org","name":"fla-org","uuid":"40835596","kind":"organization","description":"","email":"yangsl66@mit.edu","website":null,"location":null,"twitter":null,"company":null,"icon_url":"https://avatars.githubusercontent.com/u/40835596?v=4","repositories_count":1,"last_synced_at":"2024-12-29T18:32:52.617Z","metadata":{"has_sponsors_listing":false},"html_url":"https://github.com/fla-org","funding_links":[],"total_stars":1490,"followers":1,"following":0,"created_at":"2024-12-29T18:32:52.638Z","updated_at":"2024-12-29T18:32:52.638Z","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/fla-org","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/fla-org/repositories"},"packages":[{"id":11463748,"name":"flash-linear-attention","ecosystem":"pypi","description":"Fast Triton-based implementations of causal linear attention","homepage":"https://github.com/fla-org/flash-linear-attention","licenses":"MIT License\n        \n        Copyright (c) 2023-2025 Songlin Yang\n        \n        Permission is hereby granted, free of charge, to any person obtaining a copy\n        of this software and associated documentation files (the \"Software\"), to deal\n        in the Software without restriction, including without limitation the rights\n        to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n        copies of the Software, and to permit persons to whom the Software is\n        furnished to do so, subject to the following conditions:\n        \n        The above copyright notice and this permission notice shall be included in all\n        copies or substantial portions of the Software.\n        \n        THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n        IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n        FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n        AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n        LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n        OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n        SOFTWARE.\n        ","normalized_licenses":["Other"],"repository_url":"https://github.com/fla-org/flash-linear-attention","keywords_array":[],"namespace":null,"versions_count":8,"first_release_published_at":"2025-03-20T21:31:52.000Z","latest_release_published_at":"2025-08-26T20:28:24.000Z","latest_release_number":"0.3.1","last_synced_at":"2025-08-26T20:31:09.843Z","created_at":"2025-03-20T21:46:01.936Z","updated_at":"2025-08-26T20:35:44.040Z","registry_url":"https://pypi.org/project/flash-linear-attention/","install_command":"pip install flash-linear-attention --index-url https://pypi.org/simple","documentation_url":"https://flash-linear-attention.readthedocs.io/","metadata":{"funding":null,"documentation":null,"classifiers":["License :: OSI Approved :: MIT License","Operating System :: OS Independent","Programming Language :: Python :: 3","Topic :: Scientific/Engineering :: Artificial Intelligence"],"normalized_name":"flash-linear-attention","project_status":null},"repo_metadata":{"id":214928648,"uuid":"733802106","full_name":"fla-org/flash-linear-attention","owner":"fla-org","description":"🚀 Efficient implementations of state-of-the-art linear attention models","archived":false,"fork":false,"pushed_at":"2025-08-21T17:03:48.000Z","size":4906,"stargazers_count":3050,"open_issues_count":47,"forks_count":239,"subscribers_count":27,"default_branch":"main","last_synced_at":"2025-08-21T18:27:41.184Z","etag":null,"topics":["large-language-models","machine-learning-systems","natural-language-processing"],"latest_commit_sha":null,"homepage":"https://github.com/fla-org/flash-linear-attention","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/fla-org.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":"CITATION.cff","codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null,"zenodo":null}},"created_at":"2023-12-20T06:50:18.000Z","updated_at":"2025-08-21T17:03:51.000Z","dependencies_parsed_at":"2024-03-22T14:45:26.380Z","dependency_job_id":"64e99b08-87a2-4813-9f2d-df563a014eab","html_url":"https://github.com/fla-org/flash-linear-attention","commit_stats":null,"previous_names":["sustcsonglin/flash-linear-attention","fla-org/flash-linear-attention"],"tags_count":7,"template":false,"template_full_name":null,"purl":"pkg:github/fla-org/flash-linear-attention","repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/fla-org","download_url":"https://codeload.github.com/fla-org/flash-linear-attention/tar.gz/refs/heads/main","sbom_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/sbom","scorecard":null,"host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":272254155,"owners_count":24901037,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","status":"online","status_checked_at":"2025-08-26T02:00:07.904Z","response_time":60,"last_error":null,"robots_txt_status":"success","robots_txt_updated_at":"2025-07-24T06:49:26.215Z","robots_txt_url":"https://github.com/robots.txt","online":true,"can_crawl_api":true,"host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"},"owner_record":{"login":"fla-org","name":"fla-org","uuid":"40835596","kind":"organization","description":"","email":"yangsl66@mit.edu","website":null,"location":null,"twitter":null,"company":null,"icon_url":"https://avatars.githubusercontent.com/u/40835596?v=4","repositories_count":1,"last_synced_at":"2024-12-29T18:32:52.617Z","metadata":{"has_sponsors_listing":false},"html_url":"https://github.com/fla-org","funding_links":[],"total_stars":1490,"followers":1,"following":0,"created_at":"2024-12-29T18:32:52.638Z","updated_at":"2024-12-29T18:32:52.638Z","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/fla-org","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/fla-org/repositories"},"tags":[{"name":"v0.3.0","sha":"17dd5662554d46b6bcb1d1ff728cebb461c9aef9","kind":"commit","published_at":"2025-07-14T09:45:34.000Z","download_url":"https://codeload.github.com/fla-org/flash-linear-attention/tar.gz/v0.3.0","html_url":"https://github.com/fla-org/flash-linear-attention/releases/tag/v0.3.0","dependencies_parsed_at":null,"dependency_job_id":null,"purl":"pkg:github/fla-org/flash-linear-attention@v0.3.0","tag_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.3.0","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.3.0/manifests"},{"name":"v0.2.2","sha":"a46f204c1c7ac5c70ee99168adef7682e6fbf2a3","kind":"commit","published_at":"2025-06-05T12:53:18.000Z","download_url":"https://codeload.github.com/fla-org/flash-linear-attention/tar.gz/v0.2.2","html_url":"https://github.com/fla-org/flash-linear-attention/releases/tag/v0.2.2","dependencies_parsed_at":null,"dependency_job_id":null,"purl":"pkg:github/fla-org/flash-linear-attention@v0.2.2","tag_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.2.2","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.2.2/manifests"},{"name":"v0.2.1","sha":"a670dff4c2537fc1a82486584dd9569e18fba833","kind":"commit","published_at":"2025-04-23T16:54:18.000Z","download_url":"https://codeload.github.com/fla-org/flash-linear-attention/tar.gz/v0.2.1","html_url":"https://github.com/fla-org/flash-linear-attention/releases/tag/v0.2.1","dependencies_parsed_at":null,"dependency_job_id":null,"purl":"pkg:github/fla-org/flash-linear-attention@v0.2.1","tag_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.2.1","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.2.1/manifests"},{"name":"v0.2.0","sha":"6bfd5e671aa8ddba7b7bc57d0dd6a9793ccdebf9","kind":"commit","published_at":"2025-04-11T20:28:34.000Z","download_url":"https://codeload.github.com/fla-org/flash-linear-attention/tar.gz/v0.2.0","html_url":"https://github.com/fla-org/flash-linear-attention/releases/tag/v0.2.0","dependencies_parsed_at":null,"dependency_job_id":null,"purl":"pkg:github/fla-org/flash-linear-attention@v0.2.0","tag_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.2.0","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.2.0/manifests"},{"name":"v0.1.2","sha":"53b3ac7eacff5bec633f84249b528a5f8b1c36b0","kind":"commit","published_at":"2025-03-31T06:30:02.000Z","download_url":"https://codeload.github.com/fla-org/flash-linear-attention/tar.gz/v0.1.2","html_url":"https://github.com/fla-org/flash-linear-attention/releases/tag/v0.1.2","dependencies_parsed_at":null,"dependency_job_id":null,"purl":"pkg:github/fla-org/flash-linear-attention@v0.1.2","tag_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.1.2","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.1.2/manifests"},{"name":"v0.1.1","sha":"09fe6c2cbbef23c58502f04f565812636e9b5f28","kind":"commit","published_at":"2025-03-24T06:22:09.000Z","download_url":"https://codeload.github.com/fla-org/flash-linear-attention/tar.gz/v0.1.1","html_url":"https://github.com/fla-org/flash-linear-attention/releases/tag/v0.1.1","dependencies_parsed_at":null,"dependency_job_id":null,"purl":"pkg:github/fla-org/flash-linear-attention@v0.1.1","tag_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.1.1","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.1.1/manifests"},{"name":"v0.1.0","sha":"cb0d1bb0d1ebd850e0c55befd0ffb04062f5007c","kind":"commit","published_at":"2025-03-20T21:29:10.000Z","download_url":"https://codeload.github.com/fla-org/flash-linear-attention/tar.gz/v0.1.0","html_url":"https://github.com/fla-org/flash-linear-attention/releases/tag/v0.1.0","dependencies_parsed_at":null,"dependency_job_id":null,"purl":"pkg:github/fla-org/flash-linear-attention@v0.1.0","tag_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.1.0","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags/v0.1.0/manifests"}]},"repo_metadata_updated_at":"2025-08-26T20:35:44.040Z","dependent_packages_count":0,"downloads":17207,"downloads_period":"last-month","dependent_repos_count":0,"rankings":{"downloads":null,"dependent_repos_count":53.18443505036904,"dependent_packages_count":9.443570166018658,"stargazers_count":null,"forks_count":null,"docker_downloads_count":null,"average":31.31400260819385},"purl":"pkg:pypi/flash-linear-attention","advisories":[],"docker_usage_url":"https://docker.ecosyste.ms/usage/pypi/flash-linear-attention","docker_dependents_count":null,"docker_downloads_count":null,"usage_url":"https://repos.ecosyste.ms/usage/pypi/flash-linear-attention","dependent_repositories_url":"https://repos.ecosyste.ms/api/v1/usage/pypi/flash-linear-attention/dependencies","status":null,"funding_links":[],"critical":null,"issue_metadata":{"last_synced_at":"2025-08-20T11:13:23.899Z","issues_count":16,"pull_requests_count":19,"avg_time_to_close_issue":2202909.2,"avg_time_to_close_pull_request":1364363.4545454546,"issues_closed_count":10,"pull_requests_closed_count":11,"pull_request_authors_count":9,"issue_authors_count":13,"avg_comments_per_issue":3.125,"avg_comments_per_pull_request":2.8421052631578947,"merged_pull_requests_count":8,"bot_issues_count":0,"bot_pull_requests_count":0,"past_year_issues_count":16,"past_year_pull_requests_count":19,"past_year_avg_time_to_close_issue":2202909.2,"past_year_avg_time_to_close_pull_request":1364363.4545454546,"past_year_issues_closed_count":10,"past_year_pull_requests_closed_count":11,"past_year_pull_request_authors_count":9,"past_year_issue_authors_count":13,"past_year_avg_comments_per_issue":3.125,"past_year_avg_comments_per_pull_request":2.8421052631578947,"past_year_bot_issues_count":0,"past_year_bot_pull_requests_count":0,"past_year_merged_pull_requests_count":8,"issues_url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/issues","maintainers":[{"login":"zhiyuan1i","count":10,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/zhiyuan1i"},{"login":"yzhangcs","count":1,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/yzhangcs"},{"login":"sustcsonglin","count":1,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/sustcsonglin"}],"active_maintainers":[{"login":"zhiyuan1i","count":10,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/zhiyuan1i"},{"login":"sustcsonglin","count":1,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/sustcsonglin"},{"login":"yzhangcs","count":1,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/yzhangcs"}]},"versions_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/packages/flash-linear-attention/versions","version_numbers_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/packages/flash-linear-attention/version_numbers","dependent_packages_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/packages/flash-linear-attention/dependent_packages","related_packages_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/packages/flash-linear-attention/related_packages","maintainers":[{"uuid":"sonta","login":"sonta","name":null,"email":null,"url":null,"packages_count":1,"html_url":"https://pypi.org/user/sonta/","role":null,"created_at":"2025-03-20T21:51:00.809Z","updated_at":"2025-03-20T21:51:00.809Z","packages_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/maintainers/sonta/packages"}],"registry":{"name":"pypi.org","url":"https://pypi.org","ecosystem":"pypi","default":true,"packages_count":725013,"maintainers_count":308324,"namespaces_count":0,"keywords_count":238006,"github":"pypi","metadata":{"funded_packages_count":50519},"icon_url":"https://github.com/pypi.png","created_at":"2022-04-04T15:19:23.364Z","updated_at":"2025-09-05T05:40:58.860Z","packages_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/packages","maintainers_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/maintainers","namespaces_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/namespaces"}},{"id":10860058,"name":"rwkv-fla","ecosystem":"pypi","description":"Fast Triton-based implementations for RWKV","homepage":"https://github.com/TorchRWKV/flash-linear-attention","licenses":"MIT License\n        \n        Copyright (c) 2023-2025 Songlin Yang\n        \n        Permission is hereby granted, free of charge, to any person obtaining a copy\n        of this software and associated documentation files (the \"Software\"), to deal\n        in the Software without restriction, including without limitation the rights\n        to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n        copies of the Software, and to permit persons to whom the Software is\n        furnished to do so, subject to the following conditions:\n        \n        The above copyright notice and this permission notice shall be included in all\n        copies or substantial portions of the Software.\n        \n        THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n        IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n        FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n        AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n        LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n        OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n        SOFTWARE.\n        ","normalized_licenses":["Other"],"repository_url":"https://github.com/fla-org/flash-linear-attention","keywords_array":[],"namespace":null,"versions_count":104,"first_release_published_at":"2024-09-05T07:56:59.000Z","latest_release_published_at":"2025-08-22T14:13:30.000Z","latest_release_number":"0.7.202508221413","last_synced_at":"2025-08-22T14:15:55.041Z","created_at":"2024-09-05T08:03:39.309Z","updated_at":"2025-08-22T14:15:55.289Z","registry_url":"https://pypi.org/project/rwkv-fla/","install_command":"pip install rwkv-fla --index-url https://pypi.org/simple","documentation_url":"https://rwkv-fla.readthedocs.io/","metadata":{"funding":null,"documentation":null,"classifiers":["License :: OSI Approved :: MIT License","Operating System :: OS Independent","Programming Language :: Python :: 3","Topic :: Scientific/Engineering :: Artificial Intelligence"],"normalized_name":"rwkv-fla","project_status":null},"repo_metadata":{"id":249664754,"uuid":"832165709","full_name":"TorchRWKV/flash-linear-attention","owner":"TorchRWKV","description":"Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton","archived":false,"fork":true,"pushed_at":"2024-09-06T12:13:40.000Z","size":1347,"stargazers_count":2,"open_issues_count":0,"forks_count":0,"subscribers_count":0,"default_branch":"stable","last_synced_at":"2024-09-07T08:07:26.503Z","etag":null,"topics":[],"latest_commit_sha":null,"homepage":"","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":"sustcsonglin/flash-linear-attention","license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/TorchRWKV.png","metadata":{},"created_at":"2024-07-22T13:15:27.000Z","updated_at":"2024-09-06T12:13:44.000Z","dependencies_parsed_at":null,"dependency_job_id":null,"html_url":"https://github.com/TorchRWKV/flash-linear-attention","commit_stats":null,"previous_names":["torchrwkv/flash-linear-attention"],"tags_count":0,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/TorchRWKV%2Fflash-linear-attention","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/TorchRWKV%2Fflash-linear-attention/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/TorchRWKV%2Fflash-linear-attention/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/TorchRWKV%2Fflash-linear-attention/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/TorchRWKV","download_url":"https://codeload.github.com/TorchRWKV/flash-linear-attention/tar.gz/refs/heads/stable","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":217954551,"owners_count":16256514,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"},"owner_record":{"login":"TorchRWKV","name":"TorchRWKV","uuid":"172354835","kind":"organization","description":null,"email":null,"website":null,"location":null,"twitter":null,"company":null,"icon_url":"https://avatars.githubusercontent.com/u/172354835?v=4","repositories_count":1,"last_synced_at":"2024-07-20T10:23:11.363Z","metadata":{"has_sponsors_listing":false},"html_url":"https://github.com/TorchRWKV","funding_links":[],"total_stars":1,"followers":0,"following":0,"created_at":"2024-07-20T10:23:11.666Z","updated_at":"2024-07-20T10:23:11.666Z","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/TorchRWKV","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/TorchRWKV/repositories"},"tags":[]},"repo_metadata_updated_at":"2024-09-07T08:17:34.480Z","dependent_packages_count":0,"downloads":1294,"downloads_period":"last-month","dependent_repos_count":0,"rankings":{"downloads":null,"dependent_repos_count":58.45441569373787,"dependent_packages_count":10.384690886465949,"stargazers_count":null,"forks_count":null,"docker_downloads_count":null,"average":34.41955329010191},"purl":"pkg:pypi/rwkv-fla","advisories":[],"docker_usage_url":"https://docker.ecosyste.ms/usage/pypi/rwkv-fla","docker_dependents_count":null,"docker_downloads_count":null,"usage_url":"https://repos.ecosyste.ms/usage/pypi/rwkv-fla","dependent_repositories_url":"https://repos.ecosyste.ms/api/v1/usage/pypi/rwkv-fla/dependencies","status":null,"funding_links":[],"critical":null,"issue_metadata":null,"versions_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/packages/rwkv-fla/versions","version_numbers_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/packages/rwkv-fla/version_numbers","dependent_packages_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/packages/rwkv-fla/dependent_packages","related_packages_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/packages/rwkv-fla/related_packages","maintainers":[{"uuid":"uniartisan","login":"uniartisan","name":null,"email":null,"url":null,"packages_count":1,"html_url":"https://pypi.org/user/uniartisan/","role":null,"created_at":"2024-09-05T08:07:18.301Z","updated_at":"2024-09-05T08:07:18.301Z","packages_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/maintainers/uniartisan/packages"}],"registry":{"name":"pypi.org","url":"https://pypi.org","ecosystem":"pypi","default":true,"packages_count":725013,"maintainers_count":308324,"namespaces_count":0,"keywords_count":238006,"github":"pypi","metadata":{"funded_packages_count":50519},"icon_url":"https://github.com/pypi.png","created_at":"2022-04-04T15:19:23.364Z","updated_at":"2025-09-05T05:40:58.860Z","packages_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/packages","maintainers_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/maintainers","namespaces_url":"https://packages.ecosyste.ms/api/v1/registries/pypi.org/namespaces"}},{"id":11680115,"name":"github.com/fla-org/flash-linear-attention","ecosystem":"go","description":null,"homepage":null,"licenses":"mit","normalized_licenses":["MIT"],"repository_url":"https://github.com/fla-org/flash-linear-attention","keywords_array":[],"namespace":null,"versions_count":8,"first_release_published_at":"2025-03-20T21:29:10.000Z","latest_release_published_at":"2025-08-26T06:40:38.000Z","latest_release_number":"v0.3.1","last_synced_at":"2025-08-30T21:48:53.737Z","created_at":"2025-05-24T08:22:53.841Z","updated_at":"2025-08-30T21:48:53.951Z","registry_url":"https://pkg.go.dev/github.com/fla-org/flash-linear-attention","install_command":"go get github.com/fla-org/flash-linear-attention","documentation_url":"https://pkg.go.dev/github.com/fla-org/flash-linear-attention#section-documentation","metadata":{},"repo_metadata":{"id":214928648,"uuid":"733802106","full_name":"fla-org/flash-linear-attention","owner":"fla-org","description":"🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton","archived":false,"fork":false,"pushed_at":"2025-05-19T16:25:01.000Z","size":4204,"stargazers_count":2398,"open_issues_count":40,"forks_count":165,"subscribers_count":29,"default_branch":"main","last_synced_at":"2025-05-19T17:35:58.726Z","etag":null,"topics":["large-language-models","machine-learning-systems","natural-language-processing"],"latest_commit_sha":null,"homepage":"https://github.com/fla-org/flash-linear-attention","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/fla-org.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":"CITATION.cff","codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null,"zenodo":null}},"created_at":"2023-12-20T06:50:18.000Z","updated_at":"2025-05-19T16:25:05.000Z","dependencies_parsed_at":"2024-03-22T14:45:26.380Z","dependency_job_id":"64e99b08-87a2-4813-9f2d-df563a014eab","html_url":"https://github.com/fla-org/flash-linear-attention","commit_stats":null,"previous_names":["sustcsonglin/flash-linear-attention","fla-org/flash-linear-attention"],"tags_count":5,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/fla-org","download_url":"https://codeload.github.com/fla-org/flash-linear-attention/tar.gz/refs/heads/main","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":256050635,"owners_count":22328986,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"repo_metadata_updated_at":"2025-05-24T08:27:14.266Z","dependent_packages_count":0,"downloads":null,"downloads_period":null,"dependent_repos_count":0,"rankings":{"downloads":null,"dependent_repos_count":5.820448282572631,"dependent_packages_count":5.454459440716581,"stargazers_count":null,"forks_count":null,"docker_downloads_count":null,"average":5.637453861644606},"purl":"pkg:golang/github.com/fla-org/flash-linear-attention","advisories":[],"docker_usage_url":"https://docker.ecosyste.ms/usage/go/github.com/fla-org/flash-linear-attention","docker_dependents_count":null,"docker_downloads_count":null,"usage_url":"https://repos.ecosyste.ms/usage/go/github.com/fla-org/flash-linear-attention","dependent_repositories_url":"https://repos.ecosyste.ms/api/v1/usage/go/github.com/fla-org/flash-linear-attention/dependencies","status":null,"funding_links":[],"critical":null,"issue_metadata":null,"versions_url":"https://packages.ecosyste.ms/api/v1/registries/proxy.golang.org/packages/github.com%2Ffla-org%2Fflash-linear-attention/versions","version_numbers_url":"https://packages.ecosyste.ms/api/v1/registries/proxy.golang.org/packages/github.com%2Ffla-org%2Fflash-linear-attention/version_numbers","dependent_packages_url":"https://packages.ecosyste.ms/api/v1/registries/proxy.golang.org/packages/github.com%2Ffla-org%2Fflash-linear-attention/dependent_packages","related_packages_url":"https://packages.ecosyste.ms/api/v1/registries/proxy.golang.org/packages/github.com%2Ffla-org%2Fflash-linear-attention/related_packages","maintainers":[],"registry":{"name":"proxy.golang.org","url":"https://proxy.golang.org","ecosystem":"go","default":true,"packages_count":1951527,"maintainers_count":0,"namespaces_count":741275,"keywords_count":109185,"github":"golang","metadata":{"funded_packages_count":49011},"icon_url":"https://github.com/golang.png","created_at":"2022-04-04T15:19:22.939Z","updated_at":"2025-09-05T05:14:06.439Z","packages_url":"https://packages.ecosyste.ms/api/v1/registries/proxy.golang.org/packages","maintainers_url":"https://packages.ecosyste.ms/api/v1/registries/proxy.golang.org/maintainers","namespaces_url":"https://packages.ecosyste.ms/api/v1/registries/proxy.golang.org/namespaces"}}],"commits":{"message":"Repository syncing started."},"issues_stats":{"full_name":"fla-org/flash-linear-attention","html_url":"https://github.com/fla-org/flash-linear-attention","last_synced_at":"2025-09-03T16:17:46.128Z","status":null,"issues_count":164,"pull_requests_count":252,"avg_time_to_close_issue":996015.8396226416,"avg_time_to_close_pull_request":194677.45077720206,"issues_closed_count":106,"pull_requests_closed_count":193,"pull_request_authors_count":50,"issue_authors_count":65,"avg_comments_per_issue":1.7926829268292683,"avg_comments_per_pull_request":1.0476190476190477,"merged_pull_requests_count":175,"bot_issues_count":0,"bot_pull_requests_count":0,"past_year_issues_count":164,"past_year_pull_requests_count":252,"past_year_avg_time_to_close_issue":996015.8396226416,"past_year_avg_time_to_close_pull_request":194677.45077720206,"past_year_issues_closed_count":106,"past_year_pull_requests_closed_count":193,"past_year_pull_request_authors_count":50,"past_year_issue_authors_count":65,"past_year_avg_comments_per_issue":1.7926829268292683,"past_year_avg_comments_per_pull_request":1.0476190476190477,"past_year_bot_issues_count":0,"past_year_bot_pull_requests_count":0,"past_year_merged_pull_requests_count":175,"created_at":"2025-07-16T13:24:22.836Z","updated_at":"2025-09-03T16:17:46.130Z","repository_url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention","issues_url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/repositories/fla-org%2Fflash-linear-attention/issues","issue_labels_count":{"bug":82,"enhancement":55,"stale":9,"todo":3,"wontfix":3,"urgent":2,"help wanted":2,"good first issue":1,"multi-cards":1,"mix-precision":1},"pull_request_labels_count":{"enhancement":1,"bug":1},"issue_author_associations_count":{"NONE":85,"COLLABORATOR":56,"CONTRIBUTOR":16,"MEMBER":7},"pull_request_author_associations_count":{"COLLABORATOR":116,"NONE":49,"MEMBER":44,"CONTRIBUTOR":43},"issue_authors":{"sustcsonglin":44,"zhiyuan1i":12,"Triang-jyed-driung":10,"yzhangcs":7,"SmerkyG":5,"www-Ye":4,"necrophagists":4,"Chris-city":3,"Lynn-020809":3,"zmj1203":3,"mutiann":3,"zhan8855":2,"p81sunshine":2,"conceptofmind":2,"sagejiaweili":2,"JulienSiems":2,"Zazexy":2,"yangshengaa":2,"mark14wu":2,"isayoften":2,"LouChao98":2,"HanGuo97":2,"lucassunalt":2,"tesla3":1,"cleverblue":1,"MonolithFoundation":1,"Antoninnnn":1,"guoguo1314":1,"t1101675":1,"OliverShaoPT":1,"r-buitrago":1,"Weili-0234":1,"Fzkuji":1,"wxqnl":1,"johanwind":1,"Fadelis98":1,"xffxff":1,"980202006":1,"pableeto":1,"theodorblackbird":1,"zhuzeyuan":1,"kyv001":1,"EricZhang1412":1,"HallerPatrick":1,"kangyiyang":1,"Ffffffffchopin":1,"vladislavalerievich":1,"shenzhiy21":1,"sozforex":1,"Shuaizhang7":1,"void-echo":1,"KashuvY":1,"leifeng666":1,"2catycm":1,"York-Cheung":1,"YufangMo":1,"junmokane":1,"SSamDav":1,"koceja":1,"hypnopump":1,"alxndrTL":1,"changdong1687":1,"MathLover1234":1,"EricLina":1,"vanhowe":1},"pull_request_authors":{"zhiyuan1i":85,"yzhangcs":44,"uniartisan":18,"sustcsonglin":15,"Triang-jyed-driung":8,"zhixuan-lin":7,"jannalulu":7,"Pan-Yuqi":4,"HanGuo97":4,"richardodliu":3,"liqiongyu":3,"yiyousong":3,"2022tgoel":3,"phi-jkim":2,"JulienSiems":2,"yibozhong":2,"kugwzk":2,"phnazari":2,"V0XNIHILI":2,"rucnyz":2,"AwesomeSeq":2,"ridgerchu":2,"Espere-1119-Song":2,"johanwind":2,"yuweih205":1,"seanxwzhang":1,"zaydzuhri":1,"JusenD":1,"toothacher17":1,"xffxff":1,"chengshuang18":1,"jovoswald":1,"Beortext":1,"LKJacky":1,"harrisonvanderbyl":1,"jihaoh98":1,"OliverShaoPT":1,"Luther-Sparks":1,"kangyiyang":1,"vladislavalerievich":1,"ChouYuhong":1,"KevlarKanou":1,"fffffgggg54":1,"timurcarstensen":1,"zxytim":1,"exhyy":1,"WKX933":1,"HallerPatrick":1,"Sxela":1,"ahatamiz":1},"host":{"name":"GitHub","url":"https://github.com","kind":"github","last_synced_at":"2025-09-04T00:00:25.939Z","repositories_count":10072086,"issues_count":31257663,"pull_requests_count":95517173,"authors_count":10689807,"icon_url":"https://github.com/github.png","host_url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/repositories","owners_url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/owners","authors_url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors"},"past_year_issue_labels_count":{"bug":82,"enhancement":55,"stale":9,"todo":3,"wontfix":3,"urgent":2,"help wanted":2,"good first issue":1,"multi-cards":1,"mix-precision":1},"past_year_pull_request_labels_count":{"enhancement":1,"bug":1},"past_year_issue_author_associations_count":{"NONE":85,"COLLABORATOR":56,"CONTRIBUTOR":16,"MEMBER":7},"past_year_pull_request_author_associations_count":{"COLLABORATOR":116,"NONE":49,"MEMBER":44,"CONTRIBUTOR":43},"past_year_issue_authors":{"sustcsonglin":44,"zhiyuan1i":12,"Triang-jyed-driung":10,"yzhangcs":7,"SmerkyG":5,"www-Ye":4,"necrophagists":4,"zmj1203":3,"mutiann":3,"Lynn-020809":3,"Chris-city":3,"mark14wu":2,"zhan8855":2,"Zazexy":2,"HanGuo97":2,"sagejiaweili":2,"isayoften":2,"JulienSiems":2,"lucassunalt":2,"yangshengaa":2,"p81sunshine":2,"LouChao98":2,"conceptofmind":2,"vladislavalerievich":1,"vanhowe":1,"changdong1687":1,"theodorblackbird":1,"tesla3":1,"t1101675":1,"SSamDav":1,"sozforex":1,"void-echo":1,"Weili-0234":1,"Antoninnnn":1,"wxqnl":1,"xffxff":1,"York-Cheung":1,"YufangMo":1,"alxndrTL":1,"980202006":1,"zhuzeyuan":1,"2catycm":1,"Fadelis98":1,"Ffffffffchopin":1,"Fzkuji":1,"guoguo1314":1,"HallerPatrick":1,"hypnopump":1,"johanwind":1,"junmokane":1,"kangyiyang":1,"KashuvY":1,"koceja":1,"kyv001":1,"leifeng666":1,"EricZhang1412":1,"MathLover1234":1,"MonolithFoundation":1,"EricLina":1,"OliverShaoPT":1,"pableeto":1,"r-buitrago":1,"shenzhiy21":1,"Shuaizhang7":1,"cleverblue":1},"past_year_pull_request_authors":{"zhiyuan1i":85,"yzhangcs":44,"uniartisan":18,"sustcsonglin":15,"Triang-jyed-driung":8,"zhixuan-lin":7,"jannalulu":7,"HanGuo97":4,"Pan-Yuqi":4,"2022tgoel":3,"liqiongyu":3,"yiyousong":3,"richardodliu":3,"JulienSiems":2,"phi-jkim":2,"kugwzk":2,"phnazari":2,"Espere-1119-Song":2,"rucnyz":2,"yibozhong":2,"johanwind":2,"ridgerchu":2,"V0XNIHILI":2,"AwesomeSeq":2,"Luther-Sparks":1,"zxytim":1,"exhyy":1,"xffxff":1,"jihaoh98":1,"WKX933":1,"chengshuang18":1,"jovoswald":1,"OliverShaoPT":1,"yuweih205":1,"HallerPatrick":1,"kangyiyang":1,"vladislavalerievich":1,"seanxwzhang":1,"ChouYuhong":1,"Sxela":1,"KevlarKanou":1,"ahatamiz":1,"Beortext":1,"LKJacky":1,"zaydzuhri":1,"harrisonvanderbyl":1,"JusenD":1,"fffffgggg54":1,"timurcarstensen":1,"toothacher17":1},"maintainers":[{"login":"zhiyuan1i","count":97,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/zhiyuan1i"},{"login":"sustcsonglin","count":59,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/sustcsonglin"},{"login":"yzhangcs","count":51,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/yzhangcs"},{"login":"uniartisan","count":15,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/uniartisan"},{"login":"toothacher17","count":1,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/toothacher17"}],"active_maintainers":[{"login":"zhiyuan1i","count":97,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/zhiyuan1i"},{"login":"sustcsonglin","count":59,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/sustcsonglin"},{"login":"yzhangcs","count":51,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/yzhangcs"},{"login":"uniartisan","count":15,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/uniartisan"},{"login":"toothacher17","count":1,"url":"https://issues.ecosyste.ms/api/v1/hosts/GitHub/authors/toothacher17"}]},"events":null,"keywords":["large-language-models","machine-learning-systems","natural-language-processing"],"dependencies":[{"ecosystem":"actions","filepath":".github/workflows/issue.yml","sha":null,"kind":"manifest","created_at":"2024-02-09T16:03:57.362Z","updated_at":"2024-02-09T16:03:57.362Z","repository_link":"https://github.com/fla-org/flash-linear-attention/blob/main/.github/workflows/issue.yml","dependencies":[{"id":16134396168,"package_name":"actions/stale","ecosystem":"actions","requirements":"v9.0.0","direct":true,"kind":"composite","optional":false}]},{"ecosystem":"pypi","filepath":"setup.py","sha":null,"kind":"manifest","created_at":"2024-02-09T16:03:57.409Z","updated_at":"2024-02-09T16:03:57.409Z","repository_link":"https://github.com/fla-org/flash-linear-attention/blob/main/setup.py","dependencies":[{"id":16134396169,"package_name":"triton","ecosystem":"pypi","requirements":"*","direct":true,"kind":"runtime","optional":false},{"id":16134396170,"package_name":"transformers","ecosystem":"pypi","requirements":"*","direct":true,"kind":"runtime","optional":false},{"id":16134396171,"package_name":"einops","ecosystem":"pypi","requirements":"*","direct":true,"kind":"runtime","optional":false},{"id":16134396172,"package_name":"ninja","ecosystem":"pypi","requirements":"*","direct":true,"kind":"runtime","optional":false}]}],"score":null,"created_at":"2025-09-04T15:51:10.055Z","updated_at":"2025-10-07T08:15:04.061Z","avatar_url":"https://github.com/fla-org.png","language":"Python","category":null,"sub_category":null,"monthly_downloads":18501,"funding_links":[],"readme_doi_urls":[],"works":{},"citation_counts":{},"total_citations":0,"keywords_from_contributors":[],"project_url":"https://science.ecosyste.ms/api/v1/projects/36435","html_url":"https://science.ecosyste.ms/projects/36435","bibtex_url":"https://science.ecosyste.ms/projects/36435/export.bibtex","apalike_url":"https://science.ecosyste.ms/projects/36435/export.apalike"}