With the amount of new subnets being added it can be hard to get up to date information across all subnets, so data may be slightly out of date from time to time
SoundsRight, designated as Subnet 105, is dedicated to the research and development of non-proprietary speech enhancement models. As more of our daily lives revolve around consuming online content, there is growing emphasis on high-quality audio. Speech enhancement is a complex field that involves tasks like separating desired speech from background noise, which requires training sophisticated models capable of distinguishing between different audio components under various circumstances.
The fundamental challenge that SoundsRight addresses is that much of speech enhancement technology is currently hidden behind paywalls, despite all necessary components for open-source innovation being readily available. SoundsRight aims to spearhead open-source speech enhancement technology through daily fine-tuning competitions, making high-quality audio processing more accessible to the broader community.
SoundsRight, designated as Subnet 105, is dedicated to the research and development of non-proprietary speech enhancement models. As more of our daily lives revolve around consuming online content, there is growing emphasis on high-quality audio. Speech enhancement is a complex field that involves tasks like separating desired speech from background noise, which requires training sophisticated models capable of distinguishing between different audio components under various circumstances.
The fundamental challenge that SoundsRight addresses is that much of speech enhancement technology is currently hidden behind paywalls, despite all necessary components for open-source innovation being readily available. SoundsRight aims to spearhead open-source speech enhancement technology through daily fine-tuning competitions, making high-quality audio processing more accessible to the broader community.
SoundsRight operates as a specialized subnet within the Bittensor decentralized ecosystem, focusing exclusively on speech enhancement technology. The subnet creates a competitive environment where participants (miners) develop and fine-tune speech enhancement models, which are then evaluated by validators to determine the best-performing solutions.
The core function of SoundsRight is to facilitate daily fine-tuning competitions for speech enhancement models. These competitions currently focus on two primary tasks:
Each competition follows a winner-takes-all format, which incentivizes miners to submit their absolute best models rather than multiple variations. This format, combined with the validation mechanism, deters miner factions by making model duplication unviable.
How SoundsRight Works
The SoundsRight subnet operates through a well-defined workflow involving miners, validators, HuggingFace (as a model repository), the Bittensor blockchain, and the subnet’s website. Here’s a detailed breakdown of how the system functions:
Miner-Validator Architecture
There are two main entities in the subnet:
Competition Workflow
The daily competition process follows these steps:
This continuous cycle ensures that models are constantly being improved and evaluated on fresh data, driving innovation in speech enhancement technology.
Technical Architecture
The SoundsRight subnet is built on the Bittensor ecosystem, with a technical architecture designed to facilitate the competition and evaluation process efficiently.
Repository Structure
The codebase is organized into several key directories:
Core Components
Technical Implementation
The implementation uses Python with dependencies including:
Competition Metrics
The subnet currently hosts competitions at a 16 kHz sample rate, with plans to expand to 48 kHz competitions in upcoming updates. The benchmarking metrics used include:
These metrics ensure comprehensive evaluation of model performance across different aspects of speech enhancement quality.
SoundsRight operates as a specialized subnet within the Bittensor decentralized ecosystem, focusing exclusively on speech enhancement technology. The subnet creates a competitive environment where participants (miners) develop and fine-tune speech enhancement models, which are then evaluated by validators to determine the best-performing solutions.
The core function of SoundsRight is to facilitate daily fine-tuning competitions for speech enhancement models. These competitions currently focus on two primary tasks:
Each competition follows a winner-takes-all format, which incentivizes miners to submit their absolute best models rather than multiple variations. This format, combined with the validation mechanism, deters miner factions by making model duplication unviable.
How SoundsRight Works
The SoundsRight subnet operates through a well-defined workflow involving miners, validators, HuggingFace (as a model repository), the Bittensor blockchain, and the subnet’s website. Here’s a detailed breakdown of how the system functions:
Miner-Validator Architecture
There are two main entities in the subnet:
Competition Workflow
The daily competition process follows these steps:
This continuous cycle ensures that models are constantly being improved and evaluated on fresh data, driving innovation in speech enhancement technology.
Technical Architecture
The SoundsRight subnet is built on the Bittensor ecosystem, with a technical architecture designed to facilitate the competition and evaluation process efficiently.
Repository Structure
The codebase is organized into several key directories:
Core Components
Technical Implementation
The implementation uses Python with dependencies including:
Competition Metrics
The subnet currently hosts competitions at a 16 kHz sample rate, with plans to expand to 48 kHz competitions in upcoming updates. The benchmarking metrics used include:
These metrics ensure comprehensive evaluation of model performance across different aspects of speech enhancement quality.
Based on the GitHub repository, the project is maintained by the @synapsec-ai/subnet-owners team. Contributors visible in the GitHub interface include:
Based on the GitHub repository, the project is maintained by the @synapsec-ai/subnet-owners team. Contributors visible in the GitHub interface include:
The subnet uses semantic versioning (Major.Minor.Patch) with specific implications for each release type:
Major Releases (X.0.0):
Minor Releases (0.X.0):
Patch Releases (0.0.X):
Version Milestones
SoundsRight v1.0.0
SoundsRight v1.1.0
SoundsRight v2.0.0
SoundsRight v3.0.0
SoundsRight v4.0.0
Long-term Vision
The current goal for the subnet is to facilitate open-source research and development of state-of-the-art speech enhancement models. The documentation acknowledges that there is potential to create far more open-source work in this field.
The ultimate goal of the subnet is to create a monetized product in the form of an API. However, to make the product as competitive as possible, the subnet’s first goal is to create a large body of work for miners to draw their inspiration from.
The subnet uses semantic versioning (Major.Minor.Patch) with specific implications for each release type:
Major Releases (X.0.0):
Minor Releases (0.X.0):
Patch Releases (0.0.X):
Version Milestones
SoundsRight v1.0.0
SoundsRight v1.1.0
SoundsRight v2.0.0
SoundsRight v3.0.0
SoundsRight v4.0.0
Long-term Vision
The current goal for the subnet is to facilitate open-source research and development of state-of-the-art speech enhancement models. The documentation acknowledges that there is potential to create far more open-source work in this field.
The ultimate goal of the subnet is to create a monetized product in the form of an API. However, to make the product as competitive as possible, the subnet’s first goal is to create a large body of work for miners to draw their inspiration from.