--- license: apache-2.0 --- # **Introduction** **`XY-Tokenizer`** is a speech codec that simultaneously models both semantic and acoustic aspects of speech, converting audio into discrete tokens and decoding them back to high-quality audio. It achieves efficient speech representation at only 1kbps with RVQ8 quantization at 12.5Hz frame rate. - **Paper:** [Read on arXiv](https://arxiv.org/abs/2506.23325) - **Source Code:** - [GitHub Repo](https://github.com/OpenMOSS/MOSS-TTSD/tree/main/XY_Tokenizer) - [Hugging Face Repo](https://huggingface.co/spaces/fnlp/MOSS-TTSD/tree/main/XY_Tokenizer) ## 📚 Related Project: **[MOSS-TTSD](https://huggingface.co/fnlp/MOSS-TTSD-v0.5)** **`XY-Tokenizer`** serves as the underlying neural codec for **`MOSS-TTSD`**, our 1.7B Audio Language Model. \ Explore **`MOSS-TTSD`** for advanced text-to-speech and other audio generation tasks on [GitHub](https://github.com/OpenMOSS/MOSS-TTSD), [Blog](http://www.open-moss.com/en/moss-ttsd/), [博客](https://www.open-moss.com/cn/moss-ttsd/), and [Space Demo](https://huggingface.co/spaces/fnlp/MOSS-TTSD). ## ✨ Features - **Dual-channel modeling**: Simultaneously captures semantic meaning and acoustic details - **Efficient representation**: 1kbps bitrate with RVQ8 quantization at 12.5Hz - **High-quality audio tokenization**: Convert speech to discrete tokens and back with minimal quality loss - **Long audio support**: Process audio files longer than 30 seconds using chunking with overlap - **Batch processing**: Efficiently process multiple audio files in batches - **24kHz output**: Generate high-quality 24kHz audio output ## 💻 Quick Start Here's how to use **`XY-Tokenizer`** with `transformers` to encode an audio file into discrete tokens and decode it back into a waveform. ```python import torchaudio from transformers import AutoFeatureExtractor, AutoModel # 1. Load the feature extractor and the codec model model_id = "OpenMOSS-Team/XY_Tokenizer_TTSD_V0_hf" feature_extractor = AutoFeatureExtractor.from_pretrained(model_id, trust_remote_code=True) codec = AutoModel.from_pretrained(model_id, trust_remote_code=True).eval().to("cuda") # 2. Load and preprocess the audio # The model expects a 16kHz sample rate. wav_form, sampling_rate = torchaudio.load("examples/m1.wav") if sampling_rate != 16000: wav_form = torchaudio.functional.resample(wav_form, orig_freq=sampling_rate, new_freq=16000) # 3. Encode the audio into discrete codes input_features = feature_extractor(wav_form, sampling_rate=16000, return_attention_mask=True, return_tensors="pt") # The 'code' dictionary contains the discrete audio codes code = codec.encode(input_features) # 4. Decode the codes back to an audio waveform # The output is high-quality 24kHz audio. output_wav = codec.decode(code["audio_codes"], overlap_seconds=10) # 5. Save the reconstructed audio for i, audio in enumerate(output_wav["audio_values"]): torchaudio.save(f"audio_{i}.wav", audio.cpu(), 24000) ```