Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -81,6 +81,7 @@ def prepare_media(source_path_or_url: os.PathLike,
|
|
| 81 |
|
| 82 |
for line in iter(process.stdout.readline, ''):
|
| 83 |
# Filter out unnecessary messages
|
|
|
|
| 84 |
if "www.google.com/device" in line:
|
| 85 |
hl_text = line.replace("https://www.google.com/device", "\033[93mhttps://www.google.com/device\x1b[0m").split()
|
| 86 |
hl_text[-1] = "\x1b[31;1m" + hl_text[-1] + "\x1b[0m"
|
|
@@ -173,14 +174,19 @@ with gr.Blocks(theme=theme, css=css) as demo:
|
|
| 173 |
gr.Markdown(
|
| 174 |
f"""
|
| 175 |
## 🎶YourMT3+: Multi-instrument Music Transcription with Enhanced Transformer Architectures and Cross-dataset Stem Augmentation
|
| 176 |
-
## Model card:
|
| 177 |
- Model name: `{model_name}`
|
| 178 |
-
|
| 179 |
-
|
| 180 |
-
|
| 181 |
-
|
| 182 |
-
|
| 183 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 184 |
|
| 185 |
## Caution:
|
| 186 |
- For acadmic reproduction purpose, we strongly recommend to use [Colab Demo](https://colab.research.google.com/drive/1AgOVEBfZknDkjmSRA7leoa81a2vrnhBG?usp=sharing) with multiple checkpoints.
|
|
|
|
| 81 |
|
| 82 |
for line in iter(process.stdout.readline, ''):
|
| 83 |
# Filter out unnecessary messages
|
| 84 |
+
print(line)
|
| 85 |
if "www.google.com/device" in line:
|
| 86 |
hl_text = line.replace("https://www.google.com/device", "\033[93mhttps://www.google.com/device\x1b[0m").split()
|
| 87 |
hl_text[-1] = "\x1b[31;1m" + hl_text[-1] + "\x1b[0m"
|
|
|
|
| 174 |
gr.Markdown(
|
| 175 |
f"""
|
| 176 |
## 🎶YourMT3+: Multi-instrument Music Transcription with Enhanced Transformer Architectures and Cross-dataset Stem Augmentation
|
|
|
|
| 177 |
- Model name: `{model_name}`
|
| 178 |
+
<▶model details◀>
|
| 179 |
+
<summary>(Details)</summary>
|
| 180 |
+
|
| 181 |
+
| **Component** | **Details** |
|
| 182 |
+
|--------------------------|--------------------------------------------------|
|
| 183 |
+
| Encoder backbone | Perceiver-TF + Mixture of Experts (2/8) |
|
| 184 |
+
| Decoder backbone | Multi-channel T5-small |
|
| 185 |
+
| Tokenizer | MT3 tokens with Singing extension |
|
| 186 |
+
| Dataset | YourMT3 dataset |
|
| 187 |
+
| Augmentation strategy | Intra-/Cross dataset stem augment, No Pitch-shifting |
|
| 188 |
+
| FP Precision | BF16-mixed for training, FP16 for inference |
|
| 189 |
+
</details>
|
| 190 |
|
| 191 |
## Caution:
|
| 192 |
- For acadmic reproduction purpose, we strongly recommend to use [Colab Demo](https://colab.research.google.com/drive/1AgOVEBfZknDkjmSRA7leoa81a2vrnhBG?usp=sharing) with multiple checkpoints.
|