jeevster commited on
Commit
3ce0347
Β·
1 Parent(s): 93c2f63

add thumbnail, clean up description

Browse files
Files changed (5) hide show
  1. README.md +1 -0
  2. about.md +5 -3
  3. app.py +2 -2
  4. site/logo.jpeg +3 -0
  5. site/tsne.jpeg +0 -0
README.md CHANGED
@@ -1,6 +1,7 @@
1
  ---
2
  title: Carnatic Raga Classifier
3
  emoji: πŸ“ˆ
 
4
  colorFrom: pink
5
  colorTo: green
6
  sdk: gradio
 
1
  ---
2
  title: Carnatic Raga Classifier
3
  emoji: πŸ“ˆ
4
+ thumbnail: site/logo.jpeg
5
  colorFrom: pink
6
  colorTo: green
7
  sdk: gradio
about.md CHANGED
@@ -1,9 +1,11 @@
1
  ### About the Classifier
2
- The classifier is a [convolutional neural network](https://en.wikipedia.org/wiki/Convolutional_neural_network) trained on over 10,000 hours of Carnatic audio sourced from this incredible [YouTube collection](https://ramanarunachalam.github.io/Music/Carnatic/carnatic.html).
3
  ### Key Features:
4
  - Can identify **150 ragas** most commonly found on YouTube
5
- - Does not require any information about the **shruthi (tonic pitch)** of the recording.
6
- - **Compatible** with male/female vocal or instrumental recordings.
 
 
7
 
8
  ### Interpreting the Classifier:
9
  We can gain an intuitive sense for what the classifier has learned. Here is a [t-SNE](https://en.wikipedia.org/wiki/T-distributed_stochastic_neighbor_embedding) projection of the hidden activations averaged per ragam. Each point is a ragam, and relative distances between the points indicate the degree to which the classifier thinks the ragas are similar. Each ragam is color coded by the [melakartha chakra](https://en.wikipedia.org/wiki/Melakarta#Chakras) it belongs to. We observe that the classifier has learned to a representation that roughly corresponds to these chakras!
 
1
  ### About the Classifier
2
+ The classifier is a [convolutional neural network](https://en.wikipedia.org/wiki/Convolutional_neural_network) trained on over 10,000 hours of Carnatic audio sourced from this incredible [YouTube collection](https://ramanarunachalam.github.io/Music/Carnatic/carnatic.html).
3
  ### Key Features:
4
  - Can identify **150 ragas** most commonly found on YouTube
5
+ - Does not require any information about the **shruthi (tonic pitch)** of the recording
6
+ - **Compatible** with male/female vocal or instrumental recordings
7
+
8
+ For those who are interested, the inference code and model checkpoints are available at the 'Files' tab in the header.
9
 
10
  ### Interpreting the Classifier:
11
  We can gain an intuitive sense for what the classifier has learned. Here is a [t-SNE](https://en.wikipedia.org/wiki/T-distributed_stochastic_neighbor_embedding) projection of the hidden activations averaged per ragam. Each point is a ragam, and relative distances between the points indicate the degree to which the classifier thinks the ragas are similar. Each ragam is color coded by the [melakartha chakra](https://en.wikipedia.org/wiki/Melakarta#Chakras) it belongs to. We observe that the classifier has learned to a representation that roughly corresponds to these chakras!
app.py CHANGED
@@ -37,13 +37,13 @@ if __name__ == '__main__':
37
  with gr.Tab("Classifier"):
38
  gr.Interface(
39
  title="Carnatic Raga Classifier",
40
- description="**Welcome!** This is a deep-learning based raga classifier. Upload or record an audio clip to test it out. Provide at least 30 seconds of audio for best results. Wait for the audio waves to appear (and stay) before clicking Submit! \n",
41
  article = "**Get in Touch:** Feel free to reach out to [me](https://sanjeevraja.com/) via email (sanjeevr AT berkeley DOT edu) with any questions or feedback! ",
42
- fn=evaluator.inference,
43
  inputs=[
44
  gr.Slider(minimum = 1, maximum = 150, value = 5, label = "Number of displayed ragas", info = "Choose number of top predictions to display"),
45
  gr.Audio()
46
  ],
 
47
  outputs="label",
48
  allow_flagging = False
49
  )
 
37
  with gr.Tab("Classifier"):
38
  gr.Interface(
39
  title="Carnatic Raga Classifier",
40
+ description="**Welcome!** This app uses AI to recognize Carnatic ragas. Upload or record an audio clip to test it out. Provide at least 30 seconds of audio for best results. Wait for the audio waves to appear and remain before clicking Submit! \n",
41
  article = "**Get in Touch:** Feel free to reach out to [me](https://sanjeevraja.com/) via email (sanjeevr AT berkeley DOT edu) with any questions or feedback! ",
 
42
  inputs=[
43
  gr.Slider(minimum = 1, maximum = 150, value = 5, label = "Number of displayed ragas", info = "Choose number of top predictions to display"),
44
  gr.Audio()
45
  ],
46
+ fn=evaluator.inference,
47
  outputs="label",
48
  allow_flagging = False
49
  )
site/logo.jpeg ADDED

Git LFS Details

  • SHA256: 162a6e102760b79a191077f48760b3b40d2def0447d01de5178f79933bb986d0
  • Pointer size: 132 Bytes
  • Size of remote file: 7.42 MB
site/tsne.jpeg CHANGED

Git LFS Details

  • SHA256: d257d1a43ff3fb1e4ef43302b5f403c15a4808ff90036e305083d9d7badcb3c7
  • Pointer size: 131 Bytes
  • Size of remote file: 228 kB