-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FEATURE]: Integration of M5 Network Architecture for Audio Processing #1161
Comments
Hello @codingwithsurya! Thank you for submitting the Feature Request Form. We appreciate your contribution. 👋 We will look into it and provide a response as soon as possible. To work on this feature request, you can follow these branch setup instructions:
Feel free to make the necessary changes in this branch and submit a pull request when you're ready. Best regards, |
here's how i think we can incorporate the m5 network architecture. so, under training/core directory, we can make a file called m5_model.py. In the file, we can define the M5 network architecture as a class M5Net that extends nn.Module.
In the training/core/dl_model.py file, we can import the M5Net class: and, in the DLModel class in dl_model.py, we can add a new method to create an instance of M5Net:
In the training/routes/audio/audio.py file, we can use the DLModel.fromM5Net() method to create an instance of the M5 network when needed. we also need to update the training pipeline in training/core/trainer.py but this is how we can get started |
Feature Name
To enhance the Deep Learning Playground's audio data processing capabilities, we aim to integrate the M5 network architecture, inspired by the M5 network. This architecture is crucial for processing raw audio data efficiently, especially focusing on the receptive field of the first layer's filters.
Your Name
Surya Subramanian
Description
We are currently working on creating an audio trainspace in our deep learning playground. As part of this, we need to integrate a convolutional neural network to process raw audio data. The specific architecture we are looking to implement is modeled after the M5 network architecture, which is described in detail in this paper: https://arxiv.org/pdf/1610.00087.pdf.
Here is the Python code for the M5 network architecture:
(also available here: https://colab.research.google.com/github/pytorch/tutorials/blob/gh-pages/_downloads/audio_classifier_tutorial.ipynb#scrollTo=iXUe9kHdcV16)
The task is to integrate this model into our training directory. The model should be callable from the audio.py route. The files training/core/training.py and training/core/dmodel.py might be useful for this integration.
This is kind of open-ended so feel free to play around with this! Lmk if you have any questions.
So in summary:
Objectives
Neural Network Integration: Implement the M5 network architecture within the Deep Learning Playground's training module to handle audio data processing effectively.
Model Components: Define the neural network layers including convolutional, batch normalization, pooling, and fully connected layers as per the M5 architecture specifications.
Compatibility: Ensure seamless integration of the M5 network architecture with the existing training pipeline and trainers in the audio route (
audio.py
).Implementation Details
Create a new neural network class
Net
intraining/core/dl_model.py
with the M5 network architecture specifications.Include layers for convolution, batch normalization, pooling, and fully connected layers as described in the M5 architecture.
Update the training pipeline in
training/core/training.py
to utilize theNet
model for audio data processing tasks.The text was updated successfully, but these errors were encountered: