Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

tensorflow2, librosa8, better plots #8

Merged
merged 32 commits into from
Sep 6, 2021
Merged

tensorflow2, librosa8, better plots #8

merged 32 commits into from
Sep 6, 2021

Conversation

faroit
Copy link
Owner

@faroit faroit commented Oct 25, 2019

  • update tensorflow to version 0.2.1
  • update librosa to 0.7.1
  • add tensorflow.io loading functions
  • update librosa to 0.8.0
  • removing gstreamer as its too slow\
  • add soxbindings
  • add stempeg

Bug fixes

  • write 16bit wav files instead of 32bit wav files
  • change averaging according to Update for librosa 0.7 #7
  • fix tfio error handling

@faroit faroit self-assigned this Oct 25, 2019
@faroit faroit added the enhancement New feature or request label Oct 25, 2019
@faroit faroit marked this pull request as ready for review January 21, 2020 19:50
@faroit faroit mentioned this pull request Jan 21, 2020
@bmcfee
Copy link
Contributor

bmcfee commented Jul 22, 2020

Gentle nudge on this one, since the reported librosa benchmarks are now two majors out of date 😁

@faroit
Copy link
Owner Author

faroit commented Jul 22, 2020

Gentle nudge on this one, since the reported librosa benchmarks are now two majors out of date 😁

sorry again especially, since I am aware that this should be addressed soon as I never wanted to make librosa look bad in this...
That being said, I have two main issues that I could need could get some help with, as I am still lacking time to finish this up...
Basically you mentioned two things here:

  1. Use timit to launch individual file loadings to reduce warm-up bias: I will propose something, maybe you could review that.

  2. Change the plotting to swarm plots: I played with that but I wasn't that successful in getting something that brings a clear message to viewers. I could provide the pandas files so you may play with that

  3. tf.data pipelines aren't really playing nice with errors (such as when a file format is not supported) using

dataset = dataset.apply(tf.data.experimental.ignore_errors()) results in no errors at all, thus making it even even harder to track down if the data was loaded successfully. So I guess I have to check the actual return value... any other idea?

@faroit
Copy link
Owner Author

faroit commented Jul 22, 2020

@bmcfee regarding 3. I guess the easiest is to have fixed matrix of supported pairs of codecs and libs that will be tested instead of using try/except...

@bmcfee
Copy link
Contributor

bmcfee commented Jul 22, 2020

1. I will propose something, maybe you could review that.

Sure, happy to!

2\. I played with that but I wasn't that successful in getting something that brings a clear message to viewers. I could provide the pandas files so you may play with that

That's fine -- it was just a suggestion. I generally have a preference toward raw data over aggregates, but if it's too unwieldy here, there's no problem leaving it as is.

results in no errors at all, thus making it even even harder to track down if the data was loaded successfully. So I guess I have to check the actual return value... any other idea?

Hrm, indeed. I guess you could do a length check (expected # of samples) if you don't want to do a full inspection of the data.

I guess the easiest is to have fixed matrix of supported pairs of codecs and libs that will be tested instead of using try/except...

That's probably a safer option, but a lot more work I'm guessing.

@faroit
Copy link
Owner Author

faroit commented Jul 22, 2020

@bmcfee a problem with 1. is that tf.data pipelines need to be compiled for each file. That would make the comparison between frameworks (np vs torch vs tf) useless as tf would probably be last. Still it would be meaningful to give relative results: aka. which loading method is the fasted for each tensor lib. Would you say this is worth it then?

@bmcfee
Copy link
Contributor

bmcfee commented Jul 22, 2020

That would make the comparison between frameworks (np vs torch vs tf) useless as tf would probably be last. Still it would be meaningful to give relative results: aka. which loading method is the fasted for each tensor lib. Would you say this is worth it then?

Yeah, that's tricky. Maybe worth reporting two versions in that case, one of which factors out the compilation time?

@faroit faroit changed the title Tf2 librosa7 tensorflow2, librosa8, better plots Sep 6, 2021
@faroit
Copy link
Owner Author

faroit commented Sep 6, 2021

@bmcfee i think its better to merge this now than waiting another year... sorry for this.
I didn't had time to polish the plots but all libs are up to date and i am using timit to get a better average

@faroit faroit merged commit f55aab7 into master Sep 6, 2021
@faroit faroit deleted the tf2_librosa7 branch September 6, 2021 18:22
@faroit faroit mentioned this pull request Sep 6, 2021
@bmcfee
Copy link
Contributor

bmcfee commented Sep 6, 2021

Thanks @faroit ! I'm just happy to see the version updates. 😁

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants