-
Notifications
You must be signed in to change notification settings - Fork 99
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
some errors when install kaldi-python #2
Comments
As far as I remember Kaldi has to be compiled differently for Kaldi-python installation to be successful. @dmitriy-serdyuk , @janchorowski , can you please comment on that? |
Can you tell me the methods you compile kaldi.......(aslo means : how can you get the file like kaldi_matrix.a.........) From: Dzmitry Bahdanau |
Right, sorry, that I didn't mention this. Kaldi should be compile with
|
Could you please change the documentation? I guess it makes sense to do it our private repository, since we are going to make what we have there the new master pretty soon. |
After change the configure command, problem solved...... |
You can use GPU in the same way as you usually do it with Theano. Please On 23 November 2015 at 05:51, Entonytang [email protected] wrote:
|
After adding "device =gpu3" while I find GPU Process in GPU 2(device K40).....using default wsj_paper6.yaml..... |
As I measured recently, one step was taking about 6 seconds on a Titan X, K40 was a bit slower, about 8-9 seconds. So probably something goes wrong. Make sure that Theano writes something like |
Also use optimizer=fast_run in your THEANO_FLAGS On 24 November 2015 at 10:34, dmitriy-serdyuk [email protected]
|
thanks, solved..... Epoch 0, step 829 | # | Elapsed Time: 2:09:35Training status: Epoch 0, step 830 | # | Elapsed Time: 2:09:46 |
Is there exception or a core dump? Otherwise it's something wrong with your OS. |
I don't think so, I use another core and try again. the result is similar...... Epoch 0, step 84 | #| Elapsed Time: 0:09:18Training status: |
38535d8 Merge remote-tracking branch 'blocks-origin/master' into new_blocks ed481ef Merge pull request #953 from dwf/shut_up_sqlite 417dbdf Skip test_save_the_best with sqlite. f5ee622 Merge pull request #946 from vdumoulin/abstract_conv 648646d Add test for ConvolutionalTransposeActivation 87d143f Reorder constructor args to help remove duplicate code 843d3ff Replace deprecated `image_shape` kwarg with `input_shape` 65e9435 Clarify parameter description 6712e44 Make ConvolutionalTranspose inherit from Convolutional ff08eb8 Reduce repeated code in constructors. abf41f8 Make imshp not depend on input_ 7c895aa Construct grad op using constant values rather than depend on a shared variable 2a68d89 Fix superclass call 836267d Revert to not forcing compilation mode c237b51 Comment on the use of FAST_RUN mode in blocks.bricks.conv tests fa4f314 Fix conv tests failing to compile Theano functions c985b30 Fix import error ee5efc6 Fix oversights 071ad2b Add ConvolutionTranspose and ConvolutionTransposeActivation 9255647 Use Theano's new abstract_conv interface 348de31 Adapt Pooling to Theano's downsample -> pool renaming 98797a1 Merge pull request #941 from dwf/batch_norm_bricks 90cc8bb Fix StepRules to honour broadcastable. d9171ff Move __init__ logic to _allocate in SBN. a88b193 batch_normalization_updates -> get_batch_normalization_updates 725ec01 Idiomatize test. 4b39c93 Remove from 'with' for clarity. e714903 Make duplicate updates docs more explicit. 333b49c Rename W, b -> scale, shift. 09da496 Add batch_normalization_updates. d81c03c Annotate population parameters with brick. 474d3f1 Fix subtle bug with mutable metadata. b03fffc Fix typo in epsilon assignment. 6895b87 Remove needless properties. f41c286 Improve _add_batch_axis. cf4f4db Ensure correct nesting of context managers. 9805b73 Improve docs. Add example for apply_... 7efa956 Correct the assertion. 5ba9b38 Add comment and assertion following review. c70a26b Fix Scrutinizer error from 6ba3364. 2ea41a8 Improve robustness of context manager logic. e22b151 Add an assertion following review. a26e2c6 Simplify generator expression after review. d35c22b Amend error message following review. 2e7cd0e Rename save_memory -> conserve_memory. 3dc9135 More tests, move graph tests to tests/graph. ad3a84d Fix doctest. 73a6902 Comments explaining tests. 93e55f6 Big refactor to enable context manager operation. 59fd7cd Add brick filter and smoke test for bug fixed. b358849 Remove leading batch axis from BN shared variables. 02982e9 Correctly crawl the graph backward. f26c1e5 Expose save_memory in BatchNormalizedMLP. 7ede89b Make batch_normalize return the pop->minibatch map. 9ff63b0 Add tests. 6b52b35 Refactor batchnorm + graph module. 9b94116 Reorganize bricks subpackage. 0f6623f Initial batch-normalization implementation. 2046525 Merge pull request #944 from mila-udem/bound_application_error_msg 07d43ae Correct error message in filter.py. 07cec2c Merge pull request #774 from lamblin/update_doc 25c5141 Merge pull request #940 from rizar/without_namespace_packages c0640b0 Disable namespace packages suppport 83c5d2a Merge pull request #838 from matthiasreisser/docstring_additions 260296d Merge pull request #929 from dwf/conv_doc_fixes acf9c80 Merge pull request #930 from r2007/patch-1 353d525 Update bricks_overview.rst 506dfe1 Merge pull request #928 from dwf/shared_floatx_kwargs b7f5df3 Pass along kwargs in shared_like. fce7f6b Merge pull request #925 from dwf/get_output_shape_update ff70906 More robust get_output_shape call. 928dcbb Better document get_output_shape. c6cd413 Update ConvolutionalSequence docs. 5928f34 Pass shared_floatx kwargs to theano.shared. 3eddcf8 Merge pull request #921 from dwf/die_conv_layer_die 65e9952 Replace deprecated getOutputShape interface. 589543e Remove mention of `ConvolutionalLayer` from a docstring. 3cd7bfd Get rid of ConvolutionalLayer. Fixes #912. df5c55c Merge pull request #924 from dwf/fix_doctest_debugprint eedab38 Fix debugprint output for Theano/Theano#1953. 6827a9e Merge pull request #891 from dmitriy-serdyuk/release-doc 1d63fa8 Merge pull request #899 from dwf/conv_improvements 769eb70 Test for overzealous support code. 6177388 Making unpickling robust to MaxPooling change. 24a489a Merge pull request #909 from akhti/doc_fix 7734be3 Increase font size in the generator scheme 04783b7 Merge pull request #886 from sotelo/return_states_seq_gen f7866d8 Merge pull request #892 from akhti/fix_bidir 05fea66 Merge pull request #907 from dmitriy-serdyuk/fix-log-doc 7576bcc Add API sections for two log backends b87ad2e Make ignore_border=True the default. c374557 Test Pooling Bricks work in ConvolutionalSequence. 27d71a5 Tests for AveragePooling, new MaxPooling flags. f1e5908 AveragePooling Brick. ac9f095 ConvolutionalSequence: use num_output_channels. dbb407c Refactor MaxPooling. 414519d Merge pull request #903 from dwf/fix_travis_again e185f3d Hardcode a prefix for Travis Miniconda install. 7e10ed5 Merge pull request #897 from dwf/custom_scaling_dropout c4426c3 Notes on dropout usage. 07c98e2 Add dropout tests, including custom_divisor. d7d8b62 apply_dropout: support a custom divisor. 9ffc3e1 Improve apply_dropout documentation. 23d94b2 Fix label in sequence generator scheme 6750051 Merge pull request #896 from rizar/add_original_svg d7eb341 Update after review. 8950f73 Backport fixes done to Fuel's install doc 68842e7 Update developer doc to mention Blocks and Fuel ea34217 Add the original for SequenceGenerator picture 455d65f Improve docs for making new release bc4f62c Fix too long line error 46652f7 Add test for a stack of bidirectional layers 98aed8e Add instructions to make a new release 3ec837a Merge pull request #887 from rizar/fix_confpy_and_bump_version 4f74b4c Merge pull request #888 from akhti/fix_bidir 69bc613 Add get_dim to Bidirectional 0a4e3b3 Fixes conf.py and bumps version to 0.1.1 afaa45f Added the final values of states as auxiliary variables in a sequence generator so they can be reused. 12e50d9 Merge pull request #772 from adbrebs/doc_brick 92654e1 Merge pull request #881 from mila-udem/correct_main_loop_error_message 5dc2bdd brick tutorial improvements 168c7a8 Merge pull request #878 from rizar/release-0.1 f275332 Merge pull request #879 from sotelo/generator_cost_documentation e43cfbe Correct main loop message for the interrupt 6c99076 Added the missing links. c11698c Merge pull request #880 from sotelo/missing_apply_decorator_documentation 679b01c Improved the error message when an application decorator is missing. efe3585 Solved the flake8 mistake. 6ebcb37 Added documentation about the emitter cost in sequence generators. a06878f Bump version number 0f889aa Merge pull request #875 from dwf/use_bias_convolutional_activation_layer 0ca4e86 Merge pull request #873 from dwf/border_mode_conv_sequence d1af6c9 ConvolutionalSequence: propagate use_bias. 7bc11d4 Convolutional{Activation,Layer}: respect use_bias. 457049d Convolutional{Activation,Layer}: Refactor alloc. 6dec566 ConvolutionalSequence: Don't auto push border_mode. b3754dc Merge pull request #849 from mila-udem/redefine_modelr 837da08 Merge pull request #868 from lukemetz/lm/set_parameter_values 8a06204 warn if setting wrong shape parameters c843fa8 Merge pull request #867 from dwf/conv2d_impl 620bc7b Additional developer documentation. 8a8e8c1 Fix Scrutinizer complaint about whitespace. c2ebc25 Make Convolutional's conv implementation hackable. c427fa7 Merge pull request #864 from mila-udem/rtd2 377688c Add mocking back in 2fc06f3 Merge pull request #861 from mila-udem/rtd c5b9f1c Changes 743cbf3 Add a test and fix imports fe9daeb Back to items() bb81302 Call __init__ and add iteritems a9751cb Merge pull request #852 from rizar/testing_utils 5772cc3 Fix formatting 8fd7da4 Improve documentation 0cc3031 Fix imports in doctests 3100e11 Move testing utilities to make them importable from outside 5899425 Merge pull request #844 from mila-udem/mention_extras_and_examples b46dbe0 Refactor model and write a bit of documentation f6a99bc Mention examples in the documentation as well. f38881a Add forgotten period<F2> 3c1fdb5 Also refer to Fuel 96d63bf Better looking reference names bb3f6c8 Mention blocks-extras and blocks-examples in README.md 8bf07e7 typos 4b83783 Merge pull request #839 from mila-udem/deps 1436c25 Add nose 59a9553 Update deps f295d76 Added line break cba09d2 Changed docstring to plural bba2e2d added docstring to apply method of LSTM d576831 Small fixes in create your own brick tutorial bdff06b Merge pull request #1 from dmitriy-serdyuk/doc_brick 5da4696 Separate paragraph 536bf18 Make small fixes in create your own brick e377e43 Merge pull request #834 from yingzha/dev bd12f44 Rephrase docs f6e9896 Merge pull request #777 from galv/master 03a491b Add new line to comment block. ecd121d Pass input from ComputationGraph.get_theano_function() to theano.function() 53979de Example of lazy brick with get_dim 5628491 More explanations in the tutorial to create a custom brick 627da45 Refactoring. Brick example with children. ca1c7cf Very first draft of the tutorial to create a brick. 7e2535d Merge pull request #835 from johnarevalo/patch-2 e844b92 Allow uint type for lookup indexing cebec4e Replace uses of named_copy with var.copy(name=...) 7ff0f6b Merge pull request #827 from vzhong/conv-1d bc140ba fix for case in which no image size is specified for Convolution #825 9e4d0e4 Merge pull request #826 from lukemetz/lm/algorithm_roles a976b9f add roles and names to algorithm shared variables 98ed3d1 Merge pull request #823 from rizar/sgd_profile d335fd0 use parens 1f16ba5 add test a4b50e0 add theano_func_kwargs to algorithm init and use respectively 3403846 add kwargs to base class initialize func 53f292e remove unnecessary vars 3b0bbd0 add profile via kwargs to gradient descent initialize 0f6b4e0 Merge pull request #819 from rizar/fix_iterate_false 67f7388 Remove old code 2a8dff3 Arguments for SimpleRecurrent should be required 69b8ce2 Merge pull request #818 from ASalvail/master 436134a Reformated modified doc for docstrings requirements. 0e15726 Reformated modified doc for docstrings requirements. e5e14a4 Elaborate LSTM inputs argument to apply method da5d320 Merge pull request #815 from yingzha/dev f88a638 Merge pull request #2 from dwf/yingzha_ccw 79243e5 Simplify documentation. dce9c48 Merge pull request #813 from dwf/parallel_routing c7d3540 Use OrderedDict per Dima's review. b17fc13 Refactor into utils func, per Dima's request. 23fd90a Merge pull request #1 from dwf/yingzha_ccw d3c5caa Add test for argument Parallel.apply argument validation. 16c50ab Fix Parallel.apply argument routing logic. e26d04d Fix for str config with default of None. 3144fc9 remove dir argument in test_config 0fe9a92 fixed flake8/pep errors 29099d3 Override base path for temporary file creation f2c9f3e Merge pull request #798 from ASalvail/master 760b144 Merge pull request #808 from mila-udem/selector_docs d7bf7f7 Remove unnecessary import in doctest. 45c400a Eschew needless initialization. 4f3a075 Reword per Dima's review. f95f919 Improve Selector.get_parameter documentation. 325a431 Merge pull request #805 from mila-udem/cost_cost_matrix_args_kwargs 83151ca Make Cost and CostMatrix more general. 9c2a1a1 Merge pull request #803 from Tejas-Khot/Tejas-Khot-patch-1 5c1062b made line length less than 75 characters d299457 corrected trailing whitespace problem 968edbc added docstring for tied_biases 0a8cc3d Corrected docstring of Softplus. e48bbb9 Add activation tests (Softplus, Softmax, Logistic) 9ad61bc Softplus docs and reference. fab1fbe Added a Softplus brick. c72a833 Merge pull request #759 from rizar/good_softmax_for_all 4da97cd Merge pull request #790 from mila-udem/linear_docs c689353 Make Linear docstring clearer. git-subtree-dir: libs/blocks git-subtree-split: 38535d89ec166d8b4bef912d4745a51231cdc6da
41248a1 Merge remote-tracking branch 'theano-origin/master' into new_theano 4a75395 Merge pull request #4154 from abergeron/fix_blocksparse 11ad9bb Merge pull request #4152 from Theano/fix_doc d9237bf Add tests for infer_shape in blocksparse. 83d7531 Use __props__ for the gpu versions of blocksparse. 4c41ccd Replace subtensors with dimshuffles in h_softmax. 1d042d9 Add infer_shape for the blocksparse ops and pass though the broadcastable flags. 9e01f0e Fix doc build 2479cc7 Add __props__ to blocksparse ops. 4e7f550 Changes for 0.8.0rc1 release 830544b Merge pull request #4134 from nouiz/news 1086900 clean-up of NEWS_DEV 4677cd3 List of committers d9c866f More aliases 0900b66 Merge pull request #4013 from Sentient07/issue-3573 1a3f4e6 Merge pull request #3950 from aalmah/rand_weighted_select_c_impl 6ecfe0f Merge pull request #4116 from ballasn/slowimpl c421062 import xrange from six.moves for python3 compatibility 92b57e1 check border_mode in the perform methods d94cdf4 Merge pull request #4094 from abergeron/lift_ifelse e7f6b32 Merge pull request #2817 from lamblin/detect_nvcc_cast_bug bf79d0a Update NEWS_DEV.txt 50e6246 Fix the test. dbf22bb Better fix the test. 7399bc4 Add test. af3d84b Fix problems in optimization. c51b283 Merge pull request #4129 from nouiz/cleanup 04bc006 Merge pull request #4121 from nouiz/doc_contrib caf517b Deleting removed nodes' set from on_detach 28e06d1 Merge pull request #4125 from abergeron/fix_scan_bug 2c77122 change cache code 14241a7 make sure we don't get out of memory bb964aa Add optimization that verifies that no AbstractConv are present in the graph f408152 add suggested fixes 4022b2a update test 6abbbc4 update 2a0c192 Allow to easy toggle the extra infer_shape in the code. 11b755f Make FunctionGraph.disown() revert more stuff 4052c18 Remove not use attribute b12a695 Merge pull request #4127 from abergeron/fix_buildbot c549b50 Merge pull request #4118 from abergeron/fix_bin 353d86d Missing flake8 fix. c839e1b Make the add check faster and flake8 format this part. be8649b Flake8 errors from the buildbot. 522d826 Fix wrong class name. c6bfe4d Add test to make sure dots are pushed out. 0ce2958 Remove the part about pushing out dot, it's done elsewhere and this code doesn't work. d3d3a54 Do the toposort on the cloned nodes and build the client dict for client checking. 87d7a3b Check for Elemwise add in a less fragile way. f5c7aa5 Allow callers of io_toposort to get a dictionary of clients as a side effect. 5fd0a27 Doc to check for easy fix issue as a good starter 49a4624 Try to detect bug in nvcc. da9006e We need to import gof.compiledir in theano-cache. 93b4fb5 small fix to make the tests pass + flake8 ecc4e7b move debug mode test in tensor.nnet.test 63379af gpu inputs/outputs 784bab4 handle gpu input/outpus 46d6003 fix rectangular padding issue f78903a clean-up 417e6fb Fix border_mode 7065392 add DebugMode implementation for conv2d 846804d flake8 ff50c40 fixed bugs 0aa5ff7 Merge pull request #4018 from abergeron/fix_buildbot c42f938 minor fixes bb93f65 first attempt 42907a0 Merge pull request #4099 from JohnZed/master 4565218 Fix stale import of cuda_enabled in rng_mrg (issue 4098) d316bae Merge pull request #4107 from nouiz/matt-graham-cudnn_paths_fix 58135c6 Ad dnn.include_dirs by default includa cuda_root, we remove it by default instead of doing a warning. d392433 Remove empty string in lib_dirs and include_dirs 2a339bb Skip the CorrMM tests if there is no blas. 6429c27 Merge pull request #4095 from abergeron/move_flags2 8472d13 Merge pull request #3965 from harmdevries89/gpudnnpool2 8d975e6 Moved the variables into on_attach method 20e9ae0 Print the theano configuration on travis. eab9cf5 Merge pull request #4092 from abergeron/fix_buildbot2 a08fc0c Merge pull request #4069 from hantek/docassert 028459c Merge pull request #3768 from nouiz/nanguardmode f2100b0 Fix linking with mkl. 9ddb5e2 Fix format string for python < 3 3098ee0 Move leftover flags that seemed forgotten. 7d89882 Add lifter for IfElse. 7618e10 Make scipy happy across versions. b661e71 Fix toarray() call to be optional. a5da2c0 fix an example 8d68001 Merge pull request #4090 from nouiz/blas_rpath dcac871 Merge pull request #4046 from AnishShah/issue4035 5cb51d3 Fix doc syntax 145e1cf Fix opt e5ecd23 small doc fix 0912215 Fix GpuAlloc using debug_perform from GpuAllocEmpty. 44fbc1b Fix debug_perform of GpuAllocEmpty 5fce0ad These are (now?!) sparse matrices. 41607d7 Flake 8 in buildbot? 1ec5b4d Don't use the old name in tests. 1d5fe30 flake8 35df45e Merge pull request #4079 from nouiz/stack_trace e5f414f Merge pull request #4076 from nouiz/eval ae61464 Merge pull request #4089 from nouiz/gemv_broadcast fec5c70 fix e1c65a5 remove test fd278b8 fix doctest errors in opt.py c13c6d4 add doc for opt b31fd4b make doc appear 102f7b2 Try blas flags with rpath if they don't work. feaf7d4 Fix crash with gemv and some broadcast pattern b69b276 Merge pull request #4084 from matt-graham/cudnn_paths_fix bbb8f71 Merge pull request #4083 from caglar/fix_cnmemprint_msg 6b98e58 Merge pull request #4065 from abergeron/lift_dot22scalar 9625e08 added validate method to ReplaceValidate b533e26 Remove another eval that was needed due to the change of interface. 6da8726 Also changing libs + lib_dirs in gof/cmodule.py for consistency. 9d456b3 Making order of libs consistent with include_dirs and lib_dirs. b4ff2ff Addition of user specified cuDNN paths to Op compile commands. 540e7d3 Alternate strategy to not run the test in the base class. deb1f28 Add infer_shape tests for the blas ops. 7fc3d04 Added the latest changes. 0bc7d0d Test the lifted function for Dot22Scalar. dfe4049 Flake8 8215457 Merge pull request #4067 from abergeron/debugmode_empty 0f23609 Fix flake8 189069b Merge pull request #4070 from abergeron/fix_buildbot2 e482caf Better docstring 02b95b4 small docstring update 09215e4 Tell user how to make the stack trace longer f3e4e26 docstring, test and fix the number of stack level kept cf0abd4 Enable fast stack trace for python 3.5. I tested it and it work. 65e9846 rename var 42c00fb make sure to always have at least one user stack trace and don't check the stack for more then one 6b02f8c Merge pull request #4030 from hantek/fixtesterr 360ddaf Don't call eval 68880f8 Update opts using values_eq_approx to the new interface. 0ec6362 Fix typo in infer_shape methods. 3adaa82 Merge pull request #4042 from adbrebs/warn_flags d3530b0 Fix typos and wording. 2f739ed Fix the LogSoftmax tests in DebugMode. 7d036b8 Make sure that subsample is a tuple not a list. 62a7e19 Add infer_shape to the blas ops. 480ca3a Replace Dot22Scalar by a GpuGemm instead and add a test. 45753ed Made the suggested changes adb4f3b Document debug_perform(). ef01a4e Fix new opt 8b0ed6f fix comment 0185804 indentation 95f466d Doc var.tag.nan_guard_mode_check 2c1f7d8 python 3 compatibility 666b86b Merge pull request #4041 from abergeron/fix_buildbot2 9b0c950 Remove unused imports 325ce51 modify test config to check that flags are removed f2e3c8d Make AllocEmpty return constant-valued memory under DebugMode so that it won't trip up checks. 27ee537 Allow nodes to have a 'debug' behaviour for DebugMode. 622f1d5 Fix calls to as_gpuarray_variable. 90d3193 Also lift Dot22Scalar. b05017a Fix imports in tests. c386dd5 Explain more how dummy is there for backward compat and cleanup a bit. 2fc9f31 Change the base conv test class so it doesn't get tested itself. 3811ac7 Don't check inputs if it is the output of computation. If should be already tested 407f0f5 Don't check input if we told not aaab570 Fix registration of new opt b9b10ce Add optional opt that convert all *AllocEmpty to *Alloc. ef8ee79 Fix regression in this PR while unpickling. 659f639 Small file cleanup bc2f679 Remove false error by NanGuardMode due to GpuAllocEmpty 0c5014b Fix comment 11e1a82 Preserv the nan_guard_mode_check 578ad4f Don't let scan AllocEmpty cause false alarm by NanGuardMode e088e2a Merge pull request #4061 from nouiz/param fdd3150 Correct In parameter name following Param removal ca19b61 update prepare_node blas 5212d68 add pickle 4f2b268 old interface compatible bc5b983 up 0e0fa00 separate test for new pool interface b41c0a0 remove prints bbd9ff3 flake8 5c17201 gpu dnn pool takes tensor variables 494ff73 Print warnings when flags are not recognised 556500b Remove default value when flag is not found in the flag dictionary 274cf81 tip for mixed dtypes 3be031f Re-arrange order of imports in theano/__init__ a2fb4f6 fix local_bitwidth and python_int_bitwidth imports 50e0677 Merge pull request #4040 from caglar/fix_fload_multi_bug f51b38b delete empty testoutput block b5d75ef add back test in yaml file 9f46b69 Merge pull request #4045 from nouiz/corrmm 5d8526d Merge pull request #3942 from shabanian/tensor_signal_pep8 e34c32f solve more doctest errors 57aae87 flake8 1b4784a Fix flake8 and names. 7145695 Add a comment explaining the dummy param. 028aaef Move the backward compat where it should be. 51d97ce Remove name that was deprecated in 2010. This is super old. 5c4aafe fix some circular imports 1cb731e sloving test failures a262b3b Make sure the docstring is attached to the right object and reuse objects. 1d017d2 Merge pull request #4009 from AnishShah/elu 26d0f24 Fix check d9e0d1c Merge pull request #4043 from diogo149/master 435caa7 Don't use CpuCorrMM when there is no CPU BLAS c037151 Correct warning re: cuDNN and pool_2d's ignore_border argument 1647420 raise an error if the user provides an unknown flag fbb9a5a Fix import problems in configdefaults c0e864d config stuff be629ae # This is a combination of 2 commits. # The first commit's message is: 2055fee Exponential Linear Units (ELUs) 6010cad fixed the undefined variable problem. aa0982e Use working subsample values. 2c19264 Remove pool.py from the whitelist (it has no errors). 28b9927 Remove the bad compat for pickled DownsampleFactorMaxGrad and use a better one. 2d29bb3 Fix the output shape in python to match the C code for DownsampleFactorMaxGradGrad. 8a736df Merge pull request #4033 from nouiz/py3_repr 931dfec Merge pull request #3795 from aalmah/rand_weighted_select 5fc766b Test border_modes and strides together. 8f44215 fixed the multi grad bug. 20609f0 Add test for the previous fix 6c8f1a1 Refactor how the parameters are iterated over to reduce the total time of the test. (This covers less cases, but should be almost equivalent.) 332601b fix gh-4036. We need the shape and we don't compute it. It wasn't timmed, so don't try this optimization. dbb49e4 Better error message 12c4530 Make DebugMode raise the good error when no code exist for an op b236266 Merge pull request #3964 from memimo/3929 4bec565 Merge pull request #4032 from sygi/dot-a-file-in-ldlibrary 3835a61 Fix CudaNdarray repr in python 3. This fix daily buildbot error in python 3 e9425be Merge pull request #4000 from nouiz/cudnn_version 455db13 (#4024) cutting the *.a extension in ldlibrary file ef41de9 Make cnmem print the start size to make it more clear for user. 290631f Fix typo in warning 2c13f91 Fix crash during opt when we where looking for cudnn version. 89be3bc If cudnn isn't there and we use cuda back-end, print it. b7f0c30 Only raise error if convolution or pooling are used in the new back-end. 733adf0 flake8 fix and python 3 syntax fix 5df7cb2 Repair the OpenCL init code be857f2 Make CuDNN mandatory for the new back-end. b922d88 Warn if cudnn version is too recent. fix gh-3996 9f69145 Print cudnn version when staring gpuarray back-end adb02ac Print cudnn version at the same time as cnmem usage f9a68d8 reverted the requested part 9f66573 Merge pull request #4025 from nouiz/corrmm_segfault 6f4a125 Last touchups on the GPU tests. 4bba38a reverted the requested part bcc5d04 Fix segfault for very big size. I did some timming and there is no difference that is higher then noise. fix gh-4020 b46e24d Merge pull request #4017 from fvisin/fix_tutorial 238f0c8 Refactor the cuda tests. fe83c9f Disable verify_grad for CpuConv since it can fail. e14dda6 Force shape inputs to int64 in CorrMM to avoid having something that does not match the 'int' assumed type of before. 74d2d21 Raises error when removed nodes are added back 837e3ab remove unsused import 7331258 inherent from multinomial, output is by default int64, fix tests ae63503 change weighted_selection to multinomial_wo_replacement 299f0e2 Replace deleted pages with link to new pages 51b840e Merge pull request #4016 from Saxenauts/add_commit_dev_start d775ab4 minor fixes aef8cd9 fixed datatype issues e4904c8 flake8 fixes 246f740 fixed tests 32549fe fixed comment msg 431c19c removed unnecessary try-except 3f05278 add some tests 48e7107 return ints 60b3686 working 1st version 2176d21 start impl e9c56c3 Merge pull request #3931 from adbrebs/h_softmax_speedup c3b54b9 Fix tests and code so that they can be happy forever after. 9a8bb9d add an assert statement in the test of the h_softmax to checks that fun_output_tg and fun_output are consistent 5a3526c Merge pull request #4005 from nouiz/cudandarray_repr f675890 reverted the requested part dbefcd2 remove if block 573ccea typo and better example of profile output e493f9c cleup profileMode deprecation in docs 05a6369 Move the tests for AbstractConv with the implementation. 00e13c9 Added the commit instruction in dev_guide submit text 5a0d273 Merge pull request #3924 from fvisin/fix_tutorial 9ae1ab8 Merge pull request #4010 from Saxenauts/temp_disable_warn64 538300f Merge pull request #3997 from abergeron/faster_travis e851ec0 Fixed whitespace errors on the last commit, temp. disable the warn_float64 flag in Arange c635b66 Temporarily disables the warn_64 flag in Arange 5621b26 Add link to graph doc in the tutorial 1aa7664 Merge graphstructures from tutorial into extending bc6770c initial commit 121eeac Temporarily disables the warn_float64 flag in Arange Issue# 4cd0858 Temporarily disabled warn_float64 flag in Arange eba65e5 Fix PEP8 errors f649149 Fix broken links in docs 1020d27 Remove __props__ support check in example code c382385 Move doc tutorial/extending* in extending/* f9d1b59 Remove overrided implementation that is now the same as the original version. 03f13db Remove call that aren't needed anymore 6a3b192 Merge pull request #3993 from abergeron/fix_buildbot 55699ce Better error message 3f49e74 Add tp_repr to CudaNdarray. This make DebugMode error print the value instead of <CudaNdarray object at 0x7f229c2f6eb0> 5bffd8d Really skip the flake8 part in "everything but tensor". 5315127 Fix flake8 and clarify comment. 709c944 Merge pull request #3988 from nouiz/small 19b6c0a Merge pull request #3987 from SinaHonari/issue3681 472b0d7 Add back missing *. 6ff26fa Fix opts that play directly with owner.inputs. 5c31ea6 Merge pull request #3885 from abergeron/faster_travis 6296694 Fix the docstrings in abstract_conv.py abe2d24 Fix the sparse docs return types to avoid warnings. 3bfa8db Sync and fix the CompatUnpickler docstring. 2f9ec49 Fix the new doc warnings with the newer sphinx. 75cbdac Don't treat warnings as errors by default. d3dee21 Install sphinx with conda. e0b8642 Add dvipng package ac9fcb6 Add latex packages. dec46ea Fix doc again. b331652 Fix doc formatting in pool.py. a85be0a Disable documentation testing since it is broken. f233adf Fix doc formatting. 52a062b Split out the doc and flake8 tests and rebalance the tests. e5fc7cf Have docgen.py exit with an error if a call to sphinx returns one. c01215f Having two cache: entry discards the first one, which is the one we want. 349de8b Some cleanup. cbdf4fc Merge the faster parts together. 29f4f90 Add a comment 9db2dca Small doc about about fastmath and nan. dabc3dd Register in fast_compile an opt that move to the GPU. b37658f replace a make_thunk by a prepare_node d1bfd2b Convert border_mode=(0,0) to valid. 0414f66 Correctly tag for gpu scan inplace opt for GPU. d4acc24 Fix crash in theanp-cache list for subdirectory that don't have a key file. 95f0039 Make a_theano_function.get_shared() that return shared variables. b524724 Tell that fastmath can give nan and show it less. 32d2744 Merge pull request #3986 from nouiz/test_conv 10ba9ab fixing make_node param passing for conv_grad 319643c Merge pull request #3872 from SinaHonari/issue3681 1e831db Fix the Print lifter. 5f75ecd Remove useless optimizer now that mixed-type abstractconv can't happen. 3f31dc2 Normalize AbstractConv inputs types to not have nodes with mixed input types. 275ffe7 Merge pull request #3984 from nouiz/Param b99ee90 Correct mode fix 4231830 Merge pull request #3976 from nouiz/gpuarray_memset0 815e278 Fix test when mode!=DebugMode c48ee17 correcting indentation 5fe75ed Add back Param and make it deprecated. fix gh-3979. c732789 Merge pull request #3966 from nouiz/dnn d521a6c Merge pull request #3982 from abergeron/fix_nnet a4debbc removing deconv_op name 594e7de Fix Gpu......WithBias to work properly for float64. f655bba Merge pull request #3926 from nouiz/gpuarray_pickle 2f37aeb Better comment. [ci skip] 8227ec0 Merge pull request #3973 from abergeron/fix_buildbot2 178d793 Small doc update c8aa061 Merge pull request #3974 from nouiz/faster_opt 0650b0f Merge pull request #3970 from abergeron/fix_buildbot 737a4d0 Merge pull request #3969 from abergeron/fix_Wl 8276edd Merge pull request #3933 from nouiz/profile 520df6b Add some doc update that was missing in previous PR. b0b4f13 small fix following code review e486864 Make opt GpuAlloc memset0 work in the new back-end 49b79e5 small doc update 184ea12 Make the comparison on the CPU. cef892b Merge pull request #3968 from nouiz/nanguardmode_opt cd0afd4 Speed up opt by not traversing the graph again. 6a9aa55 Speed up opt. Now constant folding is done in the eq, no need to traverse the graph to get constant at each node. 9b05f82 Don't run the abstractconv tests under DebugMode. 1e2660a Fix MakeList to not make an indirect view. 673a2a7 Merge pull request #3917 from andreh7/2016-01-23-fix-for-custom-python-installations 7d2578c Make the sig workaround for unsigned ints too. ff9c346 Merge pull request #3935 from poolio/rop_bugs 752d502 Merge pull request #3961 from nouiz/test b39b15d Don't duplicate args. daab99c Have it not end up in preargs2 anyway. e205ce2 Try to fix -Wl args to nvcc. 52790dd Doc dnn.enable flag 20cf456 Enable opt in NanGuardMode. Make it use the default optimizer. Fix gh-3967. 9d1344f small code change 3135482 make test without pygpu don't crash at init 6f636a8 Fix a test where the gpuarray back-end didn't downcasted python float to float32. 50572e9 Move to a separate test file gpuarray tests that must always run, even when pygpu isn't available 768fb73 Fix error msg 08c6bbc Use config.dnn.enabled in gpuarray b755970 Add config flag dnn.enabled={auto,True,False}. 9627228 Merge pull request #3936 from nouiz/32bit d6b3dff Merge pull request #3947 from Sentient07/issue-#3944 319e0aa Fix a test in 32bit. We can't make min/max reduction on 0 sized tensor. 1ed8895 Fix code review comment ed3034a added BatchedDotOp 1d46dd1 Add d3viz in travis a3575cf Fix test in python 3 b8772f4 Fix a test with python 3 7ffd86a Fix test on GPU with mode=FAST_COMPILE aecc492 Merge pull request #3955 from pra85/patch-2 47278f7 Merge pull request #3883 from skaae/bessel dfce3ae Merge pull request #3952 from nouiz/doc 14fdf30 add j0 and j1 to if else check 6f5082f remove __eq__ and __hash__ 61c71ab Fix typo in docs dac4da8 Fix docgen as now warning are treated as error b2f9958 Better test update for python 32bit fix bdfe90b Fix comments and move code to detect those type of errors 658ba1c further refining the comments a683232 update the test assert when device=gpu 72d3a28 Renamed BatchedDotOp to GpuBatchedDot 0d82fe2 change shape dimensions int32 to int64 18319b8 Merge pull request #3943 from lamblin/fix_batcheddotop_import 73fbb21 Fix path for BatchedDot 2d75099 change shape dimensions int32 to int64 f87853c Merge pull request #3508 from cooijmanstim/batched_gemm 5565558 Make test in python 32bit. sum, average cause small difference result ~1e-17 5c02232 Fix test with python 32bit 4800a51 Merge pull request #3940 from abergeron/fix_buildbot 7fc2bcb Don't use floating-point axes. 1f98466 Remove confusing sys.stdout.flush() which displaces the printing of messages. 6e25ecf Merge pull request #3938 from carriepl/half_padding 0c17be8 Add tests for border_mode half 3ee95b8 Add border_mode 'half' to dnn convolution 1908d99 Fix test in 32 bit python 4978584 remove bincount uint64 restriction, it was a left over from the numpy implementation 817ff62 Make repeat test pass on 32bit python. We use the numpy code only for vector, not scalar 45c7b96 Make test of bincount pass with 32bit python. We don't have the numpy limit anymore 2bbcab0 Better error message 33e07c2 Update test of profiling output. 0d2f82a add missing return to GpuFromHost R_op e891fb3 Merge pull request #3918 from nouiz/tx1 45efd37 Merge pull request #3845 from f0k/add-cudnn-half-padding c46d30b Code clean up and more readable printing 5170ff7 Code refactor, make printed information more clear cc8eaed remove information not useful 06f3b03 profile_memory now print the peak if we disable inplace 62d5cbe Merge pull request #3890 from mohammadpz/remove_param_class fc5fe33 Replace sparse_block_dot by tensor_block when full output is requested. 264d8bf BatchedDot: fix reference 58cddb7 Merge pull request #3855 from ballasn/convdoc 13f8ea6 Add border_mode="half" to cuDNN convolutions 4a0add8 Remove duplicate code efd4d4b Merge pull request #3908 from vmichals/abstract_conv_keep_stacktrace b6732d2 add c code 8ace004 arguments rewritten 48f621b Alloc pickling of GpuArray object 921b8eb fix static implementation error cc80d73 removed erroneous additional string (at beginning of file) accidentally introduced in the last commit 9523985 cublas timming on tx1 1cf5fd3 Doc how to install on TX1 90c8f80 update doc 8020a38 Merge pull request #3912 from lamblin/fix_abstractconv_grad_type 529eb38 Make sure the axes array will be integer-typed. 0fe82c4 fixed travis complaints about whitespace 31f59dc fixes #3556: more robust determination of python library name and directory in function std_lib_dirs_and_libs(), now works on private python installations on Linux a324267 Avoid numpy warning about using non-integer shapes. c96c73f Fix DeprecationWarning about int division. 6bf15e5 Make sure grads of abstract conv ops have the right type 914ebf7 BatchedDot: implement c_code_cache_version c539a57 BatchedDot: move to theano/tensor/blas.py 0fde9a4 BatchedDot: simplify reference implementation 1075d83 BatchedDot: negate contiguity test for less confusion 001ddc3 BatchedDot: don't waste time copying output before overwriting it 3805e3b applied fixes mentioned by Pascal L, abstractconv unit test passes without failure a41fc8b first try 8cdf8e6 BatchedDot: satisfy flake8's perverse desires d8538e4 BatchedDot: avoid repeated indexing dd41f3e Merge pull request #3915 from Saizheng/master 0d413f9 remove the mistakenly copied file ':w' 62bbc9b Merge pull request #3870 from Saizheng/master f2a2181 Merge pull request #3907 from laurent-dinh/conv_test 2dc471f Merge pull request #3903 from matthias-k/AbstractConv2d_R_op 887e39c Merge pull request #3909 from nouiz/website 97fe07f Several optimizations possible ccf4116 Merge pull request #3899 from carriepl/log_softmax 290fc8b profile: more changes 96a257a move print_tips to profiling.py 7b8ee75 updating notations and namings 0abf03b Test for R_op of AbstractConv2d df7fdae name deep learning in git repo so that Theano show up in github search result for deep learning fdfbab3 Merge pull request #3747 from Sentient07/issue-3094 6f54979 Fix typo 7a9a94f Merge pull request #3891 from abergeron/fix_blas_default 5103e4f Merge pull request #3902 from abergeron/fix_filter 00821f9 Remove redundant code following review cbf4915 Merge pull request #3905 from nouiz/corrmm 6439d81b Merge pull request #3895 from nouiz/debug 80acc74 Flake8 fix. b88a016 Flake8 fix. 2beb17a Don't use CorrMM when there is no c compiler (no perform, it will fallback to the old conv that have perform) 578c145 R_op for AbstractConv2d 6877f6f Added optimization test 6639305 Add a test for the allow_downcast issue. 68ae1a7 add bessel of the zeroth kind 8884e69 Fix misscoped allow_downcast block. 3aa7a24 Always silently add the default library dirs, but don't append them to flags. be838ed update_val is not guarenteed to have a type attribute. ced656f Test result of theano function a341a77 indent code example f8bbacb Split local_log_softmax_dnn into two separate op lifters 034c9c3 Fix default blas flag detection by reworking try_blas_flag. 19f4a62 Made the final changes 3dc4855 Add url to gdb quick tutorial 487cf52 Add test for opt to convert LogSoftmax to DnnLogSoftmax in gpuarray backend da93970 Add opt to convert LogSoftmax to DnnLogSoftmax in gpuarray backend 25e6df0 Merge pull request #3881 from nouiz/gpu_dot22 9901219 Update doctring in gpuarray/dnn.py c40c1ed Add test for opt to convert LogSoftmax to DnnLogSoftmax in cuda backend bdcb8d2 Add opt to convert LogSoftmax to DnnLogSoftmax in cuda backend 87f5f60 Merge pull request #3478 from piotrfrankowski/with_statement 4225e32 Merge pull request #3750 from andreh7/ah-2015-12-05-more-comments-on-logistic-regression-example 171a82d Merge pull request #3896 from claymcleod/master 408568f Merge pull request #3897 from jfsantos/master 956f8dc Fixed import of local pool submodule 6253c08 removed '(constant)' from comment in doc/tutorial/examples.txt on request of @nouiz 2163e1f Bug fix: don't try to os.mkdirs on an empty directory 9a0ab54 Merge pull request #3799 from lamblin/avg_pool_v3 136c6d4 Merge pull request #3847 from jojolalpin/master ef304db Doc how to use gdb and python 4507dd1 Make nvcc.flags being using for cuda_ndarray 9cbbbd7 Make cmodule.remove_gxx_opt work for nvcc f640674 Merge pull request #3893 from abergeron/fix_buildbot 1da748d remove duplicate Ins f85c104 Param.default to In.value 279ff3b Param.default to In.value b4bd7b9 No error in v3 when padding is of 0 74b5945 Skip tests of average_exc_pad with cudnn v3 7b206fb Disable average_exc_pad pooling for cudnn v3 d93640a Get rid of a bit more warnings. 62eed15 Fix the Ananconda workaround to detect later versions. 4c01908 Fix the default return from default_blas_flags to return valid flags. ce2e15b Param class in now replaced by In f713623 Follow renaming. adf81d3 Merge pull request #3679 from shabanian/tensor_signal_pep8 15aef2f Better code 71f30a9 Fix opt warning/not moving to GPU some node. 6ce3aa7 Merge pull request #3884 from abergeron/fix_mac_blas ae444d3 Merge pull request #3861 from nouiz/version 3a0493c Merge pull request #3875 from abergeron/fix_buildbot 534e040 Small doc update. 3303c99 Flake8 fix. 0c10888 Tweak the blas flags logic so that it doesn't require certain keys in the dict. a591990 Fix brain typo. 1893c13 Use prepare_node() instead of make_thunk() in Elemwise. 710e6b4 Add a new prepare_node() method that can be used to do stuff 'just before' make_thunk and help DebugMode work correctly. 3a0de13 Add test for the elemwise issue. e98220f Make cumul be a double variable for better accumulation precision. ab40a50 Get rid of some of the deprecation warnings in the buildbot. 3dcba54 Merge pull request #3363 from fvisin/logsoftmax 9ad1ea0 Merge pull request #2 from abergeron/logsoftmax 0041ce6 Remove the try .. except construct since it masked some errors in the opt. 994c8d3 Register the logsoftmax opt in fast_compile. a5735a1 Merge pull request #3877 from abergeron/fix_segfault f986e0d Fix the cudnn context problem. 6b7a3d1 Revert "Set the proper context when destroying the cudnn handle." b1c5a67 Fix grad optimization. 91fe71d raise a warning when the old file is imported ede4874 Merge pull request #3858 from mohammadpz/try_blas_flag 7bad2e2 Merge pull request #3867 from nouiz/small 4bd7b62 deconvolution interface c282d07 Merge pull request #3869 from kmike/remove-scan-warning 19ed493 Merge pull request #3868 from kmike/py3-raise_with_op 2e660ce flake8 b7e68d2 scan: remove unconditional warning when strict=True is used. Fixes GH-3333. eb7901d PY3 fixed theano.gof.link.raise_with_op in Python 3 2588a94 blas added in flags f1eddb0 try_blas_flag return all flags fcc9944 try_flags is replaced by try_blas_flag 42dd8a8 Merge pull request #3856 from abergeron/transfer_noints 2d4f6d7 Merge pull request #3862 from saatvikshah1994/issue-3658 71a3700 Merge pull request #3853 from abergeron/fix_lift_abstract 53f2908 fixed some issues in test_pool.py 354c4a9 Revert back to transferring all inputs by default. 1838572 Set the proper context when destroying the cudnn handle. ea04c80 fixed some issues in test_pool.py 511b759 Remove unused f for flake8 compliance. 9cb9c43 Rename test_downsample.py to test_pool.py 1b55e55 Use a prefered way and clearer way to have code work with python 2 and 3. b433d70 Add version added in the library doc 60dbf4b Remove old link to now non existing page 834ab99 More information in error bd0ee86 Fix disabling of docgen.py introduced in December. 6fae1fc make docgen.py return to current directory at the end of the script. 3d82403 Link to platoon 686e85f Skip a test if cuda isn't available. 48dfc25 fixed some issues in downsample.py 372939b Rename max_pool_2d to pool_2d and Pool for the DownsampleFactorMax 8f65b44 BatchedDot: leave first dimension broadcastable if it was broadcastable for either of the inputs e527643 BatchedDot: remove reference to Dot 878633e BatchedDot: remove complex test cases as we will not support complex types for now c37a6bb Fix test so that it actually crashes on unfixed code. 25ca839 BatchedDot: upcast inputs if they are of mixed types ca630a9 Add test for shape on GPU. 16b0bc5 corrected setup of d3viz 73a7621 Bump version to 0.8.0.dev0 808296a Expand on the comment. 149d008 Merge pull request #3859 from kmike/faster-TensorConstant dde6c84 optimize TensorConstant constructor 6b4055d auxiliary function 'try_blas_flags' added. e5f7eb3 Remove the undocumented and unused tag.context_name. 1381698 Add documentation about the tag.target attribute and remove false statements about performance. 4e67e4d Ignore nodes that are already on the GPU. 7a90c78 Don't transfer int inputs to the GPU by default. 8cd678d BatchedDot: small improvements 0b89695 BatchedDot: fix more bad pointer arithmetic 78bbb56 BatchedDot: reduce duplication 77729ff update convolution documentation 1ac274a BatchedDot: fix bad pointer arithmetic 5cde49a Add test for theano.sparse.opt.sd_csc in test_opt.py c56c2f7 BatchedDot: fix potential bug fe58ada Merge pull request #3850 from dmitriy-serdyuk/dunction-dump-doc b90be7c Fix typo in function_dump doc f279798 Merge pull request #3817 from viirya/check-home-dir 80df08b Merge pull request #3814 from SuperElectric/rebroadcast_bugfix 6a1e661 Use /tmp/$USER if possibly. 09390a2 Merge pull request #3830 from Sentient07/2016 353b0fc Merge pull request #3846 from kmike/patch-1 9c7c3b8 Update opt.py 4192d13 DOC removed incorrect statement from compatibility docs 4d03d97 Merge pull request #3824 from Sentient07/issue-3686 149a9df Merge pull request #3807 from nouiz/lock 8eadfc3 Updated the copyright dates 0ef3ec3 More informative warning in InferShapeTester 493517e Following code review, better docstring and take the lock less often b9e32a0 Suggested changes made 167df2c Merge pull request #3801 from nouiz/doc cb79538 Merge pull request #3804 from nouiz/tests 6d73535 Revert some changed I didn't wanted to commit. b3d0a8b Try to be even more overflow tolerent. 1c57b67 Merge pull request #3822 from mronian/master 45d3136 Merge pull request #3828 from dwiel/patch-1 0d69ea0 Merge pull request #3831 from kmike/py3-fixes b47debf pep8 0b54e8d Update warning messge following code review 87c4e01 Merge pull request #3811 from nouiz/typed_list 2513a44 Merge pull request #3803 from nouiz/gpu 1739dda DOC fixed Python 3 compatibility issues in Tutorial and Library Reference 12aa951 PY3 fixed ProfileStats.summary() in Python 3.x 5ec1e57 PY3 fixed SeqOptimizer.print_profile in Python 3 45f55ca TST avoid numpy deprecation warning in tests under Python 3 e559fd4 Happy new year! :) cfbc73b typo in bn.py documentation d4128b7 Added test for get_conv_output_shape method 3c32963 Checks compiledir flag for prohibited characters. Fixes #3711 ac26259 Merge pull request #3819 from superbobry/master 4d69aea Fix Python3 regression in ``pydotprint`` 537b24f Check home_dir writing permission and assign temporary directory. f380577 Fixed pep8 5fa767d fixed too-stringent typecheck to accept not just bool, but also numpy.bool_ ac4d5af Add TypedListConstant. This fix the added test. 73dfccf r_shape can be None d18af18 Quick fix in __str__ ee4f2a9 Don't take the lock by unless needed at exit 1aed098 Good dtype comparison b75fc56 Removed make_node method and corrected the test e03b0ca Remove compilation warning d3dbe92 remove duplicate import 83c8299 Remove duplicate import ad980be Remove useless return 8e65af1 Small code speed up/clean up 8100d60 Update test name to the real op name 2c34d0c pep8 e082574 Skip flake8 tests on auto saved tmp file d915fb9 Tell the limitation of the pooling on the GPU. 0c6e094 Fix gh-3800 make it clear that strides must be ints 5418629 Merge pull request #3802 from f0k/fix-blas-detection 7936ca4 Remove useless comment 8c607e8 Remove opt warning with mixed dtype 4b667da Merge pull request #3796 from abergeron/gpuarray_prealloc b77115c Merge pull request #3790 from hidasib/gpu_set_subtensor_2d 554be9c Enable cudnn by default when avaialable in the new back-end 7e34c53 Speed up Elemwise.perform 3e0e53e Fix fallback in BLAS detection 91547f2 Fix DebugMode error due to overflow in comparison 9fad11d [BUG] bugfix of python code. It is triggered in DebugMode with this test theano/sandbox/tests/test_rng_mrg.py:test_multinomial_n_samples fcfa049 pep8 7197ebc pep8 a130465 Fix DebugMode when a perform raise NotImplementedError. This happen with this test theano/sandbox/cuda/tests/test_conv_cuda_ndarray.py:test_conv_grads 721666e Fix test in DebugMode with recent numpy version. AbsTester with complex input was failing. 2508997 Changed c code version from 6 to 7. Added tests to test_basic_ops.py for GpuAdvancedIncSubtensor1 (general case) and GpuAdvancedIncSubtensor1_dev20 (2d case if compute capability >=2.0). 5e1295f Merge pull request #3798 from nouiz/small 22c56b8 Skip part of tests with not compatible pydot_ng 2ba1d48 Don't crash when generating the error for some type of error f331d52 Merge pull request #3631 from hantek/doctest a5c6a13 Merge pull request #3784 from mronian/master 663c070 Cap memory preallocation to 98% of the total. 2640056 Merge pull request #3792 from abergeron/gpuarray_prealloc 0a876e2 Adds test for uint datatype for image_shape and filter_shape in theano.tensor.nnet.conv and also checks if datatype is in tensor.discrete_dtypes 54e9675 Merge pull request #3559 from abergeron/multi_gpu_doc e6ca182 Delete the cudnn handles when we no longer need them. d6b2e03 Rework the gpuarray api version check. 8fb5d66 Add the gpuarray.preallocate configuration variable. This will help fix out of memory errors in certain sequences of tests. 0f03bac BatchedDot: PyArray_Newshape complains about NPY_KEEPORDER when it shouldn't 9a3f668 BatchedDot: generalize matrix-matrix code to matrix-vector, vector-matrix, vector-vector products 4ce8a48 BatchedDot: first stab at C implementation cccef96 BatchedDot: fix comments and remove assertion a3fc110 BatchedDot: fix optimization to handle matrices 82aac7d BatchedDot: remove unnecessary __str__ a915c82 BatchedDot: remove reference to Dot 62e2bf5 BatchedDot: trust that gradients are floats cc8c9e4 BatchedDot: fix grad 2202eb4 BatchedDot: more diverse tests bee1a7a BatchedDot: catch bad inputs 91fe0a2 BatchedDot: write docstrings 1a28571 introduce optimization to move BatchedDot to GPU 7d1c991 BatchedDot: provide dtype in numpy reference implementation ed4e067 BatchedDot: test and fix grad e68999e BatchedDotOp: simplify tests f6c4864 BatchedDot: implement and test 713fd0f BatchedDotOp: implement and test infer_shape 34be0f0 solving conflicts 75b1c22 fixed all warnings in doc. added the sphinx -m flag in docgen 4ad36dd Merge pull request #3788 from nouiz/carriepl-v4 c5084ac Fix typo. 47ecce0 Remove the warning paragraph as the issues have been fixed. e28057a Fix typos. 838508a Move the warning about no speedup earlier and add a section showing the printout of the context map. 0f4f011 Fixes from review. 99e23d1 Link to PureType instead of duplicating the docstrings. f2c8575 Fix the multi gpu script to use the new target argument. 7beebd0 Add a tutorial section on how tu use the new multi-gpu functionality. d6156c6 Delete some old and useless documentation. 30617ff Rest of libdoc for gpuarray. 3bf6f4c Add a listing of ops for gpuarray. f005cc2 Add a bunch of documentation to gpuarray/type.py 418967d Link in the blocksparse docs and add a section about the gpuarray backend. 07b2896 Compile theano function and assert 61d1163 Gpu implementation of GpuAdvancedIncSubtensor1 (set mode), using CudaNdarray_CopyFromCudaNdarray. fd75d2c GPU implementation for GpuAdvancedIncSubtensor1_dev20 using atomicExch() 32e113c Flake8 ced46b7 Make dnn conv in gpuarray work with AbstractConv aafc23a Standardize documentation of the 'precision' parameter 7d44a89 Update and standardize error messages 524438b Clean up V2 code 877f421 Avoid using method cudnnGetConvolution2dDescriptor() method ebd9778 Make deprecated dnn flags raise more meaningfull error messages. 32f1155 Add 'as_input' option to dnn.conv.precision flag 7ff6621 Drop support for v2 3e4da0d Don't restrict 'small' workspace implementation of GradW to v4 and later a28cb16 Add missing syntax 757b04c Fix verification of spatial sizes in cudnn conv (gpuarray backend) fdffa58 Fix condition based on CuDNN version 6f1dc3d Call cudnnGetConvolution2dDescriptor instead of cudnnGetConvolutionNdDescriptor bf91848 Add comments and fix condition in GpuDnnConvGradW C code (gpua backend) 2a857e3 Update GpuDnnConvGradI for CuDNN v4 (gpua backend) 1e48b73 Update GpuDnnConvGradW for CuDNN v4 (gpua backend) 64439f4 Update GpuDnnConv for CuDNN V4 (gpua backend) 93f6f44 Define new V4 elements from V3 elements in cudnn_helper (gpua backend) 674cd4f Add precision param to dnnConv (gpua backend) 46c8338 Add precision param to GpuDnnConvDesc 5a015a8 Add definition for new implementation in cudnn_helper.h (cuda backend) 330da07 Retain compatibility with CuDNN v2 (cuda backend) cde0771 Add checks to make sure backward tiled-fft is not used when it shouldn't (cuda backend) b177f3a Add checks to make sure tiled-fft is not used when it shouldn't (cuda backend) cf26ae5 Limit new convolution implementations to CuDNN V4 2a84aa6 Make dnn_conv and dnn_conv3d support param precision (cuda backend) 0ed16db Add new precision parameter to GpuDnnConvDesc (cuda backend) 3914bf6 Add new flag to control CuDNN convolution precision c36f9a2 modify cuda/dnn.py to support new convolution implementations 13bad51 Deprecate algo_bwd flag in favor of algo_bwd_data and algo_bwd_filter fc7d727 fix typos e200ca2 start new cudnn v4 convolution mode 1f3c1f6 changed corresponding to the upstream e3a0aee Flake test passes 112b22f Updated tutorial/extending_theano.txt 48661f1 Reverted suggested changes 4936e63 Test added, code made complaint with pep8 standards db6fc48 mend e3430b6 Moved the mae_node method 25c0f5e58 PEP8 respect b050916 fix all doctest errors, but not turning on warning-to-error flag in sphinx build a8316c2 add back the doc test part 6d25ecd fix those paths in PART variable 159a988 remove commented lines in docgen.py; uncomment theano tests 09a5fde fix the file not found problem 7ff3ca2 add sphinx to dependencies 6cb95da remove epydoc from the repo c13f339 add epydoc in travis yaml 7b733b0 add doc test in yaml 54b662a added comments to logistic regression example in the documentation. No code changes 57a2f9c Slight improvement made to the test c4fb7e5 itypes and otypes checked and added test 633ff54 fix all doctest errors, but not turning on warning-to-error flag in sphinx build d7d6408 Do not import optimizations 16333fd Add C code for LogSoftmax e481355 Fix use crossentropy_onehot_grad instead of logsoftmax 49cf5b4 Add LogSoftmax gradient test and gradient optimization test 8b44700 Add optimization for grad f9eb767 Add test for the optimization of the fprop of log(softmax(x)) 65898f8 Remove unused optimization for log softmax afd6c4d add back the doc test part e671d03 fix those paths in PART variable 483ca0b remove commented lines in docgen.py; uncomment theano tests b66c305 fix the file not found problem 66ecd2a add sphinx to dependencies c92b36d remove epydoc from the repo 7c5c18e add epydoc in travis yaml aa32a07 add doc test in yaml f05a0c8 Add LogSoftmax python code and tests 3b9975e fix according to comments abc2876 #3429 - python 'with' statement in reset modules 54b194d #3429 - python 'with' statement in tests module git-subtree-dir: libs/Theano git-subtree-split: 41248a16a79ff1d11f8ddc68afe6571d3f80769f
ubuntu 14.04.
use thi command(.setup.py install) to setup kaldi-pthon, I have set $KALDI_ROOT already
the errors are as follows:
/usr/include/python2.7/numpy/npy_1_7_deprecated_api.h:15:2: warning: #warning "Using deprecated NumPy API, disable it by " "#defining NPY_NO_DEPRECATED_API NPY_1_7_API_VERSION" [-Wcpp]
#warning "Using deprecated NumPy API, disable it by "
^
/usr/bin/ld: /home/jtang/Kaldi/kaldi-trunk/src/matrix/kaldi-matrix.a(kaldi-matrix.o): relocation R_X86_64_32 against
.rodata' can not be used when making a shared object; recompile with -fPIC /home/jtang/Kaldi/kaldi-trunk/src/matrix/kaldi-matrix.a: error adding symbols: Bad value collect2: error: ld returned 1 exit status make[1]: *** [kaldi_io_internal.so] Error 1 make[1]: Leaving directory
/home/jtang/Attention_ASR/kaldi-python/kaldi_io'make: *** [all] Error 2
thse errors seem to happen in creating kaldi_io_internal.so, if I don't use these .a file$(KALDI_SRC)/matrix/kaldi-matrix.a $ (KALDI_SRC)/util/kaldi-util.a $(KALDI_SRC)/base/kaldi-base.a ,
kaldi_io_internal.so can create(of course this file can't be used)
The text was updated successfully, but these errors were encountered: