Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

some errors when install kaldi-python #2

Open
Entonytang opened this issue Nov 20, 2015 · 12 comments
Open

some errors when install kaldi-python #2

Entonytang opened this issue Nov 20, 2015 · 12 comments

Comments

@Entonytang
Copy link

ubuntu 14.04.
use thi command(.setup.py install) to setup kaldi-pthon, I have set $KALDI_ROOT already
the errors are as follows:

/usr/include/python2.7/numpy/npy_1_7_deprecated_api.h:15:2: warning: #warning "Using deprecated NumPy API, disable it by " "#defining NPY_NO_DEPRECATED_API NPY_1_7_API_VERSION" [-Wcpp]
#warning "Using deprecated NumPy API, disable it by "
^
/usr/bin/ld: /home/jtang/Kaldi/kaldi-trunk/src/matrix/kaldi-matrix.a(kaldi-matrix.o): relocation R_X86_64_32 against .rodata' can not be used when making a shared object; recompile with -fPIC /home/jtang/Kaldi/kaldi-trunk/src/matrix/kaldi-matrix.a: error adding symbols: Bad value collect2: error: ld returned 1 exit status make[1]: *** [kaldi_io_internal.so] Error 1 make[1]: Leaving directory/home/jtang/Attention_ASR/kaldi-python/kaldi_io'
make: *** [all] Error 2

thse errors seem to happen in creating kaldi_io_internal.so, if I don't use these .a file $(KALDI_SRC)/matrix/kaldi-matrix.a $(KALDI_SRC)/util/kaldi-util.a $(KALDI_SRC)/base/kaldi-base.a ,
kaldi_io_internal.so can create(of course this file can't be used)

@rizar
Copy link
Owner

rizar commented Nov 20, 2015

As far as I remember Kaldi has to be compiled differently for Kaldi-python installation to be successful. @dmitriy-serdyuk , @janchorowski , can you please comment on that?

@Entonytang
Copy link
Author

Can you tell me the methods you compile kaldi.......(aslo means : how can you get the file like kaldi_matrix.a.........)

[email protected]

From: Dzmitry Bahdanau
Date: 2015-11-20 23:34
To: rizar/attention-lvcsr
CC: Entonytang
Subject: Re: [attention-lvcsr] some errors when install kaldi-python (#2)
As far as I remember Kaldi has to be compiled differently for Kaldi-python installation to be successful. @dmitriy-serdyuk , @janchorowski , can you please comment on that?

Reply to this email directly or view it on GitHub.

@dmitriy-serdyuk
Copy link
Contributor

Right, sorry, that I didn't mention this. Kaldi should be compile with shared flag:

./configure --shared --use-cuda=no # No need for cuda, we don't train models with kaldi
make

@rizar
Copy link
Owner

rizar commented Nov 20, 2015

Could you please change the documentation? I guess it makes sense to do it our private repository, since we are going to make what we have there the new master pretty soon.

@Entonytang
Copy link
Author

After change the configure command, problem solved......
this steps: $LVSR/bin/run.py train wsj_paper6 $LVSR/exp/wsj/configs/wsj_paper6.yaml
this default configuration trains model using CPU.......how to use GPU instead.......

@rizar
Copy link
Owner

rizar commented Nov 23, 2015

You can use GPU in the same way as you usually do it with Theano. Please
read Theano documentation.

On 23 November 2015 at 05:51, Entonytang [email protected] wrote:

After change the configure command, problem solved......
this steps: $LVSR/bin/run.py train wsj_paper6
$LVSR/exp/wsj/configs/wsj_paper6.yaml
this default configuration trains model using CPU.......how to use GPU
instead.......


Reply to this email directly or view it on GitHub
#2 (comment)
.

@Entonytang
Copy link
Author

After adding "device =gpu3" while I find GPU Process in GPU 2(device K40).....using default wsj_paper6.yaml.....
it costs 65 seconds per steps(1 epoch = 3700 steps), I think this speed is too slow for GPU......
so this speed is right or not , what should I do for speed up the training process
and How much time one epoch?

@dmitriy-serdyuk
Copy link
Contributor

As I measured recently, one step was taking about 6 seconds on a Titan X, K40 was a bit slower, about 8-9 seconds. So probably something goes wrong.

Make sure that Theano writes something like Using gpu device 1: GeForce GTX TITAN X (CNMeM is enabled). Another suggestion is to check that you use float32, not float64. I also use optimizer_excluding=cudnn option since I had some issues with CUDNN.

@rizar
Copy link
Owner

rizar commented Nov 24, 2015

Also use optimizer=fast_run in your THEANO_FLAGS

On 24 November 2015 at 10:34, dmitriy-serdyuk [email protected]
wrote:

As I measured recently, one step was taking about 6 seconds on a Titan X,
K40 was a bit slower, about 8-9 seconds. So probably something goes wrong.

Make sure that Theano writes something like Using gpu device 1: GeForce
GTX TITAN X (CNMeM is enabled). Another suggestion is to check that you
use float32, not float64. I also use optimizer_excluding=cudnn option
since I had some issues with CUDNN.


Reply to this email directly or view it on GitHub
#2 (comment)
.

@Entonytang
Copy link
Author

thanks, solved.....
while at the 830 steps. the program stoped without any warnings......while GPU Process is still there.......the bokeh-server is also there.
and wsj_paper6.yaml doesn't seems to be the setting in end-to-end attention-based lvcsr...(250 Bi-GRUs in paper while wsj_paper6 has 320)

Epoch 0, step 829 | # | Elapsed Time: 2:09:35


Training status:
best_valid_per: 1
best_valid_sequence_log_likelihood: 503.460199693
epochs_done: 0
iterations_done: 829
Log records from the iteration 829:
gradient_norm_threshold: 239.912979126
max_attended_length: 400.0
max_attended_mask_length: 400.0
max_recording_length: 1600.0
sequence_log_likelihood: 189.054199219
time_read_data_this_batch: 0.0219719409943
time_read_data_total: 19.5282828808
time_train_this_batch: 11.5933840275
time_train_total: 7709.37198544
total_gradient_norm: 135.73147583
total_step_norm: 1.07967531681

Epoch 0, step 830 | # | Elapsed Time: 2:09:46

@dmitriy-serdyuk
Copy link
Contributor

Is there exception or a core dump? Otherwise it's something wrong with your OS.

@Entonytang
Copy link
Author

I don't think so, I use another core and try again. the result is similar......
(The best_valid_sequence_log_likelihood: 503.460199693 which is same as the result after 830 steps.)
while only pretraining_model.zip| pretraining_log.zip| pretraining.zip appear in wsj_paper6 file. and is the wsj_paper6.yaml is the right config?

Epoch 0, step 84 | #| Elapsed Time: 0:09:18


Training status:
best_valid_per: 1
best_valid_sequence_log_likelihood: 503.460199693
epochs_done: 0
iterations_done: 84
Log records from the iteration 84:
gradient_norm_threshold: 85.4330291748
max_attended_length: 248.0
max_attended_mask_length: 248.0
max_recording_length: 990.0
sequence_log_likelihood: 264.288513184
time_read_data_this_batch: 0.0211541652679
time_read_data_total: 2.17928504944
time_train_this_batch: 5.36292505264
time_train_total: 556.870803595
total_gradient_norm: 109.950737
total_step_norm: 0.572255551815
while if I use wsj_paper4.jaml, the training process seems to be no problem.......

rizar added a commit that referenced this issue Mar 23, 2016
38535d8 Merge remote-tracking branch 'blocks-origin/master' into new_blocks
ed481ef Merge pull request #953 from dwf/shut_up_sqlite
417dbdf Skip test_save_the_best with sqlite.
f5ee622 Merge pull request #946 from vdumoulin/abstract_conv
648646d Add test for ConvolutionalTransposeActivation
87d143f Reorder constructor args to help remove duplicate code
843d3ff Replace deprecated `image_shape` kwarg with `input_shape`
65e9435 Clarify parameter description
6712e44 Make ConvolutionalTranspose inherit from Convolutional
ff08eb8 Reduce repeated code in constructors.
abf41f8 Make imshp not depend on input_
7c895aa Construct grad op using constant values rather than depend on a shared variable
2a68d89 Fix superclass call
836267d Revert to not forcing compilation mode
c237b51 Comment on the use of FAST_RUN mode in blocks.bricks.conv tests
fa4f314 Fix conv tests failing to compile Theano functions
c985b30 Fix import error
ee5efc6 Fix oversights
071ad2b Add ConvolutionTranspose and ConvolutionTransposeActivation
9255647 Use Theano's new abstract_conv interface
348de31 Adapt Pooling to Theano's downsample -> pool renaming
98797a1 Merge pull request #941 from dwf/batch_norm_bricks
90cc8bb Fix StepRules to honour broadcastable.
d9171ff Move __init__ logic to _allocate in SBN.
a88b193 batch_normalization_updates -> get_batch_normalization_updates
725ec01 Idiomatize test.
4b39c93 Remove from 'with' for clarity.
e714903 Make duplicate updates docs more explicit.
333b49c Rename W, b -> scale, shift.
09da496 Add batch_normalization_updates.
d81c03c Annotate population parameters with brick.
474d3f1 Fix subtle bug with mutable metadata.
b03fffc Fix typo in epsilon assignment.
6895b87 Remove needless properties.
f41c286 Improve _add_batch_axis.
cf4f4db Ensure correct nesting of context managers.
9805b73 Improve docs. Add example for apply_...
7efa956 Correct the assertion.
5ba9b38 Add comment and assertion following review.
c70a26b Fix Scrutinizer error from 6ba3364.
2ea41a8 Improve robustness of context manager logic.
e22b151 Add an assertion following review.
a26e2c6 Simplify generator expression after review.
d35c22b Amend error message following review.
2e7cd0e Rename save_memory -> conserve_memory.
3dc9135 More tests, move graph tests to tests/graph.
ad3a84d Fix doctest.
73a6902 Comments explaining tests.
93e55f6 Big refactor to enable context manager operation.
59fd7cd Add brick filter and smoke test for bug fixed.
b358849 Remove leading batch axis from BN shared variables.
02982e9 Correctly crawl the graph backward.
f26c1e5 Expose save_memory in BatchNormalizedMLP.
7ede89b Make batch_normalize return the pop->minibatch map.
9ff63b0 Add tests.
6b52b35 Refactor batchnorm + graph module.
9b94116 Reorganize bricks subpackage.
0f6623f Initial batch-normalization implementation.
2046525 Merge pull request #944 from mila-udem/bound_application_error_msg
07d43ae Correct error message in filter.py.
07cec2c Merge pull request #774 from lamblin/update_doc
25c5141 Merge pull request #940 from rizar/without_namespace_packages
c0640b0 Disable namespace packages suppport
83c5d2a Merge pull request #838 from matthiasreisser/docstring_additions
260296d Merge pull request #929 from dwf/conv_doc_fixes
acf9c80 Merge pull request #930 from r2007/patch-1
353d525 Update bricks_overview.rst
506dfe1 Merge pull request #928 from dwf/shared_floatx_kwargs
b7f5df3 Pass along kwargs in shared_like.
fce7f6b Merge pull request #925 from dwf/get_output_shape_update
ff70906 More robust get_output_shape call.
928dcbb Better document get_output_shape.
c6cd413 Update ConvolutionalSequence docs.
5928f34 Pass shared_floatx kwargs to theano.shared.
3eddcf8 Merge pull request #921 from dwf/die_conv_layer_die
65e9952 Replace deprecated getOutputShape interface.
589543e Remove mention of `ConvolutionalLayer` from a docstring.
3cd7bfd Get rid of ConvolutionalLayer. Fixes #912.
df5c55c Merge pull request #924 from dwf/fix_doctest_debugprint
eedab38 Fix debugprint output for Theano/Theano#1953.
6827a9e Merge pull request #891 from dmitriy-serdyuk/release-doc
1d63fa8 Merge pull request #899 from dwf/conv_improvements
769eb70 Test for overzealous support code.
6177388 Making unpickling robust to MaxPooling change.
24a489a Merge pull request #909 from akhti/doc_fix
7734be3 Increase font size in the generator scheme
04783b7 Merge pull request #886 from sotelo/return_states_seq_gen
f7866d8 Merge pull request #892 from akhti/fix_bidir
05fea66 Merge pull request #907 from dmitriy-serdyuk/fix-log-doc
7576bcc Add API sections for two log backends
b87ad2e Make ignore_border=True the default.
c374557 Test Pooling Bricks work in ConvolutionalSequence.
27d71a5 Tests for AveragePooling, new MaxPooling flags.
f1e5908 AveragePooling Brick.
ac9f095 ConvolutionalSequence: use num_output_channels.
dbb407c Refactor MaxPooling.
414519d Merge pull request #903 from dwf/fix_travis_again
e185f3d Hardcode a prefix for Travis Miniconda install.
7e10ed5 Merge pull request #897 from dwf/custom_scaling_dropout
c4426c3 Notes on dropout usage.
07c98e2 Add dropout tests, including custom_divisor.
d7d8b62 apply_dropout: support a custom divisor.
9ffc3e1 Improve apply_dropout documentation.
23d94b2 Fix label in sequence generator scheme
6750051 Merge pull request #896 from rizar/add_original_svg
d7eb341 Update after review.
8950f73 Backport fixes done to Fuel's install doc
68842e7 Update developer doc to mention Blocks and Fuel
ea34217 Add the original for SequenceGenerator picture
455d65f Improve docs for making new release
bc4f62c Fix too long line error
46652f7 Add test for a stack of bidirectional layers
98aed8e Add instructions to make a new release
3ec837a Merge pull request #887 from rizar/fix_confpy_and_bump_version
4f74b4c Merge pull request #888 from akhti/fix_bidir
69bc613 Add get_dim to Bidirectional
0a4e3b3 Fixes conf.py and bumps version to 0.1.1
afaa45f Added the final values of states as auxiliary variables in a sequence generator so they can be reused.
12e50d9 Merge pull request #772 from adbrebs/doc_brick
92654e1 Merge pull request #881 from mila-udem/correct_main_loop_error_message
5dc2bdd brick tutorial improvements
168c7a8 Merge pull request #878 from rizar/release-0.1
f275332 Merge pull request #879 from sotelo/generator_cost_documentation
e43cfbe Correct main loop message for the interrupt
6c99076 Added the missing links.
c11698c Merge pull request #880 from sotelo/missing_apply_decorator_documentation
679b01c Improved the error message when an application decorator is missing.
efe3585 Solved the flake8 mistake.
6ebcb37 Added documentation about the emitter cost in sequence generators.
a06878f Bump version number
0f889aa Merge pull request #875 from dwf/use_bias_convolutional_activation_layer
0ca4e86 Merge pull request #873 from dwf/border_mode_conv_sequence
d1af6c9 ConvolutionalSequence: propagate use_bias.
7bc11d4 Convolutional{Activation,Layer}: respect use_bias.
457049d Convolutional{Activation,Layer}: Refactor alloc.
6dec566 ConvolutionalSequence: Don't auto push border_mode.
b3754dc Merge pull request #849 from mila-udem/redefine_modelr
837da08 Merge pull request #868 from lukemetz/lm/set_parameter_values
8a06204 warn if setting wrong shape parameters
c843fa8 Merge pull request #867 from dwf/conv2d_impl
620bc7b Additional developer documentation.
8a8e8c1 Fix Scrutinizer complaint about whitespace.
c2ebc25 Make Convolutional's conv implementation hackable.
c427fa7 Merge pull request #864 from mila-udem/rtd2
377688c Add mocking back in
2fc06f3 Merge pull request #861 from mila-udem/rtd
c5b9f1c Changes
743cbf3 Add a test and fix imports
fe9daeb Back to items()
bb81302 Call __init__ and add iteritems
a9751cb Merge pull request #852 from rizar/testing_utils
5772cc3 Fix formatting
8fd7da4 Improve documentation
0cc3031 Fix imports in doctests
3100e11 Move testing utilities to make them importable from outside
5899425 Merge pull request #844 from mila-udem/mention_extras_and_examples
b46dbe0 Refactor model and write a bit of documentation
f6a99bc Mention examples in the documentation as well.
f38881a Add forgotten period<F2>
3c1fdb5 Also refer to Fuel
96d63bf Better looking reference names
bb3f6c8 Mention blocks-extras and blocks-examples in README.md
8bf07e7 typos
4b83783 Merge pull request #839 from mila-udem/deps
1436c25 Add nose
59a9553 Update deps
f295d76 Added line break
cba09d2 Changed docstring to plural
bba2e2d added docstring to apply method of LSTM
d576831 Small fixes in create your own brick tutorial
bdff06b Merge pull request #1 from dmitriy-serdyuk/doc_brick
5da4696 Separate paragraph
536bf18 Make small fixes in create your own brick
e377e43 Merge pull request #834 from yingzha/dev
bd12f44 Rephrase docs
f6e9896 Merge pull request #777 from galv/master
03a491b Add new line to comment block.
ecd121d Pass input from ComputationGraph.get_theano_function() to theano.function()
53979de Example of lazy brick with get_dim
5628491 More explanations in the tutorial to create a custom brick
627da45 Refactoring. Brick example with children.
ca1c7cf Very first draft of the tutorial to create a brick.
7e2535d Merge pull request #835 from johnarevalo/patch-2
e844b92 Allow uint type for lookup indexing
cebec4e Replace uses of named_copy with var.copy(name=...)
7ff0f6b Merge pull request #827 from vzhong/conv-1d
bc140ba fix for case in which no image size is specified for Convolution #825
9e4d0e4 Merge pull request #826 from lukemetz/lm/algorithm_roles
a976b9f add roles and names to algorithm shared variables
98ed3d1 Merge pull request #823 from rizar/sgd_profile
d335fd0 use parens
1f16ba5 add test
a4b50e0 add theano_func_kwargs to algorithm init and use respectively
3403846 add kwargs to base class initialize func
53f292e remove unnecessary vars
3b0bbd0 add profile via kwargs to gradient descent initialize
0f6b4e0 Merge pull request #819 from rizar/fix_iterate_false
67f7388 Remove old code
2a8dff3 Arguments for SimpleRecurrent should be required
69b8ce2 Merge pull request #818 from ASalvail/master
436134a Reformated modified doc for docstrings requirements.
0e15726 Reformated modified doc for docstrings requirements.
e5e14a4 Elaborate LSTM inputs argument to apply method
da5d320 Merge pull request #815 from yingzha/dev
f88a638 Merge pull request #2 from dwf/yingzha_ccw
79243e5 Simplify documentation.
dce9c48 Merge pull request #813 from dwf/parallel_routing
c7d3540 Use OrderedDict per Dima's review.
b17fc13 Refactor into utils func, per Dima's request.
23fd90a Merge pull request #1 from dwf/yingzha_ccw
d3c5caa Add test for argument Parallel.apply argument validation.
16c50ab Fix Parallel.apply argument routing logic.
e26d04d Fix for str config with default of None.
3144fc9 remove dir argument in test_config
0fe9a92 fixed flake8/pep errors
29099d3 Override base path for temporary file creation
f2c9f3e Merge pull request #798 from ASalvail/master
760b144 Merge pull request #808 from mila-udem/selector_docs
d7bf7f7 Remove unnecessary import in doctest.
45c400a Eschew needless initialization.
4f3a075 Reword per Dima's review.
f95f919 Improve Selector.get_parameter documentation.
325a431 Merge pull request #805 from mila-udem/cost_cost_matrix_args_kwargs
83151ca Make Cost and CostMatrix more general.
9c2a1a1 Merge pull request #803 from Tejas-Khot/Tejas-Khot-patch-1
5c1062b made line length less than 75 characters
d299457 corrected trailing whitespace problem
968edbc added docstring for tied_biases
0a8cc3d Corrected docstring of Softplus.
e48bbb9 Add activation tests (Softplus, Softmax, Logistic)
9ad61bc Softplus docs and reference.
fab1fbe Added a Softplus brick.
c72a833 Merge pull request #759 from rizar/good_softmax_for_all
4da97cd Merge pull request #790 from mila-udem/linear_docs
c689353 Make Linear docstring clearer.

git-subtree-dir: libs/blocks
git-subtree-split: 38535d89ec166d8b4bef912d4745a51231cdc6da
rizar added a commit that referenced this issue Mar 23, 2016
41248a1 Merge remote-tracking branch 'theano-origin/master' into new_theano
4a75395 Merge pull request #4154 from abergeron/fix_blocksparse
11ad9bb Merge pull request #4152 from Theano/fix_doc
d9237bf Add tests for infer_shape in blocksparse.
83d7531 Use __props__ for the gpu versions of blocksparse.
4c41ccd Replace subtensors with dimshuffles in h_softmax.
1d042d9 Add infer_shape for the blocksparse ops and pass though the broadcastable flags.
9e01f0e Fix doc build
2479cc7 Add __props__ to blocksparse ops.
4e7f550 Changes for 0.8.0rc1 release
830544b Merge pull request #4134 from nouiz/news
1086900 clean-up of NEWS_DEV
4677cd3 List of committers
d9c866f More aliases
0900b66 Merge pull request #4013 from Sentient07/issue-3573
1a3f4e6 Merge pull request #3950 from aalmah/rand_weighted_select_c_impl
6ecfe0f Merge pull request #4116 from ballasn/slowimpl
c421062 import xrange from six.moves for python3 compatibility
92b57e1 check border_mode in the perform methods
d94cdf4 Merge pull request #4094 from abergeron/lift_ifelse
e7f6b32 Merge pull request #2817 from lamblin/detect_nvcc_cast_bug
bf79d0a Update NEWS_DEV.txt
50e6246 Fix the test.
dbf22bb Better fix the test.
7399bc4 Add test.
af3d84b Fix problems in optimization.
c51b283 Merge pull request #4129 from nouiz/cleanup
04bc006 Merge pull request #4121 from nouiz/doc_contrib
caf517b Deleting removed nodes' set from on_detach
28e06d1 Merge pull request #4125 from abergeron/fix_scan_bug
2c77122 change cache code
14241a7 make sure we don't get out of memory
bb964aa Add optimization that verifies that no AbstractConv are present in the graph
f408152 add suggested fixes
4022b2a update test
6abbbc4 update
2a0c192 Allow to easy toggle the extra infer_shape in the code.
11b755f Make FunctionGraph.disown() revert more stuff
4052c18 Remove not use attribute
b12a695 Merge pull request #4127 from abergeron/fix_buildbot
c549b50 Merge pull request #4118 from abergeron/fix_bin
353d86d Missing flake8 fix.
c839e1b Make the add check faster and flake8 format this part.
be8649b Flake8 errors from the buildbot.
522d826 Fix wrong class name.
c6bfe4d Add test to make sure dots are pushed out.
0ce2958 Remove the part about pushing out dot, it's done elsewhere and this code doesn't work.
d3d3a54 Do the toposort on the cloned nodes and build the client dict for client checking.
87d7a3b Check for Elemwise add in a less fragile way.
f5c7aa5 Allow callers of io_toposort to get a dictionary of clients as a side effect.
5fd0a27 Doc to check for easy fix issue as a good starter
49a4624 Try to detect bug in nvcc.
da9006e We need to import gof.compiledir in theano-cache.
93b4fb5 small fix to make the tests pass + flake8
ecc4e7b move debug mode test in tensor.nnet.test
63379af gpu inputs/outputs
784bab4 handle gpu input/outpus
46d6003 fix rectangular padding issue
f78903a clean-up
417e6fb Fix border_mode
7065392 add DebugMode implementation for conv2d
846804d flake8
ff50c40 fixed bugs
0aa5ff7 Merge pull request #4018 from abergeron/fix_buildbot
c42f938 minor fixes
bb93f65 first attempt
42907a0 Merge pull request #4099 from JohnZed/master
4565218 Fix stale import of cuda_enabled in rng_mrg (issue 4098)
d316bae Merge pull request #4107 from nouiz/matt-graham-cudnn_paths_fix
58135c6 Ad dnn.include_dirs by default includa cuda_root, we remove it by default instead of doing a warning.
d392433 Remove empty string in lib_dirs and include_dirs
2a339bb Skip the CorrMM tests if there is no blas.
6429c27 Merge pull request #4095 from abergeron/move_flags2
8472d13 Merge pull request #3965 from harmdevries89/gpudnnpool2
8d975e6 Moved the variables into on_attach method
20e9ae0 Print the theano configuration on travis.
eab9cf5 Merge pull request #4092 from abergeron/fix_buildbot2
a08fc0c Merge pull request #4069 from hantek/docassert
028459c Merge pull request #3768 from nouiz/nanguardmode
f2100b0 Fix linking with mkl.
9ddb5e2 Fix format string for python < 3
3098ee0 Move leftover flags that seemed forgotten.
7d89882 Add lifter for IfElse.
7618e10 Make scipy happy across versions.
b661e71 Fix toarray() call to be optional.
a5da2c0 fix an example
8d68001 Merge pull request #4090 from nouiz/blas_rpath
dcac871 Merge pull request #4046 from AnishShah/issue4035
5cb51d3 Fix doc syntax
145e1cf Fix opt
e5ecd23 small doc fix
0912215 Fix GpuAlloc using debug_perform from GpuAllocEmpty.
44fbc1b Fix debug_perform of GpuAllocEmpty
5fce0ad These are (now?!) sparse matrices.
41607d7 Flake 8 in buildbot?
1ec5b4d Don't use the old name in tests.
1d5fe30 flake8
35df45e Merge pull request #4079 from nouiz/stack_trace
e5f414f Merge pull request #4076 from nouiz/eval
ae61464 Merge pull request #4089 from nouiz/gemv_broadcast
fec5c70 fix
e1c65a5 remove test
fd278b8 fix doctest errors in opt.py
c13c6d4 add doc for opt
b31fd4b make doc appear
102f7b2 Try blas flags with rpath if they don't work.
feaf7d4 Fix crash with gemv and some broadcast pattern
b69b276 Merge pull request #4084 from matt-graham/cudnn_paths_fix
bbb8f71 Merge pull request #4083 from caglar/fix_cnmemprint_msg
6b98e58 Merge pull request #4065 from abergeron/lift_dot22scalar
9625e08 added validate method to ReplaceValidate
b533e26 Remove another eval that was needed due to the change of interface.
6da8726 Also changing libs + lib_dirs in gof/cmodule.py for consistency.
9d456b3 Making order of libs consistent with include_dirs and lib_dirs.
b4ff2ff Addition of user specified cuDNN paths to Op compile commands.
540e7d3 Alternate strategy to not run the test in the base class.
deb1f28 Add infer_shape tests for the blas ops.
7fc3d04 Added the latest changes.
0bc7d0d Test the lifted function for Dot22Scalar.
dfe4049 Flake8
8215457 Merge pull request #4067 from abergeron/debugmode_empty
0f23609 Fix flake8
189069b Merge pull request #4070 from abergeron/fix_buildbot2
e482caf Better docstring
02b95b4 small docstring update
09215e4 Tell user how to make the stack trace longer
f3e4e26 docstring, test and fix the number of stack level kept
cf0abd4 Enable fast stack trace for python 3.5. I tested it and it work.
65e9846 rename var
42c00fb make sure to always have at least one user stack trace and don't check the stack for more then one
6b02f8c Merge pull request #4030 from hantek/fixtesterr
360ddaf Don't call eval
68880f8 Update opts using values_eq_approx to the new interface.
0ec6362 Fix typo in infer_shape methods.
3adaa82 Merge pull request #4042 from adbrebs/warn_flags
d3530b0 Fix typos and wording.
2f739ed Fix the LogSoftmax tests in DebugMode.
7d036b8 Make sure that subsample is a tuple not a list.
62a7e19 Add infer_shape to the blas ops.
480ca3a Replace Dot22Scalar by a GpuGemm instead and add a test.
45753ed Made the suggested changes
adb4f3b Document debug_perform().
ef01a4e Fix new opt
8b0ed6f fix comment
0185804 indentation
95f466d Doc var.tag.nan_guard_mode_check
2c1f7d8 python 3 compatibility
666b86b Merge pull request #4041 from abergeron/fix_buildbot2
9b0c950 Remove unused imports
325ce51 modify test config to check that flags are removed
f2e3c8d Make AllocEmpty return constant-valued memory under DebugMode so that it won't trip up checks.
27ee537 Allow nodes to have a 'debug' behaviour for DebugMode.
622f1d5 Fix calls to as_gpuarray_variable.
90d3193 Also lift Dot22Scalar.
b05017a Fix imports in tests.
c386dd5 Explain more how dummy is there for backward compat and cleanup a bit.
2fc9f31 Change the base conv test class so it doesn't get tested itself.
3811ac7 Don't check inputs if it is the output of computation. If should be already tested
407f0f5 Don't check input if we told not
aaab570 Fix registration of new opt
b9b10ce Add optional opt that convert all *AllocEmpty to *Alloc.
ef8ee79 Fix regression in this PR while unpickling.
659f639 Small file cleanup
bc2f679 Remove false error by NanGuardMode due to GpuAllocEmpty
0c5014b Fix comment
11e1a82 Preserv the nan_guard_mode_check
578ad4f Don't let scan AllocEmpty cause false alarm by NanGuardMode
e088e2a Merge pull request #4061 from nouiz/param
fdd3150 Correct In parameter name following Param removal
ca19b61 update prepare_node blas
5212d68 add pickle
4f2b268 old interface compatible
bc5b983 up
0e0fa00 separate test for new pool interface
b41c0a0 remove prints
bbd9ff3 flake8
5c17201 gpu dnn pool takes tensor variables
494ff73 Print warnings when flags are not recognised
556500b Remove default value when flag is not found in the flag dictionary
274cf81 tip for mixed dtypes
3be031f Re-arrange order of imports in theano/__init__
a2fb4f6 fix local_bitwidth and python_int_bitwidth imports
50e0677 Merge pull request #4040 from caglar/fix_fload_multi_bug
f51b38b delete empty testoutput block
b5d75ef add back test in yaml file
9f46b69 Merge pull request #4045 from nouiz/corrmm
5d8526d Merge pull request #3942 from shabanian/tensor_signal_pep8
e34c32f solve more doctest errors
57aae87 flake8
1b4784a Fix flake8 and names.
7145695 Add a comment explaining the dummy param.
028aaef Move the backward compat where it should be.
51d97ce Remove name that was deprecated in 2010.  This is super old.
5c4aafe fix some circular imports
1cb731e sloving test failures
a262b3b Make sure the docstring is attached to the right object and reuse objects.
1d017d2 Merge pull request #4009 from AnishShah/elu
26d0f24 Fix check
d9e0d1c Merge pull request #4043 from diogo149/master
435caa7 Don't use CpuCorrMM when there is no CPU BLAS
c037151 Correct warning re: cuDNN and pool_2d's ignore_border argument
1647420 raise an error if the user provides an unknown flag
fbb9a5a Fix import problems in configdefaults
c0e864d config stuff
be629ae # This is a combination of 2 commits. # The first commit's message is:
2055fee Exponential Linear Units (ELUs)
6010cad fixed the undefined variable problem.
aa0982e Use working subsample values.
2c19264 Remove pool.py from the whitelist (it has no errors).
28b9927 Remove the bad compat for pickled DownsampleFactorMaxGrad and use a better one.
2d29bb3 Fix the output shape in python to match the C code for DownsampleFactorMaxGradGrad.
8a736df Merge pull request #4033 from nouiz/py3_repr
931dfec Merge pull request #3795 from aalmah/rand_weighted_select
5fc766b Test border_modes and strides together.
8f44215 fixed the multi grad bug.
20609f0 Add test for the previous fix
6c8f1a1 Refactor how the parameters are iterated over to reduce the total time of the test.  (This covers less cases, but should be almost equivalent.)
332601b fix gh-4036. We need the shape and we don't compute it. It wasn't timmed, so don't try this optimization.
dbb49e4 Better error message
12c4530 Make DebugMode raise the good error when no code exist for an op
b236266 Merge pull request #3964 from memimo/3929
4bec565 Merge pull request #4032 from sygi/dot-a-file-in-ldlibrary
3835a61 Fix CudaNdarray repr in python 3. This fix daily buildbot error in python 3
e9425be Merge pull request #4000 from nouiz/cudnn_version
455db13 (#4024) cutting the *.a extension in ldlibrary file
ef41de9 Make cnmem print the start size to make it more clear for user.
290631f Fix typo in warning
2c13f91 Fix crash during opt when we where looking for cudnn version.
89be3bc If cudnn isn't there and we use cuda back-end, print it.
b7f0c30 Only raise error if convolution or pooling are used in the new back-end.
733adf0 flake8 fix and python 3 syntax fix
5df7cb2 Repair the OpenCL init code
be857f2 Make CuDNN mandatory for the new back-end.
b922d88 Warn if cudnn version is too recent. fix gh-3996
9f69145 Print cudnn version when staring gpuarray back-end
adb02ac Print cudnn version at the same time as cnmem usage
f9a68d8 reverted the requested part
9f66573 Merge pull request #4025 from nouiz/corrmm_segfault
6f4a125 Last touchups on the GPU tests.
4bba38a reverted the requested part
bcc5d04 Fix segfault for very big size. I did some timming and there is no difference that is higher then noise. fix gh-4020
b46e24d Merge pull request #4017 from fvisin/fix_tutorial
238f0c8 Refactor the cuda tests.
fe83c9f Disable verify_grad for CpuConv since it can fail.
e14dda6 Force shape inputs to int64 in CorrMM to avoid having something that does not match the 'int' assumed type of before.
74d2d21 Raises error when removed nodes are added back
837e3ab remove unsused import
7331258 inherent from multinomial, output is by default int64, fix tests
ae63503 change weighted_selection to multinomial_wo_replacement
299f0e2 Replace deleted pages with link to new pages
51b840e Merge pull request #4016 from Saxenauts/add_commit_dev_start
d775ab4 minor fixes
aef8cd9 fixed datatype issues
e4904c8 flake8 fixes
246f740 fixed tests
32549fe fixed comment msg
431c19c removed unnecessary try-except
3f05278 add some tests
48e7107 return ints
60b3686 working 1st version
2176d21 start impl
e9c56c3 Merge pull request #3931 from adbrebs/h_softmax_speedup
c3b54b9 Fix tests and code so that they can be happy forever after.
9a8bb9d add an assert statement in the test of the h_softmax to checks that fun_output_tg and fun_output are consistent
5a3526c Merge pull request #4005 from nouiz/cudandarray_repr
f675890 reverted the requested part
dbefcd2 remove if block
573ccea typo and better example of profile output
e493f9c cleup profileMode deprecation in docs
05a6369 Move the tests for AbstractConv with the implementation.
00e13c9 Added the commit instruction in dev_guide submit text
5a0d273 Merge pull request #3924 from fvisin/fix_tutorial
9ae1ab8 Merge pull request #4010 from Saxenauts/temp_disable_warn64
538300f Merge pull request #3997 from abergeron/faster_travis
e851ec0 Fixed whitespace errors on the last commit, temp. disable the warn_float64 flag in Arange
c635b66 Temporarily disables the warn_64 flag in Arange
5621b26 Add link to graph doc in the tutorial
1aa7664 Merge graphstructures from tutorial into extending
bc6770c initial commit
121eeac Temporarily disables the warn_float64 flag in Arange Issue#
4cd0858 Temporarily disabled warn_float64 flag in Arange
eba65e5 Fix PEP8 errors
f649149 Fix broken links in docs
1020d27 Remove __props__ support check in example code
c382385 Move doc tutorial/extending* in extending/*
f9d1b59 Remove overrided implementation that is now the same as the original version.
03f13db Remove call that aren't needed anymore
6a3b192 Merge pull request #3993 from abergeron/fix_buildbot
55699ce Better error message
3f49e74 Add tp_repr to CudaNdarray. This make DebugMode error print the value instead of <CudaNdarray object at 0x7f229c2f6eb0>
5bffd8d Really skip the flake8 part in "everything but tensor".
5315127 Fix flake8 and clarify comment.
709c944 Merge pull request #3988 from nouiz/small
19b6c0a Merge pull request #3987 from SinaHonari/issue3681
472b0d7 Add back missing *.
6ff26fa Fix opts that play directly with owner.inputs.
5c31ea6 Merge pull request #3885 from abergeron/faster_travis
6296694 Fix the docstrings in abstract_conv.py
abe2d24 Fix the sparse docs return types to avoid warnings.
3bfa8db Sync and fix the CompatUnpickler docstring.
2f9ec49 Fix the new doc warnings with the newer sphinx.
75cbdac Don't treat warnings as errors by default.
d3dee21 Install sphinx with conda.
e0b8642 Add dvipng package
ac9fcb6 Add latex packages.
dec46ea Fix doc again.
b331652 Fix doc formatting in pool.py.
a85be0a Disable documentation testing since it is broken.
f233adf Fix doc formatting.
52a062b Split out the doc and flake8 tests and rebalance the tests.
e5fc7cf Have docgen.py exit with an error if a call to sphinx returns one.
c01215f Having two cache: entry discards the first one, which is the one we want.
349de8b Some cleanup.
cbdf4fc Merge the faster parts together.
29f4f90 Add a comment
9db2dca Small doc about about fastmath and nan.
dabc3dd Register in fast_compile an opt that move to the GPU.
b37658f replace a make_thunk by a prepare_node
d1bfd2b Convert border_mode=(0,0) to valid.
0414f66 Correctly tag for gpu scan inplace opt for GPU.
d4acc24 Fix crash in theanp-cache list for subdirectory that don't have a key file.
95f0039 Make a_theano_function.get_shared() that return shared variables.
b524724 Tell that fastmath can give nan and show it less.
32d2744 Merge pull request #3986 from nouiz/test_conv
10ba9ab fixing make_node param passing for conv_grad
319643c Merge pull request #3872 from SinaHonari/issue3681
1e831db Fix the Print lifter.
5f75ecd Remove useless optimizer now that mixed-type abstractconv can't happen.
3f31dc2 Normalize AbstractConv inputs types to not have nodes with mixed input types.
275ffe7 Merge pull request #3984 from nouiz/Param
b99ee90 Correct mode fix
4231830 Merge pull request #3976 from nouiz/gpuarray_memset0
815e278 Fix test when mode!=DebugMode
c48ee17 correcting indentation
5fe75ed Add back Param and make it deprecated. fix gh-3979.
c732789 Merge pull request #3966 from nouiz/dnn
d521a6c Merge pull request #3982 from abergeron/fix_nnet
a4debbc removing deconv_op name
594e7de Fix Gpu......WithBias to work properly for float64.
f655bba Merge pull request #3926 from nouiz/gpuarray_pickle
2f37aeb Better comment. [ci skip]
8227ec0 Merge pull request #3973 from abergeron/fix_buildbot2
178d793 Small doc update
c8aa061 Merge pull request #3974 from nouiz/faster_opt
0650b0f Merge pull request #3970 from abergeron/fix_buildbot
737a4d0 Merge pull request #3969 from abergeron/fix_Wl
8276edd Merge pull request #3933 from nouiz/profile
520df6b Add some doc update that was missing in previous PR.
b0b4f13 small fix following code review
e486864 Make opt GpuAlloc memset0 work in the new back-end
49b79e5 small doc update
184ea12 Make the comparison on the CPU.
cef892b Merge pull request #3968 from nouiz/nanguardmode_opt
cd0afd4 Speed up opt by not traversing the graph again.
6a9aa55 Speed up opt. Now constant folding is done in the eq, no need to traverse the graph to get constant at each node.
9b05f82 Don't run the abstractconv tests under DebugMode.
1e2660a Fix MakeList to not make an indirect view.
673a2a7 Merge pull request #3917 from andreh7/2016-01-23-fix-for-custom-python-installations
7d2578c Make the sig workaround for unsigned ints too.
ff9c346 Merge pull request #3935 from poolio/rop_bugs
752d502 Merge pull request #3961 from nouiz/test
b39b15d Don't duplicate args.
daab99c Have it not end up in preargs2 anyway.
e205ce2 Try to fix -Wl args to nvcc.
52790dd Doc dnn.enable flag
20cf456 Enable opt in NanGuardMode. Make it use the default optimizer. Fix gh-3967.
9d1344f small code change
3135482 make test without pygpu don't crash at init
6f636a8 Fix a test where the gpuarray back-end didn't downcasted python float to float32.
50572e9 Move to a separate test file gpuarray tests that must always run, even when pygpu isn't available
768fb73 Fix error msg
08c6bbc Use config.dnn.enabled in gpuarray
b755970 Add config flag dnn.enabled={auto,True,False}.
9627228 Merge pull request #3936 from nouiz/32bit
d6b3dff Merge pull request #3947 from Sentient07/issue-#3944
319e0aa Fix a test in 32bit. We can't make min/max reduction on 0 sized tensor.
1ed8895 Fix code review comment
ed3034a added BatchedDotOp
1d46dd1 Add d3viz in travis
a3575cf Fix test in python 3
b8772f4 Fix a test with python 3
7ffd86a Fix test on GPU with mode=FAST_COMPILE
aecc492 Merge pull request #3955 from pra85/patch-2
47278f7 Merge pull request #3883 from skaae/bessel
dfce3ae Merge pull request #3952 from nouiz/doc
14fdf30 add j0 and j1 to if else check
6f5082f remove __eq__ and __hash__
61c71ab Fix typo in docs
dac4da8 Fix docgen as now warning are treated as error
b2f9958 Better test update for python 32bit fix
bdfe90b Fix comments and move code to detect those type of errors
658ba1c further refining the comments
a683232 update the test assert when device=gpu
72d3a28 Renamed BatchedDotOp to GpuBatchedDot
0d82fe2 change shape dimensions int32 to int64
18319b8 Merge pull request #3943 from lamblin/fix_batcheddotop_import
73fbb21 Fix path for BatchedDot
2d75099 change shape dimensions int32 to int64
f87853c Merge pull request #3508 from cooijmanstim/batched_gemm
5565558 Make test in python 32bit. sum, average cause small difference result ~1e-17
5c02232 Fix test with python 32bit
4800a51 Merge pull request #3940 from abergeron/fix_buildbot
7fc2bcb Don't use floating-point axes.
1f98466 Remove confusing sys.stdout.flush() which displaces the printing of messages.
6e25ecf Merge pull request #3938 from carriepl/half_padding
0c17be8 Add tests for border_mode half
3ee95b8 Add border_mode 'half' to dnn convolution
1908d99 Fix test in 32 bit python
4978584 remove bincount uint64 restriction, it was a left over from the numpy implementation
817ff62 Make repeat test pass on 32bit python. We use the numpy code only for vector, not scalar
45c7b96 Make test of bincount pass with 32bit python. We don't have the numpy limit anymore
2bbcab0 Better error message
33e07c2 Update test of profiling output.
0d2f82a add missing return to GpuFromHost R_op
e891fb3 Merge pull request #3918 from nouiz/tx1
45efd37 Merge pull request #3845 from f0k/add-cudnn-half-padding
c46d30b Code clean up and more readable printing
5170ff7 Code refactor, make printed information more clear
cc8eaed remove information not useful
06f3b03 profile_memory now print the peak if we disable inplace
62d5cbe Merge pull request #3890 from mohammadpz/remove_param_class
fc5fe33 Replace sparse_block_dot by tensor_block when full output is requested.
264d8bf BatchedDot: fix reference
58cddb7 Merge pull request #3855 from ballasn/convdoc
13f8ea6 Add border_mode="half" to cuDNN convolutions
4a0add8 Remove duplicate code
efd4d4b Merge pull request #3908 from vmichals/abstract_conv_keep_stacktrace
b6732d2 add c code
8ace004 arguments rewritten
48f621b Alloc pickling of GpuArray object
921b8eb fix static implementation error
cc80d73 removed erroneous additional string (at beginning of file) accidentally introduced in the last commit
9523985 cublas timming on tx1
1cf5fd3 Doc how to install on TX1
90c8f80 update doc
8020a38 Merge pull request #3912 from lamblin/fix_abstractconv_grad_type
529eb38 Make sure the axes array will be integer-typed.
0fe82c4 fixed travis complaints about whitespace
31f59dc fixes #3556: more robust determination of python library name and directory in function std_lib_dirs_and_libs(), now works on private python installations on Linux
a324267 Avoid numpy warning about using non-integer shapes.
c96c73f Fix DeprecationWarning about int division.
6bf15e5 Make sure grads of abstract conv ops have the right type
914ebf7 BatchedDot: implement c_code_cache_version
c539a57 BatchedDot: move to theano/tensor/blas.py
0fde9a4 BatchedDot: simplify reference implementation
1075d83 BatchedDot: negate contiguity test for less confusion
001ddc3 BatchedDot: don't waste time copying output before overwriting it
3805e3b applied fixes mentioned by Pascal L, abstractconv unit test passes without failure
a41fc8b first try
8cdf8e6 BatchedDot: satisfy flake8's perverse desires
d8538e4 BatchedDot: avoid repeated indexing
dd41f3e Merge pull request #3915 from Saizheng/master
0d413f9 remove the mistakenly copied file ':w'
62bbc9b Merge pull request #3870 from Saizheng/master
f2a2181 Merge pull request #3907 from laurent-dinh/conv_test
2dc471f Merge pull request #3903 from matthias-k/AbstractConv2d_R_op
887e39c Merge pull request #3909 from nouiz/website
97fe07f Several optimizations possible
ccf4116 Merge pull request #3899 from carriepl/log_softmax
290fc8b profile: more changes
96a257a move print_tips to profiling.py
7b8ee75 updating notations and namings
0abf03b Test for R_op of AbstractConv2d
df7fdae name deep learning in git repo so that Theano show up in github search result for deep learning
fdfbab3 Merge pull request #3747 from Sentient07/issue-3094
6f54979 Fix typo
7a9a94f Merge pull request #3891 from abergeron/fix_blas_default
5103e4f Merge pull request #3902 from abergeron/fix_filter
00821f9 Remove redundant code following review
cbf4915 Merge pull request #3905 from nouiz/corrmm
6439d81b Merge pull request #3895 from nouiz/debug
80acc74 Flake8 fix.
b88a016 Flake8 fix.
2beb17a Don't use CorrMM when there is no c compiler (no perform, it will fallback to the old conv that have perform)
578c145 R_op for AbstractConv2d
6877f6f Added optimization test
6639305 Add a test for the allow_downcast issue.
68ae1a7 add bessel of the zeroth kind
8884e69 Fix misscoped allow_downcast block.
3aa7a24 Always silently add the default library dirs, but don't append them to flags.
be838ed update_val is not guarenteed to have a type attribute.
ced656f Test result of theano function
a341a77 indent code example
f8bbacb Split local_log_softmax_dnn into two separate op lifters
034c9c3 Fix default blas flag detection by reworking try_blas_flag.
19f4a62 Made the final changes
3dc4855 Add url to gdb quick tutorial
487cf52 Add test for opt to convert LogSoftmax to DnnLogSoftmax in gpuarray backend
da93970 Add opt to convert LogSoftmax to DnnLogSoftmax in gpuarray backend
25e6df0 Merge pull request #3881 from nouiz/gpu_dot22
9901219 Update doctring in gpuarray/dnn.py
c40c1ed Add test for opt to convert LogSoftmax to DnnLogSoftmax in cuda backend
bdcb8d2 Add opt to convert LogSoftmax to DnnLogSoftmax in cuda backend
87f5f60 Merge pull request #3478 from piotrfrankowski/with_statement
4225e32 Merge pull request #3750 from andreh7/ah-2015-12-05-more-comments-on-logistic-regression-example
171a82d Merge pull request #3896 from claymcleod/master
408568f Merge pull request #3897 from jfsantos/master
956f8dc Fixed import of local pool submodule
6253c08 removed '(constant)' from comment in doc/tutorial/examples.txt on request of @nouiz
2163e1f Bug fix: don't try to os.mkdirs on an empty directory
9a0ab54 Merge pull request #3799 from lamblin/avg_pool_v3
136c6d4 Merge pull request #3847 from jojolalpin/master
ef304db Doc how to use gdb and python
4507dd1 Make nvcc.flags being using for cuda_ndarray
9cbbbd7 Make cmodule.remove_gxx_opt work for nvcc
f640674 Merge pull request #3893 from abergeron/fix_buildbot
1da748d remove duplicate Ins
f85c104 Param.default to In.value
279ff3b Param.default to In.value
b4bd7b9 No error in v3 when padding is of 0
74b5945 Skip tests of average_exc_pad with cudnn v3
7b206fb Disable average_exc_pad pooling for cudnn v3
d93640a Get rid of a bit more warnings.
62eed15 Fix the Ananconda workaround to detect later versions.
4c01908 Fix the default return from default_blas_flags to return valid flags.
ce2e15b Param class in now replaced by In
f713623 Follow renaming.
adf81d3 Merge pull request #3679 from shabanian/tensor_signal_pep8
15aef2f Better code
71f30a9 Fix opt warning/not moving to GPU some node.
6ce3aa7 Merge pull request #3884 from abergeron/fix_mac_blas
ae444d3 Merge pull request #3861 from nouiz/version
3a0493c Merge pull request #3875 from abergeron/fix_buildbot
534e040 Small doc update.
3303c99 Flake8 fix.
0c10888 Tweak the blas flags logic so that it doesn't require certain keys in the dict.
a591990 Fix brain typo.
1893c13 Use prepare_node() instead of make_thunk() in Elemwise.
710e6b4 Add a new prepare_node() method that can be used to do stuff 'just before' make_thunk and help DebugMode work correctly.
3a0de13 Add test for the elemwise issue.
e98220f Make cumul be a double variable for better accumulation precision.
ab40a50 Get rid of some of the deprecation warnings in the buildbot.
3dcba54 Merge pull request #3363 from fvisin/logsoftmax
9ad1ea0 Merge pull request #2 from abergeron/logsoftmax
0041ce6 Remove the try .. except construct since it masked some errors in the opt.
994c8d3 Register the logsoftmax opt in fast_compile.
a5735a1 Merge pull request #3877 from abergeron/fix_segfault
f986e0d Fix the cudnn context problem.
6b7a3d1 Revert "Set the proper context when destroying the cudnn handle."
b1c5a67 Fix grad optimization.
91fe71d raise a warning when the old file is imported
ede4874 Merge pull request #3858 from mohammadpz/try_blas_flag
7bad2e2 Merge pull request #3867 from nouiz/small
4bd7b62 deconvolution interface
c282d07 Merge pull request #3869 from kmike/remove-scan-warning
19ed493 Merge pull request #3868 from kmike/py3-raise_with_op
2e660ce flake8
b7e68d2 scan: remove unconditional warning when strict=True is used. Fixes GH-3333.
eb7901d PY3 fixed theano.gof.link.raise_with_op in Python 3
2588a94 blas added in flags
f1eddb0 try_blas_flag return all flags
fcc9944 try_flags is replaced by try_blas_flag
42dd8a8 Merge pull request #3856 from abergeron/transfer_noints
2d4f6d7 Merge pull request #3862 from saatvikshah1994/issue-3658
71a3700 Merge pull request #3853 from abergeron/fix_lift_abstract
53f2908 fixed some issues in test_pool.py
354c4a9 Revert back to transferring all inputs by default.
1838572 Set the proper context when destroying the cudnn handle.
ea04c80 fixed some issues in test_pool.py
511b759 Remove unused f for flake8 compliance.
9cb9c43 Rename test_downsample.py to test_pool.py
1b55e55 Use a prefered way and clearer way to have code work with python 2 and 3.
b433d70 Add version added in the library doc
60dbf4b Remove old link to now non existing page
834ab99 More information in error
bd0ee86 Fix disabling of docgen.py introduced in December.
6fae1fc make docgen.py return to current directory at the end of the script.
3d82403 Link to platoon
686e85f Skip a test if cuda isn't available.
48dfc25 fixed some issues in downsample.py
372939b Rename max_pool_2d to pool_2d and Pool for the DownsampleFactorMax
8f65b44 BatchedDot: leave first dimension broadcastable if it was broadcastable for either of the inputs
e527643 BatchedDot: remove reference to Dot
878633e BatchedDot: remove complex test cases as we will not support complex types for now
c37a6bb Fix test so that it actually crashes on unfixed code.
25ca839 BatchedDot: upcast inputs if they are of mixed types
ca630a9 Add test for shape on GPU.
16b0bc5 corrected setup of d3viz
73a7621 Bump version to 0.8.0.dev0
808296a Expand on the comment.
149d008 Merge pull request #3859 from kmike/faster-TensorConstant
dde6c84 optimize TensorConstant constructor
6b4055d auxiliary function 'try_blas_flags' added.
e5f7eb3 Remove the undocumented and unused tag.context_name.
1381698 Add documentation about the tag.target attribute and remove false statements about performance.
4e67e4d Ignore nodes that are already on the GPU.
7a90c78 Don't transfer int inputs to the GPU by default.
8cd678d BatchedDot: small improvements
0b89695 BatchedDot: fix more bad pointer arithmetic
78bbb56 BatchedDot: reduce duplication
77729ff update convolution documentation
1ac274a BatchedDot: fix bad pointer arithmetic
5cde49a Add test for theano.sparse.opt.sd_csc in test_opt.py
c56c2f7 BatchedDot: fix potential bug
fe58ada Merge pull request #3850 from dmitriy-serdyuk/dunction-dump-doc
b90be7c Fix typo in function_dump doc
f279798 Merge pull request #3817 from viirya/check-home-dir
80df08b Merge pull request #3814 from SuperElectric/rebroadcast_bugfix
6a1e661 Use /tmp/$USER if possibly.
09390a2 Merge pull request #3830 from Sentient07/2016
353b0fc Merge pull request #3846 from kmike/patch-1
9c7c3b8 Update opt.py
4192d13 DOC removed incorrect statement from compatibility docs
4d03d97 Merge pull request #3824 from Sentient07/issue-3686
149a9df Merge pull request #3807 from nouiz/lock
8eadfc3 Updated the copyright dates
0ef3ec3 More informative warning in InferShapeTester
493517e Following code review, better docstring and take the lock less often
b9e32a0 Suggested changes made
167df2c Merge pull request #3801 from nouiz/doc
cb79538 Merge pull request #3804 from nouiz/tests
6d73535 Revert some changed I didn't wanted to commit.
b3d0a8b Try to be even more overflow tolerent.
1c57b67 Merge pull request #3822 from mronian/master
45d3136 Merge pull request #3828 from dwiel/patch-1
0d69ea0 Merge pull request #3831 from kmike/py3-fixes
b47debf pep8
0b54e8d Update warning messge following code review
87c4e01 Merge pull request #3811 from nouiz/typed_list
2513a44 Merge pull request #3803 from nouiz/gpu
1739dda DOC fixed Python 3 compatibility issues in Tutorial and Library Reference
12aa951 PY3 fixed ProfileStats.summary() in Python 3.x
5ec1e57 PY3 fixed SeqOptimizer.print_profile in Python 3
45f55ca TST avoid numpy deprecation warning in tests under Python 3
e559fd4 Happy new year! :)
cfbc73b typo in bn.py documentation
d4128b7 Added test for get_conv_output_shape method
3c32963 Checks compiledir flag for prohibited characters. Fixes #3711
ac26259 Merge pull request #3819 from superbobry/master
4d69aea Fix Python3 regression in ``pydotprint``
537b24f Check home_dir writing permission and assign temporary directory.
f380577 Fixed pep8
5fa767d fixed too-stringent typecheck to accept not just bool, but also numpy.bool_
ac4d5af Add TypedListConstant. This fix the added test.
73dfccf r_shape can be None
d18af18 Quick fix in __str__
ee4f2a9 Don't take the lock by unless needed at exit
1aed098 Good dtype comparison
b75fc56 Removed make_node method and corrected the test
e03b0ca Remove compilation warning
d3dbe92 remove duplicate import
83c8299 Remove duplicate import
ad980be Remove useless return
8e65af1 Small code speed up/clean up
8100d60 Update test name to the real op name
2c34d0c pep8
e082574 Skip flake8 tests on auto saved tmp file
d915fb9 Tell the limitation of the pooling on the GPU.
0c6e094 Fix gh-3800 make it clear that strides must be ints
5418629 Merge pull request #3802 from f0k/fix-blas-detection
7936ca4 Remove useless comment
8c607e8 Remove opt warning with mixed dtype
4b667da Merge pull request #3796 from abergeron/gpuarray_prealloc
b77115c Merge pull request #3790 from hidasib/gpu_set_subtensor_2d
554be9c Enable cudnn by default when avaialable in the new back-end
7e34c53 Speed up Elemwise.perform
3e0e53e Fix fallback in BLAS detection
91547f2 Fix DebugMode error due to overflow in comparison
9fad11d [BUG] bugfix of python code. It is triggered in DebugMode with this test theano/sandbox/tests/test_rng_mrg.py:test_multinomial_n_samples
fcfa049 pep8
7197ebc pep8
a130465 Fix DebugMode when a perform raise NotImplementedError. This happen with this test theano/sandbox/cuda/tests/test_conv_cuda_ndarray.py:test_conv_grads
721666e Fix test in DebugMode with recent numpy version. AbsTester with complex input was failing.
2508997 Changed c code version from 6 to 7. Added tests to test_basic_ops.py for GpuAdvancedIncSubtensor1 (general case) and GpuAdvancedIncSubtensor1_dev20 (2d case if compute capability >=2.0).
5e1295f Merge pull request #3798 from nouiz/small
22c56b8 Skip part of tests with not compatible pydot_ng
2ba1d48 Don't crash when generating the error for some type of error
f331d52 Merge pull request #3631 from hantek/doctest
a5c6a13 Merge pull request #3784 from mronian/master
663c070 Cap memory preallocation to 98% of the total.
2640056 Merge pull request #3792 from abergeron/gpuarray_prealloc
0a876e2 Adds test for uint datatype for image_shape and filter_shape in theano.tensor.nnet.conv and also checks if datatype is in tensor.discrete_dtypes
54e9675 Merge pull request #3559 from abergeron/multi_gpu_doc
e6ca182 Delete the cudnn handles when we no longer need them.
d6b2e03 Rework the gpuarray api version check.
8fb5d66 Add the gpuarray.preallocate configuration variable.  This will help fix out of memory errors in certain sequences of tests.
0f03bac BatchedDot: PyArray_Newshape complains about NPY_KEEPORDER when it shouldn't
9a3f668 BatchedDot: generalize matrix-matrix code to matrix-vector, vector-matrix, vector-vector products
4ce8a48 BatchedDot: first stab at C implementation
cccef96 BatchedDot: fix comments and remove assertion
a3fc110 BatchedDot: fix optimization to handle matrices
82aac7d BatchedDot: remove unnecessary __str__
a915c82 BatchedDot: remove reference to Dot
62e2bf5 BatchedDot: trust that gradients are floats
cc8c9e4 BatchedDot: fix grad
2202eb4 BatchedDot: more diverse tests
bee1a7a BatchedDot: catch bad inputs
91fe0a2 BatchedDot: write docstrings
1a28571 introduce optimization to move BatchedDot to GPU
7d1c991 BatchedDot: provide dtype in numpy reference implementation
ed4e067 BatchedDot: test and fix grad
e68999e BatchedDotOp: simplify tests
f6c4864 BatchedDot: implement and test
713fd0f BatchedDotOp: implement and test infer_shape
34be0f0 solving conflicts
75b1c22 fixed all warnings in doc. added the sphinx -m flag in docgen
4ad36dd Merge pull request #3788 from nouiz/carriepl-v4
c5084ac Fix typo.
47ecce0 Remove the warning paragraph as the issues have been fixed.
e28057a Fix typos.
838508a Move the warning about no speedup earlier and add a section showing the printout of the context map.
0f4f011 Fixes from review.
99e23d1 Link to PureType instead of duplicating the docstrings.
f2c8575 Fix the multi gpu script to use the new target argument.
7beebd0 Add a tutorial section on how tu use the new multi-gpu functionality.
d6156c6 Delete some old and useless documentation.
30617ff Rest of libdoc for gpuarray.
3bf6f4c Add a listing of ops for gpuarray.
f005cc2 Add a bunch of documentation to gpuarray/type.py
418967d Link in the blocksparse docs and add a section about the gpuarray backend.
07b2896 Compile theano function and assert
61d1163 Gpu implementation of GpuAdvancedIncSubtensor1 (set mode), using CudaNdarray_CopyFromCudaNdarray.
fd75d2c GPU implementation for GpuAdvancedIncSubtensor1_dev20 using atomicExch()
32e113c Flake8
ced46b7 Make dnn conv in gpuarray work with AbstractConv
aafc23a Standardize documentation of the 'precision' parameter
7d44a89 Update and standardize error messages
524438b Clean up V2 code
877f421 Avoid using method cudnnGetConvolution2dDescriptor() method
ebd9778 Make deprecated dnn flags raise more meaningfull error messages.
32f1155 Add 'as_input' option to dnn.conv.precision flag
7ff6621 Drop support for v2
3e4da0d Don't restrict 'small' workspace implementation of GradW to v4 and later
a28cb16 Add missing syntax
757b04c Fix verification of spatial sizes in cudnn conv (gpuarray backend)
fdffa58 Fix condition based on CuDNN version
6f1dc3d Call cudnnGetConvolution2dDescriptor instead of cudnnGetConvolutionNdDescriptor
bf91848 Add comments and fix condition in GpuDnnConvGradW C code (gpua backend)
2a857e3 Update GpuDnnConvGradI for CuDNN v4 (gpua backend)
1e48b73 Update GpuDnnConvGradW for CuDNN v4 (gpua backend)
64439f4 Update GpuDnnConv for CuDNN V4 (gpua backend)
93f6f44 Define new V4 elements from V3 elements in cudnn_helper  (gpua backend)
674cd4f Add precision param to dnnConv (gpua backend)
46c8338 Add precision param to GpuDnnConvDesc
5a015a8 Add definition for new implementation in cudnn_helper.h (cuda backend)
330da07 Retain compatibility with CuDNN v2 (cuda backend)
cde0771 Add checks to make sure backward tiled-fft is not used when it shouldn't (cuda backend)
b177f3a Add checks to make sure tiled-fft is not used when it shouldn't (cuda backend)
cf26ae5 Limit new convolution implementations to CuDNN V4
2a84aa6 Make dnn_conv and dnn_conv3d support param precision (cuda backend)
0ed16db Add new precision parameter to GpuDnnConvDesc (cuda backend)
3914bf6 Add new flag to control CuDNN convolution precision
c36f9a2 modify cuda/dnn.py to support new convolution implementations
13bad51 Deprecate algo_bwd flag in favor of algo_bwd_data and algo_bwd_filter
fc7d727 fix typos
e200ca2 start new cudnn v4 convolution mode
1f3c1f6 changed corresponding to the upstream
e3a0aee Flake test passes
112b22f Updated tutorial/extending_theano.txt
48661f1 Reverted suggested changes
4936e63 Test added, code made complaint with pep8 standards
db6fc48 mend
e3430b6 Moved the mae_node method
25c0f5e58 PEP8 respect
b050916 fix all doctest errors, but not turning on warning-to-error flag in sphinx build
a8316c2 add back the doc test part
6d25ecd fix those paths in PART variable
159a988 remove commented lines in docgen.py; uncomment theano tests
09a5fde fix the file not found problem
7ff3ca2 add sphinx to dependencies
6cb95da remove epydoc from the repo
c13f339 add epydoc in travis yaml
7b733b0 add doc test in yaml
54b662a added comments to logistic regression example in the documentation. No code changes
57a2f9c Slight improvement made to the test
c4fb7e5 itypes and otypes checked and added test
633ff54 fix all doctest errors, but not turning on warning-to-error flag in sphinx build
d7d6408 Do not import optimizations
16333fd Add C code for LogSoftmax
e481355 Fix use crossentropy_onehot_grad instead of logsoftmax
49cf5b4 Add LogSoftmax gradient test and gradient optimization test
8b44700 Add optimization for grad
f9eb767 Add test for the optimization of the fprop of log(softmax(x))
65898f8 Remove unused optimization for log softmax
afd6c4d add back the doc test part
e671d03 fix those paths in PART variable
483ca0b remove commented lines in docgen.py; uncomment theano tests
b66c305 fix the file not found problem
66ecd2a add sphinx to dependencies
c92b36d remove epydoc from the repo
7c5c18e add epydoc in travis yaml
aa32a07 add doc test in yaml
f05a0c8 Add LogSoftmax python code and tests
3b9975e fix according to comments
abc2876 #3429 - python 'with' statement in reset modules
54b194d #3429 - python 'with' statement in tests module

git-subtree-dir: libs/Theano
git-subtree-split: 41248a16a79ff1d11f8ddc68afe6571d3f80769f
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants