changeset 12713:47904d77c528 octave-forge

maint: move nnet package to its own hg repository.
author carandraug
date Wed, 24 Feb 2016 17:45:06 +0000
parents 7c272a2844d7
children 9d24dd9d2986
files main/nnet/COPYING main/nnet/DESCRIPTION main/nnet/INDEX main/nnet/Makefile main/nnet/NEWS main/nnet/doc/docReadme main/nnet/doc/latex/asymptote/mlp4-2-3.asy main/nnet/doc/latex/asymptote/mlp4-2-3.eps main/nnet/doc/latex/asymptote/mlp4-2-3.pdf main/nnet/doc/latex/asymptote/transferFunctions/logsig.asy main/nnet/doc/latex/asymptote/transferFunctions/logsiglogo.asy main/nnet/doc/latex/asymptote/transferFunctions/purelin.asy main/nnet/doc/latex/asymptote/transferFunctions/purelinlogo.asy main/nnet/doc/latex/asymptote/transferFunctions/tansig.asy main/nnet/doc/latex/asymptote/transferFunctions/tansiglogo.asy main/nnet/doc/latex/common/bibliography.tex main/nnet/doc/latex/common/version.tex main/nnet/doc/latex/developers/algorithm/LevenbergMarquardt.tex main/nnet/doc/latex/developers/algorithm/algorithmen.tex main/nnet/doc/latex/developers/analyzing/analyzingNewff.tex main/nnet/doc/latex/developers/analyzing/matlab.tex main/nnet/doc/latex/developers/codeConvention.tex main/nnet/doc/latex/developers/codingGuideline/codingGuideline.tex main/nnet/doc/latex/developers/examples/example1.tex main/nnet/doc/latex/developers/examples/example2.tex main/nnet/doc/latex/developers/examples/examples.tex main/nnet/doc/latex/developers/funcindex/funcindex.tex main/nnet/doc/latex/developers/funcindex/funcindexCalled.tex main/nnet/doc/latex/developers/introduction.tex main/nnet/doc/latex/developers/neuralNetworkPackageForOctaveDevelopersGuide.dvi main/nnet/doc/latex/developers/neuralNetworkPackageForOctaveDevelopersGuide.pdf main/nnet/doc/latex/developers/neuralNetworkPackageForOctaveDevelopersGuide.tcp main/nnet/doc/latex/developers/neuralNetworkPackageForOctaveDevelopersGuide.tex main/nnet/doc/latex/developers/numbering.tex main/nnet/doc/latex/developers/octave/directorys.tex main/nnet/doc/latex/developers/octave/functions/isposintOct.tex main/nnet/doc/latex/developers/octave/functions/min_max.tex main/nnet/doc/latex/developers/octave/functions/newff.tex main/nnet/doc/latex/developers/octave/functions/poststd.tex main/nnet/doc/latex/developers/octave/functions/prestd.tex main/nnet/doc/latex/developers/octave/functions/sim.tex main/nnet/doc/latex/developers/octave/functions/train.tex main/nnet/doc/latex/developers/octave/functions/trastd.tex main/nnet/doc/latex/developers/octave/octave.tex main/nnet/doc/latex/developers/tests/__analyzerows.tex main/nnet/doc/latex/developers/tests/__copycoltopos1.tex main/nnet/doc/latex/developers/tests/__optimizedatasets.tex main/nnet/doc/latex/developers/tests/__randomisecols.tex main/nnet/doc/latex/developers/tests/__rerangecolumns.tex main/nnet/doc/latex/developers/tests/isposint.tex main/nnet/doc/latex/developers/tests/min_max.tex main/nnet/doc/latex/developers/tests/newff.tex main/nnet/doc/latex/developers/tests/prestd.tex main/nnet/doc/latex/developers/tests/purelin.tex main/nnet/doc/latex/developers/tests/subset.tex main/nnet/doc/latex/developers/tests/test.tex main/nnet/doc/latex/developers/title.tex main/nnet/doc/latex/developers/varietes.tex main/nnet/doc/latex/perl/analyzeOctaveSource.pm main/nnet/doc/latex/perl/createFunctionIndexDocu.pl main/nnet/doc/latex/perl/createTestDocu.pl main/nnet/doc/latex/users/bibliography.tex main/nnet/doc/latex/users/examples/1/1.tex main/nnet/doc/latex/users/examples/1/MLP9_1_1.m_template main/nnet/doc/latex/users/examples/1/MLP9_1_1.tex main/nnet/doc/latex/users/examples/1/mData.txt main/nnet/doc/latex/users/examples/2/2.tex main/nnet/doc/latex/users/examples/2/MLP9_1_1.m_template main/nnet/doc/latex/users/examples/2/MLP9_1_1.tex main/nnet/doc/latex/users/examples/2/highlight.sty main/nnet/doc/latex/users/examples/2/mData.txt main/nnet/doc/latex/users/examples/examples.tex main/nnet/doc/latex/users/introduction.tex main/nnet/doc/latex/users/knownIncompatibilities.tex main/nnet/doc/latex/users/neuralNetworkPackageForOctaveUsersGuide.tcp main/nnet/doc/latex/users/neuralNetworkPackageForOctaveUsersGuide.tex main/nnet/doc/latex/users/neuralNetworkToolboxForOctaveUsersGuide.tcp main/nnet/doc/latex/users/neuralNetworkToolboxForOctaveUsersGuide.tex main/nnet/doc/latex/users/numbering.tex main/nnet/doc/latex/users/octave/neuroPackage/graphics/logsig.eps main/nnet/doc/latex/users/octave/neuroPackage/graphics/logsig.pdf main/nnet/doc/latex/users/octave/neuroPackage/graphics/logsiglogo.eps main/nnet/doc/latex/users/octave/neuroPackage/graphics/logsiglogo.pdf main/nnet/doc/latex/users/octave/neuroPackage/graphics/purelin.eps main/nnet/doc/latex/users/octave/neuroPackage/graphics/purelin.pdf main/nnet/doc/latex/users/octave/neuroPackage/graphics/purelinlogo.eps main/nnet/doc/latex/users/octave/neuroPackage/graphics/purelinlogo.pdf main/nnet/doc/latex/users/octave/neuroPackage/graphics/tansig.eps main/nnet/doc/latex/users/octave/neuroPackage/graphics/tansig.pdf main/nnet/doc/latex/users/octave/neuroPackage/graphics/tansiglogo.eps main/nnet/doc/latex/users/octave/neuroPackage/graphics/tansiglogo.pdf main/nnet/doc/latex/users/octave/neuroPackage/logsig.tex main/nnet/doc/latex/users/octave/neuroPackage/min_max.tex main/nnet/doc/latex/users/octave/neuroPackage/neuroPackage.tex main/nnet/doc/latex/users/octave/neuroPackage/newff.tex main/nnet/doc/latex/users/octave/neuroPackage/poststd.tex main/nnet/doc/latex/users/octave/neuroPackage/prestd.tex main/nnet/doc/latex/users/octave/neuroPackage/purelin.tex main/nnet/doc/latex/users/octave/neuroPackage/saveMLPStruct.tex main/nnet/doc/latex/users/octave/neuroPackage/sim.tex main/nnet/doc/latex/users/octave/neuroPackage/subset.tex main/nnet/doc/latex/users/octave/neuroPackage/tansig.tex main/nnet/doc/latex/users/octave/neuroPackage/train.tex main/nnet/doc/latex/users/octave/neuroPackage/trastd.tex main/nnet/doc/latex/users/octave/octave.tex main/nnet/doc/latex/users/title2.tex main/nnet/doc/pdf/neuralNetworkPackageForOctaveDevelopersGuide.pdf main/nnet/doc/pdf/neuralNetworkPackageForOctaveUsersGuide.pdf main/nnet/inst/dhardlim.m main/nnet/inst/dividerand.m main/nnet/inst/dposlin.m main/nnet/inst/dsatlin.m main/nnet/inst/dsatlins.m main/nnet/inst/hardlim.m main/nnet/inst/hardlims.m main/nnet/inst/ind2vec.m main/nnet/inst/isposint.m main/nnet/inst/logsig.m main/nnet/inst/mapstd.m main/nnet/inst/min_max.m main/nnet/inst/minmax.m main/nnet/inst/newff.m main/nnet/inst/newp.m main/nnet/inst/poslin.m main/nnet/inst/poststd.m main/nnet/inst/prestd.m main/nnet/inst/private/__analyzerows.m main/nnet/inst/private/__calcjacobian.m main/nnet/inst/private/__calcperf.m main/nnet/inst/private/__checknetstruct.m main/nnet/inst/private/__copycoltopos1.m main/nnet/inst/private/__dlogsig.m main/nnet/inst/private/__dpurelin.m main/nnet/inst/private/__dradbas.m main/nnet/inst/private/__dtansig.m main/nnet/inst/private/__getx.m main/nnet/inst/private/__init.m main/nnet/inst/private/__mae.m main/nnet/inst/private/__mse.m main/nnet/inst/private/__newnetwork.m main/nnet/inst/private/__optimizedatasets.m main/nnet/inst/private/__printAdaptFcn.m main/nnet/inst/private/__printAdaptParam.m main/nnet/inst/private/__printB.m main/nnet/inst/private/__printBiasConnect.m main/nnet/inst/private/__printBiases.m main/nnet/inst/private/__printIW.m main/nnet/inst/private/__printInitFcn.m main/nnet/inst/private/__printInitParam.m main/nnet/inst/private/__printInputConnect.m main/nnet/inst/private/__printInputWeights.m main/nnet/inst/private/__printInputs.m main/nnet/inst/private/__printLW.m main/nnet/inst/private/__printLayerConnect.m main/nnet/inst/private/__printLayerWeights.m main/nnet/inst/private/__printLayers.m main/nnet/inst/private/__printMLPHeader.m main/nnet/inst/private/__printNetworkType.m main/nnet/inst/private/__printNumInputDelays.m main/nnet/inst/private/__printNumInputs.m main/nnet/inst/private/__printNumLayerDelays.m main/nnet/inst/private/__printNumLayers.m main/nnet/inst/private/__printNumOutputs.m main/nnet/inst/private/__printNumTargets.m main/nnet/inst/private/__printOutputConnect.m main/nnet/inst/private/__printOutputs.m main/nnet/inst/private/__printPerformFcn.m main/nnet/inst/private/__printPerformParam.m main/nnet/inst/private/__printTargetConnect.m main/nnet/inst/private/__printTargets.m main/nnet/inst/private/__printTrainFcn.m main/nnet/inst/private/__printTrainParam.m main/nnet/inst/private/__randomisecols.m main/nnet/inst/private/__rerangecolumns.m main/nnet/inst/private/__setx.m main/nnet/inst/private/__trainlm.m main/nnet/inst/purelin.m main/nnet/inst/radbas.m main/nnet/inst/satlin.m main/nnet/inst/satlins.m main/nnet/inst/saveMLPStruct.m main/nnet/inst/sim.m main/nnet/inst/subset.m main/nnet/inst/tansig.m main/nnet/inst/train.m main/nnet/inst/trastd.m main/nnet/inst/vec2ind.m main/nnet/tests/MLP/MLPScripts.m main/nnet/tests/MLP/example1/mData.txt main/nnet/tests/MLP/example1/mlp9_1_1_tansig.dat main/nnet/tests/MLP/example1/mlp9_1_1_tansig.dat_orig main/nnet/tests/MLP/example1/mlp9_1_1_tansig.m main/nnet/tests/MLP/example1/mlp9_2_1_tansig.dat main/nnet/tests/MLP/example1/mlp9_2_1_tansig.m main/nnet/tests/MLP/example1/mlp9_2_2_1_tansig.dat main/nnet/tests/MLP/example1/mlp9_2_2_1_tansig.m main/nnet/tests/MLP/example1/mlp9_2_2_tansig.dat main/nnet/tests/MLP/example1/mlp9_2_2_tansig.m main/nnet/tests/MLP/example1/mlp9_2_3_tansig.dat main/nnet/tests/MLP/example1/mlp9_2_3_tansig.m main/nnet/tests/MLP/example1/mlp9_5_3_tansig.dat main/nnet/tests/MLP/example1/mlp9_5_3_tansig.m main/nnet/tests/MLP/example1/orig/mData.txt main/nnet/tests/MLP/example2/mData.txt main/nnet/tests/MLP/example2/mlp9_1_1_logsig.dat main/nnet/tests/MLP/example2/mlp9_1_1_logsig.m main/nnet/tests/MLP/example2/mlp9_2_1_logsig.dat main/nnet/tests/MLP/example2/mlp9_2_1_logsig.m main/nnet/tests/MLP/example2/mlp9_2_2_1_logsig.m main/nnet/tests/MLP/example2/mlp9_2_2_3_logsig.dat main/nnet/tests/MLP/example2/mlp9_2_2_logsig.dat main/nnet/tests/MLP/example2/mlp9_2_2_logsig.m main/nnet/tests/MLP/example2/mlp9_2_3_logsig.dat main/nnet/tests/MLP/example2/mlp9_2_3_logsig.m main/nnet/tests/MLP/example2/mlp9_5_3_logsig.dat main/nnet/tests/MLP/example2/mlp9_5_3_logsig.m main/nnet/tests/MLP/example2/orig/mData.txt main/nnet/tests/MLP/loadtestresults.m main/nnet/tests/MLP/nnetTestMLP.m main/nnet/tests/MLP/preparedata9_x_1.m main/nnet/tests/MLP/preparedata9_x_2.m main/nnet/tests/MLP/preparedata9_x_3.m main/nnet/tests/MLP/preparedata9_x_x_1.m main/nnet/tests/MLP/testExample1_1.m main/nnet/tests/MLP/testExample1_2.m main/nnet/tests/nnetTest.m main/nnet/tests/readme main/nnet/tests/test_nnet_win32.pl
diffstat 228 files changed, 0 insertions(+), 15253 deletions(-) [+]
line wrap: on
line diff
--- a/main/nnet/COPYING	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,337 +0,0 @@
-                    GNU GENERAL PUBLIC LICENSE
-                       Version 2, June 1991
-
- Copyright (C) 1989, 1991 Free Software Foundation, Inc. <http://fsf.org/>
- Everyone is permitted to copy and distribute verbatim copies
- of this license document, but changing it is not allowed.
-
-                            Preamble
-
-  The licenses for most software are designed to take away your
-freedom to share and change it.  By contrast, the GNU General Public
-License is intended to guarantee your freedom to share and change free
-software--to make sure the software is free for all its users.  This
-General Public License applies to most of the Free Software
-Foundation's software and to any other program whose authors commit to
-using it.  (Some other Free Software Foundation software is covered by
-the GNU Library General Public License instead.)  You can apply it to
-your programs, too.
-
-  When we speak of free software, we are referring to freedom, not
-price.  Our General Public Licenses are designed to make sure that you
-have the freedom to distribute copies of free software (and charge for
-this service if you wish), that you receive source code or can get it
-if you want it, that you can change the software or use pieces of it
-in new free programs; and that you know you can do these things.
-
-  To protect your rights, we need to make restrictions that forbid
-anyone to deny you these rights or to ask you to surrender the rights.
-These restrictions translate to certain responsibilities for you if you
-distribute copies of the software, or if you modify it.
-
-  For example, if you distribute copies of such a program, whether
-gratis or for a fee, you must give the recipients all the rights that
-you have.  You must make sure that they, too, receive or can get the
-source code.  And you must show them these terms so they know their
-rights.
-
-  We protect your rights with two steps: (1) copyright the software, and
-(2) offer you this license which gives you legal permission to copy,
-distribute and/or modify the software.
-
-  Also, for each author's protection and ours, we want to make certain
-that everyone understands that there is no warranty for this free
-software.  If the software is modified by someone else and passed on, we
-want its recipients to know that what they have is not the original, so
-that any problems introduced by others will not reflect on the original
-authors' reputations.
-
-  Finally, any free program is threatened constantly by software
-patents.  We wish to avoid the danger that redistributors of a free
-program will individually obtain patent licenses, in effect making the
-program proprietary.  To prevent this, we have made it clear that any
-patent must be licensed for everyone's free use or not licensed at all.
-
-  The precise terms and conditions for copying, distribution and
-modification follow.
-
-                    GNU GENERAL PUBLIC LICENSE
-   TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION
-
-  0. This License applies to any program or other work which contains
-a notice placed by the copyright holder saying it may be distributed
-under the terms of this General Public License.  The "Program", below,
-refers to any such program or work, and a "work based on the Program"
-means either the Program or any derivative work under copyright law:
-that is to say, a work containing the Program or a portion of it,
-either verbatim or with modifications and/or translated into another
-language.  (Hereinafter, translation is included without limitation in
-the term "modification".)  Each licensee is addressed as "you".
-
-Activities other than copying, distribution and modification are not
-covered by this License; they are outside its scope.  The act of
-running the Program is not restricted, and the output from the Program
-is covered only if its contents constitute a work based on the
-Program (independent of having been made by running the Program).
-Whether that is true depends on what the Program does.
-
-  1. You may copy and distribute verbatim copies of the Program's
-source code as you receive it, in any medium, provided that you
-conspicuously and appropriately publish on each copy an appropriate
-copyright notice and disclaimer of warranty; keep intact all the
-notices that refer to this License and to the absence of any warranty;
-and give any other recipients of the Program a copy of this License
-along with the Program.
-
-You may charge a fee for the physical act of transferring a copy, and
-you may at your option offer warranty protection in exchange for a fee.
-
-  2. You may modify your copy or copies of the Program or any portion
-of it, thus forming a work based on the Program, and copy and
-distribute such modifications or work under the terms of Section 1
-above, provided that you also meet all of these conditions:
-
-    a) You must cause the modified files to carry prominent notices
-    stating that you changed the files and the date of any change.
-
-    b) You must cause any work that you distribute or publish, that in
-    whole or in part contains or is derived from the Program or any
-    part thereof, to be licensed as a whole at no charge to all third
-    parties under the terms of this License.
-
-    c) If the modified program normally reads commands interactively
-    when run, you must cause it, when started running for such
-    interactive use in the most ordinary way, to print or display an
-    announcement including an appropriate copyright notice and a
-    notice that there is no warranty (or else, saying that you provide
-    a warranty) and that users may redistribute the program under
-    these conditions, and telling the user how to view a copy of this
-    License.  (Exception: if the Program itself is interactive but
-    does not normally print such an announcement, your work based on
-    the Program is not required to print an announcement.)
-
-These requirements apply to the modified work as a whole.  If
-identifiable sections of that work are not derived from the Program,
-and can be reasonably considered independent and separate works in
-themselves, then this License, and its terms, do not apply to those
-sections when you distribute them as separate works.  But when you
-distribute the same sections as part of a whole which is a work based
-on the Program, the distribution of the whole must be on the terms of
-this License, whose permissions for other licensees extend to the
-entire whole, and thus to each and every part regardless of who wrote it.
-
-Thus, it is not the intent of this section to claim rights or contest
-your rights to work written entirely by you; rather, the intent is to
-exercise the right to control the distribution of derivative or
-collective works based on the Program.
-
-In addition, mere aggregation of another work not based on the Program
-with the Program (or with a work based on the Program) on a volume of
-a storage or distribution medium does not bring the other work under
-the scope of this License.
-
-  3. You may copy and distribute the Program (or a work based on it,
-under Section 2) in object code or executable form under the terms of
-Sections 1 and 2 above provided that you also do one of the following:
-
-    a) Accompany it with the complete corresponding machine-readable
-    source code, which must be distributed under the terms of Sections
-    1 and 2 above on a medium customarily used for software interchange; or,
-
-    b) Accompany it with a written offer, valid for at least three
-    years, to give any third party, for a charge no more than your
-    cost of physically performing source distribution, a complete
-    machine-readable copy of the corresponding source code, to be
-    distributed under the terms of Sections 1 and 2 above on a medium
-    customarily used for software interchange; or,
-
-    c) Accompany it with the information you received as to the offer
-    to distribute corresponding source code.  (This alternative is
-    allowed only for noncommercial distribution and only if you
-    received the program in object code or executable form with such
-    an offer, in accord with Subsection b above.)
-
-The source code for a work means the preferred form of the work for
-making modifications to it.  For an executable work, complete source
-code means all the source code for all modules it contains, plus any
-associated interface definition files, plus the scripts used to
-control compilation and installation of the executable.  However, as a
-special exception, the source code distributed need not include
-anything that is normally distributed (in either source or binary
-form) with the major components (compiler, kernel, and so on) of the
-operating system on which the executable runs, unless that component
-itself accompanies the executable.
-
-If distribution of executable or object code is made by offering
-access to copy from a designated place, then offering equivalent
-access to copy the source code from the same place counts as
-distribution of the source code, even though third parties are not
-compelled to copy the source along with the object code.
-
-  4. You may not copy, modify, sublicense, or distribute the Program
-except as expressly provided under this License.  Any attempt
-otherwise to copy, modify, sublicense or distribute the Program is
-void, and will automatically terminate your rights under this License.
-However, parties who have received copies, or rights, from you under
-this License will not have their licenses terminated so long as such
-parties remain in full compliance.
-
-  5. You are not required to accept this License, since you have not
-signed it.  However, nothing else grants you permission to modify or
-distribute the Program or its derivative works.  These actions are
-prohibited by law if you do not accept this License.  Therefore, by
-modifying or distributing the Program (or any work based on the
-Program), you indicate your acceptance of this License to do so, and
-all its terms and conditions for copying, distributing or modifying
-the Program or works based on it.
-
-  6. Each time you redistribute the Program (or any work based on the
-Program), the recipient automatically receives a license from the
-original licensor to copy, distribute or modify the Program subject to
-these terms and conditions.  You may not impose any further
-restrictions on the recipients' exercise of the rights granted herein.
-You are not responsible for enforcing compliance by third parties to
-this License.
-
-  7. If, as a consequence of a court judgment or allegation of patent
-infringement or for any other reason (not limited to patent issues),
-conditions are imposed on you (whether by court order, agreement or
-otherwise) that contradict the conditions of this License, they do not
-excuse you from the conditions of this License.  If you cannot
-distribute so as to satisfy simultaneously your obligations under this
-License and any other pertinent obligations, then as a consequence you
-may not distribute the Program at all.  For example, if a patent
-license would not permit royalty-free redistribution of the Program by
-all those who receive copies directly or indirectly through you, then
-the only way you could satisfy both it and this License would be to
-refrain entirely from distribution of the Program.
-
-If any portion of this section is held invalid or unenforceable under
-any particular circumstance, the balance of the section is intended to
-apply and the section as a whole is intended to apply in other
-circumstances.
-
-It is not the purpose of this section to induce you to infringe any
-patents or other property right claims or to contest validity of any
-such claims; this section has the sole purpose of protecting the
-integrity of the free software distribution system, which is
-implemented by public license practices.  Many people have made
-generous contributions to the wide range of software distributed
-through that system in reliance on consistent application of that
-system; it is up to the author/donor to decide if he or she is willing
-to distribute software through any other system and a licensee cannot
-impose that choice.
-
-This section is intended to make thoroughly clear what is believed to
-be a consequence of the rest of this License.
-
-  8. If the distribution and/or use of the Program is restricted in
-certain countries either by patents or by copyrighted interfaces, the
-original copyright holder who places the Program under this License
-may add an explicit geographical distribution limitation excluding
-those countries, so that distribution is permitted only in or among
-countries not thus excluded.  In such case, this License incorporates
-the limitation as if written in the body of this License.
-
-  9. The Free Software Foundation may publish revised and/or new versions
-of the General Public License from time to time.  Such new versions will
-be similar in spirit to the present version, but may differ in detail to
-address new problems or concerns.
-
-Each version is given a distinguishing version number.  If the Program
-specifies a version number of this License which applies to it and "any
-later version", you have the option of following the terms and conditions
-either of that version or of any later version published by the Free
-Software Foundation.  If the Program does not specify a version number of
-this License, you may choose any version ever published by the Free Software
-Foundation.
-
-  10. If you wish to incorporate parts of the Program into other free
-programs whose distribution conditions are different, write to the author
-to ask for permission.  For software which is copyrighted by the Free
-Software Foundation, write to the Free Software Foundation; we sometimes
-make exceptions for this.  Our decision will be guided by the two goals
-of preserving the free status of all derivatives of our free software and
-of promoting the sharing and reuse of software generally.
-
-                            NO WARRANTY
-
-  11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY
-FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW.  EXCEPT WHEN
-OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES
-PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED
-OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
-MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE.  THE ENTIRE RISK AS
-TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU.  SHOULD THE
-PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING,
-REPAIR OR CORRECTION.
-
-  12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING
-WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR
-REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES,
-INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING
-OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED
-TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY
-YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER
-PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE
-POSSIBILITY OF SUCH DAMAGES.
-
-                     END OF TERMS AND CONDITIONS
-
-            How to Apply These Terms to Your New Programs
-
-  If you develop a new program, and you want it to be of the greatest
-possible use to the public, the best way to achieve this is to make it
-free software which everyone can redistribute and change under these terms.
-
-  To do so, attach the following notices to the program.  It is safest
-to attach them to the start of each source file to most effectively
-convey the exclusion of warranty; and each file should have at least
-the "copyright" line and a pointer to where the full notice is found.
-
-    <one line to give the program's name and a brief idea of what it does.>
-    Copyright (C) <year>  <name of author>
-
-    This program is free software; you can redistribute it and/or modify
-    it under the terms of the GNU General Public License as published by
-    the Free Software Foundation; either version 2 of the License, or
-    (at your option) any later version.
-
-    This program is distributed in the hope that it will be useful,
-    but WITHOUT ANY WARRANTY; without even the implied warranty of
-    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
-    GNU General Public License for more details.
-
-    You should have received a copy of the GNU General Public License
-    along with this program; if not, see <http://www.gnu.org/licenses/>.
-
-Also add information on how to contact you by electronic and paper mail.
-
-If the program is interactive, make it output a short notice like this
-when it starts in an interactive mode:
-
-    Gnomovision version 69, Copyright (C) year name of author
-    Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'.
-    This is free software, and you are welcome to redistribute it
-    under certain conditions; type `show c' for details.
-
-The hypothetical commands `show w' and `show c' should show the appropriate
-parts of the General Public License.  Of course, the commands you use may
-be called something other than `show w' and `show c'; they could even be
-mouse-clicks or menu items--whatever suits your program.
-
-You should also get your employer (if you work as a programmer) or your
-school, if any, to sign a "copyright disclaimer" for the program, if
-necessary.  Here is a sample; alter the names:
-
-  Yoyodyne, Inc., hereby disclaims all copyright interest in the program
-  `Gnomovision' (which makes passes at compilers) written by James Hacker.
-
-  <signature of Ty Coon>, 1 April 1989
-  Ty Coon, President of Vice
-
-This General Public License does not permit incorporating your program into
-proprietary programs.  If your program is a subroutine library, you may
-consider it more useful to permit linking proprietary applications with the
-library.  If this is what you want to do, use the GNU Library General
-Public License instead of this License.
--- a/main/nnet/DESCRIPTION	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,12 +0,0 @@
-Name: nnet
-Version: 0.1.13
-Date: 2010-12-02
-Author: various authors
-Maintainer: Octave-Forge community <octave-dev@lists.sourceforge.net>
-Title: Neural Networks
-Description: A feed forward multi-layer neural network.
-Depends: octave (>= 3.0.0)
-Autoload: no
-License: GPLv3+
-Url: http://octave.sf.net
-Url: http://octnnettb.sourceforge.net
--- a/main/nnet/INDEX	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,31 +0,0 @@
-nnet >> Neural Networks
-
-Creation
- newff
-
-Data preprocessing & postprocessing
- mapstd
- prestd
- poststd
- trastd
-
-Simulation
- sim
-
-Training
- train
-
-Transfer functions
- logsig
- purelin
- radbas
- tansig
-
-Utility
- dividerand
- ind2vec
- isposint
- min_max
- saveMLPStruct
- subset
- vec2ind
\ No newline at end of file
--- a/main/nnet/Makefile	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,12 +0,0 @@
-sinclude ../../Makeconf
-
-PKG_FILES = COPYING DESCRIPTION INDEX $(wildcard inst/*) \
-	 doc/pdf/neuralNetworkPackageForOctaveUsersGuide.pdf \
-	$(wildcard doc/latex/user/*.tex) \
-	$(wildcard doc/latex/user/examples/*.tex) \
-	$(wildcard doc/latex/user/examples/1/*.tex) \
-	$(wildcard doc/latex/user/examples/1/*.txt) \
-	$(wildcard doc/latex/user/examples/1/*.m) \
-	$(wildcard doc/latex/user/octave/neuroPackage/*.tex) \
-	$(wildcard doc/latex/user/octave/neuroPackage/graphics/*.eps) \
-	$(wildcard doc/latex/user/octave/neuroPackage/graphics/*.pdf)
--- a/main/nnet/NEWS	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,36 +0,0 @@
-Summary of important user-visible changes for Neural Networks package 0.2.0:
-----------------------------------------------------------------------------
-
- ** The following functions are new:
-
-      minmax
-
- ** The following functions have been deprecated (see their help text
-    for the recommended alternatives):
-
-      min_max
-
- ** The new function `minmax' meant to replace `min_max' can now handle single
-    row matrices and cell arrays in a Matlab compatible way.
-
- ** The following functions have been made private:
-
-      __analyzerows.m       __printBiasConnect.m     __printNumLayerDelays.m
-      __calcjacobian.m      __printBiases.m          __printNumLayers.m
-      __calcperf.m          __printB.m               __printNumOutputs.m
-      __checknetstruct.m    __printInitFcn.m         __printNumTargets.m
-      __copycoltopos1.m     __printInitParam.m       __printOutputConnect.m
-      __dlogsig.m           __printInputConnect.m    __printOutputs.m
-      __dpurelin.m          __printInputs.m          __printPerformFcn.m
-      __dradbas.m           __printInputWeights.m    __printPerformParam.m
-      __dtansig.m           __printIW.m              __printTargetConnect.m
-      __getx.m              __printLayerConnect.m    __printTargets.m
-      __init.m              __printLayers.m          __printTrainFcn.m
-      __mae.m               __printLayerWeights.m    __printTrainParam.m
-      __mse.m               __printLW.m              __randomisecols.m
-      __newnetwork.m        __printMLPHeader.m       __rerangecolumns.m
-      __optimizedatasets.m  __printNetworkType.m     __setx.m
-      __printAdaptFcn.m     __printNumInputDelays.m  __trainlm.m
-      __printAdaptParam.m   __printNumInputs.m
-
- ** Package is no longer automatically loaded.
--- a/main/nnet/doc/docReadme	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,29 +0,0 @@
-The directory nnet/doc contains two subdirectories.
-1. latex: This includes five (5) subdirectories. All of them
-          are used to create the newest version of the documentation.
-          
-		  asymptote: Includes asy-scripts to create graphics, used in
-          subdirectory users/octave/neuroToolbox/graphics. For pdf output
-          the graphic files must be in pdf format. For dvi output, the
-          graphics must be in eps format. That's why I create each time
-          both types. Used asymptote is version: 1.42 win version
-          
-          common: Contains some .tex files which are used in both documentations.
-          developer as users guide.
-          
-          developers: Contains .tex files with a lot of informations about 
-          how I did the most parts of this package. Not all of them are
-          completly written ... :-(
-          
-          perl:  Contains two files. The first one is a "modul" to put somewhere
-          in a lib directory in the perl installation. This file ends with .pm.
-          The second file (createTestDocu.pl) is a short perl script to generate
-		  some .tex files for the developer guide which contains all the test
-		  cases programmed in the function files. In the file createTestDocu.pl
-		  must be changed some variables to the local system likewise it must be
-		  copied to the "developers" directory!
-
-          users: The user's guide. The official documentation to the neural network
-          package!
-
-2. pdf:  Contains documentation generated of the latex directory.
--- a/main/nnet/doc/latex/asymptote/mlp4-2-3.asy	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,122 +0,0 @@
-import graph;
-size(100,0);
-
-// define sizes for neurons
-pair center1 = (10,10);
-pair center2 = (10,30);
-pair center3 = (10,50);
-pair center4 = (10,70);
-pair center1hidden = (50,30);
-pair center2hidden = (50,50);
-pair center1out = (90,20);
-pair center2out = (90,40);
-pair center3out = (90,60);
-real radius  = (5);
-
-
-// define circle for neurons
-path in1 = circle(center1,radius);
-path in2 = circle(center2,radius);
-path in3 = circle(center3,radius);
-path in4 = circle(center4,radius);
-path hn1 = circle(center1hidden,radius);
-path hn2 = circle(center2hidden,radius);
-path on1 = circle(center1out,radius);
-path on2 = circle(center2out,radius);
-path on3 = circle(center3out,radius);
-
-// draw neurons
-draw(in1);
-draw(in2);
-draw(in3);
-draw(in4);
-draw(hn1);
-draw(hn2);
-draw(on1);
-draw(on2);
-draw(on3);
-
-// now draw arrows from input
-// to hidden neurons
-// (starts from bottom to top)
-
-// first input
-pair z11=(16,10);
-pair z12=(45,27);
-draw(z11--z12,Arrow);
-// first input
-pair z21=(16,30);
-pair z22=(44,29);
-draw(z21--z22,Arrow);
-// first input
-pair z31=(16,50);
-pair z32=(44,31);
-draw(z31--z32,Arrow);
- // first input
-pair z41=(16,70);
-pair z42=(45,33);
-draw(z41--z42,Arrow);
-
-// second input
-pair z11=(16,10);
-pair z12=(45,47);
-draw(z11--z12,Arrow);
-// second input
-pair z21=(16,30);
-pair z22=(44,49);
-draw(z21--z22,Arrow);
-// second input
-pair z31=(16,50);
-pair z32=(44,51);
-draw(z31--z32,Arrow);
-// second input
-pair z41=(16,70);
-pair z42=(45,53);
-draw(z41--z42,Arrow);
-
-// now draw arrows from hidden
-// to output neurons
-// (starts from bottom to top)
-
-// first hidden
-pair z11=(56,30);
-pair z12=(84,19);
-draw(z11--z12,Arrow);
-// first hidden
-pair z21=(56,30);
-pair z22=(84,39);
-draw(z21--z22,Arrow);
-// first hidden
-pair z21=(56,30);
-pair z22=(84,59);
-draw(z21--z22,Arrow);
-// second hidden
-pair z31=(56,50);
-pair z32=(84,21);
-draw(z31--z32,Arrow);
- // second hidden
-pair z41=(56,50);
-pair z42=(84,41);
-draw(z41--z42,Arrow);
- // second hidden
-pair z41=(56,50);
-pair z42=(84,61);
-draw(z41--z42,Arrow);
-
-// now define text for input neurons
-// p1 for the top neuron
-label(scale(0.5)*minipage("\centering \textbf{p1}",0.5cm),(0,70));
-label(scale(0.5)*minipage("\centering \textbf{p2}",0.5cm),(0,50));
-label(scale(0.5)*minipage("\centering \textbf{p3}",0.5cm),(0,30));
-label(scale(0.5)*minipage("\centering \textbf{p4}",0.5cm),(0,10));
-
-// now define text for hidden neurons
-// n1 for the top neuron
-label(scale(0.5)*minipage("\centering \textbf{n1}",0.5cm),(50,60));
-label(scale(0.5)*minipage("\centering \textbf{n2}",0.5cm),(50,40));
-
-// now define text for output neurons
-// a1 for the top neuron
-label(scale(0.5)*minipage("\centering \textbf{a1}",0.5cm),(100,60));
-label(scale(0.5)*minipage("\centering \textbf{a2}",0.5cm),(100,40));
-label(scale(0.5)*minipage("\centering \textbf{a3}",0.5cm),(100,20));
--- a/main/nnet/doc/latex/asymptote/mlp4-2-3.eps	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,680 +0,0 @@
-%!PS-Adobe-3.0 EPSF-3.0
-%%Creator: dvips(k) 5.94b Copyright 2004 Radical Eye Software
-%%Title: mlp4-2-3_.dvi
-%%CreationDate: Sat Jan 19 20:51:42 2008
-%%Pages: 1
-%%PageOrder: Ascend
-%%BoundingBox: 255 363 356 428
-%%HiResBoundingBox: 255.5 363.026812 355.5 427.973188
-%%DocumentFonts: CMBX12
-%%EndComments
-%DVIPSWebPage: (www.radicaleye.com)
-%DVIPSCommandLine: C:\texmf\miktex\bin\dvips.exe -R -O 127.1bp,238.227bp
-%+ -T 612bp,792bp -q -o mlp4-2-3_.ps mlp4-2-3_.dvi
-%DVIPSParameters: dpi=600
-%DVIPSSource:  TeX output 2008.01.19:2051
-%%BeginProcSet: tex.pro 0 0
-%!
-/TeXDict 300 dict def TeXDict begin/N{def}def/B{bind def}N/S{exch}N/X{S
-N}B/A{dup}B/TR{translate}N/isls false N/vsize 11 72 mul N/hsize 8.5 72
-mul N/landplus90{false}def/@rigin{isls{[0 landplus90{1 -1}{-1 1}ifelse 0
-0 0]concat}if 72 Resolution div 72 VResolution div neg scale isls{
-landplus90{VResolution 72 div vsize mul 0 exch}{Resolution -72 div hsize
-mul 0}ifelse TR}if Resolution VResolution vsize -72 div 1 add mul TR[
-matrix currentmatrix{A A round sub abs 0.00001 lt{round}if}forall round
-exch round exch]setmatrix}N/@landscape{/isls true N}B/@manualfeed{
-statusdict/manualfeed true put}B/@copies{/#copies X}B/FMat[1 0 0 -1 0 0]
-N/FBB[0 0 0 0]N/nn 0 N/IEn 0 N/ctr 0 N/df-tail{/nn 8 dict N nn begin
-/FontType 3 N/FontMatrix fntrx N/FontBBox FBB N string/base X array
-/BitMaps X/BuildChar{CharBuilder}N/Encoding IEn N end A{/foo setfont}2
-array copy cvx N load 0 nn put/ctr 0 N[}B/sf 0 N/df{/sf 1 N/fntrx FMat N
-df-tail}B/dfs{div/sf X/fntrx[sf 0 0 sf neg 0 0]N df-tail}B/E{pop nn A
-definefont setfont}B/Cw{Cd A length 5 sub get}B/Ch{Cd A length 4 sub get
-}B/Cx{128 Cd A length 3 sub get sub}B/Cy{Cd A length 2 sub get 127 sub}
-B/Cdx{Cd A length 1 sub get}B/Ci{Cd A type/stringtype ne{ctr get/ctr ctr
-1 add N}if}B/CharBuilder{save 3 1 roll S A/base get 2 index get S
-/BitMaps get S get/Cd X pop/ctr 0 N Cdx 0 Cx Cy Ch sub Cx Cw add Cy
-setcachedevice Cw Ch true[1 0 0 -1 -.1 Cx sub Cy .1 sub]{Ci}imagemask
-restore}B/D{/cc X A type/stringtype ne{]}if nn/base get cc ctr put nn
-/BitMaps get S ctr S sf 1 ne{A A length 1 sub A 2 index S get sf div put
-}if put/ctr ctr 1 add N}B/I{cc 1 add D}B/bop{userdict/bop-hook known{
-bop-hook}if/SI save N @rigin 0 0 moveto/V matrix currentmatrix A 1 get A
-mul exch 0 get A mul add .99 lt{/QV}{/RV}ifelse load def pop pop}N/eop{
-SI restore userdict/eop-hook known{eop-hook}if showpage}N/@start{
-userdict/start-hook known{start-hook}if pop/VResolution X/Resolution X
-1000 div/DVImag X/IEn 256 array N 2 string 0 1 255{IEn S A 360 add 36 4
-index cvrs cvn put}for pop 65781.76 div/vsize X 65781.76 div/hsize X}N
-/p{show}N/RMat[1 0 0 -1 0 0]N/BDot 260 string N/Rx 0 N/Ry 0 N/V{}B/RV/v{
-/Ry X/Rx X V}B statusdict begin/product where{pop false[(Display)(NeXT)
-(LaserWriter 16/600)]{A length product length le{A length product exch 0
-exch getinterval eq{pop true exit}if}{pop}ifelse}forall}{false}ifelse
-end{{gsave TR -.1 .1 TR 1 1 scale Rx Ry false RMat{BDot}imagemask
-grestore}}{{gsave TR -.1 .1 TR Rx Ry scale 1 1 false RMat{BDot}
-imagemask grestore}}ifelse B/QV{gsave newpath transform round exch round
-exch itransform moveto Rx 0 rlineto 0 Ry neg rlineto Rx neg 0 rlineto
-fill grestore}B/a{moveto}B/delta 0 N/tail{A/delta X 0 rmoveto}B/M{S p
-delta add tail}B/b{S p tail}B/c{-4 M}B/d{-3 M}B/e{-2 M}B/f{-1 M}B/g{0 M}
-B/h{1 M}B/i{2 M}B/j{3 M}B/k{4 M}B/w{0 rmoveto}B/l{p -4 w}B/m{p -3 w}B/n{
-p -2 w}B/o{p -1 w}B/q{p 1 w}B/r{p 2 w}B/s{p 3 w}B/t{p 4 w}B/x{0 S
-rmoveto}B/y{3 2 roll p a}B/bos{/SS save N}B/eos{SS restore}B end
-
-%%EndProcSet
-%%BeginProcSet: texps.pro 0 0
-%!
-TeXDict begin/rf{findfont dup length 1 add dict begin{1 index/FID ne 2
-index/UniqueID ne and{def}{pop pop}ifelse}forall[1 index 0 6 -1 roll
-exec 0 exch 5 -1 roll VResolution Resolution div mul neg 0 0]/Metrics
-exch def dict begin Encoding{exch dup type/integertype ne{pop pop 1 sub
-dup 0 le{pop}{[}ifelse}{FontMatrix 0 get div Metrics 0 get div def}
-ifelse}forall Metrics/Metrics currentdict end def[2 index currentdict
-end definefont 3 -1 roll makefont/setfont cvx]cvx def}def/ObliqueSlant{
-dup sin S cos div neg}B/SlantFont{4 index mul add}def/ExtendFont{3 -1
-roll mul exch}def/ReEncodeFont{CharStrings rcheck{/Encoding false def
-dup[exch{dup CharStrings exch known not{pop/.notdef/Encoding true def}
-if}forall Encoding{]exch pop}{cleartomark}ifelse}if/Encoding exch def}
-def end
-
-%%EndProcSet
-%%BeginProcSet: special.pro 0 0
-%!
-TeXDict begin/SDict 200 dict N SDict begin/@SpecialDefaults{/hs 612 N
-/vs 792 N/ho 0 N/vo 0 N/hsc 1 N/vsc 1 N/ang 0 N/CLIP 0 N/rwiSeen false N
-/rhiSeen false N/letter{}N/note{}N/a4{}N/legal{}N}B/@scaleunit 100 N
-/@hscale{@scaleunit div/hsc X}B/@vscale{@scaleunit div/vsc X}B/@hsize{
-/hs X/CLIP 1 N}B/@vsize{/vs X/CLIP 1 N}B/@clip{/CLIP 2 N}B/@hoffset{/ho
-X}B/@voffset{/vo X}B/@angle{/ang X}B/@rwi{10 div/rwi X/rwiSeen true N}B
-/@rhi{10 div/rhi X/rhiSeen true N}B/@llx{/llx X}B/@lly{/lly X}B/@urx{
-/urx X}B/@ury{/ury X}B/magscale true def end/@MacSetUp{userdict/md known
-{userdict/md get type/dicttype eq{userdict begin md length 10 add md
-maxlength ge{/md md dup length 20 add dict copy def}if end md begin
-/letter{}N/note{}N/legal{}N/od{txpose 1 0 mtx defaultmatrix dtransform S
-atan/pa X newpath clippath mark{transform{itransform moveto}}{transform{
-itransform lineto}}{6 -2 roll transform 6 -2 roll transform 6 -2 roll
-transform{itransform 6 2 roll itransform 6 2 roll itransform 6 2 roll
-curveto}}{{closepath}}pathforall newpath counttomark array astore/gc xdf
-pop ct 39 0 put 10 fz 0 fs 2 F/|______Courier fnt invertflag{PaintBlack}
-if}N/txpose{pxs pys scale ppr aload pop por{noflips{pop S neg S TR pop 1
--1 scale}if xflip yflip and{pop S neg S TR 180 rotate 1 -1 scale ppr 3
-get ppr 1 get neg sub neg ppr 2 get ppr 0 get neg sub neg TR}if xflip
-yflip not and{pop S neg S TR pop 180 rotate ppr 3 get ppr 1 get neg sub
-neg 0 TR}if yflip xflip not and{ppr 1 get neg ppr 0 get neg TR}if}{
-noflips{TR pop pop 270 rotate 1 -1 scale}if xflip yflip and{TR pop pop
-90 rotate 1 -1 scale ppr 3 get ppr 1 get neg sub neg ppr 2 get ppr 0 get
-neg sub neg TR}if xflip yflip not and{TR pop pop 90 rotate ppr 3 get ppr
-1 get neg sub neg 0 TR}if yflip xflip not and{TR pop pop 270 rotate ppr
-2 get ppr 0 get neg sub neg 0 S TR}if}ifelse scaleby96{ppr aload pop 4
--1 roll add 2 div 3 1 roll add 2 div 2 copy TR .96 dup scale neg S neg S
-TR}if}N/cp{pop pop showpage pm restore}N end}if}if}N/normalscale{
-Resolution 72 div VResolution 72 div neg scale magscale{DVImag dup scale
-}if 0 setgray}N/psfts{S 65781.76 div N}N/startTexFig{/psf$SavedState
-save N userdict maxlength dict begin/magscale true def normalscale
-currentpoint TR/psf$ury psfts/psf$urx psfts/psf$lly psfts/psf$llx psfts
-/psf$y psfts/psf$x psfts currentpoint/psf$cy X/psf$cx X/psf$sx psf$x
-psf$urx psf$llx sub div N/psf$sy psf$y psf$ury psf$lly sub div N psf$sx
-psf$sy scale psf$cx psf$sx div psf$llx sub psf$cy psf$sy div psf$ury sub
-TR/showpage{}N/erasepage{}N/copypage{}N/p 3 def @MacSetUp}N/doclip{
-psf$llx psf$lly psf$urx psf$ury currentpoint 6 2 roll newpath 4 copy 4 2
-roll moveto 6 -1 roll S lineto S lineto S lineto closepath clip newpath
-moveto}N/endTexFig{end psf$SavedState restore}N/@beginspecial{SDict
-begin/SpecialSave save N gsave normalscale currentpoint TR
-@SpecialDefaults count/ocount X/dcount countdictstack N}N/@setspecial{
-CLIP 1 eq{newpath 0 0 moveto hs 0 rlineto 0 vs rlineto hs neg 0 rlineto
-closepath clip}if ho vo TR hsc vsc scale ang rotate rwiSeen{rwi urx llx
-sub div rhiSeen{rhi ury lly sub div}{dup}ifelse scale llx neg lly neg TR
-}{rhiSeen{rhi ury lly sub div dup scale llx neg lly neg TR}if}ifelse
-CLIP 2 eq{newpath llx lly moveto urx lly lineto urx ury lineto llx ury
-lineto closepath clip}if/showpage{}N/erasepage{}N/copypage{}N newpath}N
-/@endspecial{count ocount sub{pop}repeat countdictstack dcount sub{end}
-repeat grestore SpecialSave restore end}N/@defspecial{SDict begin}N
-/@fedspecial{end}B/li{lineto}B/rl{rlineto}B/rc{rcurveto}B/np{/SaveX
-currentpoint/SaveY X N 1 setlinecap newpath}N/st{stroke SaveX SaveY
-moveto}N/fil{fill SaveX SaveY moveto}N/ellipse{/endangle X/startangle X
-/yrad X/xrad X/savematrix matrix currentmatrix N TR xrad yrad scale 0 0
-1 startangle endangle arc savematrix setmatrix}N end
-
-%%EndProcSet
-%%BeginProcSet: color.pro 0 0
-%!
-TeXDict begin/setcmykcolor where{pop}{/setcmykcolor{dup 10 eq{pop
-setrgbcolor}{1 sub 4 1 roll 3{3 index add neg dup 0 lt{pop 0}if 3 1 roll
-}repeat setrgbcolor pop}ifelse}B}ifelse/TeXcolorcmyk{setcmykcolor}def
-/TeXcolorrgb{setrgbcolor}def/TeXcolorgrey{setgray}def/TeXcolorgray{
-setgray}def/TeXcolorhsb{sethsbcolor}def/currentcmykcolor where{pop}{
-/currentcmykcolor{currentrgbcolor 10}B}ifelse/DC{exch dup userdict exch
-known{pop pop}{X}ifelse}B/GreenYellow{0.15 0 0.69 0 setcmykcolor}DC
-/Yellow{0 0 1 0 setcmykcolor}DC/Goldenrod{0 0.10 0.84 0 setcmykcolor}DC
-/Dandelion{0 0.29 0.84 0 setcmykcolor}DC/Apricot{0 0.32 0.52 0
-setcmykcolor}DC/Peach{0 0.50 0.70 0 setcmykcolor}DC/Melon{0 0.46 0.50 0
-setcmykcolor}DC/YellowOrange{0 0.42 1 0 setcmykcolor}DC/Orange{0 0.61
-0.87 0 setcmykcolor}DC/BurntOrange{0 0.51 1 0 setcmykcolor}DC
-/Bittersweet{0 0.75 1 0.24 setcmykcolor}DC/RedOrange{0 0.77 0.87 0
-setcmykcolor}DC/Mahogany{0 0.85 0.87 0.35 setcmykcolor}DC/Maroon{0 0.87
-0.68 0.32 setcmykcolor}DC/BrickRed{0 0.89 0.94 0.28 setcmykcolor}DC/Red{
-0 1 1 0 setcmykcolor}DC/OrangeRed{0 1 0.50 0 setcmykcolor}DC/RubineRed{
-0 1 0.13 0 setcmykcolor}DC/WildStrawberry{0 0.96 0.39 0 setcmykcolor}DC
-/Salmon{0 0.53 0.38 0 setcmykcolor}DC/CarnationPink{0 0.63 0 0
-setcmykcolor}DC/Magenta{0 1 0 0 setcmykcolor}DC/VioletRed{0 0.81 0 0
-setcmykcolor}DC/Rhodamine{0 0.82 0 0 setcmykcolor}DC/Mulberry{0.34 0.90
-0 0.02 setcmykcolor}DC/RedViolet{0.07 0.90 0 0.34 setcmykcolor}DC
-/Fuchsia{0.47 0.91 0 0.08 setcmykcolor}DC/Lavender{0 0.48 0 0
-setcmykcolor}DC/Thistle{0.12 0.59 0 0 setcmykcolor}DC/Orchid{0.32 0.64 0
-0 setcmykcolor}DC/DarkOrchid{0.40 0.80 0.20 0 setcmykcolor}DC/Purple{
-0.45 0.86 0 0 setcmykcolor}DC/Plum{0.50 1 0 0 setcmykcolor}DC/Violet{
-0.79 0.88 0 0 setcmykcolor}DC/RoyalPurple{0.75 0.90 0 0 setcmykcolor}DC
-/BlueViolet{0.86 0.91 0 0.04 setcmykcolor}DC/Periwinkle{0.57 0.55 0 0
-setcmykcolor}DC/CadetBlue{0.62 0.57 0.23 0 setcmykcolor}DC
-/CornflowerBlue{0.65 0.13 0 0 setcmykcolor}DC/MidnightBlue{0.98 0.13 0
-0.43 setcmykcolor}DC/NavyBlue{0.94 0.54 0 0 setcmykcolor}DC/RoyalBlue{1
-0.50 0 0 setcmykcolor}DC/Blue{1 1 0 0 setcmykcolor}DC/Cerulean{0.94 0.11
-0 0 setcmykcolor}DC/Cyan{1 0 0 0 setcmykcolor}DC/ProcessBlue{0.96 0 0 0
-setcmykcolor}DC/SkyBlue{0.62 0 0.12 0 setcmykcolor}DC/Turquoise{0.85 0
-0.20 0 setcmykcolor}DC/TealBlue{0.86 0 0.34 0.02 setcmykcolor}DC
-/Aquamarine{0.82 0 0.30 0 setcmykcolor}DC/BlueGreen{0.85 0 0.33 0
-setcmykcolor}DC/Emerald{1 0 0.50 0 setcmykcolor}DC/JungleGreen{0.99 0
-0.52 0 setcmykcolor}DC/SeaGreen{0.69 0 0.50 0 setcmykcolor}DC/Green{1 0
-1 0 setcmykcolor}DC/ForestGreen{0.91 0 0.88 0.12 setcmykcolor}DC
-/PineGreen{0.92 0 0.59 0.25 setcmykcolor}DC/LimeGreen{0.50 0 1 0
-setcmykcolor}DC/YellowGreen{0.44 0 0.74 0 setcmykcolor}DC/SpringGreen{
-0.26 0 0.76 0 setcmykcolor}DC/OliveGreen{0.64 0 0.95 0.40 setcmykcolor}
-DC/RawSienna{0 0.72 1 0.45 setcmykcolor}DC/Sepia{0 0.83 1 0.70
-setcmykcolor}DC/Brown{0 0.81 1 0.60 setcmykcolor}DC/Tan{0.14 0.42 0.56 0
-setcmykcolor}DC/Gray{0 0 0 0.50 setcmykcolor}DC/Black{0 0 0 1
-setcmykcolor}DC/White{0 0 0 0 setcmykcolor}DC end
-
-%%EndProcSet
-%%BeginFont: CMBX12
-%!PS-AdobeFont-1.1: CMBX12 1.0
-%%CreationDate: 1991 Aug 20 16:34:54
-% Copyright (C) 1997 American Mathematical Society. All Rights Reserved.
-11 dict begin
-/FontInfo 7 dict dup begin
-/version (1.0) readonly def
-/Notice (Copyright (C) 1997 American Mathematical Society. All Rights Reserved) readonly def
-/FullName (CMBX12) readonly def
-/FamilyName (Computer Modern) readonly def
-/Weight (Bold) readonly def
-/ItalicAngle 0 def
-/isFixedPitch false def
-end readonly def
-/FontName /CMBX12 def
-/PaintType 0 def
-/FontType 1 def
-/FontMatrix [0.001 0 0 0.001 0 0] readonly def
-/Encoding 256 array
-0 1 255 {1 index exch /.notdef put} for
-dup 49 /one put
-dup 50 /two put
-dup 51 /three put
-dup 52 /four put
-dup 97 /a put
-dup 110 /n put
-dup 112 /p put
-readonly def
-/FontBBox{-53 -251 1139 750}readonly def
-/UniqueID 5000769 def
-currentdict end
-currentfile eexec
-D9D66F633B846A97B686A97E45A3D0AA052A014267B7904EB3C0D3BD0B83D891
-016CA6CA4B712ADEB258FAAB9A130EE605E61F77FC1B738ABC7C51CD46EF8171
-9098D5FEE67660E69A7AB91B58F29A4D79E57022F783EB0FBBB6D4F4EC35014F
-D2DECBA99459A4C59DF0C6EBA150284454E707DC2100C15B76B4C19B84363758
-469A6C558785B226332152109871A9883487DD7710949204DDCF837E6A8708B8
-2BDBF16FBC7512FAA308A093FE5F0364CD5660F74BEE96790DE35AFA90CCF712
-B1805DA88AE375A04D99598EADFC625BDC1F9C315B6CF28C9BD427F32C745C99
-AEBE70DAAED49EA45AF94F081934AA47894A370D698ABABDA4215500B190AF26
-7FCFB7DDA2BC68605A4EF61ECCA3D61C684B47FFB5887A3BEDE0B4D30E8EBABF
-20980C23312618EB0EAF289B2924FF4A334B85D98FD68545FDADB47F991E7390
-B10EE86A46A5AF8866C010225024D5E5862D49DEB5D8ECCB95D94283C50A363D
-68A49071445610F03CE3600945118A6BC0B3AA4593104E727261C68C4A47F809
-D77E4CF27B3681F6B6F3AC498E45361BF9E01FAF5527F5E3CC790D3084674B3E
-26296F3E03321B5C555D2458578A89E72D3166A3C5D740B3ABB127CF420C316D
-F957873DA04CF0DB25A73574A4DE2E4F2D5D4E8E0B430654CF7F341A1BDB3E26
-77C194764EAD58C585F49EF10843FE020F9FDFD9008D660DE50B9BD7A2A87299
-BC319E66D781101BB956E30643A19B93C8967E1AE4719F300BFE5866F0D6DA5E
-C55E171A24D3B707EFA325D47F473764E99BC8B1108D815CF2ACADFA6C4663E8
-30855D673CE98AB78F5F829F7FA226AB57F07B3E7D4E7CE30ED3B7EB0D3035C5
-148DA8D9FA34483414FDA8E3DC9E6C479E3EEE9A11A0547FC9085FA4631AD19C
-E936E0598E3197207FA7BB6E55CFD5EF72AEC12D9A9675241C7B00AD58FAF645
-1297991B5D01701E82228D0313FC7C66B263BC79ACDDF9AAC48A3CBF42B96E38
-583E1D059953076D68148DC8B6C9527B3A74CE7DEF788A11531F44120BDF0F61
-0B2F3ED94EEBCDE4ACD23834C242AA4314B9EF98E4BE72DB76EBDD0A028CEA9D
-B4C38C1F2D24B8FDE686832FE96204552C820E45B6BAF0C3308742AE22A8B448
-2D923E77A8BCF79945CD24B36367C61F0DF4B01008B258E19424D46B6D516A71
-DA6966E80712ADC4A6F06E4607C60763DDC43F39234F38C9ED557F31EF61A341
-91815E644BD8FA77014C082A9F0AA325C47E4F8466C6819311DBBAA950970CF2
-67CC5D08C7FFE0FC58863AF00BF5CFC6EAC3115EC60485AA725A9783ED3F0C71
-0D58973A47C0E7C6B6F94F288FD740F1E0F4E350B213CDA8E928C247BFDFB0C7
-82448FC280690112544D5D485D641DC6DDF77F4F37ED89D07FCFE92BCF44D0D5
-74AE3DA21112AADB239859682E9950323CEF2FADD468C35C41C2732FD6A6E91D
-42C63B8E10DCC65672BF00AB29E55D92CD246605EC87EE4C8BADC29D4A871411
-8F64BA1FBC10779980D52A2E88D4DD6883379B8ABE1C1488F6BC578D0108B7F3
-E8BD95D7B0AAF181F8E88EC24A737A4832B64CE4C0BE1E9FA5C9C4D6463A4A2F
-60927048AE21DA3E653D6EB01EB43F753E3462E8E34F7AB3AA8FD8617F43E9DA
-F851480AEDB2F034ACA32912795BA667D1FF4E1A09BAE04BE4C6B0A3E9C9E6E3
-A772E0E6403156903D580488CE1D79FD5F690194E91D42D9764D2DC30E699A90
-C148CB31925E4F9FBBC202AAFE17E5603A410E3CF824472E33FE64B91F619EAD
-B30997DD82F18F15811453526EF129BB49067DB80C8A9D76664F6A7920048D6C
-C287397E9BD4D4854BDBB03361321F38DD14E5431FDE3BDBF949F4F1B48D1FCA
-6B90CEBD0252C9DFB1C49A97CC5B0F8512105F42149455AA8229D03466D44772
-D376CE8B8C3334B7B3FA5F96A980716C72347B1F6B2E34D11A1E9226191C7189
-18B18BD7D04F320F15FB391DD8D3A157F76C72A81039172A71F279FA16F28485
-67801AEDA94971B863FD418047D81C5A0EA93EF9CBC7F41BA80A388B70ED925B
-3818C5F40078AD9AE5C39A03AA78B88571D8F25D2FF13827910451BB259582DB
-ED85042F073FE37C43AB6DA0C446E115496522BEDF4C6A761F9B3B203F61E3FD
-7B77561FADD8332FAD720BE7D1E47759365A3388B61750069628A87C4010895C
-E6F89A26612F86D90B8F22F37F53D24A7B2BB302E8CB4974C030D83BFAF15591
-B25D3333733AF4A16169D66226DB97CDAE6A705582ED939C7A6DBFF28BA666DC
-8D80F7F86633652BFB632F03950067621D8D885E9E41075FDC8A08A9D65B6698
-AB6A1B3FDF888F48B5B522A67B45C6A34E0281A13A8B997AB221A3FB665D2305
-2FC8D77C5BF52FEC4305A7EF84773F6BC5987E4B0ADAE80B03AAC1BF0A75EC4C
-11B801FDF9B7DF7E6272067DEC8DCB1390C42E7D3E49CA6259FDC3E6E9E30413
-CE1319B9704C3B6441845AA1409EF3B2D672558671B389457DCB5494AF59C9F4
-5A8CBBDF3023FC1469DF361A85F32A4D45AD7529096DF4EB5C6513AA34E1C82C
-597C6892813242DAC1A1AF9808101FC74228FD7DFBEC7825D584EBB129CDA5BA
-E210318F3FF5385F22A9CA4A75798BB8BDC1C3C9128995BA3797281804A3445C
-F5BDB9BB30D9ABD188BF0FF5A1180DFCBB8B94DA148125FC658690A0EFE2B600
-D584F63FEE2B7D091256B93AEFDD26968FD7BB304443F7EFBA2F9B988F5226E6
-10EAB9625AC209B734CFC4AB3CA26AA230CEA7415A9FD1C53C4830072CC955EB
-F00BDBC377004026C74671381DC9F61F57
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-cleartomark
-%%EndFont 
-TeXDict begin 40258584 52099344 1000 600 600 (mlp4-2-3_.dvi)
-@start /Fa 143[62 1[62 12[54 44[56 56 56 56 49[{}7 99.6264
-/CMBX12 rf end
-%%EndProlog
-%%BeginSetup
-%%Feature: *Resolution 600dpi
-TeXDict begin
- end
-%%EndSetup
-%%Page: 1 1
-TeXDict begin 1 0 bop Black Black 1529 2976 a @beginspecial
--3.966874 @llx 4.353313 @lly 96.033127 @urx 69.299690
-@ury 1000 @rwi @setspecial
-%%BeginDocument: mlp4-2-3_0.eps
-%!PS-Adobe-3.0 EPSF-3.0
-%%BoundingBox: -4 4 97 70
-%%HiResBoundingBox: -3.96687422 4.35331258 96.0331258 69.2996887
-%%Creator: Asymptote 1.25
-%%CreationDate: 2008.01.19 20:51:41
-%%Pages: 1
-%%EndProlog
-%%Page: 1 1
-0 setgray
- 0 0.5 dtransform truncate idtransform setlinewidth pop
-1 setlinecap
-1 setlinejoin
-newpath 13.8099377 9.20662516 moveto
- 13.8099377 11.7489645 11.7489645 13.8099377 9.20662516 13.8099377 curveto
- 6.66428582 13.8099377 4.60331258 11.7489645 4.60331258 9.20662516 curveto
- 4.60331258 6.66428582 6.66428582 4.60331258 9.20662516 4.60331258 curveto
- 11.7489645 4.60331258 13.8099377 6.66428582 13.8099377 9.20662516 curveto
-closepath
-stroke
-newpath 13.8099377 27.6198755 moveto
- 13.8099377 30.1622148 11.7489645 32.223188 9.20662516 32.223188 curveto
- 6.66428582 32.223188 4.60331258 30.1622148 4.60331258 27.6198755 curveto
- 4.60331258 25.0775361 6.66428582 23.0165629 9.20662516 23.0165629 curveto
- 11.7489645 23.0165629 13.8099377 25.0775361 13.8099377 27.6198755 curveto
-closepath
-stroke
-newpath 13.8099377 46.0331258 moveto
- 13.8099377 48.5754651 11.7489645 50.6364384 9.20662516 50.6364384 curveto
- 6.66428582 50.6364384 4.60331258 48.5754651 4.60331258 46.0331258 curveto
- 4.60331258 43.4907864 6.66428582 41.4298132 9.20662516 41.4298132 curveto
- 11.7489645 41.4298132 13.8099377 43.4907864 13.8099377 46.0331258 curveto
-closepath
-stroke
-newpath 13.8099377 64.4463761 moveto
- 13.8099377 66.9887154 11.7489645 69.0496887 9.20662516 69.0496887 curveto
- 6.66428582 69.0496887 4.60331258 66.9887154 4.60331258 64.4463761 curveto
- 4.60331258 61.9040368 6.66428582 59.8430635 9.20662516 59.8430635 curveto
- 11.7489645 59.8430635 13.8099377 61.9040368 13.8099377 64.4463761 curveto
-closepath
-stroke
-newpath 50.6364384 27.6198755 moveto
- 50.6364384 30.1622148 48.5754651 32.223188 46.0331258 32.223188 curveto
- 43.4907864 32.223188 41.4298132 30.1622148 41.4298132 27.6198755 curveto
- 41.4298132 25.0775361 43.4907864 23.0165629 46.0331258 23.0165629 curveto
- 48.5754651 23.0165629 50.6364384 25.0775361 50.6364384 27.6198755 curveto
-closepath
-stroke
-newpath 50.6364384 46.0331258 moveto
- 50.6364384 48.5754651 48.5754651 50.6364384 46.0331258 50.6364384 curveto
- 43.4907864 50.6364384 41.4298132 48.5754651 41.4298132 46.0331258 curveto
- 41.4298132 43.4907864 43.4907864 41.4298132 46.0331258 41.4298132 curveto
- 48.5754651 41.4298132 50.6364384 43.4907864 50.6364384 46.0331258 curveto
-closepath
-stroke
-newpath 87.462939 18.4132503 moveto
- 87.462939 20.9555896 85.4019657 23.0165629 82.8596264 23.0165629 curveto
- 80.3172871 23.0165629 78.2563138 20.9555896 78.2563138 18.4132503 curveto
- 78.2563138 15.870911 80.3172871 13.8099377 82.8596264 13.8099377 curveto
- 85.4019657 13.8099377 87.462939 15.870911 87.462939 18.4132503 curveto
-closepath
-stroke
-newpath 87.462939 36.8265006 moveto
- 87.462939 39.36884 85.4019657 41.4298132 82.8596264 41.4298132 curveto
- 80.3172871 41.4298132 78.2563138 39.36884 78.2563138 36.8265006 curveto
- 78.2563138 34.2841613 80.3172871 32.223188 82.8596264 32.223188 curveto
- 85.4019657 32.223188 87.462939 34.2841613 87.462939 36.8265006 curveto
-closepath
-stroke
-newpath 87.462939 55.2397509 moveto
- 87.462939 57.7820903 85.4019657 59.8430635 82.8596264 59.8430635 curveto
- 80.3172871 59.8430635 78.2563138 57.7820903 78.2563138 55.2397509 curveto
- 78.2563138 52.6974116 80.3172871 50.6364384 82.8596264 50.6364384 curveto
- 85.4019657 50.6364384 87.462939 52.6974116 87.462939 55.2397509 curveto
-closepath
-stroke
-newpath 34.9595784 21.0649917 moveto
- 28.2165857 17.1122028 21.473593 13.159414 14.7306002 9.20662516 curveto
-stroke
-newpath 41.4298132 24.8578879 moveto
- 33.9432749 22.7986859 lineto
- 35.9758819 19.3312975 lineto
- 37.793859 21.1734943 39.6118361 23.0156911 41.4298132 24.8578879 curveto
- 41.4298132 24.8578879 41.4298132 24.8578879 41.4298132 24.8578879 curveto
-closepath
-fill
-newpath 41.4298132 24.8578879 moveto
- 33.9432749 22.7986859 lineto
- 35.9758819 19.3312975 lineto
- 37.793859 21.1734943 39.6118361 23.0156911 41.4298132 24.8578879 curveto
- 41.4298132 24.8578879 41.4298132 24.8578879 41.4298132 24.8578879 curveto
-closepath
-stroke
-newpath 33.0139293 26.9668994 moveto
- 26.9194863 27.1845581 20.8250433 27.4022168 14.7306002 27.6198755 curveto
-stroke
-newpath 40.5091507 26.699213 moveto
- 33.0856557 28.975238 lineto
- 32.9422029 24.9585609 lineto
- 35.4645188 25.5387783 37.9868348 26.1189956 40.5091507 26.699213 curveto
- 40.5091507 26.699213 40.5091507 26.699213 40.5091507 26.699213 curveto
-closepath
-fill
-newpath 40.5091507 26.699213 moveto
- 33.0856557 28.975238 lineto
- 32.9422029 24.9585609 lineto
- 35.4645188 25.5387783 37.9868348 26.1189956 40.5091507 26.699213 curveto
- 40.5091507 26.699213 40.5091507 26.699213 40.5091507 26.699213 curveto
-closepath
-stroke
-newpath 34.3030824 32.7517986 moveto
- 27.7789217 37.1789077 21.254761 41.6060167 14.7306002 46.0331258 curveto
-stroke
-newpath 40.5091507 28.540538 moveto
- 35.4314863 34.4147096 lineto
- 33.1746785 31.0888876 lineto
- 35.6195026 30.2394377 38.0643266 29.3899879 40.5091507 28.540538 curveto
- 40.5091507 28.540538 40.5091507 28.540538 40.5091507 28.540538 curveto
-closepath
-fill
-newpath 40.5091507 28.540538 moveto
- 35.4314863 34.4147096 lineto
- 33.1746785 31.0888876 lineto
- 35.6195026 30.2394377 38.0643266 29.3899879 40.5091507 28.540538 curveto
- 40.5091507 28.540538 40.5091507 28.540538 40.5091507 28.540538 curveto
-closepath
-stroke
-newpath 36.8032007 36.2847824 moveto
- 29.4456672 45.6719803 22.0881337 55.0591782 14.7306002 64.4463761 curveto
-stroke
-newpath 41.4298132 30.381863 moveto
- 38.3848832 37.5244795 lineto
- 35.2215182 35.0450853 lineto
- 37.2909499 33.4906779 39.3603816 31.9362704 41.4298132 30.381863 curveto
- 41.4298132 30.381863 41.4298132 30.381863 41.4298132 30.381863 curveto
-closepath
-fill
-newpath 41.4298132 30.381863 moveto
- 38.3848832 37.5244795 lineto
- 35.2215182 35.0450853 lineto
- 37.2909499 33.4906779 39.3603816 31.9362704 41.4298132 30.381863 curveto
- 41.4298132 30.381863 41.4298132 30.381863 41.4298132 30.381863 curveto
-closepath
-stroke
-newpath 36.8032007 37.3682189 moveto
- 29.4456672 27.981021 22.0881337 18.5938231 14.7306002 9.20662516 curveto
-stroke
-newpath 41.4298132 43.2711382 moveto
- 35.2215182 38.6079159 lineto
- 38.3848832 36.1285218 lineto
- 39.3998599 38.5093939 40.4148365 40.8902661 41.4298132 43.2711382 curveto
- 41.4298132 43.2711382 41.4298132 43.2711382 41.4298132 43.2711382 curveto
-closepath
-fill
-newpath 41.4298132 43.2711382 moveto
- 35.2215182 38.6079159 lineto
- 38.3848832 36.1285218 lineto
- 39.3998599 38.5093939 40.4148365 40.8902661 41.4298132 43.2711382 curveto
- 41.4298132 43.2711382 41.4298132 43.2711382 41.4298132 43.2711382 curveto
-closepath
-stroke
-newpath 34.3030824 40.9012026 moveto
- 27.7789217 36.4740936 21.254761 32.0469845 14.7306002 27.6198755 curveto
-stroke
-newpath 40.5091507 45.1124633 moveto
- 33.1746785 42.5641136 lineto
- 35.4314863 39.2382916 lineto
- 37.1240411 41.1963488 38.8165959 43.1544061 40.5091507 45.1124633 curveto
- 40.5091507 45.1124633 40.5091507 45.1124633 40.5091507 45.1124633 curveto
-closepath
-fill
-newpath 40.5091507 45.1124633 moveto
- 33.1746785 42.5641136 lineto
- 35.4314863 39.2382916 lineto
- 37.1240411 41.1963488 38.8165959 43.1544061 40.5091507 45.1124633 curveto
- 40.5091507 45.1124633 40.5091507 45.1124633 40.5091507 45.1124633 curveto
-closepath
-stroke
-newpath 33.0139293 46.6861018 moveto
- 26.9194863 46.4684431 20.8250433 46.2507845 14.7306002 46.0331258 curveto
-stroke
-newpath 40.5091507 46.9537883 moveto
- 32.9422029 48.6944403 lineto
- 33.0856557 44.6777633 lineto
- 35.560154 45.4364383 38.0346523 46.1951133 40.5091507 46.9537883 curveto
- 40.5091507 46.9537883 40.5091507 46.9537883 40.5091507 46.9537883 curveto
-closepath
-fill
-newpath 40.5091507 46.9537883 moveto
- 32.9422029 48.6944403 lineto
- 33.0856557 44.6777633 lineto
- 35.560154 45.4364383 38.0346523 46.1951133 40.5091507 46.9537883 curveto
- 40.5091507 46.9537883 40.5091507 46.9537883 40.5091507 46.9537883 curveto
-closepath
-stroke
-newpath 34.9595784 52.5880096 moveto
- 28.2165857 56.5407984 21.473593 60.4935873 14.7306002 64.4463761 curveto
-stroke
-newpath 41.4298132 48.7951133 moveto
- 35.9758819 54.3217038 lineto
- 33.9432749 50.8543154 lineto
- 36.4387877 50.1679147 38.9343004 49.481514 41.4298132 48.7951133 curveto
- 41.4298132 48.7951133 41.4298132 48.7951133 41.4298132 48.7951133 curveto
-closepath
-fill
-newpath 41.4298132 48.7951133 moveto
- 35.9758819 54.3217038 lineto
- 33.9432749 50.8543154 lineto
- 36.4387877 50.1679147 38.9343004 49.481514 41.4298132 48.7951133 curveto
- 41.4298132 48.7951133 41.4298132 48.7951133 41.4298132 48.7951133 curveto
-closepath
-stroke
-newpath 70.3550151 20.2349806 moveto
- 64.0890437 22.6966122 57.8230723 25.1582438 51.5571009 27.6198755 curveto
-stroke
-newpath 77.3356513 17.4925878 moveto
- 71.089837 22.1054364 lineto
- 69.6201932 18.3645248 lineto
- 72.1920126 18.0738791 74.7638319 17.7832334 77.3356513 17.4925878 curveto
- 77.3356513 17.4925878 77.3356513 17.4925878 77.3356513 17.4925878 curveto
-closepath
-fill
-newpath 77.3356513 17.4925878 moveto
- 71.089837 22.1054364 lineto
- 69.6201932 18.3645248 lineto
- 72.1920126 18.0738791 74.7638319 17.7832334 77.3356513 17.4925878 curveto
- 77.3356513 17.4925878 77.3356513 17.4925878 77.3356513 17.4925878 curveto
-closepath
-stroke
-newpath 70.1954371 33.6107693 moveto
- 63.9826584 31.6138047 57.7698796 29.6168401 51.5571009 27.6198755 curveto
-stroke
-newpath 77.3356513 35.9058381 moveto
- 69.5804753 35.5239839 lineto
- 70.8103989 31.6975546 lineto
- 72.9854831 33.1003158 75.1605672 34.5030769 77.3356513 35.9058381 curveto
- 77.3356513 35.9058381 77.3356513 35.9058381 77.3356513 35.9058381 curveto
-closepath
-fill
-newpath 77.3356513 35.9058381 moveto
- 69.5804753 35.5239839 lineto
- 70.8103989 31.6975546 lineto
- 72.9854831 33.1003158 75.1605672 34.5030769 77.3356513 35.9058381 curveto
- 77.3356513 35.9058381 77.3356513 35.9058381 77.3356513 35.9058381 curveto
-closepath
-stroke
-newpath 72.1261925 48.9235775 moveto
- 65.2698286 41.8223435 58.4134647 34.7211095 51.5571009 27.6198755 curveto
-stroke
-newpath 77.3356513 54.3190884 moveto
- 70.6804697 50.3194478 lineto
- 73.5719153 47.5277072 lineto
- 74.8264939 49.7915009 76.0810726 52.0552947 77.3356513 54.3190884 curveto
- 77.3356513 54.3190884 77.3356513 54.3190884 77.3356513 54.3190884 curveto
-closepath
-fill
-newpath 77.3356513 54.3190884 moveto
- 70.6804697 50.3194478 lineto
- 73.5719153 47.5277072 lineto
- 74.8264939 49.7915009 76.0810726 52.0552947 77.3356513 54.3190884 curveto
- 77.3356513 54.3190884 77.3356513 54.3190884 77.3356513 54.3190884 curveto
-closepath
-stroke
-newpath 72.1261925 24.7294238 moveto
- 65.2698286 31.8306578 58.4134647 38.9318918 51.5571009 46.0331258 curveto
-stroke
-newpath 77.3356513 19.3339128 moveto
- 73.5719153 26.1252941 lineto
- 70.6804697 23.3335535 lineto
- 72.8988635 22.0003399 75.1172574 20.6671264 77.3356513 19.3339128 curveto
- 77.3356513 19.3339128 77.3356513 19.3339128 77.3356513 19.3339128 curveto
-closepath
-fill
-newpath 77.3356513 19.3339128 moveto
- 73.5719153 26.1252941 lineto
- 70.6804697 23.3335535 lineto
- 72.8988635 22.0003399 75.1172574 20.6671264 77.3356513 19.3339128 curveto
- 77.3356513 19.3339128 77.3356513 19.3339128 77.3356513 19.3339128 curveto
-closepath
-stroke
-newpath 70.1954371 40.042232 moveto
- 63.9826584 42.0391966 57.7698796 44.0361612 51.5571009 46.0331258 curveto
-stroke
-newpath 77.3356513 37.7471631 moveto
- 70.8103989 41.9554466 lineto
- 69.5804753 38.1290174 lineto
- 72.1655339 38.0017326 74.7505926 37.8744479 77.3356513 37.7471631 curveto
- 77.3356513 37.7471631 77.3356513 37.7471631 77.3356513 37.7471631 curveto
-closepath
-fill
-newpath 77.3356513 37.7471631 moveto
- 70.8103989 41.9554466 lineto
- 69.5804753 38.1290174 lineto
- 72.1655339 38.0017326 74.7505926 37.8744479 77.3356513 37.7471631 curveto
- 77.3356513 37.7471631 77.3356513 37.7471631 77.3356513 37.7471631 curveto
-closepath
-stroke
-newpath 70.3550151 53.4180207 moveto
- 64.0890437 50.956389 57.8230723 48.4947574 51.5571009 46.0331258 curveto
-stroke
-newpath 77.3356513 56.1604134 moveto
- 69.6201932 55.2884765 lineto
- 71.089837 51.5475648 lineto
- 73.1717751 53.085181 75.2537132 54.6227972 77.3356513 56.1604134 curveto
- 77.3356513 56.1604134 77.3356513 56.1604134 77.3356513 56.1604134 curveto
-closepath
-fill
-newpath 77.3356513 56.1604134 moveto
- 69.6201932 55.2884765 lineto
- 71.089837 51.5475648 lineto
- 73.1717751 53.085181 75.2537132 54.6227972 77.3356513 56.1604134 curveto
- 77.3356513 56.1604134 77.3356513 56.1604134 77.3356513 56.1604134 curveto
-closepath
-stroke
-showpage
-%%EOF
-
-%%EndDocument
- @endspecial 0.000000 TeXcolorgray 1562 2475 a
- gsave currentpoint currentpoint translate [0.500000 -0.000000 -0.000000
-0.500000 0 0] concat neg exch neg exch translate
- 1562 2475
-a 1503 2498 a Fa(p1)1621 2475 y
- currentpoint grestore moveto
- 1621 2475 a 1562 2629
-a
- gsave currentpoint currentpoint translate [0.500000 -0.000000 -0.000000
-0.500000 0 0] concat neg exch neg exch translate
- 1562 2629 a 1503 2651 a Fa(p2)1621 2629 y
- currentpoint grestore moveto
- 1621 2629
-a 1562 2782 a
- gsave currentpoint currentpoint translate [0.500000 -0.000000 -0.000000
-0.500000 0 0] concat neg exch neg exch translate
- 1562 2782 a 1503 2805 a Fa(p3)1621 2782
-y
- currentpoint grestore moveto
- 1621 2782 a 1562 2936 a
- gsave currentpoint currentpoint translate [0.500000 -0.000000 -0.000000
-0.500000 0 0] concat neg exch neg exch translate
- 1562 2936 a 1503 2958 a Fa(p4)1621
-2936 y
- currentpoint grestore moveto
- 1621 2936 a 1946 2552 a
- gsave currentpoint currentpoint translate [0.500000 -0.000000 -0.000000
-0.500000 0 0] concat neg exch neg exch translate
- 1946 2552 a 1887 2584
-a Fa(n1)2005 2552 y
- currentpoint grestore moveto
- 2005 2552 a 1946 2706 a
- gsave currentpoint currentpoint translate [0.500000 -0.000000 -0.000000
-0.500000 0 0] concat neg exch neg exch translate
- 1946 2706
-a 1887 2738 a Fa(n2)2005 2706 y
- currentpoint grestore moveto
- 2005 2706 a 2330 2552
-a
- gsave currentpoint currentpoint translate [0.500000 -0.000000 -0.000000
-0.500000 0 0] concat neg exch neg exch translate
- 2330 2552 a 2275 2584 a Fa(a1)2388 2552 y
- currentpoint grestore moveto
- 2388 2552
-a 2330 2706 a
- gsave currentpoint currentpoint translate [0.500000 -0.000000 -0.000000
-0.500000 0 0] concat neg exch neg exch translate
- 2330 2706 a 2275 2738 a Fa(a2)2388 2706
-y
- currentpoint grestore moveto
- 2388 2706 a 2330 2859 a
- gsave currentpoint currentpoint translate [0.500000 -0.000000 -0.000000
-0.500000 0 0] concat neg exch neg exch translate
- 2330 2859 a 2275 2891 a Fa(a3)2388
-2859 y
- currentpoint grestore moveto
- 2388 2859 a Black 0.000000 TeXcolorgray eop end
-%%Trailer
-
-userdict /end-hook known{end-hook}if
-%%EOF
Binary file main/nnet/doc/latex/asymptote/mlp4-2-3.pdf has changed
--- a/main/nnet/doc/latex/asymptote/transferFunctions/logsig.asy	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,18 +0,0 @@
-import graph;
-size(100,0);
-
-real f(real x) {return 1/(1+exp(-x));}
-pair F(real x) {return (x,f(x));}
-
-
-xaxis("$n$",EndArrow);
-yaxis("$a$",-1.75,1.75,EndArrow);
-
-draw(graph(f,-2.5,2.5,operator ..));
-draw((-2.5,-1)--(2.5,-1),currentpen+dashed);
-draw((-2.5,1)--(2.5,1),currentpen+dashed);
-
-label("$a = logsig(n) $",(0,-2.00));
-label("$0$",(0.2,-0.3));
-label("$-1$",(0.6,-1.35));
-label("$+1$",(0.75,1.35));
\ No newline at end of file
--- a/main/nnet/doc/latex/asymptote/transferFunctions/logsiglogo.asy	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,17 +0,0 @@
-// logsig symbol for nnet
-
-// define size of outer square = 1cm
-unitsize(1cm);
-draw(unitsquare);
-
-// in the middle one short line from left to right
-draw((0.1,0.3)--(0.9,0.3));
-
-// now draw logsig
-import graph;
-
-real f(real x) {return tanh(x);}
-draw(shift(0.5,0.5)*((scale(0.2)*graph(f,-2.0,2.0,operator ..))));
-//shift(2,1);
-
-//scale(real 0.5);
\ No newline at end of file
--- a/main/nnet/doc/latex/asymptote/transferFunctions/purelin.asy	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,19 +0,0 @@
-// purelin
-import graph;
-size(100,0);
-
-real f(real x) {return 1*x;}
-pair F(real x) {return (x,f(x));}
-
-
-xaxis("$n$",EndArrow);
-yaxis("$a$",-1.75,1.75,EndArrow);
-
-draw(graph(f,-2.5,2.5,operator ..));
-draw((-2.5,-1)--(2.5,-1),currentpen+dashed);
-draw((-2.5,1)--(2.5,1),currentpen+dashed);
-
-label("$a = purelin(n) $",(0,-2.00));
-label("$0$",(0.2,-0.3));
-label("$-1$",(0.6,-1.35));
-label("$+1$",(0.75,1.35));
\ No newline at end of file
--- a/main/nnet/doc/latex/asymptote/transferFunctions/purelinlogo.asy	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,17 +0,0 @@
-// purelin symbol for nnet
-
-// define size of outer square = 1cm
-unitsize(1cm);
-draw(unitsquare);
-
-// in the middle one short line from left to right
-draw((0.1,0.5)--(0.9,0.5));
-
-// now draw purelin
-import graph;
-
-real f(real x) {return 1*x;}
-draw(shift(0.5,0.5)*((scale(0.2)*graph(f,-2.0,2.0,operator ..))));
-//shift(2,1);
-
-//scale(real 0.5);
\ No newline at end of file
--- a/main/nnet/doc/latex/asymptote/transferFunctions/tansig.asy	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,18 +0,0 @@
-import graph;
-size(100,0);
-
-real f(real x) {return tanh(x);}
-pair F(real x) {return (x,f(x));}
-
-
-xaxis("$n$",EndArrow);
-yaxis("$a$",-1.75,1.75,EndArrow);
-
-draw(graph(f,-2.5,2.5,operator ..));
-draw((-2.5,-1)--(2.5,-1),currentpen+dashed);
-draw((-2.5,1)--(2.5,1),currentpen+dashed);
-
-label("$a = tansig(n) $",(0,-2.00));
-label("$0$",(0.2,-0.3));
-label("$-1$",(0.6,-1.35));
-label("$+1$",(0.75,1.35));
\ No newline at end of file
--- a/main/nnet/doc/latex/asymptote/transferFunctions/tansiglogo.asy	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,17 +0,0 @@
-// tansig symbol for nnet
-
-// define size of outer square = 1cm
-unitsize(1cm);
-draw(unitsquare);
-
-// in the middle one short line from left to right
-draw((0.1,0.5)--(0.9,0.5));
-
-// now draw tansig
-import graph;
-
-real f(real x) {return tanh(x);}
-draw(shift(0.5,0.5)*((scale(0.2)*graph(f,-2.0,2.0,operator ..))));
-//shift(2,1);
-
-//scale(real 0.5);
\ No newline at end of file
--- a/main/nnet/doc/latex/common/bibliography.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,39 +0,0 @@
-% Preamble
-
-%\documentclass[a4paper]{report}
-
-%\usepackage[ngerman]{babel}
-%\usepackage[T1]{fontenc}
-%\usepackage[ansinew]{inputenc}
-
-
-%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
-% start text here!!
-
-%\begin{document}
-
-\begin{thebibliography}{XXXXXXX}
-
-\bibitem [1]{1} John W. Eaton
-
-GNU Octave Manual, Edition 3, PDF-Version, February 1997
-
-\bibitem [2]{2} The MathWorks, Inc.
-
-MATLAB Online-Help
-
-\bibitem [3]{3} Steven W. Smith
-
-The Scientist and Engineer's Guide to Digital Signal Processing
-ISBN 0-9660176-3-3, California Technical Publishing, 1997
-
-\bibitem [4]{4} Martin T. Hagan, Howard B. Demuth, Mark Beale
-
-Neural Network Design, ISBN 0971732108, PWS Publishing Company, USA, Boston, 1996
-
-
-
-
-
-\end{thebibliography}
-%\end{document}
\ No newline at end of file
--- a/main/nnet/doc/latex/common/version.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,1 +0,0 @@
-Version: 0.1.9.1
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/algorithm/LevenbergMarquardt.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,39 +0,0 @@
-\section{Levenberg Marquardt}
-This algorithm will be programed with \cite{4}.
-
-\subsection{Sensitivity Matrix}
-How does this looks like?\\
-\begin{enumerate}
-	\item for a 1-1-1 MLP
-	\item for a 2-1-1 MLP
-	\item for a 1-2-1 MLP
-\end{enumerate}
-
-\subsubsection{1-1-1 MLP}
-In this case, the MLP holds one input neuron, one hidden neuron and one output neuron. The number of weights needed for this MLP is 4 (2 weights, 2 biases).\\
-
-It needs two sensitivity matrices because the two layers. Each sensitivity matrix will hold 1 element.
-This is taken from \cite{4}, example P12.5 page 12-44. Attention, in this example are two data sets used, this is the reason for the 4 elements...!
-
-\subsubsection{2-1-1 MLP}
-In this case, the MLP holds two input neurons, one hidden neuron and one output neuron. The number of weights needed for this MLP is 5 (3 weights, 2 biases).\\
-
-It needs also two sensitivity matrices because the two layers. Actually, the input is not only a scalar, it is a vector with 2 elements. Even though, again after \cite{4}. I think the sensitivity matrices will hold only 1 element. So the number of elements will bi proportional to the number of hidden neurons and the number of output neurons.
-
-\subsubsection{1-2-1 MLP}
-In this case, the MLP holds one input neuron, two hidden neurons and one output neuron. The number of weights needed for this MLP is 7 (4 weights, 3 biases).\\
-
-It needs also two sensitivity matrices because the two layers. Actually, the input is again only a scalar. 
-Now calculating $n_1^1$ will result in a row vector with 2 elements. $n_1^2$ will hold only one element and so we have 3 elements in the sensitivity matrix.\\
-
-We can say, the number of hidden neurons is responsible for the dimension of the sensitivity matrices.
-For example, a 4-3-1 MLP with 100 data sets will generate following sensitivity matrix for the first layer:
-$\tilde{\textbf{S}}^1 = [\tilde{\textbf{S}}^1_1 | \tilde{\textbf{S}}^1_2 | ... | \tilde{\textbf{S}}^1_{100}]$\\
-\noindent $\tilde{\textbf{S}}^1_1$ will hold 3 elements  $\tilde{\textbf{S}}^1_1 = [\tilde{\textbf{S}}^1_{1,1} ~ \tilde{\textbf{S}}^1_{2,1} ~ \tilde{\textbf{S}}^1_{3,1}]^T$;
-$\tilde{\textbf{S}}^1_2 = [\tilde{\textbf{S}}^1_{1,2} ~ \tilde{\textbf{S}}^1_{2,2} ~ \tilde{\textbf{S}}^1_{3,2}]^T$ and so on. So matrix will have a size of 3x100 for $\tilde{\textbf{S}}^1_{1}$
-and a size of 1x100 for $\tilde{\textbf{S}}^1_{2}$.\\
-
-By the way, the jacobian matrix will be a 100x20 matrix ..
-
-
-
--- a/main/nnet/doc/latex/developers/algorithm/algorithmen.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,5 +0,0 @@
-\chapter{Algorithm}
-Here are some general thoughts about calculating parts are used in algorithm.
-
-\input{algorithm/LevenbergMarquardt}
-
--- a/main/nnet/doc/latex/developers/analyzing/analyzingNewff.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,231 +0,0 @@
-\section{analyzing newff}
-First, \textit{newff} will be analyzed for a X-X-X mlp. This means, maximum 3 layers, including the input layer. Or in words, one input- one hidden- and one output-layer. The number of neurons are choosable.
-
-Following command will be used, to create a new feed-forward neural network:\\
-\noindent MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],...\newline
-\{'tansig','purelin'\},'trainlm','learngdm','mse');\\
-
-newff is the matlab command, mMinMaxElements is a $Rx2$-Matrix with minimum and maximum values of the inputs. $R$ is equal to the number of input neurons. [nHiddenNeurons nOutputNeurons] are the scalar values, to define the number of neurons in the hidden and output layer. One value, for each layer. \{'tansig','purelin'\} are the transfer functions, for each layer. This means, 'tansig' for the hidden layer and 'purelin' for the output layer. 'trainlm' is the training algorithm, in this case, Levenberg-Marquardt. 'learngdm' is the learn algorithm and 'mse' is the performance function, \textbf{m}ean-\textbf{s}quare-\textbf{e}rror.\\
-MLPnet will be a structure with following content:
-
-\begin{verbatim}
-Neural Network object:
-
-    architecture:
-
-         numInputs: 1
-         numLayers: 2
-       biasConnect: [1; 1]
-      inputConnect: [1; 0]
-      layerConnect: [0 0; 1 0]
-     outputConnect: [0 1]
-     targetConnect: [0 1]
-
-        numOutputs: 1  (read-only)
-        numTargets: 1  (read-only)
-    numInputDelays: 0  (read-only)
-    numLayerDelays: 0  (read-only)
-
-    subobject structures:
-
-            inputs: {1x1 cell} of inputs
-            layers: {2x1 cell} of layers
-           outputs: {1x2 cell} containing 1 output
-           targets: {1x2 cell} containing 1 target
-            biases: {2x1 cell} containing 2 biases
-      inputWeights: {2x1 cell} containing 1 input weight
-      layerWeights: {2x2 cell} containing 1 layer weight
-
-    functions:
-
-          adaptFcn: 'trains'
-           initFcn: 'initlay'
-        performFcn: 'mse'
-          trainFcn: 'trainlm'
-
-    parameters:
-
-        adaptParam: .passes
-         initParam: (none)
-      performParam: (none)
-        trainParam: .epochs, .goal, .max_fail, .mem_reduc, 
-                    .min_grad, .mu, .mu_dec, .mu_inc, 
-                    .mu_max, .show, .time
-
-    weight and bias values:
-
-                IW: {2x1 cell} containing 1 input weight matrix
-                LW: {2x2 cell} containing 1 layer weight matrix
-                 b: {2x1 cell} containing 2 bias vectors
-
-    other:
-
-          userdata: (user stuff)
-\end{verbatim}
-\textit{numInputs: 1}: one input layer\\
-\noindent \textit{numLayers: 2}: one hidden and one output layer\\
-\noindent \textit{biasConnect: [1; 1]}: unknown till now!!\\
-\noindent \textit{inputConnect: [1; 0]}: unknown till now!!\\
-\noindent \textit{layerConnect: [0 0; 1 0]}: unknown till now!!\\
-\noindent \textit{outputConnect: [0 1]}: unknown till now!!\\
-\noindent \textit{targetConnect: [0 1]}: unknown till now!!\\
-\noindent \textit{numOutputs: 1  (read-only)}: unknown till now!!\\
-\noindent \textit{numTargets: 1  (read-only)}: unknown till now!!\\
-\noindent \textit{numInputDelays: 0  (read-only)}: unknown till now!!\\
-\noindent \textit{numLayerDelays: 0  (read-only)}: unknown till now!!\\
-\noindent \textit{inputs: {1x1 cell} of inputs}: input layer definition\\
-Because we have defined only one input layer, you can see the detailed definition with
-following command in the matlab prompt:\\
-\begin{verbatim}
-	MLPnet.inputs{1}
-
-ans = 
-
-       range: [26x2 double]
-        size: 26
-    userdata: [1x1 struct]
-\end{verbatim}
-range are the min. and max. values of the inputs. size is the number of input neurons and userdata are user specified inputs...!\\
-\noindent \textit{layers: {2x1 cell} of layers}: hidden and output layer definition\\
-The dimension of $2x1 cell$ is because we have one hidden and one output layer. So too see the details of the hidden layer definitions, we have to enter:
-\begin{verbatim}
-K>> MLPnet.layers{1}
-
-ans = 
-
-     dimensions: 2
-    distanceFcn: ''
-      distances: []
-        initFcn: 'initnw'
-    netInputFcn: 'netsum'
-      positions: [0 1]
-           size: 2
-    topologyFcn: 'hextop'
-    transferFcn: 'tansig'
-       userdata: [1x1 struct]
-\end{verbatim}
-and for the output layer:
-\begin{verbatim}
-K>> MLPnet.layers{2}
-
-ans = 
-
-     dimensions: 1
-    distanceFcn: ''
-      distances: []
-        initFcn: 'initnw'
-    netInputFcn: 'netsum'
-      positions: 0
-           size: 1
-    topologyFcn: 'hextop'
-    transferFcn: 'purelin'
-       userdata: [1x1 struct]
-\end{verbatim}
-
-\noindent \textit{outputs: {1x2 cell} containing 1 output}: output layer definitions\\
-\begin{verbatim}
-K>> MLPnet.outputs
-
-ans = 
-
-     []    [1x1 struct]
-\end{verbatim}
-How knows, why this is a $1x2 cell$? The next command will also show the detailed definition! Of course, realy simple.
-\begin{verbatim}
-K>> MLPnet.outputs{2}
-
-ans = 
-
-        size: 1
-    userdata: [1x1 struct]
-\end{verbatim} 
-
-\noindent \textit{targets: {1x2 cell} containing 1 target}: unknow till now\\
-
-\noindent \textit{biases: {2x1 cell} containing 2 biases}: detailed definitions, for the biases\\
-\begin{verbatim}
-K>> MLPnet.biases
-
-ans = 
-
-    [1x1 struct]
-    [1x1 struct]
-
-K>> MLPnet.biases{1}
-
-ans = 
-
-       initFcn: ''
-         learn: 1
-      learnFcn: 'learngdm'
-    learnParam: [1x1 struct]
-          size: 2
-      userdata: [1x1 struct]
-
-K>> MLPnet.biases{2}
-
-ans = 
-
-       initFcn: ''
-         learn: 1
-      learnFcn: 'learngdm'
-    learnParam: [1x1 struct]
-          size: 1
-      userdata: [1x1 struct]
-\end{verbatim}
-      inputWeights: {2x1 cell} containing 1 input weight
-      layerWeights: {2x2 cell} containing 1 layer weight
-
-
-\paragraph{weight and bias values:}
-
-\subparagraph{IW:}
-\begin{verbatim}
-K>> MLPnet.IW
-
-ans = 
-
-    [2x26 double]
-               []
-\end{verbatim}
-
-\subparagraph{LW:}
-\begin{verbatim}
-K>> MLPnet.LW
-
-ans = 
-
-              []     []
-    [1x2 double]     []
-\end{verbatim}
-
-\subparagraph{b:}
-\begin{verbatim}
-K>> MLPnet.b
-
-ans = 
-
-    [2x1 double]
-    [   -0.3908]
-\end{verbatim}
-
-
-\paragraph{net.trainParam:}
-Output for the Levenberg-Marquardt train algorithm.
-\begin{verbatim}
-K>> MLPnet.trainParam
-
-ans = 
-
-       epochs: 100
-         goal: 0
-     max_fail: 5
-    mem_reduc: 1
-     min_grad: 1.0000e-010
-           mu: 0.0010
-       mu_dec: 0.1000
-       mu_inc: 10
-       mu_max: 1.0000e+010
-         show: 25
-         time: Inf
-\end{verbatim}
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/analyzing/matlab.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,4 +0,0 @@
-\chapter{analyzing matlab functions}
-
-\input{analyzing/analyzingNewff}
-\input{analyzing/analyzingNewp}
--- a/main/nnet/doc/latex/developers/codeConvention.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,12 +0,0 @@
-\section{Code convention}
-
-The main function of this toolbox will be programed with help of the book in \cite{4}.
-So the variables will have the same names like in the book with one exception: Variables, only with one letter, will have two letters, e.g. 
-\begin{itemize}
-	\item $X \rightarrow Xx$
-	\item $Q \rightarrow Qq$
-	\item $a \rightarrow aa$
-\end{itemize}
-and so on ...\\
-
-This is only to make it possible to search for variable names.
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/codingGuideline/codingGuideline.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,51 +0,0 @@
-\chapter{Coding Guideline}
-Some genereal descriptions why a variable has a chosen name. This is valid for the complete
-toolbox... or so :-)\\
-Here is only the description of variable names, which aren't visible to the user. Visible names are
-described in the User's Guide!\\
-The variable identifiers are taken from \cite{4}. One difference is purposeful added. If a variable has only one letter, a second small letter will be added to make it searchable. Who has ever tried to search a variable called "t"?
-
-\section{Variable identifier}
-
-\begin{tabbing}
-\hspace*{1em} \= \textcolor{blue}{Identifier} \hspace*{3em}\= \textcolor{blue}{Description:} \\ 
-  \textbf{Aa} \> \> hold the network values after transfer function.\\
-  blf \>  \> \textbf{b}atch \textbf{l}earning \textbf{f}unction \\
-  btf  \>  \> \textbf{b}atch \textbf{t}rainings \textbf{f}unction \\
-  \textbf{Jj} \>  \> Jacobi matrix \\
-  \textbf{Nn}  \> \> hold the network values before transfer function.\\  						
-  net	\> \> structure which holds the neural network informations \\
-  pf \>  \> \textbf{p}erformance \textbf{f}unction \\
-  \textbf{Pp}				\>						\>input matrix; nInputs x nDataSets  \\
-  \textbf{Pr}	\>		\> input range, this is a Nx2 matrix, that's why the capitalized P \\
-  trf \>  \> \textbf{tr}ansfer \textbf{f}unction \\
-  \textbf{Tt} \>  \> target matrix, nTargets x nDataSets\\
-  \textbf{ss}	\> \> row vector with numbers of neurons in the layers, for each layer, one entry \\
-  \textbf{vE}	\> \> row vector with errors... size depends on network structure. \\
-  VV \>  \> Validation structure \\
-  \textbf{xx} \>  \> Weight vector in column format\\
-  
-\end{tabbing}
-
-\subsection{Nn}
-\textbf{Nn} is a cell array and has one entry for each layer. In reality, this will have 2 or 3 layers.\\
-In \textbf{Nn\{1,1\}} are the values for the first hidden layer. The size of this matrix depends
-on the number of neurons used for this layer.\\
-In \textbf{Nn\{2,1\}} are the values for the second hidden layer or the output layer. The size of this matrix depends
-on the number of neurons used for this layer and so on ...\\
-\textbf{Nn\{x,1\}} where \textbf{x} can be $\infty$.\\
-
-\subsection{Aa}
-\textbf{Aa} is a cell array and has one entry for each layer.\\
-In \textbf{Aa\{1,1\}} are the values for the first hidden layer. The size of this matrix depends
-on the number of neurons used for this layer.\\
-In \textbf{Aa\{2,1\}} are the values for the second hidden layer or the output layer. The size of this matrix depends
-on the number of neurons used for this layer.\\
-See \textbf{Nn} for a more detailed description.\\
-
-\subsection{vE}
-\textbf{vE} is also a cell array which holds in the last (second) element the error vector. It's not completly clear, why in the last (second) element.\\
-The number of rows depends on the number of output neurons. For one output neuron, \textbf{vE} holds only one row, for 2 output neurons, this holds of course 2 rows, and so on. 
-
-\subsection{Jj}
-This is the short term for the Jacobi matrix.
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/examples/example1.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,40 +0,0 @@
-\section{Example 1}
-
-This MLP is designed with 2-2-1. This is not a complete example but it will help to understand the
-dimensions of all matrices and vectores are used inside the Levenberg-Marquardt algorithm.
-
-\subsection{Data matrices}
-The input matrix will be defined like in equation \eqref{equ:mInput} and the output matrix like in 
-equation \eqref{equ:mOutput}.
-
-\begin{equation}
-  mInput = \left[ \begin{array}{c c c c}
-												1 & 2 & 3 & 1   \\
-												1	& 1 & 1 & 2 	\\
-												1 & 2 & 1 & 2		\\																					
-												\end{array}
-					\right]
-		\label{equ:mInput}
-\end{equation}
-
-\begin{equation}
-  mOutput = \left[ \begin{array}{c c c c}
-												1 & 1.5 & 2 & 3   \\																					
-									 \end{array}
-					\right]
-		\label{equ:mOutput}
-\end{equation}
-
-
-
-
-\subsection{Weight matrices}
-The first layer matrix will hold 2x3 weights. The second layer matrix will hold 1x2 weights.
-The first bias holds 3x1 weights and the second holds only a scalar element.
-
-\subsection{Sensitivity Matrices}
-This part is right now not so clear in my mind. What is the dimension of these two matrices?
-The first layer sensitivity matrix should be about 2x71. Number of hidden neurons in the rows and number of train data sets in the columns.\\
-
-In the actual version, the dimension is about 71x71 .. so it seems to have a mistake inside the algorithm :-(
-
--- a/main/nnet/doc/latex/developers/examples/example2.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,15 +0,0 @@
-\section{Example 2}
-
-This MLP is designed with 26-2-1. Why starting easy if it's possible to start difficult?
-The input matrix which will be given to the training algorithm has the dimension of 26x71. The jacobian matrix should reach a size of 71x57.
-
-\subsection{Weight matrices}
-The first layer matrix will hold 2x26 weights. The second layer matrix will hold 1x2 weights.
-The first bias holds 26x1 weights and the second holds only a scalar element.
-
-\subsection{Sensitivity Matrices}
-This part is right now not so clear in my mind. What is the dimension of these two matrices?
-The first layer sensitivity matrix should be about 2x71. Number of hidden neurons in the rows and number of train data sets in the columns.\\
-
-In the actual version, the dimension is about 71x71 .. so it seems to have a mistake inside the algorithm :-(
-
--- a/main/nnet/doc/latex/developers/examples/examples.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,4 +0,0 @@
-\chapter{Examples}
-
-
-\input{examples/example1}
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/funcindex/funcindex.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,5 +0,0 @@
-\chapter{Function Index}
-
-\section{Who calles who}
-\input{funcindex/funcindexCalled}
-
--- a/main/nnet/doc/latex/developers/funcindex/funcindexCalled.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,321 +0,0 @@
-\begin{verbatim}
-Function-File: isposint
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: logsig
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: min_max
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: newff
-=============================
-__init
-__newnetwork
-tansig
-train
-\end{verbatim}
-\begin{verbatim}
-Function-File: poststd
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: prestd
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: purelin
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: saveMLPStruct
-=============================
-__checknetstruct
-__printAdaptFcn
-__printAdaptParam
-__printB
-__printBiasConnect
-__printBiases
-__printIW
-__printInitFcn
-__printInitParam
-__printInputConnect
-__printInputWeights
-__printInputs
-__printLW
-__printLayerConnect
-__printLayerWeights
-__printLayers
-__printMLPHeader
-__printNetworkType
-__printNumInputDelays
-__printNumInputs
-__printNumLayerDelays
-__printNumLayers
-__printNumOutputs
-__printNumTargets
-__printOutputConnect
-__printOutputs
-__printPerformFcn
-__printPerformParam
-__printTargetConnect
-__printTargets
-__printTrainFcn
-__printTrainParam
-\end{verbatim}
-\begin{verbatim}
-Function-File: sim
-=============================
-__checknetstruct
-logsig
-purelin
-tansig
-\end{verbatim}
-\begin{verbatim}
-Function-File: subset
-=============================
-__optimizedatasets
-\end{verbatim}
-\begin{verbatim}
-Function-File: tansig
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: train
-=============================
-__checknetstruct
-__trainlm
-\end{verbatim}
-\begin{verbatim}
-Function-File: trastd
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __analyzerows
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __calcjacobian
-=============================
-__dlogsig
-__dpurelin
-__dtansig
-logsig
-purelin
-tansig
-\end{verbatim}
-\begin{verbatim}
-Function-File: __calcperf
-=============================
-__mse
-logsig
-purelin
-tansig
-\end{verbatim}
-\begin{verbatim}
-Function-File: __checknetstruct
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __copycoltopos1
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __dlogsig
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __dpurelin
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __dtansig
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __getx
-=============================
-__checknetstruct
-\end{verbatim}
-\begin{verbatim}
-Function-File: __init
-=============================
-__checknetstruct
-newff
-\end{verbatim}
-\begin{verbatim}
-Function-File: __mae
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __mse
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __newnetwork
-=============================
-__checknetstruct
-isposint
-newff
-train
-\end{verbatim}
-\begin{verbatim}
-Function-File: __optimizedatasets
-=============================
-__analyzerows
-__randomisecols
-__rerangecolumns
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printAdaptFcn
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printAdaptParam
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printB
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printBiasConnect
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printBiases
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printInitFcn
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printInitParam
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printInputConnect
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printInputs
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printInputWeights
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printIW
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printLayerConnect
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printLayers
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printLayerWeights
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printLW
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printMLPHeader
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printNetworkType
-=============================
-newff
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printNumInputDelays
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printNumInputs
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printNumLayerDelays
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printNumLayers
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printNumOutputs
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printNumTargets
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printOutputConnect
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printOutputs
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printPerformFcn
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printPerformParam
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printTargetConnect
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printTargets
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printTrainFcn
-=============================
-train
-\end{verbatim}
-\begin{verbatim}
-Function-File: __printTrainParam
-=============================
-train
-\end{verbatim}
-\begin{verbatim}
-Function-File: __randomisecols
-=============================
-\end{verbatim}
-\begin{verbatim}
-Function-File: __rerangecolumns
-=============================
-__copycoltopos1
-\end{verbatim}
-\begin{verbatim}
-Function-File: __setx
-=============================
-__checknetstruct
-\end{verbatim}
-\begin{verbatim}
-Function-File: __trainlm
-=============================
-__calcjacobian
-__calcperf
-__getx
-__setx
-isposint
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/introduction.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,28 +0,0 @@
-\chapter{Introduction}
-This documentation isn't a well defined and structured docu for the neural network toolbox.
-It's more a \textit{collection of my ideas and minds}.
-
-
-\section{Installed system}
-I'm developing and testing the actual version of the neural network toolbox on following 
-program versions:
-
-\begin{itemize}
-  \item Octave 2.9.5
-  \item octave-forge-2006.01.28
-  \item OctPlot svn version
-\end{itemize}
-
-.. and on another system
-
-\begin{itemize}
-  \item Octave 2.9.12
-  \item octave-forge packages ...
-  \item OctPlot svn version
-\end{itemize}
-
-
-\input{numbering}
-\input{codeConvention}
-
-
Binary file main/nnet/doc/latex/developers/neuralNetworkPackageForOctaveDevelopersGuide.dvi has changed
Binary file main/nnet/doc/latex/developers/neuralNetworkPackageForOctaveDevelopersGuide.pdf has changed
--- a/main/nnet/doc/latex/developers/neuralNetworkPackageForOctaveDevelopersGuide.tcp	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,12 +0,0 @@
-[FormatInfo]
-Type=TeXnicCenterProjectInformation
-Version=4
-
-[ProjectInfo]
-MainFile=neuralNetworkPackageForOctaveDevelopersGuide.tex
-UseBibTeX=0
-UseMakeIndex=1
-ActiveProfile=LaTeX => PDF
-ProjectLanguage=de
-ProjectDialect=DE
-
--- a/main/nnet/doc/latex/developers/neuralNetworkPackageForOctaveDevelopersGuide.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,83 +0,0 @@
-% Preambel
-\documentclass[a4paper,openany]{report}
-
-
-\usepackage{a4wide}
-\usepackage[ansinew]{inputenc}
-\usepackage[T1]{fontenc}
-\RequirePackage{ifpdf}
-
-\usepackage{hyperref}
-	\hypersetup{%
-  colorlinks=true,   % activates colored references
-  pdfpagemode=None,  % PDF-Viewer starts without content et.al.
-  pdfstartview=FitH, % PDF-Viewer uses a defined page width
-  %linkbordercolor=111,
-  % citebordercolor=111,
-  citecolor=blue,
-  linkcolor=blue}
-
-\ifpdf
-  \usepackage[pdftex]{graphicx}
-	  \DeclareGraphicsExtensions{.pdf}
-\else
-  \usepackage[dvips]{graphicx}
-	  \DeclareGraphicsExtensions{.eps}
-\fi
-	
-\usepackage{fancyhdr}
-\usepackage{booktabs}
-%\usepackage[dvips]{rotating}
-\usepackage{multirow}
-\usepackage{multicol}
-
-\usepackage{color}
-\usepackage{amsmath}
-\usepackage{alltt}
-%\usepackage{array}
-%\usepackage{colortbl}
-
-\clubpenalty = 10000
-\widowpenalty = 10000 \displaywidowpenalty = 10000
-
-\definecolor{hellgrau}{gray}{0.95}
-\definecolor{dunkelgrau}{gray}{0.55}
-
-
-
-\renewcommand{\headrulewidth}{0pt} % no head rule
-\renewcommand{\footrulewidth}{0pt} % no foot rule
-
-
-
-
-%\nointend
-%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
-% start text here!!
-
-
-
-\begin{document}
-\pagenumbering{roman}
-\input{title}
-
-\tableofcontents
-\pagenumbering{arabic}
-
-\include{introduction}
-\include{octave/octave}
-\include{codingGuideline/codingGuideline}
-\include{algorithm/algorithmen}
-\include{funcindex/funcindex}
-\include{tests/test}
-\include{analyzing/matlab}
-
-\appendix
-\include{examples/examples}
-\include{../common/bibliography}
-
-
-
-\end{document}
-
-
--- a/main/nnet/doc/latex/developers/numbering.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,9 +0,0 @@
-\section{Version numbers of the neural network toolbox}
-
-The first number describes the major release. Version number V1.0 will be the first toolbox release which should have the same functions like the Matlab R14 SP3 neural network Toolbox.\\
-
-The second number defines the finished functions. So to start, only the MLPs will realised and so this will be the number V0.1.0.\\
-
-The third number defines the status of the actual development and function. V0.0.1 means no release with MLP, actually, everything under construction... ;-D.\\
-
-Right now it's version V0.1.3 which means MLP works and currently the transfer function logsig is added.
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/octave/directorys.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,59 +0,0 @@
-\section{Functions location}
-
-\subsection{Directory \textit{nnet/nnet}}
-
-\noindent Available functions:
-\begin{enumerate}
-	\item min\_max
-	\item network
-  \item newff
-\end{enumerate}
-
-
-\subsection{Directory \textit{nnet/nnutils}}
-This directory holds a further subdirectory: \textit{saveMLPStructure}.\\
-
-\noindent Available functions:
-\begin{enumerate}
-  \item saveMLPStruct
-	
-\end{enumerate}
-
-Function \textit{saveMLPStruct} doesn't exist in MATLAB(TM). This is a helper function to print a neural network architecture.
-
-\subsubsection{saveMLPStructure}
-Inside this directory are a lot of different functions which will help to print a
-text file to display the neural network architecture. The look will be the same like
-you would open the type \textit{network} in MATLAB(TM).
-
-\begin{itemize}
-  \item printAdaptFcn
-  \item printAdaptParam
-  \item printB
-  \item printBiasConnect
-  \item printBiases
-  \item printInitFcn
-  \item printInputConnect
-  \item printInputs
-  \item printInputWeights
-  \item printIW
-  \item printLayerConnect
-  \item printLayers
-  \item printLayerWeights
-  \item printLW
-  \item printMLPHeader
-  \item printNumInputDelays
-  \item printNumInputs
-  \item printNumLayerDelays
-  \item printNumLayers
-  \item printNumOutputs
-  \item printNumTargets
-  \item printOutputConnect
-  \item printOutputs
-  \item printPerformFcn
-  \item printPerformParam
-  \item printTargetConnect
-  \item printTargets
-  \item printTrainFcn
-  \item printTrainParam
-\end{itemize}
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/octave/functions/isposintOct.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,28 +0,0 @@
-\section{isposint}
-Checks if value is a positive integer.\\
-
-\subsubsection{Syntax:}
-
-$f = isposint(val)$\\
-$f = 0$ if val is negative integer or float and is 0 if val is a positive float.\\
-
-\subsubsection{Example 1:}
-val = 5.3;\\
-isposint(val)\\
-ans = 0\\
-
-\subsubsection{Example 2:}
-val = -5;\\
-isposint(val)\\
-ans = 0\\
-
-\subsubsection{Example 3:}
-val = 0;\\
-isposint(val)\\
-ans = 1\\
-
-\subsubsection{Example 4:}
-val = 5;\\
-isposint(val)\\
-ans = 1\\
-
--- a/main/nnet/doc/latex/developers/octave/functions/min_max.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,14 +0,0 @@
-\subsection{min\_max}
-Checks for minimal and maximal values of an input matrix for \textbf{newff}.\\
-
-\subsubsection{Syntax:}
-
-$pr = min\_max(mInputs)$\\
-
-\subsubsection{Description:}
-\textit{mInputs} must be a matrix with input training data sets. This means in the case, for a 9-2-1 MLP
-(this means 9 input-, 2 hidden- and 1 output-neuron) with 100 input training data sets, the matrix must be
-an 9x100 matrix. \textit{pr} will then be a 9x2 matrix with minimal values in the first column and maximal values in the second column. If a row holds 2 zeros, a warning will appear (no information in this row!).
-
-\subsubsection{Important:}
-The equival function in MATLAB(TM) is called \textit{minmax}. This is not possible because the functions \textit{min} and \textit{max} in Octave are programed in minmax.cc!
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/octave/functions/newff.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,14 +0,0 @@
-\subsection{min\_max}
-Checks for minimal and maximal values of an input matrix for \textbf{newff}.\\
-
-\subsubsection{Syntax:}
-
-$pr = min\_max(mInputs)$\\
-
-\subsubsection{Description:}
-\textit{mInputs} must be a matrix with input training data sets. This means in the case, for a 9-2-1 MLP
-(this means 9 input-, 2 hidden- and 1 output-neuron) with 100 input training data sets, the matrix must be
-an 9x100 matrix. \textit{pr} will then be a 9x2 matrix with minimal values in the first column and maximal values in the second column. If a row holds 2 zeros, a warning will appear (no information in this row!).
-
-\subsubsection{Important:}
-The equival function in MATLAB(TM) is called \textit{minmax}. This is not possible because the functions \textit{min} and \textit{max} in Octave are programed in minmax.cc!
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/octave/functions/poststd.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,14 +0,0 @@
-\subsection{min\_max}
-Checks for minimal and maximal values of an input matrix for \textbf{newff}.\\
-
-\subsubsection{Syntax:}
-
-$pr = min\_max(mInputs)$\\
-
-\subsubsection{Description:}
-\textit{mInputs} must be a matrix with input training data sets. This means in the case, for a 9-2-1 MLP
-(this means 9 input-, 2 hidden- and 1 output-neuron) with 100 input training data sets, the matrix must be
-an 9x100 matrix. \textit{pr} will then be a 9x2 matrix with minimal values in the first column and maximal values in the second column. If a row holds 2 zeros, a warning will appear (no information in this row!).
-
-\subsubsection{Important:}
-The equival function in MATLAB(TM) is called \textit{minmax}. This is not possible because the functions \textit{min} and \textit{max} in Octave are programed in minmax.cc!
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/octave/functions/prestd.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,14 +0,0 @@
-\subsection{min\_max}
-Checks for minimal and maximal values of an input matrix for \textbf{newff}.\\
-
-\subsubsection{Syntax:}
-
-$pr = min\_max(mInputs)$\\
-
-\subsubsection{Description:}
-\textit{mInputs} must be a matrix with input training data sets. This means in the case, for a 9-2-1 MLP
-(this means 9 input-, 2 hidden- and 1 output-neuron) with 100 input training data sets, the matrix must be
-an 9x100 matrix. \textit{pr} will then be a 9x2 matrix with minimal values in the first column and maximal values in the second column. If a row holds 2 zeros, a warning will appear (no information in this row!).
-
-\subsubsection{Important:}
-The equival function in MATLAB(TM) is called \textit{minmax}. This is not possible because the functions \textit{min} and \textit{max} in Octave are programed in minmax.cc!
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/octave/functions/sim.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,14 +0,0 @@
-\subsection{min\_max}
-Checks for minimal and maximal values of an input matrix for \textbf{newff}.\\
-
-\subsubsection{Syntax:}
-
-$pr = min\_max(mInputs)$\\
-
-\subsubsection{Description:}
-\textit{mInputs} must be a matrix with input training data sets. This means in the case, for a 9-2-1 MLP
-(this means 9 input-, 2 hidden- and 1 output-neuron) with 100 input training data sets, the matrix must be
-an 9x100 matrix. \textit{pr} will then be a 9x2 matrix with minimal values in the first column and maximal values in the second column. If a row holds 2 zeros, a warning will appear (no information in this row!).
-
-\subsubsection{Important:}
-The equival function in MATLAB(TM) is called \textit{minmax}. This is not possible because the functions \textit{min} and \textit{max} in Octave are programed in minmax.cc!
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/octave/functions/train.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,14 +0,0 @@
-\subsection{min\_max}
-Checks for minimal and maximal values of an input matrix for \textbf{newff}.\\
-
-\subsubsection{Syntax:}
-
-$pr = min\_max(mInputs)$\\
-
-\subsubsection{Description:}
-\textit{mInputs} must be a matrix with input training data sets. This means in the case, for a 9-2-1 MLP
-(this means 9 input-, 2 hidden- and 1 output-neuron) with 100 input training data sets, the matrix must be
-an 9x100 matrix. \textit{pr} will then be a 9x2 matrix with minimal values in the first column and maximal values in the second column. If a row holds 2 zeros, a warning will appear (no information in this row!).
-
-\subsubsection{Important:}
-The equival function in MATLAB(TM) is called \textit{minmax}. This is not possible because the functions \textit{min} and \textit{max} in Octave are programed in minmax.cc!
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/octave/functions/trastd.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,14 +0,0 @@
-\subsection{min\_max}
-Checks for minimal and maximal values of an input matrix for \textbf{newff}.\\
-
-\subsubsection{Syntax:}
-
-$pr = min\_max(mInputs)$\\
-
-\subsubsection{Description:}
-\textit{mInputs} must be a matrix with input training data sets. This means in the case, for a 9-2-1 MLP
-(this means 9 input-, 2 hidden- and 1 output-neuron) with 100 input training data sets, the matrix must be
-an 9x100 matrix. \textit{pr} will then be a 9x2 matrix with minimal values in the first column and maximal values in the second column. If a row holds 2 zeros, a warning will appear (no information in this row!).
-
-\subsubsection{Important:}
-The equival function in MATLAB(TM) is called \textit{minmax}. This is not possible because the functions \textit{min} and \textit{max} in Octave are programed in minmax.cc!
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/octave/octave.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,4 +0,0 @@
-\chapter{Octave's available functions}
-
-\section{Available functions}
-\input{octave/functions/min_max}
\ No newline at end of file
--- a/main/nnet/doc/latex/developers/tests/__analyzerows.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,23 +0,0 @@
-\begin{verbatim}
-%!shared b, retmat
-%! disp("testing __analyzerows")
-%! b = [1 0 0 1; 1 0 0 0; 1 2 0 1];
-%! retmat = __analyzerows(b);
-%!assert(retmat(1,1)==1);#%!assert(retmat(1,1)==1);
-%!assert(retmat(2,1)==1);
-%!assert(retmat(3,1)==0);
-%! b = [1 0 0 2; 1 0 0 0; 1 1 1 1];
-%! retmat = __analyzerows(b);
-%!assert(retmat(1,2)==0);
-%!assert(retmat(2,2)==0);
-%!assert(retmat(3,2)==1);
-%! b = [1 0 0 2; 1 0 0 0; 1 1 1 1];
-%! retmat = __analyzerows(b);
-%!assert(retmat(1,3)==2);
-%!assert(retmat(2,3)==0);
-%!assert(retmat(3,3)==0);
-%! retmat = __analyzerows(b);
-%!assert(retmat(1,4)==1);
-%!assert(retmat(2,4)==0);
-%!assert(retmat(3,4)==0);
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/__copycoltopos1.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,11 +0,0 @@
-\begin{verbatim}
-%!shared a, retmat
-%! disp("testing __copycoltopos1")
-%! a = [0 1 2 3 4; 5 6 7 8 9];
-%! retmat = __copycoltopos1(a,3);
-%!assert(retmat(1,1)==2);
-%!assert(retmat(2,1)==7);
-%! retmat = __copycoltopos1(a,5);
-%!assert(retmat(1,1)==4);
-%!assert(retmat(2,1)==9);
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/__optimizedatasets.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,23 +0,0 @@
-\begin{verbatim}
-%!shared retmatrix, matrix
-%! disp("testing __optimizedatasets")
-%! matrix = [1 2 3 2 1 2 3 0 5 4 3 2 2 2 2 2 2; \
-%!			 0 1 1 0 0 0 0 0 0 0 0 0 0 0 1 1 0; \
-%!			-1 3 2 4 9 1 1 1 1 1 9 1 1 1 9 9 0; \
-%!			 2 3 2 3 2 2 2 2 3 3 3 3 1 1 1 1 1];
-%! ## The last row is equal to the neural network targets
-%! retmatrix = __optimizedatasets(matrix,9,1);
-%! ## the above statement can't be tested with assert!
-%! ## it contains random values! So pass a "success" message
-%!assert(1==1);
-%! matrix = [1 2 3 2 1 2 3 0 5 4 3 2 2 2 2 2 2; \
-%!			 0 1 1 0 0 0 0 0 0 0 0 0 0 0 1 1 0; \
-%!			-1 3 2 4 9 1 1 1 1 1 9 1 1 1 9 9 0; \
-%!			 2 3 2 3 2 2 2 2 3 3 3 3 1 1 1 1 1];
-%! ## The last row is equal to the neural network targets
-%! retmatrix = __optimizedatasets(matrix,9,1,0);
-%!assert(retmatrix(1,1)==5);
-%!assert(retmatrix(2,1)==0);
-%!assert(retmatrix(3,1)==1);
-%!assert(retmatrix(4,1)==3);
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/__randomisecols.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,4 +0,0 @@
-\begin{verbatim}
-%!# no test possible, contains randperm which is using
-%!# some randome functions
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/__rerangecolumns.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,32 +0,0 @@
-\begin{verbatim}
-%!shared matrix,analyzeMatrix,nTrainSets, returnmatrix
-%! disp("testing __rerangecolumns")
-%! matrix = [0 1 0 0 0 0 1 0 1 1;  \
-%!			 4 4 4 4 4 4 4 4 4 4;  \
-%!        -1.1 -1.1 2 3 4 3.2 1 8 9 10; \
-%!           0 1.1 3 4 5 2 10 10 2 3; \
-%!          -1 1 1 1 1 2 3 4 1 5];
-%! analyzeMatrix = [1 0 0 0; 0 1 0 0; 0 0 2 1; 0 0 1 2; 0 0 1 1];
-%! nTrainSets = 8;
-%! returnmatrix = __rerangecolumns(matrix,analyzeMatrix,nTrainSets);
-%!assert(returnmatrix(1,1)==1);
-%!assert(returnmatrix(2,1)==4);
-%!assert(returnmatrix(3,1)==1);
-%!assert(returnmatrix(4,1)==10);
-%!assert(returnmatrix(5,1)==3);
-%! matrix = [0 1 0 0 0 0 1 0 1 1; 			\
-%!			 4 4 4 4 4 4 4 4 4 4; 			\
-%!          -1.1 -1.1 2 3 4 3.2 1 8 9 10; 	\
-%!           0 1.1 3 4 5 2 10 10 2 3; 		\
-%!          -1 1 1 1 1 2 3 4 1 5;     		\
-%!			 0 1 2 1 2 1 2 3 4 5;];  # the last row is euqal to the nnet targets
-%! analyzeMatrix = [1 0 0 0; 0 1 0 0; 0 0 2 1; 0 0 1 2; 0 0 1 1];
-%! nTrainSets = 8;
-%! returnmatrix = __rerangecolumns(matrix,analyzeMatrix,nTrainSets);
-%!assert(returnmatrix(1,1)==1);
-%!assert(returnmatrix(2,1)==4);
-%!assert(returnmatrix(3,1)==1);
-%!assert(returnmatrix(4,1)==10);
-%!assert(returnmatrix(5,1)==3);
-%!assert(returnmatrix(6,1)==2);
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/isposint.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,11 +0,0 @@
-\begin{verbatim}
-%!shared
-%! disp("testing isposint")
-%!assert(isposint(1)) # this should pass
-%!assert(isposint(0.5),0) # should return zero
-%!assert(isposint(-1),0) # should return zero
-%!assert(isposint(-1.5),0) # should return zero
-%!assert(isposint(0),0) # should return zero
-%!fail("isposint([0 0])","Input argument must not be a vector, only scalars are allowed!")
-%!fail("isposint('testString')","Input argument must not be a vector, only scalars are allowed!")
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/min_max.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,9 +0,0 @@
-\begin{verbatim}
-%!shared
-%! disp("testing min_max")
-%!test fail("min_max(1)","Argument must be a matrix.")
-%!test fail("min_max('testString')","Argument must be a matrix.")
-%!test fail("min_max(cellA{1}=1)","Argument must be a matrix.")
-%!test fail("min_max([1+1i, 2+2i])","Argument must be a matrix.")
-%!test fail("min_max([1+1i, 2+2i; 3+1i, 4+2i])","Argument has illegal type.")
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/newff.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,32 +0,0 @@
-\begin{verbatim}
-%!shared
-%! disp("testing newff")
-%!test
-%! Pr = [1;2];
-%! fail("newff(Pr,[1 1],{'tansig','purelin'},'trainlm','unused','mse')","Input ranges must be a two column matrix!")
-%!test
-%! Pr = [1 2 ; 4  6];
-%! assert(__checknetstruct(newff(Pr,[1 1],{'tansig','purelin'},'trainlm','unused','mse')))
-%!test
-%! Pr = [1 2 3; 4 5 6];
-%! fail("newff(Pr,[1 1],{'tansig','purelin'},'trainlm','unused','mse')","Input ranges must be a two column matrix!")
-%!test
-%! Pr = [5 3; 4 5];
-%! fail("newff(Pr,[1 1],{'tansig','purelin'},'trainlm','unused','mse')",\
-%!  "Input ranges has values in the second column larger as in the same row of the first column.")
-%!test
-%! Pr = [1 2 ; 4 6];
-%! fail("newff(Pr,[1 1; 2 3],{'tansig','purelin'},'trainlm','unused','mse')",\
-%!  "Layer sizes is not a row vector.")
-%!test
-%! Pr = [1 2 ; 4 6];
-%! assert(__checknetstruct(newff(Pr,[ 2 3],{'tansig','purelin'},'trainlm','unused','mse')))
-%!test
-%! Pr = [1 2 ; 4 6];
-%! fail("newff(Pr,[1],{'tansig','purelin'},'trainlm','unused','mse')",\
-%!  "There must be at least one hidden layer and one output layer!")
-%!test
-%! Pr = [1 2 ; 4 6];
-%! fail("newff(Pr,[-1 1],{'tansig','purelin'},'trainlm','unused','mse')",\
-%!  "Layer sizes is not a row vector of positive integers.")
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/prestd.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,7 +0,0 @@
-\begin{verbatim}
-%!shared Pp, Tt, pn
-%!  Pp = [1 2 3 4; -1 3 2 -1];
-%!  Tt = [3 4 5 6];
-%!  [pn,meanp,stdp] = prestd(Pp);
-%!assert(pn,[-1.16190 -0.38730 0.38730 1.16190; -0.84887 1.09141 0.60634 -0.84887],0.00001);
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/purelin.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,7 +0,0 @@
-\begin{verbatim}
-%!assert(purelin(2),2);
-%!assert(purelin(-2),-2);
-%!assert(purelin(0),0);
-%!error  # this test must throw an error!
-%! assert(purelin(2),1);
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/subset.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,67 +0,0 @@
-\begin{verbatim}
-%!shared matrix, nTargets, mTrain, mTest, mVali
-%! disp("testing subset")
-%! matrix = [1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 18 20; \
-%!			 0 2 4 1 3 5 3 4 1 -1 -2 -9 -1 10 12 20 11 11 11 11; \
-%!			-2 2 2 2 2 0 0 0 0  0 10 12 13 12 13 44 33 32 98 11; \
-%!			 0 0 0 0 1 1 1 1 0  0  1  1  1  0  0  1  1  1  0  0; \
-%!           4 4 4 4 4 4 4 4 4  4  4  4  4  4  4  4  4  4  4  4; \
-%!           1 2 3 4 5 6 7 8 9 10 11 12 13 33 44 55 66 77 88 99];
-%! nTargets = 1; # the last row is equivalent to the target values.
-%! [mTrain, mTest, mVali] = subset(matrix,nTargets);  ############################
-%!assert(size(mTrain,2)==10);# 50% of 20
-%!assert(size(mTest,2)==6);# 1/3 of 20 = 6 (floor)
-%!assert(size(mVali,2)==4);# 1/6 of 20 = 4 (floor)
-%! # It's not possible to test the column order with this call!
-%! # randomizing is used! But all max and min values should be
-%! # in the training set
-%!assert(max(mTrain(1,:))==max(matrix(1,:)));
-%!assert(min(mTrain(1,:))==min(matrix(1,:)));
-%!assert(max(mTrain(2,:))==max(matrix(2,:)));
-%!assert(min(mTrain(2,:))==min(matrix(2,:)));
-%!assert(max(mTrain(3,:))==max(matrix(3,:)));
-%!assert(min(mTrain(3,:))==min(matrix(3,:)));
-%!assert(max(mTrain(4,:))==max(matrix(4,:)));
-%!assert(min(mTrain(4,:))==min(matrix(4,:)));
-%!
-%!
-%! [mTrain, mTest, mVali] = subset(matrix,nTargets,0);  ############################
-%!assert(size(mTrain,2)==10);# 50% of 20
-%!assert(size(mTest,2)==6);# 1/3 of 20 = 6 (floor)
-%!assert(size(mVali,2)==4);# 1/6 of 20 = 4 (floor)
-%!assert(mTrain==matrix(:,1:10));
-%!assert(mTest==matrix(:,11:16));
-%!assert(mVali==matrix(:,17:20));
-%!
-%!
-%! [mTrain, mTest, mVali] = subset(matrix,nTargets,2);  ############################
-%!assert(size(mTrain,2)==10);# 50% of 20
-%!assert(size(mTest,2)==6);# 1/3 of 20 = 6 (floor)
-%!assert(size(mVali,2)==4);# 1/6 of 20 = 4 (floor)
-%!assert(max(mTrain(1,:))==max(matrix(1,:)));
-%!assert(min(mTrain(1,:))==min(matrix(1,:)));
-%!assert(max(mTrain(2,:))==max(matrix(2,:)));
-%!assert(min(mTrain(2,:))==min(matrix(2,:)));
-%!assert(max(mTrain(3,:))==max(matrix(3,:)));
-%!assert(min(mTrain(3,:))==min(matrix(3,:)));
-%!assert(max(mTrain(4,:))==max(matrix(4,:)));
-%!assert(min(mTrain(4,:))==min(matrix(4,:)));
-%!
-%!
-%! ## next test ... optimize twice
-%! matrix = [1 2 3 4 5 6 7 20 8 10 11 12 13 14 15 16 17 18 18 9; \
-%!			 0 2 4 1 3 5 3 4 1 -1 -2 -9 -1 10 12 20 11 11 11 11; \
-%!			-2 2 2 2 2 0 0 0 0  0 10 12 13 12 13 44 33 32 98 11; \
-%!			 0 0 0 0 1 1 1 1 0  0  1  1  1  0  0  1  1  1  0  0; \
-%!           4 4 4 4 4 4 4 4 4  4  4  4  4  4  4  4  4  4  4  4; \
-%!           1 2 3 4 5 6 7 8 9 10 11 12 13 33 44 55 66 77 88 99];
-%! [mTrain, mTest, mVali] = subset(matrix,nTargets,2);  ############################
-%!assert(max(mTrain(1,:))==max(matrix(1,:)));
-%!assert(min(mTrain(1,:))==min(matrix(1,:)));
-%!assert(max(mTrain(2,:))==max(matrix(2,:)));
-%!assert(min(mTrain(2,:))==min(matrix(2,:)));
-%!assert(max(mTrain(3,:))==max(matrix(3,:)));
-%!assert(min(mTrain(3,:))==min(matrix(3,:)));
-%!assert(max(mTrain(4,:))==max(matrix(4,:)));
-%!assert(min(mTrain(4,:))==min(matrix(4,:)));
-\end{verbatim}
--- a/main/nnet/doc/latex/developers/tests/test.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,24 +0,0 @@
-\chapter{Test}
-
-\section{isposint}
-\input{tests/isposint}
-\section{min\_max}
-\input{tests/min_max}
-\section{newff}
-\input{tests/newff}
-\section{prestd}
-\input{tests/prestd}
-\section{purelin}
-\input{tests/purelin}
-\section{subset}
-\input{tests/subset}
-\section{\_\_analyzerows}
-\input{tests/__analyzerows}
-\section{\_\_copycoltopos1}
-\input{tests/__copycoltopos1}
-\section{\_\_optimizedatasets}
-\input{tests/__optimizedatasets}
-\section{\_\_randomisecols}
-\input{tests/__randomisecols}
-\section{\_\_rerangecolumns}
-\input{tests/__rerangecolumns}
--- a/main/nnet/doc/latex/developers/title.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,7 +0,0 @@
-
-\title{A neural network toolbox for Octave\\
-			Developer's Guide\\
-			  \input{../common/version}}
-
-\author{M. Schmid}
-\maketitle
--- a/main/nnet/doc/latex/developers/varietes.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,96 +0,0 @@
-\section{Varietes}
-\label{chap:intro:sec:varietes}
-\subsection{Object-oriented programming}
-Some of the functions for Octave will have not the same name, like the matlab ones has.
-This difference is because the object-oriented programming technology. The object-oriented functions will have a name postfix in Octave.\\
-As example: \textit{isposint\_netw\_priv} means, \textit{isposint} is a private function from the matlab function \textit{network}.\\
-This is a kind of "`simulation"' for the object-oriented programming technology.\\
-
-\subsubsection{Data-Typ \textit{network} and m-file \textit{subsasgn}}
-Matlab has a further data type \textit{network}. A basic neural network will be initialized in a
-structure and than changed to the \textit{network} type with the \textit{class} command. The class command from Octave doesn't support the creation of this network type!
-In the Matlab m-file \textit{network.m} on row 256, the network type will be created! From this moment, structure subscription assignment will call the file subsasgn from the \textit{@network} directory. Back in the \textit{newff} m-file on row 134, the @network subsasgn will be called the first time.
-
-\paragraph{newff row 134; net.biasConnect} net=subsasgn(net,subscripts,v) will hold the following values:
-\begin{itemize}
-	\item net: the whole net structure
-	\item subscripts: 1x1 structure;\\
-		subscripts.type = '.'\\
-		subscripts.subs = 'biasConnect'
-	\item v: 2x1 double [1; 1]
-\end{itemize}
-
-\paragraph{newff row 135; net.inputConnect} net=subsasgn(net,subscripts,v) will hold the following values:
-\begin{itemize}
-	\item net: the whole net structure
-	\item subscripts: 1x2 structure;\\
-		subscripts(1).type = '.'\\
-		subscripts(1).subs = 'inputConnect'\\
-		subscripts(2).type = '()'\\
-		subscripts(2).subs = [1] [1]
-	\item v: 1x1 double 1
-\end{itemize}
-\textcolor{red}{bis hier sollte es erledigt sein...!}
-
-\paragraph{newff row 137; net.layerConnect} net=subsasgn(net,subscripts,v) will hold the following values:
-\begin{itemize}
-	\item net: the whole net structure
-	\item subscripts: 1x1 structure;\\
-		subscripts(1).type = '.'\\
-		subscripts(1).subs = 'layerConnect'
-	\item v: 2x2 logical [0 0; 1 0]
-\end{itemize}
-
-\paragraph{newff row 138; net.outputConnect} net=subsasgn(net,subscripts,v) will hold the following values:
-\begin{itemize}
-	\item net: the whole net structure
-	\item subscripts: 1x2 structure;\\
-		subscripts(1).type = '.'\\
-		subscripts(1).subs = 'outputConnect'\\
-		subscripts(2).type = '()'\\
-		subscripts(2).subs = '[2]'
-	\item v: 1x1 double = 1
-\end{itemize}
-
-\paragraph{newff row 139; net.targetConnect} net=subsasgn(net,subscripts,v) will hold the following values:
-\begin{itemize}
-	\item net: the whole net structure
-	\item subscripts: 1x2 structure;\\
-		subscripts(1).type = '.'\\
-		subscripts(1).subs = 'targetConnect'\\
-		subscripts(2).type = '()'\\
-		subscripts(2).subs = '[2]'
-	\item v: 1x1 double = 1
-\end{itemize}
-
-
-\subsection{Data types}
-
-\begin{table}
-	\centering
-	\begin{tabular}{c c c}
-		\toprule
-									& Matlab										& Octave\\
-		\midrule
-  								& double 											& double \\
-									& int8  											& \\
-									& int16 											& \\
-									& int32 											& \\
-									& int64 											& \\
-			Numeric			& single											& \\
-									& uint8 											& \\
-									& uint16											& \\
-									& uint32 											& \\ 
-									& uint64 											& \\ 
-									& complex 										& complex\\
-		\midrule
-			characters	& x														&  \\
-		\midrule
-			string			&															& \\
-		\midrule
-			cell				&	x														& x \\
-		\midrule
-			structure		& x 												  & x  \\
-		\bottomrule
-	\end{tabular}
-\end{table}
\ No newline at end of file
--- a/main/nnet/doc/latex/perl/analyzeOctaveSource.pm	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,66 +0,0 @@
-package analyzeOctaveSource;
-
-use File::Find;
-
-sub new {
- my $Objekt = shift;
- my $Referenz = {};
- bless($Referenz,$Objekt);
- return($Referenz);
-}
-
-sub readDirTree{
-	my $Objekt = shift;
-	my $Dir = shift;
-	my $fileExt = shift;
-	my $File;
-	# read directory
-	my @dirArray = ();
-
-	find sub { push @dirArray, $File::Find::name }, $Dir;
-
-	# remove all files except with file ending $fileExt
-	@DirArray = grep /.+\.$fileExt/ , @DirArray;
-	my $nFiles = @dirArray;
-	if ($nFiles==0){
-		print "No Octave files found.\n";
-		print "Octave files must end with *.m\n";
-		die " ==========================\n";
-	}
-	return @dirArray;
-}
-
-sub searchFilesContainTestLines{
-	my $Objekt = shift;
-	my @dirArray = shift;
-
-	#
-	my $fileName = "";
-	my @fileContent = ();
-	my @fileArray = ();
-	my @temp = ();
-	my $nTemp = "";
-
-	foreach (@dirArray){
-	  # open file and search for lines beginning with
-	  # !%
-	  if (-d $_){
-        # if directory, do nothing
-      }else{
-	    open(FILE,$_) or die "File $_ not found!\n";
-	    @fileContent = <FILE>;
-	    close FILE;
-        @temp = grep /^!%/, @fileContent;
-        $nTemp = @temp;
-        if ($nTemp>0){
-          @fileArray = $_
-        }
-	  }
-	}
-
-	return @fileArray;
-}
-
-
-1;
-
--- a/main/nnet/doc/latex/perl/createFunctionIndexDocu.pl	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,172 +0,0 @@
-#!/usr/bin/perl -w
-
-use strict;
-use diagnostics;# Force verbose warning diagnostics.
-use warnings;
-use English;
-
-#
-# Modules from the Standard Perl Library.
-
-
-#
-# Own modules
-use analyzeOctaveSource;
-
-
-#--- DEFINE VARIABLES -------------------------------
-my $Dir = "D:/daten/octave/neuroPackage/0.1.9/nnet/inst";
-my $fileExt = "m";
-my $funcIndDir = "D:/daten/octave/neuroPackage/0.1.9/nnet/doc/latex/developers/funcindex";
-my $relFuncDir = "funcindex/";
-my $funcFileExt = "tex";
-my $mainLatexFuncIndexFile = "funcindexCalled.tex";
-my $chapter = "Function Index";
-#--- END DEFINE VARIABLES ---------------------------
-
-
-my @filesArray = ();
-my @filesName = ();
-my $fileName = "";
-my $savePath = "";
-my $nTestLines = 0;
-my $m = 0;
-
-# analyze directory structure
-my $obj = analyzeOctaveSource->new();
-my @DirArray = $obj->readDirTree($Dir,$fileExt);
-
-# in @DirArray should be only names, ending with *.m
-# but if there is a directory, we will remove it at
-# the next line of code
-my @FuncReferences = grep /.+\.$fileExt/ , @DirArray; #/
-my @FuncNames = @FuncReferences;
-# now I have to remove the path and file extension
-foreach (@FuncNames)
-{
- 	s/\.m//g;   # removes file ending
- 	s/$Dir\///g; # removes any parts of the file & directory path
-}
-my @input = ();
-my @calledFunction = ();
-my $deleteFile = 1;
-my $anzahl_elemente = 0;
-# now analyze functions to see which other functions are called
-foreach my $FuncRef (@FuncReferences)
-{
-
-  open(FILE,$FuncRef); # opens e.g. 'subset.m'
-  @input = <FILE>; # read the complete content of the file to @input
-  # now remove all comment and test lines ..
-  @input = grep s/^\s+//, @input; # removes white-space characters at the
-  								  # beginning of a line
-  @input = grep /^[^#|%]/ , @input; # removes lines starting with # or %
-
-#   foreach (@input)
-#   {
-#     print "$_";
-#     sleep(1);
-#   }
-      my $actFuncName = "";
-  foreach my $FuncName (@FuncNames)
-  {
-
-    if ($FuncRef !~/$FuncName/)   # returns true if pattern is not found
-    {
-      # now search for each $FuncName
-      # inside of the @input array
-      # if one is found, put them to a list
-      # \todo ERRORs are still available
-      # if the $FuncName occures in another context!!
-      # such lines should be deleted!
-	  if (grep /$FuncName/,@input)
-	  {
-          push (@calledFunction, "$FuncName");
-      }
-    }else{
-      $actFuncName = $FuncName;
-    }
-  }
-  # now remove double entries of @calledFunction
-  undef my %saw;
-  @saw{@calledFunction} = ();
-  @calledFunction = sort keys %saw;  # remove sort if undesired
-  if (-e "$funcIndDir" . "/" . "$mainLatexFuncIndexFile")
-  {
-    # if the file exist, delete it
-    if ($deleteFile){
-      unlink("$funcIndDir" . "/" . "$mainLatexFuncIndexFile");
-      $deleteFile = 0;
-      open(DAT,">$funcIndDir" . "/" . "$mainLatexFuncIndexFile");
-      print DAT "\\begin{longtable}{ll}\n";
-      print DAT "\\textbf{main function}	&	\\textbf{called function} \\\\ \n";
-      print DAT "\\hline\n";
-	  $_ = $actFuncName;
-	  s/_/\\_/g;     # put a backslash for each underscore
-      print DAT "$_	";
-      $anzahl_elemente = @calledFunction;
-      if ($anzahl_elemente > 0)
-      {
-      	foreach (@calledFunction){          
-          	  s/_/\\_/g;  # put a backslash for each underscore
-	    	print DAT " &	$_\\\\ \n";
-  	  	}
-	  }else{
-        print DAT " & \\\\ \n";
-      }
-
-  	  close(DAT);
-
-    }else{
-      # file doesn't have to be deleted
-      open(DAT,">>$funcIndDir" . "/" . "$mainLatexFuncIndexFile");
-      print DAT "\\hline\n";
-      	  $_ = $actFuncName;
-	  s/_/\\_/g;     # put a backslash for each underscore
-      print DAT "$_	";
-      $anzahl_elemente = @calledFunction;
-      if ($anzahl_elemente > 0)
-      {
-      	foreach (@calledFunction){
-          s/_/\\_/g;  # put a backslash for each underscore
-	    	print DAT " &	$_\\\\ \n";
-  	  	}
-	  }else{
-        print DAT " & \\\\ \n";
-      }
-  	    close(DAT);
-    }
-
-  }else{
-    # File doesn't exist yet
-    open(DAT,">$funcIndDir" . "/" . "$mainLatexFuncIndexFile");
-    print DAT "\\begin{longtable}{l l}\n";
-    print DAT "\\textbf{main function}	&	\\textbf{called function} \\\\ \n";
-    print DAT "\\hline\n";
-          	  $_ = $actFuncName;
-	  s/_/\\_/g;    # put a backslash for each underscore
-      print DAT "$_	";
-      $anzahl_elemente = @calledFunction;
-      if ($anzahl_elemente > 0)
-      {
-      	foreach (@calledFunction){
-            s/_/\\_/g;  # put a backslash for each underscore
-	    	print DAT " &	$_\\\\ \n";
-  	  	}
-	  }else{
-        print DAT " & \\\\ \n";
-      }
-  	  close(DAT);
-  }
-#  print DAT "Function-File: $actFuncName\n";
-#  print DAT "=============================\n";
-#   foreach (@calledFunction){
-# 	print DAT "$_\n";
-#   }
-
-  @calledFunction = ();
-
-}
-  open(DAT,">>$funcIndDir" . "/" . "$mainLatexFuncIndexFile");
-  print DAT "\\end{longtable}\n";
-  close(DAT);
--- a/main/nnet/doc/latex/perl/createTestDocu.pl	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,145 +0,0 @@
-#!/usr/bin/perl -w
-
-use strict;
-use diagnostics;# Force verbose warning diagnostics.
-use warnings;
-use English;
-
-#
-# Modules from the Standard Perl Library.
-
-
-#
-# Own modules
-use analyzeOctaveSource;
-
-
-#--- DEFINE VARIABLES -------------------------------
-my $Dir = "D:\\daten\\octave\\neuroPackage\\0.1.9\\nnet\\inst";
-my $fileExt = "m";
-my $testDir = "D:\\daten\\octave\\neuroPackage\\0.1.9\\nnet/doc/latex/developers/tests";
-my $relTestDir = "tests/";
-my $testFileExt = "tex";
-my $mainLatexTestFile = "test.tex";
-my $chapter = "Test";
-#--- END DEFINE VARIABLES ---------------------------
-
-
-my @filesArray = ();
-my @filesName = ();
-my $fileName = "";
-my $savePath = "";
-my $nTestLines = 0;
-my $m = 0;
-
-# analyze directory structure
-my $obj = analyzeOctaveSource->new();
-my @DirArray = $obj->readDirTree($Dir,$fileExt);
-
-# analyze file structure
-my $nFiles = @DirArray;
-if ($nFiles>=1){ # if $nFiles==0 readDirTree will die
-  # if we are in this branch, we should check the files
-  # to found those which content tests (lines beginning with
-  # !% )
-  foreach (@DirArray){
-	  # open file and search for lines beginning with
-	  # !%
-	  if (-d $_){
-        # if directory, do nothing
-      }else{
-      	print "$_\n";
-      	sleep(0.1);
-      	
-	    open(FILE,$_) or die "File $_ not found!\n";
-	    my @fileContent = <FILE>;
-	    chomp(@fileContent);
-	    $nTestLines = @fileContent;
-        my @temp = grep /^%!/, @fileContent;
-        my $nTemp = @temp;
-        if ($nTemp>0){ # this means, 
-                       # test lines are available
-          # now create latex files without header
-          # take the same name like the *.m file
-          # and save in specified directory $testDir
-          # with file extens in $testFileExt
-          # use verbatim environment
-          @filesName = split(/\//, $_);
-          $fileName = $filesName[$#filesName];
-          # now remove file extension .m
-          @filesName = split(/\./,$fileName);
-          $savePath = ("$testDir" . "\\\\" . "$filesName[0]." . "$testFileExt");
-          open(OUTFILE, ">$savePath");
-          my $i = 0;
-          print OUTFILE "\\begin{verbatim}\n";
-          while ($i < $nTestLines){
-            if ($fileContent[$i]=~/^%!/){
-              print OUTFILE "$fileContent[$i]\n";            
-            }
-            $i++;
-		  } # END while ($i <= $#fileContent)
-		  print OUTFILE "\\end{verbatim}\n";
-		  close OUTFILE;
-		  
-		  ## now set entries in the main test latex file ..
-          my $mainTestFile = ("$testDir" . "\\\\" . "$mainLatexTestFile");
-          if ($m==0){
-            open(TESTFILE,">$mainTestFile") or die "Could not found file $mainTestFile!\n";
-            print TESTFILE "\\chapter{$chapter}\n\n";
-            
-			# test if back slash needed
-            # a back slash is needed if the sign "underscore"
-            # is used in the file name. This happens at each
-            # "sub function". There are two underscores!
-            my $tempString = "";
-            my $oldString = $filesName[0];
-			$_ = $filesName[0];
-			s/_/\\_/g; # s/ : search & replace pattern (everything between / /)
-					   # here: search underscore
-					   # if found, replace with \_ (two back slashes are needed
-					   # to get \_ as sign)
-					   # /g means: each occurence of pattern, otherwise, only one _
-					   # will be replaced
-
-			print "test file name: $_\n";
-            print TESTFILE "\\section{$_}\n";
-            $tempString = $relTestDir . $oldString;
-            print TESTFILE "\\input{$tempString}\n";
-          }else{
-            open(TESTFILE,">>$mainTestFile") or die "Could not found file $mainTestFile!\n";
-            # test if back slash needed
-            my $tempString = "";
-            my $oldString = $filesName[0];
-            # test if back slash needed
-            # a back slash is needed if the sign "underscore"
-            # is used in the file name. This happens at each
-            # "sub function". There are two underscores!
-            my $tempString = "";
-            my $oldString = $filesName[0];
-			$_ = $filesName[0];
-			s/_/\\_/g; # s/ : search & replace pattern (everything between / /)
-					   # here: search underscore
-					   # if found, replace with \_ (two back slashes are needed
-					   # to get \_ as sign)
-					   # /g means: each occurence of pattern, otherwise, only one _
-					   # will be replaced
-
-            print TESTFILE "\\section{$_}\n";
-            $tempString = $relTestDir . $oldString;
-            print TESTFILE "\\input{$tempString}\n";
-          }
-          $m++;
-          close TESTFILE;
-  
-		  
-        }# END if($nTemp>0)
-        close FILE;
-               
-	  }# END if(-d $_)
-	  
-	}# END foreach (@DirArray)
-
-}else{ # if $nFiles==0
-  print "No file found with valid file extension: .$fileExt.\n";
-  die;
-}
\ No newline at end of file
--- a/main/nnet/doc/latex/users/bibliography.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,35 +0,0 @@
-% Preamble
-
-%\documentclass[a4paper]{report}
-
-%\usepackage[ngerman]{babel}
-%\usepackage[T1]{fontenc}
-%\usepackage[ansinew]{inputenc}
-
-
-%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
-% start text here!!
-
-%\begin{document}
-
-\begin{thebibliography}{XXXXXXX}
-
-\bibitem [1]{1} John W. Eaton
-
-GNU Octave Manual, Edition 3, PDF-Version, February 1997
-
-\bibitem [2]{2} The MathWorks, Inc.
-
-MATLAB Help, MATLAB Version 7.1 (R14SP3), Neural Network Toolbox Version 4.0.6 (R14SP3) 
-
-\bibitem [3]{3} Christopher M. Bishop
-
-Neural Networks for Pattern Recognition, OXFORD University Press, Great Clarendon Streed, Oxford OX2 6DP,
-ISBN 0-19-853864-2, 2002
-
-\bibitem [4]{4} Martin T. Hagen, Howard B. Demuth, Mark H. Beale
-
-NEURAL NETWORK DESIGN, PWS Publishing Company, 20 Park Plaza, Boston, MA 02116-4324, ISBN 053494332-2, 1996
-
-\end{thebibliography}
-%\end{document}
\ No newline at end of file
--- a/main/nnet/doc/latex/users/examples/1/1.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,34 +0,0 @@
-\section{Example 1}
-You can find this example in the \textit{tests/MLP} directory of each release or from the subversion repository. I will do (more or less) a line by line walkthrough, so after this should be everything clear. I assume that you have some experience with multilayer perceptrons.
-
-\subsection{Introduction}
-Our problem can be solved with a monotonically increasing or decreasing surface. An input vector \textbf{p} (with 9 values) should be mapped onto one output value. Because we know that it can be solved with a monotonically increasing or decreasing surface, we can choose a 9-1-1 multi-layer perceptron (short: MLP).
-This means an MLP with 9 input neurons, only 1 hidden neuron and with 1 output neuron.
-
-\subsection{Code m-file}
-\input{examples/1/MLP9_1_1}
-
-\subsection{Walkthrough}
-Till line number 0023 there is realy nothing interesting.\\
-On line 0023 \& 0024 data will be loaded. This data matrix contains 13 columns. Column 4, 8 and 12 won't be used (this is because the datas are of a real world problem). Column 13 contains the target values.
-So on the lines 0049 till 0051 this will be splittet into the corresponding peaces. A short repetition about the datas: Each line is a data set with 9 input values and one target value. On line 0038 and 0039 the datas are transposed. So we have now in each column one data set.\\
-
-Now let's split the data matrix again in 3 pieces. The biggest part is for training the network. The second part for testing the trained network to be sure it's still possible to generalize with the net. And the third part, and the smallest one, for validate during training. This splitting happens on the lines 0041 till 0061.\\
-
-Line 0063 is the first special command from this toolbox. This command will be used to pre-standardize the input datas. Do it ever! Non linear transfer functions will squash the whole input range to an small second range e.g. the transfer function \textit{tansig} will squash the datas between -1 and +1.\\
-
-On line 0069 the next toolbox command will be used. This command \textit{min\_max} creates a $Rx2$ matrix of the complete input matrix. Don't ask me for what MATLAB(TM) this is using. I couldn't figure out it. One part is the number of input neurons, but for this, the range would not be needed. Who cares ;-)\\
-
-Now it's time to create a structure which holds the informations about the neural network. The command \textbf{newff} can do it for us. See the complete line and actually, please use it only on this way, each other try will fail! This means, you can change the number of input neurons, the number of hidden neurons and the number of output neurons of course. But don't change the train algorithm or the performance function.\\
-
-\textbf{saveMLPStruct} on line 0083 is a command which doesn't exist in MATLAB(TM). This will save the structure with the same informations you can see in MATLAB(TM) if you try to open the net-type.\\
-
-The validation part on line 0086 \& 0087 is important. The naming convention is for MATLAB(TM) compatibility. For validate, you have to define a structure with the name \textbf{VV}. Inside this structure you have to define actually \textbf{VV.P} \& \textbf{VV.T} for validate inputs and validate targets. Bye the way, you have to pre-standardize them like the training input matrix. Use for this the command \textbf{trastd} like on line 0090.\\
-
-\textbf{train} is the next toolbox command and of course one of the most important. Please also use this command like on line 0092. Nothing else will work.\\
-
-The second last step is to standardize again datas. This time the test datas. See line 0096 for this and the last step. Simulate the network. This can be done with the command \textbf{sim}. This will be a critical part if someone else will write a toolbox with this command name!\\
-
-I hope this short walkthrough will help for first steps. In next time, I will try to improve this documentation and of course, the toolbox commands. But time is realy rare.
-
-
--- a/main/nnet/doc/latex/users/examples/1/MLP9_1_1.m_template	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,96 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## Author: Michel D. Schmid
-
-
-## load data
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-	# this file contains 13 columns.
-	# The first 12 columns are the inputs
-	# the last column is the output,
-	# remove column 4, 8 and 12!
-	# 89 rows.
-
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-
-# now split the data matrix in 3 pieces, train data, test data and validate data
-# the proportion should be about 1/2 train, 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights, for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = [];
-mOutput(:,1:nValiSets) = [];
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = [];
-mOutput(:,1:nTestSets) = [];
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); # input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 1;
-nOutputNeurons = 1;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],\
-		{"tansig","purelin"},"trainlm","","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(:) = 1.5;
-MLPnet.LW{2,1}(:) = 0.5;
-MLPnet.b{1,1}(:) = 1.5;
-MLPnet.b{2,1}(:) = 0.5;
-
-saveMLPStruct(MLPnet,"MLP3test.txt");
-
-## define validation data new, for matlab compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-
-# make preparations for net test and test MLPnet
-#  standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-[simOut] = sim(net,mTestInputN);
-simOut
--- a/main/nnet/doc/latex/users/examples/1/MLP9_1_1.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,103 +0,0 @@
-\noindent
-\ttfamily
-\hlstd{}\hlline{00001\ }\hlslc{\#\# Copyright (C) 2006 Michel D. Schmid}\hlstd{\ \ }\hlslc{$<$michael.schmid@plexso.com$>$}\\
-\hlline{00002\ }\hlstd{}\hlslc{\#\#}\\
-\hlline{00003\ }\hlstd{}\hlslc{\#\#}\\
-\hlline{00004\ }\hlstd{}\hlslc{\#\# This program is free software; you can redistribute it and/or modify it}\\
-\hlline{00005\ }\hlstd{}\hlslc{\#\# under the terms of the GNU General Public License as published by}\\
-\hlline{00006\ }\hlstd{}\hlslc{\#\# the Free Software Foundation; either version 2, or (at your option)}\\
-\hlline{00007\ }\hlstd{}\hlslc{\#\# any later version.}\\
-\hlline{00008\ }\hlstd{}\hlslc{\#\#}\\
-\hlline{00009\ }\hlstd{}\hlslc{\#\# This program is distributed in the hope that it will be useful, but}\\
-\hlline{00010\ }\hlstd{}\hlslc{\#\# WITHOUT ANY WARRANTY; without even the implied warranty of}\\
-\hlline{00011\ }\hlstd{}\hlslc{\#\# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.}\hlstd{\ \ }\hlslc{See the GNU}\\
-\hlline{00012\ }\hlstd{}\hlslc{\#\# General Public License for more details.}\\
-\hlline{00013\ }\hlstd{}\hlslc{\#\#}\\
-\hlline{00014\ }\hlstd{}\hlslc{\#\# You should have received a copy of the GNU General Public License}\\
-\hlline{00015\ }\hlstd{}\hlslc{\#\# along with this program; see the file COPYING.}\hlstd{\ \ }\hlslc{If not, write to the Free}\\
-\hlline{00016\ }\hlstd{}\hlslc{\#\# Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA}\\
-\hlline{00017\ }\hlstd{}\hlslc{\#\# 02110{-}1301, USA.}\\
-\hlline{00018\ }\hlstd{}\\
-\hlline{00019\ }\hlslc{\#\# Author: Michel D. Schmid}\\
-\hlline{00020\ }\hlstd{}\\
-\hlline{00021\ }\\
-\hlline{00022\ }\hlslc{\#\# load data}\\
-\hlline{00023\ }\hlstd{mData }\hlsym{= }\hlstd{}\hlkwc{load}\hlstd{}\hlsym{(}\hlstd{}\hlstr{"mData.txt"}\hlstd{}\hlsym{,}\hlstd{}\hlstr{"mData"}\hlstd{}\hlsym{);}\\
-\hlline{00024\ }\hlstd{mData }\hlsym{= }\hlstd{mData.mData}\hlsym{;}\\
-\hlline{00025\ }\hlstd{}\hlsym{{[}}\hlstd{nRows}\hlsym{, }\hlstd{nColumns}\hlsym{{]} = }\hlstd{}\hlkwc{size}\hlstd{}\hlsym{(}\hlstd{mData}\hlsym{);}\\
-\hlline{00026\ }\hlstd{}\hlstd{\ \ \ \ }\hlstd{}\hlslc{\# this file contains 13 columns.}\\
-\hlline{00027\ }\hlstd{}\hlstd{\ \ \ \ }\hlstd{}\hlslc{\# The first 12 columns are the inputs}\\
-\hlline{00028\ }\hlstd{}\hlstd{\ \ \ \ }\hlstd{}\hlslc{\# the last column is the output,}\\
-\hlline{00029\ }\hlstd{}\hlstd{\ \ \ \ }\hlstd{}\hlslc{\# remove column 4, 8 and 12!}\\
-\hlline{00030\ }\hlstd{}\hlstd{\ \ \ \ }\hlstd{}\hlslc{\# 89 rows.}\\
-\hlline{00031\ }\hlstd{\\
-\hlline{00032\ }\\
-\hlline{00033\ }mOutput }\hlsym{= }\hlstd{}\hlstd{mData}\hlstd{}\hlsym{(:,}\hlstd{}\hlkwa{end}\hlstd{}\hlsym{);}\\
-\hlline{00034\ }\hlstd{mInput }\hlsym{= }\hlstd{}\hlstd{mData}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{}\hlkwa{end}\hlstd{}\hlsym{{-}}\hlstd{}\hlnum{1}\hlstd{}\hlsym{);}\\
-\hlline{00035\ }\hlstd{}\hlstd{mInput}\hlstd{}\hlsym{(:,{[}}\hlstd{}\hlnum{4 8 12}\hlstd{}\hlsym{{]}) = {[}{]}; }\hlstd{}\hlslc{\# delete column 4, 8 and 12}\\
-\hlline{00036\ }\hlstd{}\\
-\hlline{00037\ }\hlslc{\#\# now prepare data}\\
-\hlline{00038\ }\hlstd{mInput }\hlsym{= }\hlstd{mInput}\hlstd{';}\\
-\hlline{00039\ }\hlstd{mOutput = mOutput'}\hlstd{}\hlsym{;}\\
-\hlline{00040\ }\hlstd{}\\
-\hlline{00041\ }\hlslc{\# now split the data matrix in 3 pieces, train data, test data and validate data}\\
-\hlline{00042\ }\hlstd{}\hlslc{\# the proportion should be about 1/2 train, 1/3 test and 1/6 validate data}\\
-\hlline{00043\ }\hlstd{}\hlslc{\# in this neural network we have 12 weights, for each weight at least 3 train sets..}\\
-\hlline{00044\ }\hlstd{}\hlslc{\# (that's a rule of thumb like 1/2, 1/3 and 1/6)}\\
-\hlline{00045\ }\hlstd{}\hlslc{\# 1/2 of 89 = 44.5; let's take 44 for training}\\
-\hlline{00046\ }\hlstd{nTrainSets }\hlsym{= }\hlstd{}\hlkwc{floor}\hlstd{}\hlsym{(}\hlstd{nRows}\hlsym{/}\hlstd{}\hlnum{2}\hlstd{}\hlsym{);}\\
-\hlline{00047\ }\hlstd{}\hlslc{\# now the rest of the sets are again 100\%}\\
-\hlline{00048\ }\hlstd{}\hlslc{\# ==$>$ 2/3 for test sets and 1/3 for validate sets}\\
-\hlline{00049\ }\hlstd{nTestSets }\hlsym{= (}\hlstd{nRows}\hlsym{{-}}\hlstd{nTrainSets}\hlsym{)/}\hlstd{}\hlnum{3}\hlstd{}\hlsym{{*}}\hlstd{}\hlnum{2}\hlstd{}\hlsym{;}\\
-\hlline{00050\ }\hlstd{nValiSets }\hlsym{= }\hlstd{nRows}\hlsym{{-}}\hlstd{nTrainSets}\hlsym{{-}}\hlstd{nTestSets}\hlsym{;}\\
-\hlline{00051\ }\hlstd{\\
-\hlline{00052\ }mValiInput }\hlsym{= }\hlstd{}\hlstd{mInput}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{nValiSets}\hlsym{);}\\
-\hlline{00053\ }\hlstd{mValliOutput }\hlsym{= }\hlstd{}\hlstd{mOutput}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{nValiSets}\hlsym{);}\\
-\hlline{00054\ }\hlstd{}\hlstd{mInput}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{nValiSets}\hlsym{) = {[}{]};}\\
-\hlline{00055\ }\hlstd{}\hlstd{mOutput}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{nValiSets}\hlsym{) = {[}{]};}\\
-\hlline{00056\ }\hlstd{mTestInput }\hlsym{= }\hlstd{}\hlstd{mInput}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{nTestSets}\hlsym{);}\\
-\hlline{00057\ }\hlstd{mTestOutput }\hlsym{= }\hlstd{}\hlstd{mOutput}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{nTestSets}\hlsym{);}\\
-\hlline{00058\ }\hlstd{}\hlstd{mInput}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{nTestSets}\hlsym{) = {[}{]};}\\
-\hlline{00059\ }\hlstd{}\hlstd{mOutput}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{nTestSets}\hlsym{) = {[}{]};}\\
-\hlline{00060\ }\hlstd{mTrainInput }\hlsym{= }\hlstd{}\hlstd{mInput}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{nTrainSets}\hlsym{);}\\
-\hlline{00061\ }\hlstd{mTrainOutput }\hlsym{= }\hlstd{}\hlstd{mOutput}\hlstd{}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{nTrainSets}\hlsym{);}\\
-\hlline{00062\ }\hlstd{}\\
-\hlline{00063\ }\hlsym{{[}}\hlstd{mTrainInputN}\hlsym{,}\hlstd{cMeanInput}\hlsym{,}\hlstd{cStdInput}\hlsym{{]} = }\hlstd{}\hlkwc{prestd}\hlstd{}\hlsym{(}\hlstd{mTrainInput}\hlsym{);}\hlstd{}\hlslc{\# standardize inputs}\\
-\hlline{00064\ }\hlstd{}\\
-\hlline{00065\ }\hlslc{\#\# comments: there is no reason to standardize the outputs because we have only}\\
-\hlline{00066\ }\hlstd{}\hlslc{\# one output ...}\\
-\hlline{00067\ }\hlstd{}\\
-\hlline{00068\ }\hlslc{\# define the max and min inputs for each row}\\
-\hlline{00069\ }\hlstd{mMinMaxElements }\hlsym{= }\hlstd{}\hlkwc{min\textunderscore max}\hlstd{}\hlsym{(}\hlstd{mTrainInputN}\hlsym{); }\hlstd{}\hlslc{\# input matrix with (R x 2)...}\\
-\hlline{00070\ }\hlstd{}\\
-\hlline{00071\ }\hlslc{\#\# define network}\\
-\hlline{00072\ }\hlstd{nHiddenNeurons }\hlsym{= }\hlstd{}\hlnum{1}\hlstd{}\hlsym{;}\\
-\hlline{00073\ }\hlstd{nOutputNeurons }\hlsym{= }\hlstd{}\hlnum{1}\hlstd{}\hlsym{;}\\
-\hlline{00074\ }\hlstd{\\
-\hlline{00075\ }MLPnet }\hlsym{= }\hlstd{}\hlkwc{newff}\hlstd{}\hlsym{(}\hlstd{mMinMaxElements}\hlsym{,{[}}\hlstd{nHiddenNeurons nOutputNeurons}\hlsym{{]},}\hlstd{$\backslash$\\
-\hlline{00076\ }}\hlstd{\ \ \ \ \ \ \ \ }\hlstd{}\hlsym{\{}\hlstd{}\hlstr{"tansig"}\hlstd{}\hlsym{,}\hlstd{}\hlstr{"purelin"}\hlstd{}\hlsym{\},}\hlstd{}\hlstr{"trainlm"}\hlstd{}\hlsym{,}\hlstd{}\hlstr{""}\hlstd{}\hlsym{,}\hlstd{}\hlstr{"mse"}\hlstd{}\hlsym{);}\\
-\hlline{00077\ }\hlstd{}\hlslc{\#\# for test purpose, define weights by hand}\\
-\hlline{00078\ }\hlstd{MLPnet.IW}\hlsym{\{}\hlstd{}\hlnum{1}\hlstd{}\hlsym{,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{\}(:) = }\hlstd{}\hlnum{1.5}\hlstd{}\hlsym{;}\\
-\hlline{00079\ }\hlstd{MLPnet.LW}\hlsym{\{}\hlstd{}\hlnum{2}\hlstd{}\hlsym{,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{\}(:) = }\hlstd{}\hlnum{0.5}\hlstd{}\hlsym{;}\\
-\hlline{00080\ }\hlstd{MLPnet.b}\hlsym{\{}\hlstd{}\hlnum{1}\hlstd{}\hlsym{,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{\}(:) = }\hlstd{}\hlnum{1.5}\hlstd{}\hlsym{;}\\
-\hlline{00081\ }\hlstd{MLPnet.b}\hlsym{\{}\hlstd{}\hlnum{2}\hlstd{}\hlsym{,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{\}(:) = }\hlstd{}\hlnum{0.5}\hlstd{}\hlsym{;}\\
-\hlline{00082\ }\hlstd{}\\
-\hlline{00083\ }\hlkwc{saveMLPStruct}\hlstd{}\hlsym{(}\hlstd{MLPnet}\hlsym{,}\hlstd{}\hlstr{"MLP3test.txt"}\hlstd{}\hlsym{);}\\
-\hlline{00084\ }\hlstd{}\\
-\hlline{00085\ }\hlslc{\#\# define validation data new, for matlab compatibility}\\
-\hlline{00086\ }\hlstd{VV.P }\hlsym{= }\hlstd{mValiInput}\hlsym{;}\\
-\hlline{00087\ }\hlstd{VV.T }\hlsym{= }\hlstd{mValliOutput}\hlsym{;}\\
-\hlline{00088\ }\hlstd{}\\
-\hlline{00089\ }\hlslc{\#\# standardize also the validate data}\\
-\hlline{00090\ }\hlstd{VV.P }\hlsym{= }\hlstd{}\hlkwc{trastd}\hlstd{}\hlsym{(}\hlstd{VV.P}\hlsym{,}\hlstd{cMeanInput}\hlsym{,}\hlstd{cStdInput}\hlsym{);}\\
-\hlline{00091\ }\hlstd{}\\
-\hlline{00092\ }\hlsym{{[}}\hlstd{net}\hlsym{{]} = }\hlstd{}\hlkwc{train}\hlstd{}\hlsym{(}\hlstd{MLPnet}\hlsym{,}\hlstd{mTrainInputN}\hlsym{,}\hlstd{mTrainOutput}\hlsym{,{[}{]},{[}{]},}\hlstd{VV}\hlsym{);}\\
-\hlline{00093\ }\hlstd{}\\
-\hlline{00094\ }\hlslc{\# make preparations for net test and test MLPnet}\\
-\hlline{00095\ }\hlstd{}\hlslc{\#}\hlstd{\ \ }\hlslc{standardise input \& output test data}\\
-\hlline{00096\ }\hlstd{}\hlsym{{[}}\hlstd{mTestInputN}\hlsym{{]} = }\hlstd{}\hlkwc{trastd}\hlstd{}\hlsym{(}\hlstd{mTestInput}\hlsym{,}\hlstd{cMeanInput}\hlsym{,}\hlstd{cStdInput}\hlsym{);}\\
-\hlline{00097\ }\hlstd{}\\
-\hlline{00098\ }\hlsym{{[}}\hlstd{simOut}\hlsym{{]} = }\hlstd{}\hlkwc{sim}\hlstd{}\hlsym{(}\hlstd{net}\hlsym{,}\hlstd{mTestInputN}\hlsym{);}\\
-\hlline{00099\ }\hlstd{simOut}\\
-\mbox{}
-\normalfont
\ No newline at end of file
--- a/main/nnet/doc/latex/users/examples/1/mData.txt	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,94 +0,0 @@
-# Created by Octave 2.9.5, Wed May 24 10:33:36 2006     <sim@TL3124>
-# name: mData
-# type: matrix
-# rows: 89
-# columns: 13
- 306 286 2 0 12 61 2 0 3 28 4 0 2
- 368 188 1 0 6 49 0 0 3 37 0 0 1
- 511 73 0 0 40 21 0 0 16 22 0 0 1
- 418 43 0 0 34 30 1 0 9 27 5 0 1
- 299 173 1 0 8 63 1 0 1 37 3 0 1
- 312 253 0 0 2 63 2 0 0 35 3 0 1
- 492 98 0 0 7 23 0 0 13 36 0 0 2
- 506 64 0 0 32 23 0 0 13 29 0 0 3
- 476 41 0 0 32 5 0 0 19 26 0 0 3
- 483 66 0 0 17 16 0 0 10 28 2 0 3
- 429 44 0 0 37 19 0 0 23 19 0 0 1
- 521 137 0 0 16 17 0 0 7 24 0 0 2
- 340 163 1 0 16 72 3 0 3 31 1 0 1
- 323 177 0 0 8 68 4 0 0 37 2 0 1
- 344 240 2 0 4 43 1 0 4 39 6 0 2
- 459 22 0 0 46 14 0 0 27 17 0 0 1
- 487 36 0 0 34 10 0 0 19 27 0 0 3
- 331 169 0 0 19 66 1 0 0 34 1 0 1
- 541 269 1 0 12 5 0 0 6 20 0 0 3
- 475 23 0 0 37 5 0 0 26 14 0 0 3
- 475 186 1 0 15 28 0 0 5 35 0 0 2
- 496 319 0 0 2 4 0 0 2 14 0 0 3
- 525 41 0 0 38 9 0 0 37 18 0 0 3
- 484 37 0 0 50 13 0 0 29 18 0 0 2
- 511 55 0 0 32 15 0 0 23 21 0 0 2
- 515 44 0 0 29 16 0 0 19 31 0 0 2
- 478 101 0 0 15 34 0 0 11 40 0 0 2
- 429 433 3 0 8 11 0 0 5 15 0 0 3
- 471 8 0 0 58 2 0 0 25 13 0 0 1
- 303 269 3 0 10 66 2 0 1 32 7 0 2
- 445 74 0 0 19 30 0 0 4 43 0 0 1
- 488 83 0 0 34 18 0 0 15 34 0 0 3
- 264 298 4 0 7 68 10 0 1 30 7 0 2
- 489 44 0 0 21 12 0 0 29 21 0 0 1
- 475 34 0 0 38 13 0 0 28 18 0 0 3
- 492 14 0 0 44 4 0 0 40 7 0 0 1
- 454 173 1 0 14 21 0 0 2 42 0 0 2
- 306 285 5 0 5 68 3 0 1 33 1 0 2
- 508 64 0 0 31 14 0 0 17 33 0 0 2
- 477 61 0 0 36 13 0 0 22 15 0 0 2
- 349 224 1 0 3 66 3 0 2 41 1 0 1
- 447 189 0 0 11 34 0 0 5 45 0 0 2
- 496 34 0 0 49 6 0 0 31 11 0 0 3
- 484 222 0 0 16 12 0 0 9 24 0 0 2
- 412 50 0 0 25 47 0 0 11 29 0 0 1
- 316 184 0 0 13 65 3 0 5 34 3 0 1
- 345 163 1 0 17 57 2 0 2 32 1 0 1
- 285 273 2 0 7 50 13 0 1 28 10 0 1
- 317 179 0 0 11 64 0 0 4 35 0 0 1
- 500 68 0 0 23 15 0 0 18 30 0 0 1
- 495 34 0 0 39 4 0 0 21 21 0 0 3
- 387 294 1 0 10 37 0 0 4 30 5 0 2
- 258 236 2 0 2 72 7 0 0 26 8 0 1
- 423 25 0 0 26 16 0 0 16 29 0 0 3
- 501 32 0 0 40 11 0 0 24 20 0 0 3
- 459 37 0 0 46 4 0 0 32 16 0 0 3
- 511 48 0 0 35 7 0 0 27 15 0 0 2
- 295 271 3 0 6 71 5 0 3 29 4 0 1
- 502 34 0 0 25 9 0 0 23 11 0 0 3
- 458 36 0 0 12 7 0 0 24 19 0 0 2
- 470 273 1 0 9 17 0 0 2 32 0 0 3
- 477 30 0 0 24 14 0 0 18 26 0 0 3
- 406 77 1 0 15 24 2 0 6 37 1 0 1
- 291 251 2 0 8 53 2 0 0 43 2 0 1
- 407 51 0 0 28 47 0 0 5 35 0 0 1
- 390 47 0 0 17 45 1 0 3 33 2 0 1
- 347 123 1 0 7 52 7 0 4 38 3 0 1
- 300 175 0 0 8 71 2 0 0 35 0 0 1
- 473 59 0 0 44 15 0 0 22 23 0 0 1
- 487 35 0 0 34 12 0 0 29 24 0 0 3
- 532 57 0 0 19 6 0 0 28 15 0 0 2
- 286 207 1 0 5 74 2 0 1 38 4 0 1
- 405 199 0 0 13 37 0 0 6 45 0 0 2
- 337 177 0 0 11 47 1 0 4 39 3 0 1
- 527 20 0 0 32 6 0 0 23 16 0 0 1
- 326 218 1 0 4 71 2 0 1 33 1 0 1
- 486 14 0 0 41 4 0 0 40 8 0 0 1
- 420 43 0 0 26 30 9 0 9 24 4 0 1
- 286 293 3 0 7 83 9 0 1 22 5 0 1
- 395 59 0 0 30 33 0 0 4 36 0 0 1
- 506 24 0 0 34 12 0 0 26 17 0 0 3
- 396 217 0 0 16 43 2 0 6 30 0 0 2
- 457 15 0 0 44 1 0 0 30 13 0 0 1
- 472 139 0 0 16 25 0 0 11 38 0 0 2
- 493 21 0 0 33 16 0 0 22 19 0 0 3
- 311 236 0 0 6 66 4 0 2 24 9 0 1
- 490 23 0 0 34 6 0 0 30 14 0 0 1
- 485 29 0 0 33 7 0 0 14 20 0 0 3
- 481 43 0 0 38 11 0 0 21 24 0 0 2
--- a/main/nnet/doc/latex/users/examples/2/2.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,33 +0,0 @@
-\section{Example 2}
-You can find this example in this directory but renamed to \textit{MLP9\_1\_1.m\_template}.
-I will explain only differencies to the example1, so please read it first, if you haven't.
-
-\subsection{Introduction}
-Our problem can be solved with a monotonically increasing or decreasing surface. An input vector \textbf{p}
-(with 9 values) should be mapped onto one output value.
-Because we know that it can be solved with a monotonically increasing or decreasing surface,
-we can choose a 9-1-1 multi-layer perceptron (short: MLP).
-This means an MLP with 9 input neurons, only 1 hidden neuron and with 1 output neuron.
-
-\subsection{Code m-file}
-\input{examples/2/MLP9_1_1}
-
-\subsection{Walkthrough}
-The difference to the example1 starts below the line number 0035.\\
-The difference concerns only the pre-processing section where the data set is splittet into 
-the three subsets. This time, the command \textbf{subset} is used, which makes the complete
-example about 19 lines shorter!
-On line 0023 \& 0024 data will be loaded. This data matrix contains 13 columns.
-Column 4, 8 and 12 won't be used (this is because the datas are of a real world problem).
-Column 13 contains the target values.\\
-
-Now on line 35 we have to merge the input and output targets again. Subset will take the complete
-matrix as argument! On line 42 happens the complete magic :-). Subset will return three 
-subsets containing each time the input and output arguments. So this part must be splitet once more!
-But this is very easy and happens at some specific positions below.\\
-
-That's it, \textbf{subset} will help you to write short scripts!
-
-
-
-
--- a/main/nnet/doc/latex/users/examples/2/MLP9_1_1.m_template	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,78 +0,0 @@
-## Copyright (C) 2008 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## Author: Michel D. Schmid
-
-
-## load data
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-	# this file contains 13 columns.
-	# The first 12 columns are the inputs
-	# the last column is the output,
-	# remove column 4, 8 and 12!
-	# 89 rows.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-mData = [mInput mOutput];
-
-# now split the data matrix in 3 pieces, train data, test data and validate data
-# the proportion should be about 1/2 train, 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights, for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-[mTrain,mTest,mVali] = subset(mData',1);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrain(1:end-1,:));# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); # input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 1;
-nOutputNeurons = 1;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],\
-		{"tansig","purelin"},"trainlm","","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(:) = 1.5;
-MLPnet.LW{2,1}(:) = 0.5;
-MLPnet.b{1,1}(:) = 1.5;
-MLPnet.b{2,1}(:) = 0.5;
-
-saveMLPStruct(MLPnet,"MLP3test.txt");
-
-## define validation data new, for matlab compatibility
-VV.P = mVali(1:end-1,:);
-VV.T = mVali(end,:);
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-[net] = train(MLPnet,mTrainInputN,mTrain(end,:),[],[],VV);
-
-# make preparations for net test and test MLPnet
-#  standardise input & output test data
-[mTestInputN] = trastd(mTest(1:end-1,:),cMeanInput,cStdInput);
-
-[simOut] = sim(net,mTestInputN);
-simOut
--- a/main/nnet/doc/latex/users/examples/2/MLP9_1_1.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,89 +0,0 @@
-\noindent
-\ttfamily
-\hlstd{}\hlline{00001\ }\hlslc{\#\# Copyright (C) 2008 Michel D. Schmid}\hlstd{\ \ }\hlslc{$<$michael.schmid@plexso.com$>$}\\
-\hlline{00002\ }\hlstd{}\hlslc{\#\#}\\
-\hlline{00003\ }\hlstd{}\hlslc{\#\#}\\
-\hlline{00004\ }\hlstd{}\hlslc{\#\#}\\
-\hlline{00005\ }\hlstd{}\hlslc{\#\# This program is free software;you can redistribute it and/or modify it}\\
-\hlline{00006\ }\hlstd{}\hlslc{\#\# under the terms of the GNU General Public License as published by}\\
-\hlline{00007\ }\hlstd{}\hlslc{\#\# the Free Software Foundation; either version 2, or (at your option)}\\
-\hlline{00008\ }\hlstd{}\hlslc{\#\# any later version.}\\
-\hlline{00009\ }\hlstd{}\hlslc{\#\#}\\
-\hlline{00010\ }\hlstd{}\hlslc{\#\# This program is distributed in the hope that it will be useful, but}\\
-\hlline{00011\ }\hlstd{}\hlslc{\#\# WITHOUT ANY WARRANTY; without even the implied warranty of}\\
-\hlline{00012\ }\hlstd{}\hlslc{\#\# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.}\hlstd{\ \ }\hlslc{See the GNU}\\
-\hlline{00013\ }\hlstd{}\hlslc{\#\# General Public License for more details.}\\
-\hlline{00014\ }\hlstd{}\hlslc{\#\#\\}
-\hlline{00015\ }\hlstd{}\hlslc{\#\# You should have received a copy of the GNU General Public License}\\
-\hlline{00016\ }\hlstd{}\hlslc{\#\# along with this program; see the file COPYING.}\hlstd{\ \ }\hlslc{If not, see}\\
-\hlline{00017\ }\hlstd{}\hlslc{\#\# http://www.gnu.org/licenses.}\\
-\hlline{00018\ }\hlstd{}\\
-\hlline{00019\ }\hlstd{}\hlslc{\#\# Author: Michel D. Schmid}\\
-\hlline{00020\ }\hlstd{}\\
-\hlline{00021\ }\hlstd{}\\
-\hlline{00022\ }\hlstd{}\hlslc{\#\# load data}\\
-\hlline{00023\ }\hlstd{mData }\hlsym{= }\hlstd{}\hlkwc{load}\hlstd{}\hlsym{(}\hlstd{}\hlstr{"mData.txt"}\hlstd{}\hlsym{,}\hlstd{}\hlstr{"mData"}\hlstd{}\hlsym{);}\\
-\hlline{00024\ }\hlstd{mData }\hlsym{= }\hlstd{mData.mData}\hlsym{;}\\
-\hlline{00025\ }\hlstd{}\hlsym{{[}}\hlstd{nRows}\hlsym{, }\hlstd{nColumns}\hlsym{{]} = }\hlstd{}\hlkwa{size}\hlstd{}\hlsym{(}\hlstd{mData}\hlsym{);}\\
-\hlline{00026\ }\hlstd{}\hlstd{\ \ \ \ }\hlslc{\# this file contains 13 columns.}\\
-\hlline{00027\ }\hlstd{\ \ \ \ }\hlslc{\# The first 12 columns are the inputs}\\
-\hlline{00028\ }\hlstd{\ \ \ \ }\hlslc{\# the last column is the output,}\\
-\hlline{00029\ }\hlstd{}\hlstd{\ \ \ \ }\hlslc{\# remove column 4, 8 and 12!}\\
-\hlline{00030\ }\hlstd{}\hlstd{\ \ \ \ }\hlslc{\# 89 rows.}\\
-\hlline{00031\ }\\
-\hlline{00032\ }\hlstd{mOutput }\hlsym{= }\hlstd{mData}\hlsym{(:,}\hlstd{}\hlkwa{end}\hlstd{}\hlsym{);}\\
-\hlline{00033\ }\hlstd{mInput }\hlsym{= }\hlstd{mData}\hlsym{(:,}\hlstd{}\hlnum{1}\hlstd{}\hlsym{:}\hlstd{}\hlkwa{end}\hlstd{}\hlsym{{-}}\hlstd{}\hlnum{1}\hlstd{}\hlsym{);}\\
-\hlline{00034\ }\hlstd{mInput}\hlsym{(:,{[}}\hlstd{}\hlnum{4 8 12}\hlstd{}\hlsym{{]}) = {[}{]}; }\hlslc{\# delete column 4, 8 and 12}\\
-\hlline{00035\ }\hlstd{\\
-\hlline{00036\ }mData }\hlsym{= {[}}\hlstd{mInput mOutput}\hlsym{{]};}\\
-\hlline{00037\ }\hlstd{}\\
-\hlline{00038\ }\hlstd{}\hlslc{\# now split the data matrix in 3 pieces, train data, test data and validate}\\
-\hlline{00039\ }\hlstd{}\hlslc{data}\\
-\hlline{00040\ }\hlstd{}\hlslc{\# the proportion should be about 1/2 train, 1/3 test and 1/6 validate data}\\
-\hlline{00041\ }\hlstd{}\hlslc{\# in this neural network we have 12 weights, for each weight at least 3}\\
-\hlline{00042\ }\hlstd{}\hlslc{train sets..}\\ 
-\hlline{00043\ }\hlstd{}\hlslc{\# that's a rule of thumb like 1/2, 1/3 and 1/6}\\
-\hlline{00044\ }\hlslc{\# 1/2 of 89 = 44.5; let's take 44 for training}\\
-\hlline{00045\ }\hlsym{{[}}\hlstd{mTrain}\hlsym{,}\hlstd{mTest}\hlsym{,}\hlstd{mVali}\hlsym{{]} = }\hlkwc{subset}\hlsym{(}\hlstd{mData',}\hlnum{1}\hlstd{);}\\
-\hlline{00046\ }\hlstd{}\\
-\hlline{00047\ }\hlstd{[mTrainInputN,cMeanInput,cStdInput] =\hlkwc{ prestd}(mTrain(\hlnum{1}:\hlkwa{end}-\hlnum{1},:));}\\
-\hlline{00048\ }\hlstd{\ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ }\hlslc{\#standardize inputs}\\
-\hlline{00049\ }\hlstd{}\\
-\hlline{00050\ }\hlslc{\#\# comments: there is no reason to standardize the outputs because we have}\\
-\hlline{00051\ }\hlslc{only}\\
-\hlline{00052\ }\hlslc{\# one output ...}\\
-\hlline{00053\ }\hlstd{}\\
-\hlline{00054\ }\hlstd{}\hlslc{\# define the max and min inputs for each row}\\
-\hlline{00055\ }\hlstd{mMinMaxElements = \hlkwc{min\textunderscore max}(mTrainInputN);} \hlslc{\# input matrix with (R x 2)...}\\
-\hlline{00056\ }\hlstd{}\\
-\hlline{00057\ }\hlstd{}\hlslc{\#\# define network}\\
-\hlline{00058\ }\hlstd{nHiddenNeurons = 1;}\\
-\hlline{00059\ }\hlstd{nOutputNeurons = 1;}\\
-\hlline{00060\ }\hlstd{}\\
-\hlline{00061\ }\hlstd{MLPnet = \hlkwc{newff}(mMinMaxElements,[nHiddenNeurons nOutputNeurons],$\backslash$}\\
-\hlline{00062\ }\hlstd{}\hlstd{\ \ \ \ \ \ \ \ }\hlstd{\{\hlstr{"tansig"},\hlstr{"purelin"}\},\hlstr{"trainlm"},\hlstr{""},\hlstr{"mse"});}\\
-\hlline{00063\ }\hlslc{\#\# for test purpose, define weights by hand}\\
-\hlline{00064\ }\hlstd{MLPnet.IW\{1,1\}(:) = 1.5;}\\
-\hlline{00065\ }\hlstd{MLPnet.LW\{2,1\}(:) = 0.5;}\\
-\hlline{00066\ }\hlstd{MLPnet.b\{1,1\}(:) = 1.5;}\\
-\hlline{00067\ }\hlstd{MLPnet.b\{2,1\}(:) = 0.5;}\\
-\hlline{00068\ }\hlstd{}\\
-\hlline{00069\ }\hlkwc{saveMLPStruct}\hlstd{(MLPnet,"MLP3test.txt");}\\
-\hlline{00070\ }\hlstd{}\\
-\hlline{00071\ }\hlslc{\#\# define validation data new, for matlab compatibility}\\
-\hlline{00072\ }\hlstd{VV.P = mVali(1:end{-}1,:);}\\
-\hlline{00073\ }\hlstd{VV.T = mVali(end,:);}\\
-\hlline{00074\ }\hlstd{}\\
-\hlline{00075\ }\hlslc{\#\# standardize also the validate data}\\
-\hlline{00076\ }\hlstd{VV.P = trastd(VV.P,cMeanInput,cStdInput);}\\
-\hlline{00077\ }\hlstd{}\\
-\hlline{00078\ }\hlstd{[net] = \hlkwc{train}(MLPnet,mTrainInputN,mTrain(end,:),[],[],VV);}\\
-\hlline{00079\ }\hlstd{}\\
-\hlline{00080\ }\hlslc{\# make preparations for net test and test MLPnet}\\
-\hlline{00081\ }\hlslc{\#\hlstd{\ \ }standardise input \& output test data}\\
-\hlline{00082\ }\hlstd{[mTestInputN] = \hlkwc{trastd}(mTest(1:end-1,:),cMeanInput,cStdInput);}\\
-\hlline{00083\ }\hlstd{}\\
-\hlline{00084\ }\hlstd{[simOut] = \hlkwc{sim}(net,mTestInputN);}\\
-\hlline{00085\ }\hlstd{simOut}\hlstd{}\\
-\mbox{}
-\normalfont
--- a/main/nnet/doc/latex/users/examples/2/highlight.sty	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,20 +0,0 @@
-% Style definition file generated by highlight 2.6.0, http://www.andre-simon.de/ 
-
-% Highlighting theme definition: 
-
-\newcommand{\hlstd}[1]{\textcolor[rgb]{0,0,0}{#1}}
-\newcommand{\hlnum}[1]{\textcolor[rgb]{0.5,0,0.5}{\bf{#1}}}
-\newcommand{\hlesc}[1]{\textcolor[rgb]{1,0,1}{\bf{#1}}}
-\newcommand{\hlstr}[1]{\textcolor[rgb]{0.65,0.52,0}{#1}}
-\newcommand{\hldstr}[1]{\textcolor[rgb]{0,0,1}{#1}}
-\newcommand{\hlslc}[1]{\textcolor[rgb]{0.95,0.47,0}{#1}}
-\newcommand{\hlcom}[1]{\textcolor[rgb]{1,0.5,0}{#1}}
-\newcommand{\hldir}[1]{\textcolor[rgb]{0,0.5,0.75}{\bf{#1}}}
-\newcommand{\hlsym}[1]{\textcolor[rgb]{1,0,0.5}{\bf{#1}}}
-\newcommand{\hlline}[1]{\textcolor[rgb]{0.19,0.19,0.19}{#1}}
-\newcommand{\hlkwa}[1]{\textcolor[rgb]{0.73,0.47,0.47}{\bf{#1}}}
-\newcommand{\hlkwb}[1]{\textcolor[rgb]{0.5,0.5,0.75}{\bf{#1}}}
-\newcommand{\hlkwc}[1]{\textcolor[rgb]{0,0.5,0.75}{#1}}
-\newcommand{\hlkwd}[1]{\textcolor[rgb]{0,0.27,0.4}{#1}}
-\definecolor{bgcolor}{rgb}{0.93,0.93,0.93}
-
--- a/main/nnet/doc/latex/users/examples/2/mData.txt	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,94 +0,0 @@
-# Created by Octave 2.9.5, Wed May 24 10:33:36 2006     <sim@TL3124>
-# name: mData
-# type: matrix
-# rows: 89
-# columns: 13
- 306 286 2 0 12 61 2 0 3 28 4 0 2
- 368 188 1 0 6 49 0 0 3 37 0 0 1
- 511 73 0 0 40 21 0 0 16 22 0 0 1
- 418 43 0 0 34 30 1 0 9 27 5 0 1
- 299 173 1 0 8 63 1 0 1 37 3 0 1
- 312 253 0 0 2 63 2 0 0 35 3 0 1
- 492 98 0 0 7 23 0 0 13 36 0 0 2
- 506 64 0 0 32 23 0 0 13 29 0 0 3
- 476 41 0 0 32 5 0 0 19 26 0 0 3
- 483 66 0 0 17 16 0 0 10 28 2 0 3
- 429 44 0 0 37 19 0 0 23 19 0 0 1
- 521 137 0 0 16 17 0 0 7 24 0 0 2
- 340 163 1 0 16 72 3 0 3 31 1 0 1
- 323 177 0 0 8 68 4 0 0 37 2 0 1
- 344 240 2 0 4 43 1 0 4 39 6 0 2
- 459 22 0 0 46 14 0 0 27 17 0 0 1
- 487 36 0 0 34 10 0 0 19 27 0 0 3
- 331 169 0 0 19 66 1 0 0 34 1 0 1
- 541 269 1 0 12 5 0 0 6 20 0 0 3
- 475 23 0 0 37 5 0 0 26 14 0 0 3
- 475 186 1 0 15 28 0 0 5 35 0 0 2
- 496 319 0 0 2 4 0 0 2 14 0 0 3
- 525 41 0 0 38 9 0 0 37 18 0 0 3
- 484 37 0 0 50 13 0 0 29 18 0 0 2
- 511 55 0 0 32 15 0 0 23 21 0 0 2
- 515 44 0 0 29 16 0 0 19 31 0 0 2
- 478 101 0 0 15 34 0 0 11 40 0 0 2
- 429 433 3 0 8 11 0 0 5 15 0 0 3
- 471 8 0 0 58 2 0 0 25 13 0 0 1
- 303 269 3 0 10 66 2 0 1 32 7 0 2
- 445 74 0 0 19 30 0 0 4 43 0 0 1
- 488 83 0 0 34 18 0 0 15 34 0 0 3
- 264 298 4 0 7 68 10 0 1 30 7 0 2
- 489 44 0 0 21 12 0 0 29 21 0 0 1
- 475 34 0 0 38 13 0 0 28 18 0 0 3
- 492 14 0 0 44 4 0 0 40 7 0 0 1
- 454 173 1 0 14 21 0 0 2 42 0 0 2
- 306 285 5 0 5 68 3 0 1 33 1 0 2
- 508 64 0 0 31 14 0 0 17 33 0 0 2
- 477 61 0 0 36 13 0 0 22 15 0 0 2
- 349 224 1 0 3 66 3 0 2 41 1 0 1
- 447 189 0 0 11 34 0 0 5 45 0 0 2
- 496 34 0 0 49 6 0 0 31 11 0 0 3
- 484 222 0 0 16 12 0 0 9 24 0 0 2
- 412 50 0 0 25 47 0 0 11 29 0 0 1
- 316 184 0 0 13 65 3 0 5 34 3 0 1
- 345 163 1 0 17 57 2 0 2 32 1 0 1
- 285 273 2 0 7 50 13 0 1 28 10 0 1
- 317 179 0 0 11 64 0 0 4 35 0 0 1
- 500 68 0 0 23 15 0 0 18 30 0 0 1
- 495 34 0 0 39 4 0 0 21 21 0 0 3
- 387 294 1 0 10 37 0 0 4 30 5 0 2
- 258 236 2 0 2 72 7 0 0 26 8 0 1
- 423 25 0 0 26 16 0 0 16 29 0 0 3
- 501 32 0 0 40 11 0 0 24 20 0 0 3
- 459 37 0 0 46 4 0 0 32 16 0 0 3
- 511 48 0 0 35 7 0 0 27 15 0 0 2
- 295 271 3 0 6 71 5 0 3 29 4 0 1
- 502 34 0 0 25 9 0 0 23 11 0 0 3
- 458 36 0 0 12 7 0 0 24 19 0 0 2
- 470 273 1 0 9 17 0 0 2 32 0 0 3
- 477 30 0 0 24 14 0 0 18 26 0 0 3
- 406 77 1 0 15 24 2 0 6 37 1 0 1
- 291 251 2 0 8 53 2 0 0 43 2 0 1
- 407 51 0 0 28 47 0 0 5 35 0 0 1
- 390 47 0 0 17 45 1 0 3 33 2 0 1
- 347 123 1 0 7 52 7 0 4 38 3 0 1
- 300 175 0 0 8 71 2 0 0 35 0 0 1
- 473 59 0 0 44 15 0 0 22 23 0 0 1
- 487 35 0 0 34 12 0 0 29 24 0 0 3
- 532 57 0 0 19 6 0 0 28 15 0 0 2
- 286 207 1 0 5 74 2 0 1 38 4 0 1
- 405 199 0 0 13 37 0 0 6 45 0 0 2
- 337 177 0 0 11 47 1 0 4 39 3 0 1
- 527 20 0 0 32 6 0 0 23 16 0 0 1
- 326 218 1 0 4 71 2 0 1 33 1 0 1
- 486 14 0 0 41 4 0 0 40 8 0 0 1
- 420 43 0 0 26 30 9 0 9 24 4 0 1
- 286 293 3 0 7 83 9 0 1 22 5 0 1
- 395 59 0 0 30 33 0 0 4 36 0 0 1
- 506 24 0 0 34 12 0 0 26 17 0 0 3
- 396 217 0 0 16 43 2 0 6 30 0 0 2
- 457 15 0 0 44 1 0 0 30 13 0 0 1
- 472 139 0 0 16 25 0 0 11 38 0 0 2
- 493 21 0 0 33 16 0 0 22 19 0 0 3
- 311 236 0 0 6 66 4 0 2 24 9 0 1
- 490 23 0 0 34 6 0 0 30 14 0 0 1
- 485 29 0 0 33 7 0 0 14 20 0 0 3
- 481 43 0 0 38 11 0 0 21 24 0 0 2
--- a/main/nnet/doc/latex/users/examples/examples.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,7 +0,0 @@
-\chapter{Examples}
-
-
-
-
-\input{examples/1/1}
-\input{examples/2/2}
\ No newline at end of file
--- a/main/nnet/doc/latex/users/introduction.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,16 +0,0 @@
-\chapter{Introduction}
-
-\section{Compatibility to Matlab's \texttrademark Neural Network Toolbox}
-The compatibility is one of the strongest targets during developing this toolbox.
-If I have to develope an incompatibility e.g. in naming the functions, it will be descriped
-in this documentation. Even though it should be clear that I can't make a one to one copy.  First,
-the m-files are copyrighted and second, octave doesn't yet support the object oriented-programming techonology.\\
-
-If you find a bug, any not described incompatibility or have some suggestions, please write me at
-michael.schmid@plexso.com. This will help improving this toolbox.
-
-
-\input{numbering}
-
-\input{knownIncompatibilities}
-
--- a/main/nnet/doc/latex/users/knownIncompatibilities.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,11 +0,0 @@
-\section{Known incompatibilities}
-\label{chap:intro:sec:knownIncompatibilities}
-
-
-\subsection{Function names}
-
-\subsubsection{minmax}
-\textit{minmax} is in this toolbox called \textit{min\_max}. This is because Octave already has
-a function whichs name is \textit{minmax}. This is a c file and the functions \textit{min} and \textit{max} are therein realized.
-
-
--- a/main/nnet/doc/latex/users/neuralNetworkPackageForOctaveUsersGuide.tcp	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,12 +0,0 @@
-[FormatInfo]
-Type=TeXnicCenterProjectInformation
-Version=4
-
-[ProjectInfo]
-MainFile=neuralNetworkPackageForOctaveUsersGuide.tex
-UseBibTeX=0
-UseMakeIndex=1
-ActiveProfile=LaTeX => PDF
-ProjectLanguage=de
-ProjectDialect=DE
-
--- a/main/nnet/doc/latex/users/neuralNetworkPackageForOctaveUsersGuide.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,102 +0,0 @@
-% Preambel
-\documentclass[a4paper,openany]{report}
-
-
-%\usepackage{a4wide}
-\usepackage[ansinew]{inputenc}
-\usepackage[T1]{fontenc}
-\RequirePackage{ifpdf}
-
-\usepackage{hyperref}
-	\hypersetup{%
-  colorlinks=true,   % activates colored references
-  pdfpagemode=None,  % PDF-Viewer starts without content et.al.
-  pdfstartview=FitH, % PDF-Viewer uses a defined page width
-  %linkbordercolor=111,
-  % citebordercolor=111,
-  citecolor=blue,
-  linkcolor=blue}
-
-\ifpdf
-  \usepackage[pdftex]{graphicx}
-	  \DeclareGraphicsExtensions{.pdf}
-\else
-  \usepackage[dvips]{graphicx}
-	  \DeclareGraphicsExtensions{.eps}
-\fi
-
-%\usepackage{asymptote}	
-\usepackage{fancyhdr}
-%\usepackage{supertabular}
-\usepackage{booktabs}
-%\usepackage{longtable}
-%\usepackage[dvips]{rotating}
-\usepackage{multirow}
-\usepackage{multicol}
-
-\usepackage{color}
-\usepackage{amsmath}
-\usepackage{alltt}
-%\usepackage{array}
-%\usepackage{colortbl}
-
-%%%%%%%%%%%%%%%% will be used to defined a color scheme for
-%%%%%%%%%%%%%%%% latex pages converted with "Highlight"
-\newcommand{\hlstd}[1]{\textcolor[rgb]{0,0,0}{#1}}
-\newcommand{\hlnum}[1]{\textcolor[rgb]{0.75,0,0.35}{#1}}
-\newcommand{\hlesc}[1]{\textcolor[rgb]{0.42,0.35,0.8}{#1}}
-\newcommand{\hlstr}[1]{\textcolor[rgb]{0.75,0,0.35}{#1}}
-\newcommand{\hldstr}[1]{\textcolor[rgb]{0.75,0,0.35}{#1}}
-\newcommand{\hlslc}[1]{\textcolor[rgb]{0.25,0.38,0.56}{#1}}
-\newcommand{\hlcom}[1]{\textcolor[rgb]{0.25,0.38,0.56}{#1}}
-\newcommand{\hldir}[1]{\textcolor[rgb]{0.8,0,0.8}{#1}}
-\newcommand{\hlsym}[1]{\textcolor[rgb]{0,0,0}{#1}}
-\newcommand{\hlline}[1]{\textcolor[rgb]{0.25,0.38,0.56}{#1}}
-\newcommand{\hlkwa}[1]{\textcolor[rgb]{0.65,0.16,0.16}{\bf{#1}}}
-\newcommand{\hlkwb}[1]{\textcolor[rgb]{0.18,0.55,0.34}{\bf{#1}}}
-\newcommand{\hlkwc}[1]{\textcolor[rgb]{0.15,0.37,0.93}{\bf{#1}}}
-\newcommand{\hlkwd}[1]{\textcolor[rgb]{0.32,0.11,0.78}{#1}}
-\definecolor{bgcolor}{rgb}{1,0.85,0.73}
-\oddsidemargin -3mm
-\textwidth 165,2truemm
-\topmargin 0truept
-\headheight 0truept
-\headsep 0truept
-\textheight 230truemm
-%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
-
-\clubpenalty = 10000
-\widowpenalty = 10000 \displaywidowpenalty = 10000
-
-\definecolor{hellgrau}{gray}{0.95}
-\definecolor{dunkelgrau}{gray}{0.55}
-
-\definecolor{brown}{rgb}{0.75,0.004,0.3}
-
-
-\renewcommand{\headrulewidth}{0pt} % no head rule
-\renewcommand{\footrulewidth}{0pt} % no footer rule
-
-%\nointend
-%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
-% start text here!!
-
-
-\begin{document}
-\pagenumbering{roman}
-\input{title2}
-
-\tableofcontents
-\pagenumbering{arabic}
-
-\include{introduction}
-\include{octave/neuroPackage/neuroPackage}
-\include{examples/examples}
-
-\include{bibliography}
-
-
-
-\end{document}
-
-
--- a/main/nnet/doc/latex/users/neuralNetworkToolboxForOctaveUsersGuide.tcp	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,12 +0,0 @@
-[FormatInfo]
-Type=TeXnicCenterProjectInformation
-Version=4
-
-[ProjectInfo]
-MainFile=neuralNetworkToolboxForOctaveUsersGuide.tex
-UseBibTeX=0
-UseMakeIndex=1
-ActiveProfile=LaTeX => PDF
-ProjectLanguage=de
-ProjectDialect=DE
-
--- a/main/nnet/doc/latex/users/neuralNetworkToolboxForOctaveUsersGuide.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,102 +0,0 @@
-% Preambel
-\documentclass[a4paper,openany]{report}
-
-
-\usepackage{a4wide}
-\usepackage[ansinew]{inputenc}
-\usepackage[T1]{fontenc}
-\RequirePackage{ifpdf}
-
-\usepackage{hyperref}
-	\hypersetup{%
-  colorlinks=true,   % activates colored references
-  pdfpagemode=None,  % PDF-Viewer starts without content et.al.
-  pdfstartview=FitH, % PDF-Viewer uses a defined page width
-  %linkbordercolor=111,
-  % citebordercolor=111,
-  citecolor=blue,
-  linkcolor=blue}
-
-\ifpdf
-  \usepackage[pdftex]{graphicx}
-	  \DeclareGraphicsExtensions{.pdf}
-\else
-  \usepackage[dvips]{graphicx}
-	  \DeclareGraphicsExtensions{.eps}
-\fi
-
-\usepackage{asymptote}	
-\usepackage{fancyhdr}
-%\usepackage{supertabular}
-\usepackage{booktabs}
-%\usepackage{longtable}
-\usepackage[dvips]{rotating}
-\usepackage{multirow}
-\usepackage{multicol}
-
-\usepackage{color}
-\usepackage{amsmath}
-\usepackage{alltt}
-%\usepackage{array}
-%\usepackage{colortbl}
-
-%%%%%%%%%%%%%%%% will be used to defined a color scheme for
-%%%%%%%%%%%%%%%% latex pages converted with "Highlight"
-\newcommand{\hlstd}[1]{\textcolor[rgb]{0,0,0}{#1}}
-\newcommand{\hlnum}[1]{\textcolor[rgb]{0.75,0,0.35}{#1}}
-\newcommand{\hlesc}[1]{\textcolor[rgb]{0.42,0.35,0.8}{#1}}
-\newcommand{\hlstr}[1]{\textcolor[rgb]{0.75,0,0.35}{#1}}
-\newcommand{\hldstr}[1]{\textcolor[rgb]{0.75,0,0.35}{#1}}
-\newcommand{\hlslc}[1]{\textcolor[rgb]{0.25,0.38,0.56}{#1}}
-\newcommand{\hlcom}[1]{\textcolor[rgb]{0.25,0.38,0.56}{#1}}
-\newcommand{\hldir}[1]{\textcolor[rgb]{0.8,0,0.8}{#1}}
-\newcommand{\hlsym}[1]{\textcolor[rgb]{0,0,0}{#1}}
-\newcommand{\hlline}[1]{\textcolor[rgb]{0.25,0.38,0.56}{#1}}
-\newcommand{\hlkwa}[1]{\textcolor[rgb]{0.65,0.16,0.16}{\bf{#1}}}
-\newcommand{\hlkwb}[1]{\textcolor[rgb]{0.18,0.55,0.34}{\bf{#1}}}
-\newcommand{\hlkwc}[1]{\textcolor[rgb]{0.15,0.37,0.93}{\bf{#1}}}
-\newcommand{\hlkwd}[1]{\textcolor[rgb]{0.32,0.11,0.78}{#1}}
-\definecolor{bgcolor}{rgb}{1,0.85,0.73}
-\oddsidemargin -3mm
-\textwidth 165,2truemm
-\topmargin 0truept
-\headheight 0truept
-\headsep 0truept
-\textheight 230truemm
-%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
-
-\clubpenalty = 10000
-\widowpenalty = 10000 \displaywidowpenalty = 10000
-
-\definecolor{hellgrau}{gray}{0.95}
-\definecolor{dunkelgrau}{gray}{0.55}
-
-\definecolor{brown}{rgb}{0.75,0.004,0.3}
-
-
-\renewcommand{\headrulewidth}{0pt} % no head rule
-\renewcommand{\footrulewidth}{0pt} % no footer rule
-
-%\nointend
-%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
-% start text here!!
-
-
-\begin{document}
-\pagenumbering{roman}
-\input{title2}
-
-\tableofcontents
-\pagenumbering{arabic}
-
-\include{introduction}
-\include{octave/neuroToolbox/neuroToolbox}
-\include{examples/examples}
-
-\include{bibliography}
-
-
-
-\end{document}
-
-
--- a/main/nnet/doc/latex/users/numbering.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,7 +0,0 @@
-\section{Version numbers}
-
-The first number describes the major release. Version number V1.0 will be the first toolbox release which should have the same functions like the Matlab R14 SP3 neural network Toolbox.\\
-
-The second number defines the finished functions. So to start, only the MLPs will realised and so this will be the number V0.1.0.\\
-
-The third number defines the status of the actual development and function. V0.1.0 means a first release with MLP. Actually it works only with Levenberg-Marquardt algorithm and Mean-Square-Error as performance function.
\ No newline at end of file
--- a/main/nnet/doc/latex/users/octave/neuroPackage/graphics/logsig.eps	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,702 +0,0 @@
-%!PS-Adobe-3.0 EPSF-3.0
-%%Creator: dvips(k) 5.94b Copyright 2004 Radical Eye Software
-%%Title: logsig_.dvi
-%%CreationDate: Mon Jul 02 22:01:20 2007
-%%Pages: 1
-%%PageOrder: Ascend
-%%BoundingBox: 255 354 356 437
-%%HiResBoundingBox: 255.5 354.623708 355.5 436.376292
-%%DocumentFonts: CMMI12 CMR12 CMSY10
-%%EndComments
-%DVIPSWebPage: (www.radicaleye.com)
-%DVIPSCommandLine: C:\texmf\miktex\bin\dvips.exe -R -O 127.1bp,229.824bp
-%+ -T 612bp,792bp -q -o logsig_.ps logsig_.dvi
-%DVIPSParameters: dpi=600
-%DVIPSSource:  TeX output 2007.07.02:2201
-%%BeginProcSet: tex.pro 0 0
-%!
-/TeXDict 300 dict def TeXDict begin/N{def}def/B{bind def}N/S{exch}N/X{S
-N}B/A{dup}B/TR{translate}N/isls false N/vsize 11 72 mul N/hsize 8.5 72
-mul N/landplus90{false}def/@rigin{isls{[0 landplus90{1 -1}{-1 1}ifelse 0
-0 0]concat}if 72 Resolution div 72 VResolution div neg scale isls{
-landplus90{VResolution 72 div vsize mul 0 exch}{Resolution -72 div hsize
-mul 0}ifelse TR}if Resolution VResolution vsize -72 div 1 add mul TR[
-matrix currentmatrix{A A round sub abs 0.00001 lt{round}if}forall round
-exch round exch]setmatrix}N/@landscape{/isls true N}B/@manualfeed{
-statusdict/manualfeed true put}B/@copies{/#copies X}B/FMat[1 0 0 -1 0 0]
-N/FBB[0 0 0 0]N/nn 0 N/IEn 0 N/ctr 0 N/df-tail{/nn 8 dict N nn begin
-/FontType 3 N/FontMatrix fntrx N/FontBBox FBB N string/base X array
-/BitMaps X/BuildChar{CharBuilder}N/Encoding IEn N end A{/foo setfont}2
-array copy cvx N load 0 nn put/ctr 0 N[}B/sf 0 N/df{/sf 1 N/fntrx FMat N
-df-tail}B/dfs{div/sf X/fntrx[sf 0 0 sf neg 0 0]N df-tail}B/E{pop nn A
-definefont setfont}B/Cw{Cd A length 5 sub get}B/Ch{Cd A length 4 sub get
-}B/Cx{128 Cd A length 3 sub get sub}B/Cy{Cd A length 2 sub get 127 sub}
-B/Cdx{Cd A length 1 sub get}B/Ci{Cd A type/stringtype ne{ctr get/ctr ctr
-1 add N}if}B/CharBuilder{save 3 1 roll S A/base get 2 index get S
-/BitMaps get S get/Cd X pop/ctr 0 N Cdx 0 Cx Cy Ch sub Cx Cw add Cy
-setcachedevice Cw Ch true[1 0 0 -1 -.1 Cx sub Cy .1 sub]{Ci}imagemask
-restore}B/D{/cc X A type/stringtype ne{]}if nn/base get cc ctr put nn
-/BitMaps get S ctr S sf 1 ne{A A length 1 sub A 2 index S get sf div put
-}if put/ctr ctr 1 add N}B/I{cc 1 add D}B/bop{userdict/bop-hook known{
-bop-hook}if/SI save N @rigin 0 0 moveto/V matrix currentmatrix A 1 get A
-mul exch 0 get A mul add .99 lt{/QV}{/RV}ifelse load def pop pop}N/eop{
-SI restore userdict/eop-hook known{eop-hook}if showpage}N/@start{
-userdict/start-hook known{start-hook}if pop/VResolution X/Resolution X
-1000 div/DVImag X/IEn 256 array N 2 string 0 1 255{IEn S A 360 add 36 4
-index cvrs cvn put}for pop 65781.76 div/vsize X 65781.76 div/hsize X}N
-/p{show}N/RMat[1 0 0 -1 0 0]N/BDot 260 string N/Rx 0 N/Ry 0 N/V{}B/RV/v{
-/Ry X/Rx X V}B statusdict begin/product where{pop false[(Display)(NeXT)
-(LaserWriter 16/600)]{A length product length le{A length product exch 0
-exch getinterval eq{pop true exit}if}{pop}ifelse}forall}{false}ifelse
-end{{gsave TR -.1 .1 TR 1 1 scale Rx Ry false RMat{BDot}imagemask
-grestore}}{{gsave TR -.1 .1 TR Rx Ry scale 1 1 false RMat{BDot}
-imagemask grestore}}ifelse B/QV{gsave newpath transform round exch round
-exch itransform moveto Rx 0 rlineto 0 Ry neg rlineto Rx neg 0 rlineto
-fill grestore}B/a{moveto}B/delta 0 N/tail{A/delta X 0 rmoveto}B/M{S p
-delta add tail}B/b{S p tail}B/c{-4 M}B/d{-3 M}B/e{-2 M}B/f{-1 M}B/g{0 M}
-B/h{1 M}B/i{2 M}B/j{3 M}B/k{4 M}B/w{0 rmoveto}B/l{p -4 w}B/m{p -3 w}B/n{
-p -2 w}B/o{p -1 w}B/q{p 1 w}B/r{p 2 w}B/s{p 3 w}B/t{p 4 w}B/x{0 S
-rmoveto}B/y{3 2 roll p a}B/bos{/SS save N}B/eos{SS restore}B end
-
-%%EndProcSet
-%%BeginProcSet: texps.pro 0 0
-%!
-TeXDict begin/rf{findfont dup length 1 add dict begin{1 index/FID ne 2
-index/UniqueID ne and{def}{pop pop}ifelse}forall[1 index 0 6 -1 roll
-exec 0 exch 5 -1 roll VResolution Resolution div mul neg 0 0]/Metrics
-exch def dict begin Encoding{exch dup type/integertype ne{pop pop 1 sub
-dup 0 le{pop}{[}ifelse}{FontMatrix 0 get div Metrics 0 get div def}
-ifelse}forall Metrics/Metrics currentdict end def[2 index currentdict
-end definefont 3 -1 roll makefont/setfont cvx]cvx def}def/ObliqueSlant{
-dup sin S cos div neg}B/SlantFont{4 index mul add}def/ExtendFont{3 -1
-roll mul exch}def/ReEncodeFont{CharStrings rcheck{/Encoding false def
-dup[exch{dup CharStrings exch known not{pop/.notdef/Encoding true def}
-if}forall Encoding{]exch pop}{cleartomark}ifelse}if/Encoding exch def}
-def end
-
-%%EndProcSet
-%%BeginProcSet: special.pro 0 0
-%!
-TeXDict begin/SDict 200 dict N SDict begin/@SpecialDefaults{/hs 612 N
-/vs 792 N/ho 0 N/vo 0 N/hsc 1 N/vsc 1 N/ang 0 N/CLIP 0 N/rwiSeen false N
-/rhiSeen false N/letter{}N/note{}N/a4{}N/legal{}N}B/@scaleunit 100 N
-/@hscale{@scaleunit div/hsc X}B/@vscale{@scaleunit div/vsc X}B/@hsize{
-/hs X/CLIP 1 N}B/@vsize{/vs X/CLIP 1 N}B/@clip{/CLIP 2 N}B/@hoffset{/ho
-X}B/@voffset{/vo X}B/@angle{/ang X}B/@rwi{10 div/rwi X/rwiSeen true N}B
-/@rhi{10 div/rhi X/rhiSeen true N}B/@llx{/llx X}B/@lly{/lly X}B/@urx{
-/urx X}B/@ury{/ury X}B/magscale true def end/@MacSetUp{userdict/md known
-{userdict/md get type/dicttype eq{userdict begin md length 10 add md
-maxlength ge{/md md dup length 20 add dict copy def}if end md begin
-/letter{}N/note{}N/legal{}N/od{txpose 1 0 mtx defaultmatrix dtransform S
-atan/pa X newpath clippath mark{transform{itransform moveto}}{transform{
-itransform lineto}}{6 -2 roll transform 6 -2 roll transform 6 -2 roll
-transform{itransform 6 2 roll itransform 6 2 roll itransform 6 2 roll
-curveto}}{{closepath}}pathforall newpath counttomark array astore/gc xdf
-pop ct 39 0 put 10 fz 0 fs 2 F/|______Courier fnt invertflag{PaintBlack}
-if}N/txpose{pxs pys scale ppr aload pop por{noflips{pop S neg S TR pop 1
--1 scale}if xflip yflip and{pop S neg S TR 180 rotate 1 -1 scale ppr 3
-get ppr 1 get neg sub neg ppr 2 get ppr 0 get neg sub neg TR}if xflip
-yflip not and{pop S neg S TR pop 180 rotate ppr 3 get ppr 1 get neg sub
-neg 0 TR}if yflip xflip not and{ppr 1 get neg ppr 0 get neg TR}if}{
-noflips{TR pop pop 270 rotate 1 -1 scale}if xflip yflip and{TR pop pop
-90 rotate 1 -1 scale ppr 3 get ppr 1 get neg sub neg ppr 2 get ppr 0 get
-neg sub neg TR}if xflip yflip not and{TR pop pop 90 rotate ppr 3 get ppr
-1 get neg sub neg 0 TR}if yflip xflip not and{TR pop pop 270 rotate ppr
-2 get ppr 0 get neg sub neg 0 S TR}if}ifelse scaleby96{ppr aload pop 4
--1 roll add 2 div 3 1 roll add 2 div 2 copy TR .96 dup scale neg S neg S
-TR}if}N/cp{pop pop showpage pm restore}N end}if}if}N/normalscale{
-Resolution 72 div VResolution 72 div neg scale magscale{DVImag dup scale
-}if 0 setgray}N/psfts{S 65781.76 div N}N/startTexFig{/psf$SavedState
-save N userdict maxlength dict begin/magscale true def normalscale
-currentpoint TR/psf$ury psfts/psf$urx psfts/psf$lly psfts/psf$llx psfts
-/psf$y psfts/psf$x psfts currentpoint/psf$cy X/psf$cx X/psf$sx psf$x
-psf$urx psf$llx sub div N/psf$sy psf$y psf$ury psf$lly sub div N psf$sx
-psf$sy scale psf$cx psf$sx div psf$llx sub psf$cy psf$sy div psf$ury sub
-TR/showpage{}N/erasepage{}N/copypage{}N/p 3 def @MacSetUp}N/doclip{
-psf$llx psf$lly psf$urx psf$ury currentpoint 6 2 roll newpath 4 copy 4 2
-roll moveto 6 -1 roll S lineto S lineto S lineto closepath clip newpath
-moveto}N/endTexFig{end psf$SavedState restore}N/@beginspecial{SDict
-begin/SpecialSave save N gsave normalscale currentpoint TR
-@SpecialDefaults count/ocount X/dcount countdictstack N}N/@setspecial{
-CLIP 1 eq{newpath 0 0 moveto hs 0 rlineto 0 vs rlineto hs neg 0 rlineto
-closepath clip}if ho vo TR hsc vsc scale ang rotate rwiSeen{rwi urx llx
-sub div rhiSeen{rhi ury lly sub div}{dup}ifelse scale llx neg lly neg TR
-}{rhiSeen{rhi ury lly sub div dup scale llx neg lly neg TR}if}ifelse
-CLIP 2 eq{newpath llx lly moveto urx lly lineto urx ury lineto llx ury
-lineto closepath clip}if/showpage{}N/erasepage{}N/copypage{}N newpath}N
-/@endspecial{count ocount sub{pop}repeat countdictstack dcount sub{end}
-repeat grestore SpecialSave restore end}N/@defspecial{SDict begin}N
-/@fedspecial{end}B/li{lineto}B/rl{rlineto}B/rc{rcurveto}B/np{/SaveX
-currentpoint/SaveY X N 1 setlinecap newpath}N/st{stroke SaveX SaveY
-moveto}N/fil{fill SaveX SaveY moveto}N/ellipse{/endangle X/startangle X
-/yrad X/xrad X/savematrix matrix currentmatrix N TR xrad yrad scale 0 0
-1 startangle endangle arc savematrix setmatrix}N end
-
-%%EndProcSet
-%%BeginProcSet: color.pro 0 0
-%!
-TeXDict begin/setcmykcolor where{pop}{/setcmykcolor{dup 10 eq{pop
-setrgbcolor}{1 sub 4 1 roll 3{3 index add neg dup 0 lt{pop 0}if 3 1 roll
-}repeat setrgbcolor pop}ifelse}B}ifelse/TeXcolorcmyk{setcmykcolor}def
-/TeXcolorrgb{setrgbcolor}def/TeXcolorgrey{setgray}def/TeXcolorgray{
-setgray}def/TeXcolorhsb{sethsbcolor}def/currentcmykcolor where{pop}{
-/currentcmykcolor{currentrgbcolor 10}B}ifelse/DC{exch dup userdict exch
-known{pop pop}{X}ifelse}B/GreenYellow{0.15 0 0.69 0 setcmykcolor}DC
-/Yellow{0 0 1 0 setcmykcolor}DC/Goldenrod{0 0.10 0.84 0 setcmykcolor}DC
-/Dandelion{0 0.29 0.84 0 setcmykcolor}DC/Apricot{0 0.32 0.52 0
-setcmykcolor}DC/Peach{0 0.50 0.70 0 setcmykcolor}DC/Melon{0 0.46 0.50 0
-setcmykcolor}DC/YellowOrange{0 0.42 1 0 setcmykcolor}DC/Orange{0 0.61
-0.87 0 setcmykcolor}DC/BurntOrange{0 0.51 1 0 setcmykcolor}DC
-/Bittersweet{0 0.75 1 0.24 setcmykcolor}DC/RedOrange{0 0.77 0.87 0
-setcmykcolor}DC/Mahogany{0 0.85 0.87 0.35 setcmykcolor}DC/Maroon{0 0.87
-0.68 0.32 setcmykcolor}DC/BrickRed{0 0.89 0.94 0.28 setcmykcolor}DC/Red{
-0 1 1 0 setcmykcolor}DC/OrangeRed{0 1 0.50 0 setcmykcolor}DC/RubineRed{
-0 1 0.13 0 setcmykcolor}DC/WildStrawberry{0 0.96 0.39 0 setcmykcolor}DC
-/Salmon{0 0.53 0.38 0 setcmykcolor}DC/CarnationPink{0 0.63 0 0
-setcmykcolor}DC/Magenta{0 1 0 0 setcmykcolor}DC/VioletRed{0 0.81 0 0
-setcmykcolor}DC/Rhodamine{0 0.82 0 0 setcmykcolor}DC/Mulberry{0.34 0.90
-0 0.02 setcmykcolor}DC/RedViolet{0.07 0.90 0 0.34 setcmykcolor}DC
-/Fuchsia{0.47 0.91 0 0.08 setcmykcolor}DC/Lavender{0 0.48 0 0
-setcmykcolor}DC/Thistle{0.12 0.59 0 0 setcmykcolor}DC/Orchid{0.32 0.64 0
-0 setcmykcolor}DC/DarkOrchid{0.40 0.80 0.20 0 setcmykcolor}DC/Purple{
-0.45 0.86 0 0 setcmykcolor}DC/Plum{0.50 1 0 0 setcmykcolor}DC/Violet{
-0.79 0.88 0 0 setcmykcolor}DC/RoyalPurple{0.75 0.90 0 0 setcmykcolor}DC
-/BlueViolet{0.86 0.91 0 0.04 setcmykcolor}DC/Periwinkle{0.57 0.55 0 0
-setcmykcolor}DC/CadetBlue{0.62 0.57 0.23 0 setcmykcolor}DC
-/CornflowerBlue{0.65 0.13 0 0 setcmykcolor}DC/MidnightBlue{0.98 0.13 0
-0.43 setcmykcolor}DC/NavyBlue{0.94 0.54 0 0 setcmykcolor}DC/RoyalBlue{1
-0.50 0 0 setcmykcolor}DC/Blue{1 1 0 0 setcmykcolor}DC/Cerulean{0.94 0.11
-0 0 setcmykcolor}DC/Cyan{1 0 0 0 setcmykcolor}DC/ProcessBlue{0.96 0 0 0
-setcmykcolor}DC/SkyBlue{0.62 0 0.12 0 setcmykcolor}DC/Turquoise{0.85 0
-0.20 0 setcmykcolor}DC/TealBlue{0.86 0 0.34 0.02 setcmykcolor}DC
-/Aquamarine{0.82 0 0.30 0 setcmykcolor}DC/BlueGreen{0.85 0 0.33 0
-setcmykcolor}DC/Emerald{1 0 0.50 0 setcmykcolor}DC/JungleGreen{0.99 0
-0.52 0 setcmykcolor}DC/SeaGreen{0.69 0 0.50 0 setcmykcolor}DC/Green{1 0
-1 0 setcmykcolor}DC/ForestGreen{0.91 0 0.88 0.12 setcmykcolor}DC
-/PineGreen{0.92 0 0.59 0.25 setcmykcolor}DC/LimeGreen{0.50 0 1 0
-setcmykcolor}DC/YellowGreen{0.44 0 0.74 0 setcmykcolor}DC/SpringGreen{
-0.26 0 0.76 0 setcmykcolor}DC/OliveGreen{0.64 0 0.95 0.40 setcmykcolor}
-DC/RawSienna{0 0.72 1 0.45 setcmykcolor}DC/Sepia{0 0.83 1 0.70
-setcmykcolor}DC/Brown{0 0.81 1 0.60 setcmykcolor}DC/Tan{0.14 0.42 0.56 0
-setcmykcolor}DC/Gray{0 0 0 0.50 setcmykcolor}DC/Black{0 0 0 1
-setcmykcolor}DC/White{0 0 0 0 setcmykcolor}DC end
-
-%%EndProcSet
-%%BeginFont: CMSY10
-%!PS-AdobeFont-1.1: CMSY10 1.0
-%%CreationDate: 1991 Aug 15 07:20:57
-% Copyright (C) 1997 American Mathematical Society. All Rights Reserved.
-11 dict begin
-/FontInfo 7 dict dup begin
-/version (1.0) readonly def
-/Notice (Copyright (C) 1997 American Mathematical Society. All Rights Reserved) readonly def
-/FullName (CMSY10) readonly def
-/FamilyName (Computer Modern) readonly def
-/Weight (Medium) readonly def
-/ItalicAngle -14.035 def
-/isFixedPitch false def
-end readonly def
-/FontName /CMSY10 def
-/PaintType 0 def
-/FontType 1 def
-/FontMatrix [0.001 0 0 0.001 0 0] readonly def
-/Encoding 256 array
-0 1 255 {1 index exch /.notdef put} for
-dup 0 /minus put
-readonly def
-/FontBBox{-29 -960 1116 775}readonly def
-/UniqueID 5000820 def
-currentdict end
-currentfile eexec
-D9D66F633B846A97B686A97E45A3D0AA052F09F9C8ADE9D907C058B87E9B6964
-7D53359E51216774A4EAA1E2B58EC3176BD1184A633B951372B4198D4E8C5EF4
-A213ACB58AA0A658908035BF2ED8531779838A960DFE2B27EA49C37156989C85
-E21B3ABF72E39A89232CD9F4237FC80C9E64E8425AA3BEF7DED60B122A52922A
-221A37D9A807DD01161779DDE7D31FF2B87F97C73D63EECDDA4C49501773468A
-27D1663E0B62F461F6E40A5D6676D1D12B51E641C1D4E8E2771864FC104F8CBF
-5B78EC1D88228725F1C453A678F58A7E1B7BD7CA700717D288EB8DA1F57C4F09
-0ABF1D42C5DDD0C384C7E22F8F8047BE1D4C1CC8E33368FB1AC82B4E96146730
-DE3302B2E6B819CB6AE455B1AF3187FFE8071AA57EF8A6616B9CB7941D44EC7A
-71A7BB3DF755178D7D2E4BB69859EFA4BBC30BD6BB1531133FD4D9438FF99F09
-4ECC068A324D75B5F696B8688EEB2F17E5ED34CCD6D047A4E3806D000C199D7C
-515DB70A8D4F6146FE068DC1E5DE8BC5703711DA090312BA3FC00A08C453C609
-C627A8B1550654AD5E22C5F3F3CC8C1C0A6C7ADDAB55016A76EC46213FD9BAAF
-03F7A5FD261BF647FCA5049118033F809370A84AC3ADA3D5BE032CBB494D7851
-A6242E785CCC20D81FC5EE7871F1E588DA3E31BD321C67142C5D76BC6AC708DF
-C21616B4CC92F0F8B92BD37A4AB83E066D1245FAD89B480CB0AC192D4CAFA6AD
-241BD8DF7AD566A2022FBC67364AB89F33608554113D210FE5D27F8FB1B2B78A
-F22EC999DBAAFC9C60017101D5FB2A3B6E2BF4BE47B8E5E4662B8C41AB471DFC
-A31EE1
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-cleartomark
-%%EndFont 
-%%BeginFont: CMR12
-%!PS-AdobeFont-1.1: CMR12 1.0
-%%CreationDate: 1991 Aug 20 16:38:05
-% Copyright (C) 1997 American Mathematical Society. All Rights Reserved.
-11 dict begin
-/FontInfo 7 dict dup begin
-/version (1.0) readonly def
-/Notice (Copyright (C) 1997 American Mathematical Society. All Rights Reserved) readonly def
-/FullName (CMR12) readonly def
-/FamilyName (Computer Modern) readonly def
-/Weight (Medium) readonly def
-/ItalicAngle 0 def
-/isFixedPitch false def
-end readonly def
-/FontName /CMR12 def
-/PaintType 0 def
-/FontType 1 def
-/FontMatrix [0.001 0 0 0.001 0 0] readonly def
-/Encoding 256 array
-0 1 255 {1 index exch /.notdef put} for
-dup 40 /parenleft put
-dup 41 /parenright put
-dup 43 /plus put
-dup 48 /zero put
-dup 49 /one put
-dup 61 /equal put
-readonly def
-/FontBBox{-34 -251 988 750}readonly def
-/UniqueID 5000794 def
-currentdict end
-currentfile eexec
-D9D66F633B846A97B686A97E45A3D0AA052A014267B7904EB3C0D3BD0B83D891
-016CA6CA4B712ADEB258FAAB9A130EE605E61F77FC1B738ABC7C51CD46EF8171
-9098D5FEE67660E69A7AB91B58F29A4D79E57022F783EB0FBBB6D4F4EC35014F
-D2DECBA99459A4C59DF0C6EBA150284454E707DC2100C15B76B4C19B84363758
-469A6C558785B226332152109871A9883487DD7710949204DDCF837E6A8708B8
-2BDBF16FBC7512FAA308A093FE5CF4E9D2405B169CD5365D6ECED5D768D66D6C
-68618B8C482B341F8CA38E9BB9BAFCFAAD9C2F3FD033B62690986ED43D9C9361
-3645B82392D5CAE11A7CB49D7E2E82DCD485CBA04C77322EB2E6A79D73DC194E
-59C120A2DABB9BF72E2CF256DD6EB54EECBA588101ABD933B57CE8A3A0D16B28
-51D7494F73096DF53BDC66BBF896B587DF9643317D5F610CD9088F9849126F23
-DDE030F7B277DD99055C8B119CAE9C99158AC4E150CDFC2C66ED92EBB4CC092A
-AA078CE16247A1335AD332DAA950D20395A7384C33FF72EAA31A5B89766E635F
-45C4C068AD7EE867398F0381B07CB94D29FF097D59FF9961D195A948E3D87C31
-821E9295A56D21875B41988F7A16A1587050C3C71B4E4355BB37F255D6B237CE
-96F25467F70FA19E0F85785FF49068949CCC79F2F8AE57D5F79BB9C5CF5EED5D
-9857B9967D9B96CDCF73D5D65FF75AFABB66734018BAE264597220C89FD17379
-26764A9302D078B4EB0E29178C878FD61007EEA2DDB119AE88C57ECFEF4B71E4
-140A34951DDC3568A84CC92371A789021A103A1A347050FDA6ECF7903F67D213
-1D0C7C474A9053866E9C88E65E6932BA87A73686EAB0019389F84D159809C498
-1E7A30ED942EB211B00DBFF5BCC720F4E276C3339B31B6EABBB078430E6A09BB
-377D3061A20B1EB98796B8607EECBC699445EAA866C38E02DF59F5EDD378303A
-0733B90E7835C0AAF32BA04F1566D8161EA89CD4D14DDB953F8B910BFC8A7F03
-5020F55EF8FC2640ADADA156F6CF8F2EB6610F7EE8874A26CBE7CD154469B9F4
-ED76886B3FB679FFDEB59BB6C55AF7087BA48B75EE2FB374B19BCC421A963E15
-FE05ECAAF9EECDF4B2715010A320102E6F8CCAA342FA11532671C8926C9ED415
-D9C320876265E289F7FA41B4BB6252B17463EF2AC4A92D616D39E58816A6F8F2
-367DBF4EC567A70AF0E7BD49173056591769FB20BD5048CA92C6B1994457323B
-9950B5F84037A826CC226EE233EF4D0E893CEE5C1F652F4F3E71E7CEA4A01879
-EA41FAB023FC06B7ABCF70C48E5F934B765298142FF142EBCEB4A96DD478F51E
-C4923850A838B1A21DAA720558EA0B46AA90175AC1413FC2AE9729C8D0A0AE60
-8308EF0474B68ECC49D2BDD08E003D38DD06EB2B4BFF2D670CB67075B26D39CD
-2E06571D410CAFEB8D5A5CD85316AC3480FFD6F13332CB610F821594247A8160
-A75CE2C3B81601604174C634417F1F8214BA467438F6A1AA72DF3D30195BA544
-B7EBE7B387D15C9135A3DFC67392964E192909B8F78DC39D458A5E8B6EB9EB97
-2946FE6D7A91BCED70DF5CC879A0D3386BD4A0446ACE5500A45F3976C8AE60C5
-4B18CE7283C9763C179A02BD59631825B95740BAB616858ED5FEC11D6590D4C5
-B1EBC7E78DD271A45AB212BD86297B706DDFACEE146F388A20EE30F1378F1E5C
-C4F5743EDECCF4C256A1FE53A655553DF1783C2BC6768C3A24F5D691C962691C
-2E1870D8BB49455851A5CFFFAD7D5B4045D66236FEB0F318D83788BC166A4C9E
-4EE1636CDFBB59BD8A1A6520D9F31AE3DD129D3F81B4786A82CA43B9E6BAFB55
-EED33714E2CBADEE7BF01BD2B560A3A70577D6BD9B5F05B9DA70FB0CA5676C53
-A2385727BFD5F471D5570F40FBE5F1A6BF76C0A17EBE6D468BFDB2FCE1BF1EC5
-3351B5EA44A54BF405AC94DED3DE28EFE253678550056DDEA892DB08E90541EE
-935DE706E8D1CB155DD4EB762A3E18CC7D3E7DEE85C1775A082DCA68BC4FA433
-B81F7E608FB86D6D8F30A67003DF771ACE5DA00293F1FF4137CD87ECC5713309
-E4FD2DCF054A7301462C5AB3C024CD16E8311BE610034610B13911C14A457E0E
-528345ECED9B063EF7D5C69A73CE9799CCC9A23DAC7C90C4FF29DC70025EC2D0
-736EB59000F02F27F3AD6F645B28C5C69A43EF1537E4FA44EDDE536AF5C6C5B5
-763111E88F29B86B9783623ED39EA704B38B193F6DCDF202A1AF04FCFFFDA2DC
-DF887BEA50F5800C3C821388EF3E3189067FE0541BE609FCF6E5A0DAD8C4FC1B
-EB51267D02E3CEC620AB85D8D624DB85FC04005C1AE9DCE7A209A3CD3BCF89C5
-5B3CA84ADA7CA6E3DAFB07C5E46DF7AF29F31346B395E839F074D8B889C60837
-842024F7E6A7A5C50A54AD97D89F5DCBD671B6735D6D1D4E9AA95111449EA839
-4A642ACA
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-cleartomark
-%%EndFont 
-%%BeginFont: CMMI12
-%!PS-AdobeFont-1.1: CMMI12 1.100
-%%CreationDate: 1996 Jul 27 08:57:55
-% Copyright (C) 1997 American Mathematical Society. All Rights Reserved.
-11 dict begin
-/FontInfo 7 dict dup begin
-/version (1.100) readonly def
-/Notice (Copyright (C) 1997 American Mathematical Society. All Rights Reserved) readonly def
-/FullName (CMMI12) readonly def
-/FamilyName (Computer Modern) readonly def
-/Weight (Medium) readonly def
-/ItalicAngle -14.04 def
-/isFixedPitch false def
-end readonly def
-/FontName /CMMI12 def
-/PaintType 0 def
-/FontType 1 def
-/FontMatrix [0.001 0 0 0.001 0 0] readonly def
-/Encoding 256 array
-0 1 255 {1 index exch /.notdef put} for
-dup 97 /a put
-dup 103 /g put
-dup 105 /i put
-dup 108 /l put
-dup 110 /n put
-dup 111 /o put
-dup 115 /s put
-readonly def
-/FontBBox{-30 -250 1026 750}readonly def
-/UniqueID 5087386 def
-currentdict end
-currentfile eexec
-D9D66F633B846A97B686A97E45A3D0AA0529731C99A784CCBE85B4993B2EEBDE
-3B12D472B7CF54651EF21185116A69AB1096ED4BAD2F646635E019B6417CC77B
-532F85D811C70D1429A19A5307EF63EB5C5E02C89FC6C20F6D9D89E7D91FE470
-B72BEFDA23F5DF76BE05AF4CE93137A219ED8A04A9D7D6FDF37E6B7FCDE0D90B
-986423E5960A5D9FBB4C956556E8DF90CBFAEC476FA36FD9A5C8175C9AF513FE
-D919C2DDD26BDC0D99398B9F4D03D6A8F05B47AF95EF28A9C561DBDC98C47CF5
-5250011D19E9366EB6FD153D3A100CAA6212E3D5D93990737F8D326D347B7EDC
-4391C9DF440285B8FC159D0E98D4258FC57892DCC57F7903449E07914FBE9E67
-3C15C2153C061EB541F66C11E7EE77D5D77C0B11E1AC55101DA976CCACAB6993
-EED1406FBB7FF30EAC9E90B90B2AF4EC7C273CA32F11A5C1426FF641B4A2FB2F
-4E68635C93DB835737567FAF8471CBC05078DCD4E40E25A2F4E5AF46C234CF59
-2A1CE8F39E1BA1B2A594355637E474167EAD4D97D51AF0A899B44387E1FD933A
-323AFDA6BA740534A510B4705C0A15647AFBF3E53A82BF320DD96753639BE49C
-2F79A1988863EF977B800C9DB5B42039C23EB86953713F730E03EA22FF7BB2C1
-D97D33FD77B1BDCC2A60B12CF7805CFC90C5B914C0F30A673DF9587F93E47CEA
-5932DD1930560C4F0D97547BCD805D6D854455B13A4D7382A22F562D7C55041F
-0FD294BDAA1834820F894265A667E5C97D95FF152531EF97258F56374502865D
-A1E7C0C5FB7C6FB7D3C43FEB3431095A59FBF6F61CEC6D6DEE09F4EB0FD70D77
-2A8B0A4984C6120293F6B947944BE23259F6EB64303D627353163B6505FC8A60
-00681F7A3968B6CBB49E0420A691258F5E7B07B417157803FCBE9B9FB1F80FD8
-CA0BD2E774E4D04F1F0CB9AD88152DF9799FB90EC43955871EB7F0338141CF69
-3A94F81431168EFFF7462ABF70F1AAD9909E0183601E417073F4EC7DF0180A48
-73C309956ED2BC852965D7D4EF3F2A3F2A798CD61AE418D9573497D3911F5323
-ED3496F6AEBE685EE322F58EA7402EF6A7B6EB9E433EB7D0F6E3C3BDAD24F983
-AC4415A43C9687642E3BF1E4F4A99F03FA39177E5FFF4A9205E20954906ACE66
-1BF1C9E2E43707530FF446F58B37C73CF2857A7ABB3355DC42F2E66AAA8E40FB
-4F9A575B9C83CF9529A2AF30DA023468630AF059A7DC07EFF8041298B7AAEE9F
-010E4C93C08FCDA085657E92D98E9B33E1A28D3DA18FCBCBC7839C0744DD5CE0
-17FCC070EFE545CB2387F92A4B74262D7729B2DD458248397176142195B59718
-AA5429ED39CDE4F9CD1F92837B1EDAC168765EDD6395239B7C1CC552A6EC2A8A
-76E87AE3D015F874FECEF9406C030BE3732916C975F583FC660BE945F1A3EEFA
-A3B4E315BC32CF5EC239A9CC1B8ACB2C09540B1A42B6D057F6EC11DC7BD2F474
-72592808C08B7725B4F629671C96961BEA8F3C44C56A09C74FEE732584F36B00
-27977D6B37B2827E64FF0CA96215E62E3A5C325700D9B26E3550CFE92EB1ADB8
-E291B92E4BDEB32E539CD690B41B639E21B547FCF698B77B18417E645C6DCD63
-3FD68D26835FB59036B3EC45D58EB879F03FD8DF16CB785948643D059790CE79
-3BA847D6F75BE113B64E703A059B090ED349D382B2A73506C004B8A6D183AE18
-5AD305146A6DA14E3A7A16E3C5F095B249A8BE5CD1CC5BE1E0FADEDE5FB469A3
-CF8DE193CD5E42769D1F86F927B9752A982E8E42365FAAA3E3C33421D78CE39F
-F56E3C711136B926D7ADD91A6CA8BD527B0F0A28C1D16720B0E2F4FEB2BA12D8
-81BE8B788A6D8C42A8EB37E0E58C7D92BD698A8D7D3B66BDF5A2BD6A74F7702B
-6D1CB4079564BC57D6D97B91665DD0526534563A2F601924CBC4AB5CB1DE34C0
-CF279622FE084010F052050E4D8229203E2A612B14A4A810514B69AC94D6CB43
-A63348C975958E5C08A587E463698EC92112B8A0C457BE505ED9CBF680F10A26
-A3388315FD7FBE56F1491EA4F9E366E735DE1746C85C9DF1F0D398E024384A1C
-E121570A958A29ACE99A125B98DD8F1F9A6EF3B2DCC231A14F870678253FD456
-99A01B058D94D9238574F9E10D069435EA993C80EF2009FB96B76013938C3104
-46429B9D2E66109A0D2EA05B16747F76A485C5B5FC5A91FE664434BA89BA8D2C
-EF551234D25BD2B6C31FBB16CD7C54E3DC535E13B316475166B53258338A0DEF
-BC352DAC94DDE2136B52E9A89BA6EA54BB74300C2750B49A40F52DD788A95CFA
-9042A65C0AA3C6AC5D203025765DC62D86A4C1C47481AC94D9636C3AF005D335
-85B9AB275C1CDCD86C37E63C4ED751C74C3B73D128BFA3DCEE7FFEE6F2E3C1DE
-E1F209B6B02CD8B64745EC02E4726E7A754D76F95BB588AE29B9D99F9221DC3E
-DF691CE70D2CC4F3CB05440ADD8E0F321CF212219F3DDB0ACFD1C81CB066DE3F
-7CF8E9538E9FF435C1DC25FD9D0AD32D13F00BF9A82B365594FA7792523369C3
-15C96B33BCF35CCFD176EC284BF3535AC3185906E3245BB3C59B8B5008F837A0
-F0A1C59BA22894294110D1D9102DC8BAD3DB8EEC2088FF1297D59B8518AC6A7D
-41517424189468C2467DB32CB736E0D25F426F05316C1D15463DA7104EC6554A
-E5F2AFC00DBF91E935F1EAE0DBE76034EA4062A4AECE90A97C870C87FDDEC3FF
-5CC86D7B5FFBDC9B0956D4B538C24C27FDD093646B54AA82C9AD969558B1C1B4
-C8251768A7080C4576653F766E63ACA2A489611E17693C53BCFC66FA8C4E6A0C
-6639EB7C5200D4EF22A8F7C0224119FB6BDBBEF5264FF0D91F3CD92C2D7F90D9
-7F1AC8193730C263D6E3D52A1BE239E6877F2366574557190070CC2E904EE00A
-BBB965711ED8395246173D672CCB58D31554C04C5DF8B5D39DA597693947F919
-6841708C07D3F769447F3680F1E28DA4323E7DD0293FD8E6948775EA47D27FB8
-9C7E147BB5C262AC2AFF46E0414B33926F0FB75E3E16A4DC102640D361553139
-C5C7E01A3B852C4CFE76B8B4312FEBE82D0FD74D264C6F40E572184B10A32502
-025846A58918A736334219F6721F69E4F1616012C7D7D3D3B7F51D03E89A2A3B
-EEF04B0C4C09584A37833EEC6C2C14AC660780DCE7985DD9E01A54500FBBB172
-93CFAA9CDD88DD9ED22232D1EED42FC17216BE2F88FE588DFE2AF9DF080CF593
-F2D484B7EA6302A5F2AC8BB9E353B6564EBD492DC6330033BC93F6AFFB4A566E
-BCA1FFA22C8D23EBA7C702F733130DD421B11D60AAA5A7401B85758D543490E6
-025CCC438EFCDA0A1AAE014ECBB36D441039306103493D3AB367045BADA1D4A7
-A85A46375FFEE6609951B465B9A3AD3E8F8A8B1A40DEC0971F00E2D06E6CC26A
-1B31D8235953CF8C3CC99D7ADEC0A90020CB8EB1B7C0DBF182890F6568B8D785
-D45D2457964E47E2462CB7CF9CD2EF9458AD15F64C09BDB7CF9C9CF92B0CDBD8
-856F154D910BA11EB62FE6871D1C691B9F0E4FC040506C19A8141939EFB0E9BE
-E2F16FCCC1ECA41B71B8116E1B7A6EF2DE1DFBD1B6C6AEBC9C7D78DBF5317573
-206781C03DD9CBC67A69B7E54226C6ECC91B1F20564F6CD4A5F1FCD2D0300BD8
-3363C9F999B08E3DB1DA0284B68B2C40DE8FCB6B052B10DC9FFD3D76BC14FE3B
-353E5EA29403EC0E8B2C6D2D5F42DC3996A54A8A567B78B7BE09A1DB66DF7C34
-713C791447FE0B1DD0C22C6822D0C077CC95E349E6AF88685DE002B359792A85
-0C88E84E033990306CDAD934EED291897E6F02D2931754BEC1C2AD0C7DA87D60
-CBBE82CD900D8400E04A37D32A9BB55E0C8B8BBDC6082F7A37C9A6691333DBA8
-F2FDF1C0599A591F5D1D66597C91293F7632E970689D0EB54312922C5FD55BCB
-8239E07DA2A4276FDCCDD8DEB39C2D9A0AC274051F8932CE78733665D3
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-cleartomark
-%%EndFont 
-TeXDict begin 40258584 52099344 1000 600 600 (logsig_.dvi)
-@start /Fa 255[77{}1 99.6264 /CMSY10 rf /Fb 194[76 11[49
-49 4[76 1[38 38 40[{}6 99.6264 /CMR12 rf /Fc 140[46 3[47
-58 1[29 2[33 1[47 5[51 97[{}7 99.6264 /CMMI12 rf end
-%%EndProlog
-%%BeginSetup
-%%Feature: *Resolution 600dpi
-TeXDict begin
- end
-%%EndSetup
-%%Page: 1 1
-TeXDict begin 1 0 bop Black Black 1529 3045 a @beginspecial
--50 @llx -46.677586 @lly 50 @urx 35.075001 @ury 1000
-@rwi @setspecial
-%%BeginDocument: logsig_0.eps
-%!PS-Adobe-3.0 EPSF-3.0
-%%BoundingBox: -51 -47 51 36
-%%HiResBoundingBox: -50 -46.6775841 50 35.075
-%%Creator: Asymptote 1.25
-%%CreationDate: 2007.07.02 22:01:20
-%%Pages: 1
-%%EndProlog
-%%Page: 1 1
-0 setgray
- 0 0.5 dtransform truncate idtransform setlinewidth pop
-1 setlinecap
-1 setlinejoin
-gsave
- 0 0 translate
-newpath 42.25 0 moveto
- 11.5833333 0 -19.0833333 0 -49.75 0 curveto
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-gsave
- 0 0 translate
-newpath 49.75 1.5959456e-16 moveto
- 42.25 2.00961894 lineto
- 42.25 -2.00961894 lineto
- 44.75 -1.33974596 47.25 -0.669872981 49.75 -1.5959456e-16 curveto
- 49.75 -1.15843382e-16 49.75 1.15843382e-16 49.75 1.5959456e-16 curveto
-closepath
-fill
-grestore
-gsave
- 0 0 translate
-newpath 49.75 1.5959456e-16 moveto
- 42.25 2.00961894 lineto
- 42.25 -2.00961894 lineto
- 44.75 -1.33974596 47.25 -0.669872981 49.75 -1.5959456e-16 curveto
- 49.75 -1.15843382e-16 49.75 1.15843382e-16 49.75 1.5959456e-16 curveto
-closepath
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-gsave
- 0 0 translate
-newpath 0 27.325 moveto
- 0 6.60833333 0 -14.1083333 0 -34.825 curveto
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-gsave
- 0 0 translate
-newpath -4.23272528e-16 34.825 moveto
- -2.00961894 27.325 lineto
- 2.00961894 27.325 lineto
- 1.33974596 29.825 0.669872981 32.325 -1.50053581e-16 34.825 curveto
- -1.87503637e-16 34.825 -3.85822471e-16 34.825 -4.23272528e-16 34.825 curveto
-closepath
-fill
-grestore
-gsave
- 0 0 translate
-newpath -4.23272528e-16 34.825 moveto
- -2.00961894 27.325 lineto
- 2.00961894 27.325 lineto
- 1.33974596 29.825 0.669872981 32.325 -1.50053581e-16 34.825 curveto
- -1.87503637e-16 34.825 -3.85822471e-16 34.825 -4.23272528e-16 34.825 curveto
-closepath
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-newpath -49.75 1.50957778 moveto
- -49.4182975 1.5328238 -49.0866304 1.55657361 -48.755 1.58082713 curveto
- -48.4232954 1.60508607 -48.0916278 1.62984893 -47.76 1.65513666 curveto
- -47.4282923 1.68043049 -47.0966248 1.70624942 -46.765 1.73260887 curveto
- -46.4332892 1.75897516 -46.1016216 1.7858822 -45.77 1.81334692 curveto
- -45.4382859 1.8408193 -45.1066183 1.86884962 -44.775 1.89745435 curveto
- -44.4432825 1.92606763 -44.1116149 1.95525561 -43.78 1.98503473 curveto
- -43.4482789 2.01482339 -43.1166112 2.04520352 -42.785 2.07619134 curveto
- -42.4532752 2.10718977 -42.1216074 2.13879626 -41.79 2.17102674 curveto
- -41.4582712 2.20326902 -41.1266034 2.2361357 -40.795 2.26964239 curveto
- -40.4632672 2.30316215 -40.1315993 2.33732234 -39.8 2.37213815 curveto
- -39.4682629 2.40696841 -39.1365951 2.44245476 -38.805 2.47861183 curveto
- -38.4732586 2.51478485 -38.1415907 2.55162912 -37.81 2.58915864 curveto
- -37.4782541 2.62670573 -37.1465862 2.66493863 -36.815 2.70387065 curveto
- -36.4832496 2.74282195 -36.1515816 2.78247298 -35.82 2.82283619 curveto
- -35.4882449 2.86322052 -35.156577 2.90431769 -34.825 2.94613924 curveto
- -34.4932402 2.98798385 -34.1615723 3.03055353 -33.83 3.07385878 curveto
- -33.4982355 3.11718913 -33.1665676 3.16125579 -32.835 3.2060681 curveto
- -32.5032308 3.25090765 -32.1715629 3.29649365 -31.84 3.34283414 curveto
- -31.5082261 3.38920411 -31.1765583 3.43632942 -30.845 3.48421674 curveto
- -30.5132215 3.53213585 -30.1815537 3.58081786 -29.85 3.63026792 curveto
- -29.5182171 3.67975217 -29.1865493 3.7300054 -28.855 3.78103116 curveto
- -28.5232128 3.83209354 -28.1915451 3.88392941 -27.86 3.93654062 curveto
- -27.5282087 3.98919088 -27.1965411 4.04261748 -26.865 4.0968204 curveto
- -26.5332049 4.15106483 -26.2015375 4.20608663 -25.87 4.26188384 curveto
- -25.5382015 4.31772499 -25.2065341 4.37434262 -24.875 4.43173276 curveto
- -24.5431984 4.48916921 -24.2115312 4.54737926 -23.88 4.60635681 curveto
- -23.5481957 4.66538294 -23.2165286 4.72517769 -22.885 4.78573275 curveto
- -22.5531936 4.84633856 -22.2215266 4.90770579 -21.89 4.9698239 curveto
- -21.5581919 5.03199474 -21.2265252 5.09491757 -20.895 5.15857951 curveto
- -20.5631909 5.22229596 -20.2315244 5.28675264 -19.9 5.35193429 curveto
- -19.5681905 5.41717199 -19.2365242 5.48313576 -18.905 5.54980796 curveto
- -18.5731908 5.61653746 -18.2415248 5.68397646 -17.91 5.7521049 curveto
- -17.5781919 5.82029157 -17.246526 5.88916871 -16.915 5.95871386 curveto
- -16.5831937 6.02831781 -16.2515281 6.09859075 -15.92 6.16950783 curveto
- -15.5881963 6.24048385 -15.256531 6.31210494 -14.925 6.38434389 curveto
- -14.5931998 6.45664149 -14.2615347 6.52955782 -13.93 6.60306333 curveto
- -13.5982041 6.67662676 -13.2665393 6.75078018 -12.935 6.82549179 curveto
- -12.6032092 6.90026009 -12.2715447 6.97558731 -11.94 7.05143951 curveto
- -11.6082152 7.12734664 -11.2765509 7.2037794 -10.945 7.28070174 curveto
- -10.613222 7.35767677 -10.281558 7.43514196 -9.95 7.51305931 curveto
- -9.61822962 7.59102658 -9.2865658 7.66944651 -8.955 7.74827924 curveto
- -8.623238 7.82715863 -8.2915744 7.90645125 -7.96 7.98611556 curveto
- -7.6282471 8.06582276 -7.2965837 8.145902 -6.965 8.22631018 curveto
- -6.63325686 8.30675703 -6.30159365 8.38753309 -5.97 8.46859392 curveto
- -5.63826722 8.54968875 -5.30660418 8.63106854 -4.975 8.71268763 curveto
- -4.64327811 8.79433569 -4.31161521 8.8762232 -3.98 8.95830345 curveto
- -3.64828945 9.04040731 -3.31662666 9.12270402 -2.985 9.20514608 curveto
- -2.65330114 9.28760608 -2.32163844 9.37021149 -1.99 9.45291417 curveto
- -1.65831309 9.53562893 -1.32665045 9.618441 -0.995 9.70130181 curveto
- -0.66332521 9.7841687 -0.331662605 9.86708435 0 9.95 curveto
- 0.331662605 10.0329156 0.66332521 10.1158313 0.995 10.1986982 curveto
- 1.32665045 10.281559 1.65831309 10.3643711 1.99 10.4470858 curveto
- 2.32163844 10.5297885 2.65330114 10.6123939 2.985 10.6948539 curveto
- 3.31662666 10.777296 3.64828945 10.8595927 3.98 10.9416965 curveto
- 4.31161521 11.0237768 4.64327811 11.1056643 4.975 11.1873124 curveto
- 5.30660418 11.2689315 5.63826722 11.3503113 5.97 11.4314061 curveto
- 6.30159365 11.5124669 6.63325686 11.593243 6.965 11.6736898 curveto
- 7.2965837 11.754098 7.6282471 11.8341772 7.96 11.9138844 curveto
- 8.2915744 11.9935487 8.623238 12.0728414 8.955 12.1517208 curveto
- 9.2865658 12.2305535 9.61822962 12.3089734 9.95 12.3869407 curveto
- 10.281558 12.464858 10.613222 12.5423232 10.945 12.6192983 curveto
- 11.2765509 12.6962206 11.6082152 12.7726534 11.94 12.8485605 curveto
- 12.2715447 12.9244127 12.6032092 12.9997399 12.935 13.0745082 curveto
- 13.2665393 13.1492198 13.5982041 13.2233732 13.93 13.2969367 curveto
- 14.2615347 13.3704422 14.5931998 13.4433585 14.925 13.5156561 curveto
- 15.256531 13.5878951 15.5881963 13.6595162 15.92 13.7304922 curveto
- 16.2515281 13.8014093 16.5831937 13.8716822 16.915 13.9412861 curveto
- 17.246526 14.0108313 17.5781919 14.0797084 17.91 14.1478951 curveto
- 18.2415248 14.2160235 18.5731908 14.2834625 18.905 14.350192 curveto
- 19.2365242 14.4168642 19.5681905 14.482828 19.9 14.5480657 curveto
- 20.2315244 14.6132474 20.5631909 14.677704 20.895 14.7414205 curveto
- 21.2265252 14.8050824 21.5581919 14.8680053 21.89 14.9301761 curveto
- 22.2215266 14.9922942 22.5531936 15.0536614 22.885 15.1142672 curveto
- 23.2165286 15.1748223 23.5481957 15.2346171 23.88 15.2936432 curveto
- 24.2115312 15.3526207 24.5431984 15.4108308 24.875 15.4682672 curveto
- 25.2065341 15.5256574 25.5382015 15.582275 25.87 15.6381162 curveto
- 26.2015375 15.6939134 26.5332049 15.7489352 26.865 15.8031796 curveto
- 27.1965411 15.8573825 27.5282087 15.9108091 27.86 15.9634594 curveto
- 28.1915451 16.0160706 28.5232128 16.0679065 28.855 16.1189688 curveto
- 29.1865493 16.1699946 29.5182171 16.2202478 29.85 16.2697321 curveto
- 30.1815537 16.3191821 30.5132215 16.3678641 30.845 16.4157833 curveto
- 31.1765583 16.4636706 31.5082261 16.5107959 31.84 16.5571659 curveto
- 32.1715629 16.6035064 32.5032308 16.6490923 32.835 16.6939319 curveto
- 33.1665676 16.7387442 33.4982355 16.7828109 33.83 16.8261412 curveto
- 34.1615723 16.8694465 34.4932402 16.9120162 34.825 16.9538608 curveto
- 35.156577 16.9956823 35.4882449 17.0367795 35.82 17.0771638 curveto
- 36.1515816 17.117527 36.4832496 17.157178 36.815 17.1961293 curveto
- 37.1465862 17.2350614 37.4782541 17.2732943 37.81 17.3108414 curveto
- 38.1415907 17.3483709 38.4732586 17.3852151 38.805 17.4213882 curveto
- 39.1365951 17.4575452 39.4682629 17.4930316 39.8 17.5278619 curveto
- 40.1315993 17.5626777 40.4632672 17.5968379 40.795 17.6303576 curveto
- 41.1266034 17.6638643 41.4582712 17.696731 41.79 17.7289733 curveto
- 42.1216074 17.7612037 42.4532752 17.7928102 42.785 17.8238087 curveto
- 43.1166112 17.8547965 43.4482789 17.8851766 43.78 17.9149653 curveto
- 44.1116149 17.9447444 44.4432825 17.9739324 44.775 18.0025456 curveto
- 45.1066183 18.0311504 45.4382859 18.0591807 45.77 18.0866531 curveto
- 46.1016216 18.1141178 46.4332892 18.1410248 46.765 18.1673911 curveto
- 47.0966248 18.1937506 47.4282923 18.2195695 47.76 18.2448633 curveto
- 48.0916278 18.2701511 48.4232954 18.2949139 48.755 18.3191729 curveto
- 49.0866304 18.3434264 49.4182975 18.3671762 49.75 18.3904222 curveto
-stroke
-[3.98 3.98 ] 0 setdash
-newpath -49.75 -19.9 moveto
- 49.75 -19.9 lineto
-stroke
-newpath -49.75 19.9 moveto
- 49.75 19.9 lineto
-stroke
-showpage
-%%EOF
-
-%%EndDocument
- @endspecial 0.000000 TeXcolorgray 2324 2686 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 2324 2686
-a 2295 2729 a Fc(n)2353 2686 y
- currentpoint grestore moveto
- 2353 2686 a 1916 2395
-a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 1916 2395 a 1865 2416 a Fc(a)1916 2395 y
- currentpoint grestore moveto
- 1916 2395
-a 1946 2987 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 1946 2987 a 1659 3012 a Fc(a)28 b Fb(=)f
-Fc(l)r(og)t(sig)t Fb(\()p Fc(n)p Fb(\))2233 2987 y
- currentpoint grestore moveto
- 2233
-2987 a 1979 2706 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 1979 2706 a 1955 2738 a Fb(0)2004
-2706 y
- currentpoint grestore moveto
- 2004 2706 a 2045 2880 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 2045 2880 a 1982 2908
-a Fa(\000)p Fb(1)2109 2880 y
- currentpoint grestore moveto
- 2109 2880 a 2070 2432 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
-
-2070 2432 a 2008 2460 a Fb(+1)2133 2432 y
- currentpoint grestore moveto
- 2133 2432 a
-Black 0.000000 TeXcolorgray eop end
-%%Trailer
-
-userdict /end-hook known{end-hook}if
-%%EOF
Binary file main/nnet/doc/latex/users/octave/neuroPackage/graphics/logsig.pdf has changed
--- a/main/nnet/doc/latex/users/octave/neuroPackage/graphics/logsiglogo.eps	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,129 +0,0 @@
-%!PS-Adobe-3.0 EPSF-3.0
-%%BoundingBox: 291 381 320 410
-%%HiResBoundingBox: 291.0768 381.0768 319.9232 409.9232
-%%Creator: Asymptote 1.25
-%%CreationDate: 2007.07.02 21:46:43
-%%Pages: 1
-%%EndProlog
-%%Page: 1 1
-gsave
- 291.3268 381.3268 translate
-0 setgray
- 0 0.5 dtransform truncate idtransform setlinewidth pop
-1 setlinecap
-1 setlinejoin
-newpath 0 0 moveto
- 28.3464009 0 lineto
- 28.3464009 28.3464009 lineto
- 0 28.3464009 lineto
- 0 0 lineto
-closepath
-stroke
-newpath 2.83464009 8.50392027 moveto
- 25.5117608 8.50392027 lineto
-stroke
-newpath 2.83464009 8.707858 moveto
- 2.91024628 8.71319256 2.98583699 8.71874422 3.0614113 8.72451292 curveto
- 3.13701931 8.73028419 3.21261059 8.73627266 3.2881825 8.74249885 curveto
- 3.36379355 8.74872827 3.43938482 8.75519563 3.51495371 8.76191723 curveto
- 3.59056799 8.76864286 3.66615943 8.775623 3.74172492 8.78287639 curveto
- 3.81734299 8.79013482 3.89293458 8.79766684 3.96849613 8.80549188 curveto
- 4.04411854 8.81332323 4.11971029 8.82144801 4.19526733 8.82988682 curveto
- 4.27089472 8.8383335 4.34648667 8.84709473 4.42203854 8.85619215 curveto
- 4.49767161 8.86529935 4.57326377 8.87474338 4.64880975 8.88454693 curveto
- 4.72444931 8.89436263 4.8000417 8.90453862 4.87558095 8.9150986 curveto
- 4.95122789 8.92567364 5.02682055 8.93663363 5.10235216 8.94800318 curveto
- 5.17800746 8.95939135 5.25360041 8.97119027 5.32912337 8.9834254 curveto
- 5.40478813 8.9956835 5.48038138 9.00837924 5.55589458 9.02153882 curveto
- 5.63156998 9.03472667 5.70716356 9.0483801 5.78266578 9.06252585 curveto
- 5.85835313 9.07670629 5.93394706 9.09138118 6.00943699 9.10657764 curveto
- 6.08513768 9.12181652 6.16073197 9.13757952 6.2362082 9.1538939 curveto
- 6.31192371 9.17025999 6.38751838 9.18718053 6.4629794 9.2046826 curveto
- 6.5387113 9.2222475 6.61430634 9.24039759 6.68975061 9.2591595 curveto
- 6.76550051 9.27799742 6.84109592 9.29745149 6.91652182 9.31754749 curveto
- 6.99229137 9.33773505 7.06788712 9.35856966 7.14329303 9.3800758 curveto
- 7.21908388 9.40169172 7.29467993 9.42398518 7.37006423 9.44697891 curveto
- 7.445878 9.47010363 7.5214743 9.49393561 7.59683544 9.51849532 curveto
- 7.67267362 9.5432105 7.7482701 9.56866149 7.82360665 9.59486599 curveto
- 7.89947061 9.62125395 7.97506716 9.64840467 8.05037785 9.67633253 curveto
- 8.12626874 9.70447554 8.20186525 9.73340621 8.27714906 9.76313505 curveto
- 8.35306772 9.79311459 8.42866406 9.82390411 8.50392027 9.85550979 curveto
- 8.5798672 9.88740556 8.65546321 9.92013064 8.73069148 9.95368635 curveto
- 8.80666676 9.98757528 8.88226226 10.0223093 8.95746268 10.0578846 curveto
- 9.0334659 10.0938397 9.10906071 10.1306518 9.18423389 10.1683114 curveto
- 9.26026405 10.2064004 9.33585799 10.2453539 9.4110051 10.2851567 curveto
- 9.48706063 10.3254408 9.56265352 10.3665922 9.6377763 10.4085899 curveto
- 9.71385501 10.4511221 9.78944666 10.4945196 9.86454751 10.5387555 curveto
- 9.94064654 10.5835794 10.0162368 10.6292613 10.0913187 10.6757688 curveto
- 10.1674346 10.7229167 10.2430233 10.7709101 10.3180899 10.8197113 curveto
- 10.3942186 10.8692028 10.4698057 10.9195222 10.5448611 10.9706266 curveto
- 10.6209979 11.0224672 10.6965833 11.0751128 10.7716323 11.128516 curveto
- 10.8477721 11.1826954 10.9233559 11.2376518 10.9984035 11.2933341 curveto
- 11.0745409 11.3498248 11.1501229 11.40706 11.2251748 11.4649852 curveto
- 11.3013039 11.5237419 11.3768843 11.583206 11.451946 11.6433204 curveto
- 11.528061 11.7042784 11.6036398 11.7659027 11.6787172 11.8281342 curveto
- 11.7548121 11.8912092 11.8303894 11.9549058 11.9054884 12.0191633 curveto
- 11.9815573 12.0842508 12.0571333 12.1499121 12.1322596 12.2160854 curveto
- 12.2082968 12.283061 12.2838716 12.3505596 12.3590308 12.418519 curveto
- 12.4350308 12.4872386 12.5106045 12.5564281 12.585802 12.6260251 curveto
- 12.6617598 12.6963258 12.7373326 12.7670412 12.8125732 12.8381091 curveto
- 12.8884843 12.9098102 12.9640563 12.9818694 13.0393444 13.0542245 curveto
- 13.1152047 13.1271295 13.1907762 13.2003344 13.2661156 13.2737776 curveto
- 13.3419218 13.3476759 13.4174928 13.4218149 13.4928868 13.4961336 curveto
- 13.5686362 13.5708026 13.6442069 13.6456528 13.719658 13.7206231 curveto
- 13.7953487 13.7958314 13.8709191 13.8711606 13.9464292 13.9465501 curveto
- 14.0220599 14.0220599 14.0976302 14.0976302 14.1732004 14.1732004 curveto
- 14.2487707 14.2487707 14.324341 14.324341 14.3999717 14.3998508 curveto
- 14.4754818 14.4752403 14.5510522 14.5505695 14.6267429 14.6257778 curveto
- 14.702194 14.7007481 14.7777647 14.7755982 14.8535141 14.8502673 curveto
- 14.9289081 14.924586 15.0044791 14.998725 15.0802853 15.0726233 curveto
- 15.1556247 15.1460665 15.2311961 15.2192714 15.3070565 15.2921764 curveto
- 15.3823446 15.3645315 15.4579166 15.4365906 15.5338277 15.5082918 curveto
- 15.6090682 15.5793597 15.6846411 15.6500751 15.7605989 15.7203758 curveto
- 15.8357963 15.7899728 15.9113701 15.8591623 15.9873701 15.9278819 curveto
- 16.0625293 15.9958413 16.1381041 16.0633399 16.2141413 16.1303155 curveto
- 16.2892676 16.1964888 16.3648436 16.26215 16.4409125 16.3272376 curveto
- 16.5160115 16.3914951 16.5915888 16.4551917 16.6676837 16.5182667 curveto
- 16.7427611 16.5804982 16.8183399 16.6421225 16.8944549 16.7030805 curveto
- 16.9695166 16.7631949 17.045097 16.822659 17.1212261 16.8814157 curveto
- 17.196278 16.9393408 17.27186 16.9965761 17.3479973 17.0530668 curveto
- 17.423045 17.1087491 17.4986287 17.1637055 17.5747686 17.2178849 curveto
- 17.6498176 17.2712881 17.725403 17.3239336 17.8015398 17.3757743 curveto
- 17.8765952 17.4268787 17.9521823 17.4771981 18.028311 17.5266896 curveto
- 18.1033776 17.5754908 18.1789663 17.6234842 18.2550822 17.6706321 curveto
- 18.3301641 17.7171396 18.4057544 17.7628215 18.4818534 17.8076454 curveto
- 18.5569542 17.8518813 18.6325459 17.8952788 18.7086246 17.9378109 curveto
- 18.7837474 17.9798087 18.8593403 18.0209601 18.9353958 18.0612442 curveto
- 19.0105429 18.101047 19.0861368 18.1400005 19.162167 18.1780895 curveto
- 19.2373402 18.2157491 19.312935 18.2525612 19.3889382 18.2885163 curveto
- 19.4641386 18.3240916 19.5397341 18.3588256 19.6157094 18.3927145 curveto
- 19.6909377 18.4262703 19.7665337 18.4589953 19.8424806 18.4908911 curveto
- 19.9177368 18.5224968 19.9933332 18.5532863 20.0692518 18.5832658 curveto
- 20.1445356 18.6129947 20.2201322 18.6419253 20.296023 18.6700684 curveto
- 20.3713337 18.6979962 20.4469303 18.7251469 20.5227942 18.7515349 curveto
- 20.5981308 18.7777394 20.6737273 18.8031904 20.7495655 18.8279056 curveto
- 20.8249266 18.8524653 20.9005229 18.8762973 20.9763367 18.899422 curveto
- 21.051721 18.9224157 21.127317 18.9447092 21.2031079 18.9663251 curveto
- 21.2785138 18.9878312 21.3541095 19.0086658 21.4298791 19.0288534 curveto
- 21.505305 19.0489494 21.5809004 19.0684035 21.6566503 19.0872414 curveto
- 21.7320946 19.1060033 21.8076896 19.1241534 21.8834215 19.1417183 curveto
- 21.9588825 19.1592204 22.0344772 19.1761409 22.1101927 19.192507 curveto
- 22.1856689 19.2088214 22.2612632 19.2245844 22.3369639 19.2398233 curveto
- 22.4124538 19.2550197 22.4880478 19.2696946 22.5637351 19.283875 curveto
- 22.6392373 19.2980208 22.7148309 19.3116742 22.7905063 19.3248621 curveto
- 22.8660195 19.3380216 22.9416128 19.3507174 23.0172775 19.3629755 curveto
- 23.0928005 19.3752106 23.1683934 19.3870095 23.2440487 19.3983977 curveto
- 23.3195803 19.4097673 23.395173 19.4207273 23.4708199 19.4313023 curveto
- 23.5463592 19.4418623 23.6219516 19.4520383 23.6975911 19.461854 curveto
- 23.7731371 19.4716575 23.8487293 19.4811015 23.9243624 19.4902087 curveto
- 23.9999142 19.4993062 24.0755062 19.5080674 24.1511336 19.5165141 curveto
- 24.2266906 19.5249529 24.3022824 19.5330777 24.3779048 19.540909 curveto
- 24.4534663 19.548734 24.5290579 19.5562661 24.604676 19.5635245 curveto
- 24.6802415 19.5707779 24.7558329 19.577758 24.8314472 19.5844837 curveto
- 24.9070161 19.5912053 24.9826073 19.5976726 25.0582184 19.603902 curveto
- 25.1337903 19.6101282 25.2093816 19.6161167 25.2849896 19.621888 curveto
- 25.3605639 19.6276567 25.4361546 19.6332083 25.5117608 19.6385429 curveto
-stroke
-grestore
-showpage
-%%EOF
Binary file main/nnet/doc/latex/users/octave/neuroPackage/graphics/logsiglogo.pdf has changed
--- a/main/nnet/doc/latex/users/octave/neuroPackage/graphics/purelin.eps	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,713 +0,0 @@
-%!PS-Adobe-3.0 EPSF-3.0
-%%Creator: dvips(k) 5.94b Copyright 2004 Radical Eye Software
-%%Title: purelin_.dvi
-%%CreationDate: Mon Jul 02 22:13:33 2007
-%%Pages: 1
-%%PageOrder: Ascend
-%%BoundingBox: 255 345 356 446
-%%HiResBoundingBox: 255.5 345.5 355.5 445.5
-%%DocumentFonts: CMMI12 CMR12 CMSY10
-%%EndComments
-%DVIPSWebPage: (www.radicaleye.com)
-%DVIPSCommandLine: C:\texmf\miktex\bin\dvips.exe -R -O 127.1bp,220.7bp
-%+ -T 612bp,792bp -q -o purelin_.ps purelin_.dvi
-%DVIPSParameters: dpi=600
-%DVIPSSource:  TeX output 2007.07.02:2213
-%%BeginProcSet: tex.pro 0 0
-%!
-/TeXDict 300 dict def TeXDict begin/N{def}def/B{bind def}N/S{exch}N/X{S
-N}B/A{dup}B/TR{translate}N/isls false N/vsize 11 72 mul N/hsize 8.5 72
-mul N/landplus90{false}def/@rigin{isls{[0 landplus90{1 -1}{-1 1}ifelse 0
-0 0]concat}if 72 Resolution div 72 VResolution div neg scale isls{
-landplus90{VResolution 72 div vsize mul 0 exch}{Resolution -72 div hsize
-mul 0}ifelse TR}if Resolution VResolution vsize -72 div 1 add mul TR[
-matrix currentmatrix{A A round sub abs 0.00001 lt{round}if}forall round
-exch round exch]setmatrix}N/@landscape{/isls true N}B/@manualfeed{
-statusdict/manualfeed true put}B/@copies{/#copies X}B/FMat[1 0 0 -1 0 0]
-N/FBB[0 0 0 0]N/nn 0 N/IEn 0 N/ctr 0 N/df-tail{/nn 8 dict N nn begin
-/FontType 3 N/FontMatrix fntrx N/FontBBox FBB N string/base X array
-/BitMaps X/BuildChar{CharBuilder}N/Encoding IEn N end A{/foo setfont}2
-array copy cvx N load 0 nn put/ctr 0 N[}B/sf 0 N/df{/sf 1 N/fntrx FMat N
-df-tail}B/dfs{div/sf X/fntrx[sf 0 0 sf neg 0 0]N df-tail}B/E{pop nn A
-definefont setfont}B/Cw{Cd A length 5 sub get}B/Ch{Cd A length 4 sub get
-}B/Cx{128 Cd A length 3 sub get sub}B/Cy{Cd A length 2 sub get 127 sub}
-B/Cdx{Cd A length 1 sub get}B/Ci{Cd A type/stringtype ne{ctr get/ctr ctr
-1 add N}if}B/CharBuilder{save 3 1 roll S A/base get 2 index get S
-/BitMaps get S get/Cd X pop/ctr 0 N Cdx 0 Cx Cy Ch sub Cx Cw add Cy
-setcachedevice Cw Ch true[1 0 0 -1 -.1 Cx sub Cy .1 sub]{Ci}imagemask
-restore}B/D{/cc X A type/stringtype ne{]}if nn/base get cc ctr put nn
-/BitMaps get S ctr S sf 1 ne{A A length 1 sub A 2 index S get sf div put
-}if put/ctr ctr 1 add N}B/I{cc 1 add D}B/bop{userdict/bop-hook known{
-bop-hook}if/SI save N @rigin 0 0 moveto/V matrix currentmatrix A 1 get A
-mul exch 0 get A mul add .99 lt{/QV}{/RV}ifelse load def pop pop}N/eop{
-SI restore userdict/eop-hook known{eop-hook}if showpage}N/@start{
-userdict/start-hook known{start-hook}if pop/VResolution X/Resolution X
-1000 div/DVImag X/IEn 256 array N 2 string 0 1 255{IEn S A 360 add 36 4
-index cvrs cvn put}for pop 65781.76 div/vsize X 65781.76 div/hsize X}N
-/p{show}N/RMat[1 0 0 -1 0 0]N/BDot 260 string N/Rx 0 N/Ry 0 N/V{}B/RV/v{
-/Ry X/Rx X V}B statusdict begin/product where{pop false[(Display)(NeXT)
-(LaserWriter 16/600)]{A length product length le{A length product exch 0
-exch getinterval eq{pop true exit}if}{pop}ifelse}forall}{false}ifelse
-end{{gsave TR -.1 .1 TR 1 1 scale Rx Ry false RMat{BDot}imagemask
-grestore}}{{gsave TR -.1 .1 TR Rx Ry scale 1 1 false RMat{BDot}
-imagemask grestore}}ifelse B/QV{gsave newpath transform round exch round
-exch itransform moveto Rx 0 rlineto 0 Ry neg rlineto Rx neg 0 rlineto
-fill grestore}B/a{moveto}B/delta 0 N/tail{A/delta X 0 rmoveto}B/M{S p
-delta add tail}B/b{S p tail}B/c{-4 M}B/d{-3 M}B/e{-2 M}B/f{-1 M}B/g{0 M}
-B/h{1 M}B/i{2 M}B/j{3 M}B/k{4 M}B/w{0 rmoveto}B/l{p -4 w}B/m{p -3 w}B/n{
-p -2 w}B/o{p -1 w}B/q{p 1 w}B/r{p 2 w}B/s{p 3 w}B/t{p 4 w}B/x{0 S
-rmoveto}B/y{3 2 roll p a}B/bos{/SS save N}B/eos{SS restore}B end
-
-%%EndProcSet
-%%BeginProcSet: texps.pro 0 0
-%!
-TeXDict begin/rf{findfont dup length 1 add dict begin{1 index/FID ne 2
-index/UniqueID ne and{def}{pop pop}ifelse}forall[1 index 0 6 -1 roll
-exec 0 exch 5 -1 roll VResolution Resolution div mul neg 0 0]/Metrics
-exch def dict begin Encoding{exch dup type/integertype ne{pop pop 1 sub
-dup 0 le{pop}{[}ifelse}{FontMatrix 0 get div Metrics 0 get div def}
-ifelse}forall Metrics/Metrics currentdict end def[2 index currentdict
-end definefont 3 -1 roll makefont/setfont cvx]cvx def}def/ObliqueSlant{
-dup sin S cos div neg}B/SlantFont{4 index mul add}def/ExtendFont{3 -1
-roll mul exch}def/ReEncodeFont{CharStrings rcheck{/Encoding false def
-dup[exch{dup CharStrings exch known not{pop/.notdef/Encoding true def}
-if}forall Encoding{]exch pop}{cleartomark}ifelse}if/Encoding exch def}
-def end
-
-%%EndProcSet
-%%BeginProcSet: special.pro 0 0
-%!
-TeXDict begin/SDict 200 dict N SDict begin/@SpecialDefaults{/hs 612 N
-/vs 792 N/ho 0 N/vo 0 N/hsc 1 N/vsc 1 N/ang 0 N/CLIP 0 N/rwiSeen false N
-/rhiSeen false N/letter{}N/note{}N/a4{}N/legal{}N}B/@scaleunit 100 N
-/@hscale{@scaleunit div/hsc X}B/@vscale{@scaleunit div/vsc X}B/@hsize{
-/hs X/CLIP 1 N}B/@vsize{/vs X/CLIP 1 N}B/@clip{/CLIP 2 N}B/@hoffset{/ho
-X}B/@voffset{/vo X}B/@angle{/ang X}B/@rwi{10 div/rwi X/rwiSeen true N}B
-/@rhi{10 div/rhi X/rhiSeen true N}B/@llx{/llx X}B/@lly{/lly X}B/@urx{
-/urx X}B/@ury{/ury X}B/magscale true def end/@MacSetUp{userdict/md known
-{userdict/md get type/dicttype eq{userdict begin md length 10 add md
-maxlength ge{/md md dup length 20 add dict copy def}if end md begin
-/letter{}N/note{}N/legal{}N/od{txpose 1 0 mtx defaultmatrix dtransform S
-atan/pa X newpath clippath mark{transform{itransform moveto}}{transform{
-itransform lineto}}{6 -2 roll transform 6 -2 roll transform 6 -2 roll
-transform{itransform 6 2 roll itransform 6 2 roll itransform 6 2 roll
-curveto}}{{closepath}}pathforall newpath counttomark array astore/gc xdf
-pop ct 39 0 put 10 fz 0 fs 2 F/|______Courier fnt invertflag{PaintBlack}
-if}N/txpose{pxs pys scale ppr aload pop por{noflips{pop S neg S TR pop 1
--1 scale}if xflip yflip and{pop S neg S TR 180 rotate 1 -1 scale ppr 3
-get ppr 1 get neg sub neg ppr 2 get ppr 0 get neg sub neg TR}if xflip
-yflip not and{pop S neg S TR pop 180 rotate ppr 3 get ppr 1 get neg sub
-neg 0 TR}if yflip xflip not and{ppr 1 get neg ppr 0 get neg TR}if}{
-noflips{TR pop pop 270 rotate 1 -1 scale}if xflip yflip and{TR pop pop
-90 rotate 1 -1 scale ppr 3 get ppr 1 get neg sub neg ppr 2 get ppr 0 get
-neg sub neg TR}if xflip yflip not and{TR pop pop 90 rotate ppr 3 get ppr
-1 get neg sub neg 0 TR}if yflip xflip not and{TR pop pop 270 rotate ppr
-2 get ppr 0 get neg sub neg 0 S TR}if}ifelse scaleby96{ppr aload pop 4
--1 roll add 2 div 3 1 roll add 2 div 2 copy TR .96 dup scale neg S neg S
-TR}if}N/cp{pop pop showpage pm restore}N end}if}if}N/normalscale{
-Resolution 72 div VResolution 72 div neg scale magscale{DVImag dup scale
-}if 0 setgray}N/psfts{S 65781.76 div N}N/startTexFig{/psf$SavedState
-save N userdict maxlength dict begin/magscale true def normalscale
-currentpoint TR/psf$ury psfts/psf$urx psfts/psf$lly psfts/psf$llx psfts
-/psf$y psfts/psf$x psfts currentpoint/psf$cy X/psf$cx X/psf$sx psf$x
-psf$urx psf$llx sub div N/psf$sy psf$y psf$ury psf$lly sub div N psf$sx
-psf$sy scale psf$cx psf$sx div psf$llx sub psf$cy psf$sy div psf$ury sub
-TR/showpage{}N/erasepage{}N/copypage{}N/p 3 def @MacSetUp}N/doclip{
-psf$llx psf$lly psf$urx psf$ury currentpoint 6 2 roll newpath 4 copy 4 2
-roll moveto 6 -1 roll S lineto S lineto S lineto closepath clip newpath
-moveto}N/endTexFig{end psf$SavedState restore}N/@beginspecial{SDict
-begin/SpecialSave save N gsave normalscale currentpoint TR
-@SpecialDefaults count/ocount X/dcount countdictstack N}N/@setspecial{
-CLIP 1 eq{newpath 0 0 moveto hs 0 rlineto 0 vs rlineto hs neg 0 rlineto
-closepath clip}if ho vo TR hsc vsc scale ang rotate rwiSeen{rwi urx llx
-sub div rhiSeen{rhi ury lly sub div}{dup}ifelse scale llx neg lly neg TR
-}{rhiSeen{rhi ury lly sub div dup scale llx neg lly neg TR}if}ifelse
-CLIP 2 eq{newpath llx lly moveto urx lly lineto urx ury lineto llx ury
-lineto closepath clip}if/showpage{}N/erasepage{}N/copypage{}N newpath}N
-/@endspecial{count ocount sub{pop}repeat countdictstack dcount sub{end}
-repeat grestore SpecialSave restore end}N/@defspecial{SDict begin}N
-/@fedspecial{end}B/li{lineto}B/rl{rlineto}B/rc{rcurveto}B/np{/SaveX
-currentpoint/SaveY X N 1 setlinecap newpath}N/st{stroke SaveX SaveY
-moveto}N/fil{fill SaveX SaveY moveto}N/ellipse{/endangle X/startangle X
-/yrad X/xrad X/savematrix matrix currentmatrix N TR xrad yrad scale 0 0
-1 startangle endangle arc savematrix setmatrix}N end
-
-%%EndProcSet
-%%BeginProcSet: color.pro 0 0
-%!
-TeXDict begin/setcmykcolor where{pop}{/setcmykcolor{dup 10 eq{pop
-setrgbcolor}{1 sub 4 1 roll 3{3 index add neg dup 0 lt{pop 0}if 3 1 roll
-}repeat setrgbcolor pop}ifelse}B}ifelse/TeXcolorcmyk{setcmykcolor}def
-/TeXcolorrgb{setrgbcolor}def/TeXcolorgrey{setgray}def/TeXcolorgray{
-setgray}def/TeXcolorhsb{sethsbcolor}def/currentcmykcolor where{pop}{
-/currentcmykcolor{currentrgbcolor 10}B}ifelse/DC{exch dup userdict exch
-known{pop pop}{X}ifelse}B/GreenYellow{0.15 0 0.69 0 setcmykcolor}DC
-/Yellow{0 0 1 0 setcmykcolor}DC/Goldenrod{0 0.10 0.84 0 setcmykcolor}DC
-/Dandelion{0 0.29 0.84 0 setcmykcolor}DC/Apricot{0 0.32 0.52 0
-setcmykcolor}DC/Peach{0 0.50 0.70 0 setcmykcolor}DC/Melon{0 0.46 0.50 0
-setcmykcolor}DC/YellowOrange{0 0.42 1 0 setcmykcolor}DC/Orange{0 0.61
-0.87 0 setcmykcolor}DC/BurntOrange{0 0.51 1 0 setcmykcolor}DC
-/Bittersweet{0 0.75 1 0.24 setcmykcolor}DC/RedOrange{0 0.77 0.87 0
-setcmykcolor}DC/Mahogany{0 0.85 0.87 0.35 setcmykcolor}DC/Maroon{0 0.87
-0.68 0.32 setcmykcolor}DC/BrickRed{0 0.89 0.94 0.28 setcmykcolor}DC/Red{
-0 1 1 0 setcmykcolor}DC/OrangeRed{0 1 0.50 0 setcmykcolor}DC/RubineRed{
-0 1 0.13 0 setcmykcolor}DC/WildStrawberry{0 0.96 0.39 0 setcmykcolor}DC
-/Salmon{0 0.53 0.38 0 setcmykcolor}DC/CarnationPink{0 0.63 0 0
-setcmykcolor}DC/Magenta{0 1 0 0 setcmykcolor}DC/VioletRed{0 0.81 0 0
-setcmykcolor}DC/Rhodamine{0 0.82 0 0 setcmykcolor}DC/Mulberry{0.34 0.90
-0 0.02 setcmykcolor}DC/RedViolet{0.07 0.90 0 0.34 setcmykcolor}DC
-/Fuchsia{0.47 0.91 0 0.08 setcmykcolor}DC/Lavender{0 0.48 0 0
-setcmykcolor}DC/Thistle{0.12 0.59 0 0 setcmykcolor}DC/Orchid{0.32 0.64 0
-0 setcmykcolor}DC/DarkOrchid{0.40 0.80 0.20 0 setcmykcolor}DC/Purple{
-0.45 0.86 0 0 setcmykcolor}DC/Plum{0.50 1 0 0 setcmykcolor}DC/Violet{
-0.79 0.88 0 0 setcmykcolor}DC/RoyalPurple{0.75 0.90 0 0 setcmykcolor}DC
-/BlueViolet{0.86 0.91 0 0.04 setcmykcolor}DC/Periwinkle{0.57 0.55 0 0
-setcmykcolor}DC/CadetBlue{0.62 0.57 0.23 0 setcmykcolor}DC
-/CornflowerBlue{0.65 0.13 0 0 setcmykcolor}DC/MidnightBlue{0.98 0.13 0
-0.43 setcmykcolor}DC/NavyBlue{0.94 0.54 0 0 setcmykcolor}DC/RoyalBlue{1
-0.50 0 0 setcmykcolor}DC/Blue{1 1 0 0 setcmykcolor}DC/Cerulean{0.94 0.11
-0 0 setcmykcolor}DC/Cyan{1 0 0 0 setcmykcolor}DC/ProcessBlue{0.96 0 0 0
-setcmykcolor}DC/SkyBlue{0.62 0 0.12 0 setcmykcolor}DC/Turquoise{0.85 0
-0.20 0 setcmykcolor}DC/TealBlue{0.86 0 0.34 0.02 setcmykcolor}DC
-/Aquamarine{0.82 0 0.30 0 setcmykcolor}DC/BlueGreen{0.85 0 0.33 0
-setcmykcolor}DC/Emerald{1 0 0.50 0 setcmykcolor}DC/JungleGreen{0.99 0
-0.52 0 setcmykcolor}DC/SeaGreen{0.69 0 0.50 0 setcmykcolor}DC/Green{1 0
-1 0 setcmykcolor}DC/ForestGreen{0.91 0 0.88 0.12 setcmykcolor}DC
-/PineGreen{0.92 0 0.59 0.25 setcmykcolor}DC/LimeGreen{0.50 0 1 0
-setcmykcolor}DC/YellowGreen{0.44 0 0.74 0 setcmykcolor}DC/SpringGreen{
-0.26 0 0.76 0 setcmykcolor}DC/OliveGreen{0.64 0 0.95 0.40 setcmykcolor}
-DC/RawSienna{0 0.72 1 0.45 setcmykcolor}DC/Sepia{0 0.83 1 0.70
-setcmykcolor}DC/Brown{0 0.81 1 0.60 setcmykcolor}DC/Tan{0.14 0.42 0.56 0
-setcmykcolor}DC/Gray{0 0 0 0.50 setcmykcolor}DC/Black{0 0 0 1
-setcmykcolor}DC/White{0 0 0 0 setcmykcolor}DC end
-
-%%EndProcSet
-%%BeginFont: CMSY10
-%!PS-AdobeFont-1.1: CMSY10 1.0
-%%CreationDate: 1991 Aug 15 07:20:57
-% Copyright (C) 1997 American Mathematical Society. All Rights Reserved.
-11 dict begin
-/FontInfo 7 dict dup begin
-/version (1.0) readonly def
-/Notice (Copyright (C) 1997 American Mathematical Society. All Rights Reserved) readonly def
-/FullName (CMSY10) readonly def
-/FamilyName (Computer Modern) readonly def
-/Weight (Medium) readonly def
-/ItalicAngle -14.035 def
-/isFixedPitch false def
-end readonly def
-/FontName /CMSY10 def
-/PaintType 0 def
-/FontType 1 def
-/FontMatrix [0.001 0 0 0.001 0 0] readonly def
-/Encoding 256 array
-0 1 255 {1 index exch /.notdef put} for
-dup 0 /minus put
-readonly def
-/FontBBox{-29 -960 1116 775}readonly def
-/UniqueID 5000820 def
-currentdict end
-currentfile eexec
-D9D66F633B846A97B686A97E45A3D0AA052F09F9C8ADE9D907C058B87E9B6964
-7D53359E51216774A4EAA1E2B58EC3176BD1184A633B951372B4198D4E8C5EF4
-A213ACB58AA0A658908035BF2ED8531779838A960DFE2B27EA49C37156989C85
-E21B3ABF72E39A89232CD9F4237FC80C9E64E8425AA3BEF7DED60B122A52922A
-221A37D9A807DD01161779DDE7D31FF2B87F97C73D63EECDDA4C49501773468A
-27D1663E0B62F461F6E40A5D6676D1D12B51E641C1D4E8E2771864FC104F8CBF
-5B78EC1D88228725F1C453A678F58A7E1B7BD7CA700717D288EB8DA1F57C4F09
-0ABF1D42C5DDD0C384C7E22F8F8047BE1D4C1CC8E33368FB1AC82B4E96146730
-DE3302B2E6B819CB6AE455B1AF3187FFE8071AA57EF8A6616B9CB7941D44EC7A
-71A7BB3DF755178D7D2E4BB69859EFA4BBC30BD6BB1531133FD4D9438FF99F09
-4ECC068A324D75B5F696B8688EEB2F17E5ED34CCD6D047A4E3806D000C199D7C
-515DB70A8D4F6146FE068DC1E5DE8BC5703711DA090312BA3FC00A08C453C609
-C627A8B1550654AD5E22C5F3F3CC8C1C0A6C7ADDAB55016A76EC46213FD9BAAF
-03F7A5FD261BF647FCA5049118033F809370A84AC3ADA3D5BE032CBB494D7851
-A6242E785CCC20D81FC5EE7871F1E588DA3E31BD321C67142C5D76BC6AC708DF
-C21616B4CC92F0F8B92BD37A4AB83E066D1245FAD89B480CB0AC192D4CAFA6AD
-241BD8DF7AD566A2022FBC67364AB89F33608554113D210FE5D27F8FB1B2B78A
-F22EC999DBAAFC9C60017101D5FB2A3B6E2BF4BE47B8E5E4662B8C41AB471DFC
-A31EE1
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-cleartomark
-%%EndFont 
-%%BeginFont: CMR12
-%!PS-AdobeFont-1.1: CMR12 1.0
-%%CreationDate: 1991 Aug 20 16:38:05
-% Copyright (C) 1997 American Mathematical Society. All Rights Reserved.
-11 dict begin
-/FontInfo 7 dict dup begin
-/version (1.0) readonly def
-/Notice (Copyright (C) 1997 American Mathematical Society. All Rights Reserved) readonly def
-/FullName (CMR12) readonly def
-/FamilyName (Computer Modern) readonly def
-/Weight (Medium) readonly def
-/ItalicAngle 0 def
-/isFixedPitch false def
-end readonly def
-/FontName /CMR12 def
-/PaintType 0 def
-/FontType 1 def
-/FontMatrix [0.001 0 0 0.001 0 0] readonly def
-/Encoding 256 array
-0 1 255 {1 index exch /.notdef put} for
-dup 40 /parenleft put
-dup 41 /parenright put
-dup 43 /plus put
-dup 48 /zero put
-dup 49 /one put
-dup 61 /equal put
-readonly def
-/FontBBox{-34 -251 988 750}readonly def
-/UniqueID 5000794 def
-currentdict end
-currentfile eexec
-D9D66F633B846A97B686A97E45A3D0AA052A014267B7904EB3C0D3BD0B83D891
-016CA6CA4B712ADEB258FAAB9A130EE605E61F77FC1B738ABC7C51CD46EF8171
-9098D5FEE67660E69A7AB91B58F29A4D79E57022F783EB0FBBB6D4F4EC35014F
-D2DECBA99459A4C59DF0C6EBA150284454E707DC2100C15B76B4C19B84363758
-469A6C558785B226332152109871A9883487DD7710949204DDCF837E6A8708B8
-2BDBF16FBC7512FAA308A093FE5CF4E9D2405B169CD5365D6ECED5D768D66D6C
-68618B8C482B341F8CA38E9BB9BAFCFAAD9C2F3FD033B62690986ED43D9C9361
-3645B82392D5CAE11A7CB49D7E2E82DCD485CBA04C77322EB2E6A79D73DC194E
-59C120A2DABB9BF72E2CF256DD6EB54EECBA588101ABD933B57CE8A3A0D16B28
-51D7494F73096DF53BDC66BBF896B587DF9643317D5F610CD9088F9849126F23
-DDE030F7B277DD99055C8B119CAE9C99158AC4E150CDFC2C66ED92EBB4CC092A
-AA078CE16247A1335AD332DAA950D20395A7384C33FF72EAA31A5B89766E635F
-45C4C068AD7EE867398F0381B07CB94D29FF097D59FF9961D195A948E3D87C31
-821E9295A56D21875B41988F7A16A1587050C3C71B4E4355BB37F255D6B237CE
-96F25467F70FA19E0F85785FF49068949CCC79F2F8AE57D5F79BB9C5CF5EED5D
-9857B9967D9B96CDCF73D5D65FF75AFABB66734018BAE264597220C89FD17379
-26764A9302D078B4EB0E29178C878FD61007EEA2DDB119AE88C57ECFEF4B71E4
-140A34951DDC3568A84CC92371A789021A103A1A347050FDA6ECF7903F67D213
-1D0C7C474A9053866E9C88E65E6932BA87A73686EAB0019389F84D159809C498
-1E7A30ED942EB211B00DBFF5BCC720F4E276C3339B31B6EABBB078430E6A09BB
-377D3061A20B1EB98796B8607EECBC699445EAA866C38E02DF59F5EDD378303A
-0733B90E7835C0AAF32BA04F1566D8161EA89CD4D14DDB953F8B910BFC8A7F03
-5020F55EF8FC2640ADADA156F6CF8F2EB6610F7EE8874A26CBE7CD154469B9F4
-ED76886B3FB679FFDEB59BB6C55AF7087BA48B75EE2FB374B19BCC421A963E15
-FE05ECAAF9EECDF4B2715010A320102E6F8CCAA342FA11532671C8926C9ED415
-D9C320876265E289F7FA41B4BB6252B17463EF2AC4A92D616D39E58816A6F8F2
-367DBF4EC567A70AF0E7BD49173056591769FB20BD5048CA92C6B1994457323B
-9950B5F84037A826CC226EE233EF4D0E893CEE5C1F652F4F3E71E7CEA4A01879
-EA41FAB023FC06B7ABCF70C48E5F934B765298142FF142EBCEB4A96DD478F51E
-C4923850A838B1A21DAA720558EA0B46AA90175AC1413FC2AE9729C8D0A0AE60
-8308EF0474B68ECC49D2BDD08E003D38DD06EB2B4BFF2D670CB67075B26D39CD
-2E06571D410CAFEB8D5A5CD85316AC3480FFD6F13332CB610F821594247A8160
-A75CE2C3B81601604174C634417F1F8214BA467438F6A1AA72DF3D30195BA544
-B7EBE7B387D15C9135A3DFC67392964E192909B8F78DC39D458A5E8B6EB9EB97
-2946FE6D7A91BCED70DF5CC879A0D3386BD4A0446ACE5500A45F3976C8AE60C5
-4B18CE7283C9763C179A02BD59631825B95740BAB616858ED5FEC11D6590D4C5
-B1EBC7E78DD271A45AB212BD86297B706DDFACEE146F388A20EE30F1378F1E5C
-C4F5743EDECCF4C256A1FE53A655553DF1783C2BC6768C3A24F5D691C962691C
-2E1870D8BB49455851A5CFFFAD7D5B4045D66236FEB0F318D83788BC166A4C9E
-4EE1636CDFBB59BD8A1A6520D9F31AE3DD129D3F81B4786A82CA43B9E6BAFB55
-EED33714E2CBADEE7BF01BD2B560A3A70577D6BD9B5F05B9DA70FB0CA5676C53
-A2385727BFD5F471D5570F40FBE5F1A6BF76C0A17EBE6D468BFDB2FCE1BF1EC5
-3351B5EA44A54BF405AC94DED3DE28EFE253678550056DDEA892DB08E90541EE
-935DE706E8D1CB155DD4EB762A3E18CC7D3E7DEE85C1775A082DCA68BC4FA433
-B81F7E608FB86D6D8F30A67003DF771ACE5DA00293F1FF4137CD87ECC5713309
-E4FD2DCF054A7301462C5AB3C024CD16E8311BE610034610B13911C14A457E0E
-528345ECED9B063EF7D5C69A73CE9799CCC9A23DAC7C90C4FF29DC70025EC2D0
-736EB59000F02F27F3AD6F645B28C5C69A43EF1537E4FA44EDDE536AF5C6C5B5
-763111E88F29B86B9783623ED39EA704B38B193F6DCDF202A1AF04FCFFFDA2DC
-DF887BEA50F5800C3C821388EF3E3189067FE0541BE609FCF6E5A0DAD8C4FC1B
-EB51267D02E3CEC620AB85D8D624DB85FC04005C1AE9DCE7A209A3CD3BCF89C5
-5B3CA84ADA7CA6E3DAFB07C5E46DF7AF29F31346B395E839F074D8B889C60837
-842024F7E6A7A5C50A54AD97D89F5DCBD671B6735D6D1D4E9AA95111449EA839
-4A642ACA
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-cleartomark
-%%EndFont 
-%%BeginFont: CMMI12
-%!PS-AdobeFont-1.1: CMMI12 1.100
-%%CreationDate: 1996 Jul 27 08:57:55
-% Copyright (C) 1997 American Mathematical Society. All Rights Reserved.
-11 dict begin
-/FontInfo 7 dict dup begin
-/version (1.100) readonly def
-/Notice (Copyright (C) 1997 American Mathematical Society. All Rights Reserved) readonly def
-/FullName (CMMI12) readonly def
-/FamilyName (Computer Modern) readonly def
-/Weight (Medium) readonly def
-/ItalicAngle -14.04 def
-/isFixedPitch false def
-end readonly def
-/FontName /CMMI12 def
-/PaintType 0 def
-/FontType 1 def
-/FontMatrix [0.001 0 0 0.001 0 0] readonly def
-/Encoding 256 array
-0 1 255 {1 index exch /.notdef put} for
-dup 97 /a put
-dup 101 /e put
-dup 105 /i put
-dup 108 /l put
-dup 110 /n put
-dup 112 /p put
-dup 114 /r put
-dup 117 /u put
-readonly def
-/FontBBox{-30 -250 1026 750}readonly def
-/UniqueID 5087386 def
-currentdict end
-currentfile eexec
-D9D66F633B846A97B686A97E45A3D0AA0529731C99A784CCBE85B4993B2EEBDE
-3B12D472B7CF54651EF21185116A69AB1096ED4BAD2F646635E019B6417CC77B
-532F85D811C70D1429A19A5307EF63EB5C5E02C89FC6C20F6D9D89E7D91FE470
-B72BEFDA23F5DF76BE05AF4CE93137A219ED8A04A9D7D6FDF37E6B7FCDE0D90B
-986423E5960A5D9FBB4C956556E8DF90CBFAEC476FA36FD9A5C8175C9AF513FE
-D919C2DDD26BDC0D99398B9F4D03D6A8F05B47AF95EF28A9C561DBDC98C47CF5
-5250011D19E9366EB6FD153D3A100CAA6212E3D5D93990737F8D326D347B7EDC
-4391C9DF440285B8FC159D0E98D4258FC57892DCC57F7903449E07914FBE9E67
-3C15C2153C061EB541F66C11E7EE77D5D77C0B11E1AC55101DA976CCACAB6993
-EED1406FBB7FF30EAC9E90B90B2AF4EC7C273CA32F11A5C1426FF641B4A2FB2F
-4E68635C93DB835737567FAF8471CBC05078DCD4E40E25A2F4E5AF46C234CF59
-2A1CE8F39E1BA1B2A594355637E474167EAD4D97D51AF0A899B44387E1FD933A
-323AFDA6BA740534A510B4705C0A15647AFBF3E53A82BF320DD96753639BE49C
-2F79A1988863EF977B800C9DB5B42039C23EB86953713F730E03EA22FF7BB2C1
-D97D33FD77B1BDCC2A60B12CF7805CFC90C5B914C0F30A673DF9587F93E47CEA
-5932DD1930560C4F0D97547BCD805D6D854455B13A4D7382A22F562D7C55041F
-0FD294BDAA1834820F894265A667E5C97D95FF152531EF97258F56374502865D
-A1E7C0C5FB7C6FB7D3C43FEB3431095A59FBF6F61CEC6D6DEE09F4EB0FD70D77
-2A8B0A4984C6120293F6B947944BE23259F6EB64303D627353163B6505FC8A60
-00681F7A3968B6CBB49E0420A691258F5E7B07B417157803FCBE9B9FB1F80FD8
-CA0BD3B53F9FD95670A878A64E913C339F38088BB76CF1458CB383805AB5359E
-BA3A3537C88B7A8CAB42412D8D645B9ACA6D433E26192F237AF00ABC7BC13DA3
-0697FB05CC83DBCA72AA53D2025675BA5E2DF9E2C4A2038DCF160C2BB20B02CF
-572199F77731E00587F025D73A3C33B647F7BF6664241320667428AD844C9D84
-7A7747AB70245A91B895E0183954C8A93855793FE2798EE5543D775A3A54F98A
-822ACE46FEB009049940F99D24D7050BB8B6CE2AC2990521ED35A9488016DC83
-4C3F251F5A4E3AD99E4E5550830124778E065A4ADD58BCC72CAFAED1EB8AF229
-119A2752AA2AD62BA6AB975E71220D5E3D23E5D9175BB3BA70303F2C006EAA33
-5A906E068C3FB7CC40418179190D683917F11722A259F7809F3C7774D128B51B
-C4CC20850F9E1A5C5856D41D8FB730E04B0E747CFA9674F943E7EC7A73805701
-F9D4D8E4170264A2A8E591C7AA3445D0F86AF773B960B1F5B7604001FA594EF9
-99C3FEA429669960EEE33E1150A0FF41733D2E25CD1E5BB0CD6685444271EA2B
-F01587CE8E7034B2C4C167594B1216C2D9C75A57207AAADF134F4E9CD6AAC616
-784D91B39AD85200341B4B33127E9AD1D7C24466E5410DD289DF78C335D2BE90
-B5C581D6E02F001742B779171454EE86CBBA0E84F9060C62A29E941B69FD970E
-1FACADBA00A4434305C1B6C0FBC00C995A427BA9C84833870E3626DA568710EF
-FDCAC2756E95C19EA4A64D445BECC73A40B91077D3816D3A2E7E2804C6E43200
-1C0938D3EDA049E3430FB8AB80908250B638359944962E09541FBB7A00956D1A
-CA7D8C235656180BEB1725360319CD3B6C808BB29CE0999C081133B7A874723F
-B40B681502C7C9B4D5548FE64E93A3F9C7ED7AF3B5A8798B3C99BB075EA712B8
-7F894E9535CFAC44997D918C5FAAA3D12F1A1E761EB0B6484BC961437DC6966E
-6FB7E79FA21574B3860BCFC50C49A6A00E0623EC34F7DDE1E7F7DD84D435EF6B
-D527B0FF0B862CCAE6C59BD53FBB4AB6A40B9037BA85F5FB55D9F669168D6C63
-836FEEDD9D3D7780758992C74ADC93F937422F6389512258D3179F81831ACA5D
-9ED1AFFD9BE0B9DA241FE7A5F87430AB2D502A574FABFF76CFF2CB6B5CD36CA7
-3AEACC8546AC8F3E8B9294D341CFF798B4BB0159748B87962404D35AFB7FE0F0
-975B78ACE5052B63524281267544FFFC033ABE44B7211EA8B683D1A1E111F5B2
-8B054946786834A782663AB0281BD98890614D628FF2D0AE8E4872356D1ADEAC
-B28D8FE91B40597E2A93C2FAD5865F6433565020DCFAD67966FDDE238A806CFC
-4AA1CF35C85C8482A5DCB777B67DB60C75DFDC90A443BA4A0C1AB204F9071FB5
-9A86B1433E73AD8C89456D1C7A397C3107E613093D10FF7C0C2A333F6D4989FE
-55315C8D6D2F9AE7D64EE802DB04537177730D9C45EB86923D14ACCFB25931A3
-85CAEF6BB6C87F57B726D35C7CEB2B7E7B383134991DE010D37CDBFDCDE6560D
-C92ECE847A8DF5B49047AC03FB832343070C1A107446F5A8C3AC0F1CD98D13F0
-72DFCE06907EB8B006EAFECC668A2069171E180E7D6C4D7B3667A88DC4B897F9
-3E8EC186316C6A063C1685BC719A22755D9187AABA29C85981DDCCB36749FB4E
-5147BDB6996CD4484189F19899677EAF93E2CF7B8AFD7E9025A489F61EAF76A3
-116296EEF674F490A6E5B8C99933EDAA5AD69C26D778861DAED43F13EFCAA5B4
-F6BA2A4F99A2367CBB565642E33D1EDFB93B9CCB6A8DE9E6EBAA1284795A8AE6
-E713821DE81E216D9DA2B77E8320C39017C065EC1E291BAB4C0970E681777E43
-B02B86A4771D6CF8A2E35996835D42568ECB5021CB79E869FA8A1C6DA36B3528
-81C182EF2D162D42C9E2A461B9846F69CAAD16F8E1D0B362E93D8B0C4F98EDFC
-E80CEF27BAE611D5BDD9A4E9604E3EEC9AD4D6E78F2C6126BB6C0367528BA99A
-616BC94E08E7DC9AD5704A20357651C97F11C1CC8382ADE9372FF10059153F2C
-570640CB6F59F38F5EFCCDB25F8204A32B2B5C625EACFE2710F654456F162BBD
-4A1102628841C7D7F1CCDE34BE07EDFCABC0074EBB5905314F5495B76932ABF7
-439344B44F20C39036C428AC624539BBD90D6A18472E49396CF626E8A36278EB
-EEC0495B90BB335B57B21295647BED95E73DA15E8F0F58274807744FCC7A1D80
-8D32048390164C0E98B15BAA426FF95E38A2E3F2C4F7071AD15E6F9E29A06F97
-627254DA105043685FA4EA619D396B05A5262E0C0D825B761229E45E2502F258
-0096999F06AA62DB9F0F5D045073F46FBE2C2E24D61819DD4E95772B752F5D68
-DC8FE0BD5D469C390D307CFB740EFDCC4DFE7D880122544A26D7F74B8BF08170
-5AE2A74247F88C2554BD1390D9046AE1C3EA2714852599E5D5502EC332E8CF8D
-6DF662061DADFE0C5CF47A8C49EB5BE73480F869CB3FB796D32075F0FCC62826
-78B99A248E5643313546494FBC81204EEFE48FA74DE080ACFA414D320EDB0CC8
-9561F52B084B518D18E6FA395656239A185C31524D1DC4B38DF0AD976B16A126
-F3C3E561BB8D823D86809188D3E53C26A211F4A48269F42FF9CD7CAF18D353FD
-1C818F47121EB0F9F7BF5297F0B0D64A8A6F0E5F40EF2D53E705B8E3A0708DD0
-498F1699826F8AE9A4091BEFEDBCF29B634364D2099A7563A9C94AB5C70FD96B
-6B50962996662E878CFCD78EFCC04DE160A310A12E1041C710EDCDB52411104D
-1D06928D15ED7E9EF47BFA6D3E94ED3D08DB8C51C46D1FAAF631508A04E0B8B8
-9D522F986DA912A614CA81CC04D29904A778ECECD24F1E6B2B770D7C233063DB
-339990AA4C6790A9D18BB067893392CD19DD1A64A4BBC801AAA1E4EEFAA73165
-BADC6CC545C4CF3032A8B2C989ABD29A23321DD2EC7284D50EF6ECC9C58FAFDF
-E49693AC570EC7F13553C7F22979480366F20DA425F6F57B13D7E0DDDB9D9709
-9FA2143704485D981F17D5D60CA5D1D11DE1706DD50AAE79A9F6E4064503E9A3
-7B167C3E3BF165E15B35CDF3DF5F6CDC1905FDDC37FA3BB80F64264D7B7A5572
-0859CE5EC4743F63B0F5DFFECE1033B69D980597D2462C6EE201C9D3C58976EB
-3AD4085EE042C0764B91658631D48E94421DB1B944BE3E7DC289F31421B9CDD7
-5D542765492E873DF269BB4321843898F1E9C4268B9215DAE306805E3A808BA8
-B25B7C7A5C5BE84DEC2D7EBCD2882F91E5097C6444D4C0A026AEC0C6DE719C36
-2C2F7ECB45C39494DA998533AB69C15B1CA3F6EBE133AB28548603067A23478D
-A914502B1F12CB69DE9607507D2FDD8F16D1B158CBC8AAEFEA496B7A5E88EE8F
-9C1B7BBD7009FA5913DD08E54374D678BAEB52BFB90939B0628FA57ADA06FF2B
-538156FCC4F86631374D508D0FDFF287B1FD185FA6D734356503F339E4299F8B
-CB7FE2E118737DA4262B64ED3A07E0BF58242837E462F8C6
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-cleartomark
-%%EndFont 
-TeXDict begin 40258584 52099344 1000 600 600 (purelin_.dvi)
-@start /Fa 255[77{}1 99.6264 /CMSY10 rf /Fb 194[76 11[49
-49 4[76 1[38 38 40[{}6 99.6264 /CMR12 rf /Fc 138[56 2[44
-1[49 1[58 1[29 2[33 3[45 3[51 97[{}8 99.6264 /CMMI12
-rf end
-%%EndProlog
-%%BeginSetup
-%%Feature: *Resolution 600dpi
-TeXDict begin
- end
-%%EndSetup
-%%Page: 1 1
-TeXDict begin 1 0 bop Black Black 1529 3121 a @beginspecial
--50 @llx -50 @lly 50 @urx 50 @ury 1000 @rwi @setspecial
-%%BeginDocument: purelin_0.eps
-%!PS-Adobe-3.0 EPSF-3.0
-%%BoundingBox: -51 -50 51 50
-%%HiResBoundingBox: -50 -50 50 50
-%%Creator: Asymptote 1.25
-%%CreationDate: 2007.07.02 22:13:33
-%%Pages: 1
-%%EndProlog
-%%Page: 1 1
-0 setgray
- 0 0.5 dtransform truncate idtransform setlinewidth pop
-1 setlinecap
-1 setlinejoin
-gsave
- 0 0 translate
-newpath 42.25 0 moveto
- 11.5833333 0 -19.0833333 0 -49.75 0 curveto
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-gsave
- 0 0 translate
-newpath 49.75 1.5959456e-16 moveto
- 42.25 2.00961894 lineto
- 42.25 -2.00961894 lineto
- 44.75 -1.33974596 47.25 -0.669872981 49.75 -1.5959456e-16 curveto
- 49.75 -1.15843382e-16 49.75 1.15843382e-16 49.75 1.5959456e-16 curveto
-closepath
-fill
-grestore
-gsave
- 0 0 translate
-newpath 49.75 1.5959456e-16 moveto
- 42.25 2.00961894 lineto
- 42.25 -2.00961894 lineto
- 44.75 -1.33974596 47.25 -0.669872981 49.75 -1.5959456e-16 curveto
- 49.75 -1.15843382e-16 49.75 1.15843382e-16 49.75 1.5959456e-16 curveto
-closepath
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-gsave
- 0 0 translate
-newpath 0 27.325 moveto
- 0 6.60833333 0 -14.1083333 0 -34.825 curveto
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-gsave
- 0 0 translate
-newpath -4.23272528e-16 34.825 moveto
- -2.00961894 27.325 lineto
- 2.00961894 27.325 lineto
- 1.33974596 29.825 0.669872981 32.325 -1.50053581e-16 34.825 curveto
- -1.87503637e-16 34.825 -3.85822471e-16 34.825 -4.23272528e-16 34.825 curveto
-closepath
-fill
-grestore
-gsave
- 0 0 translate
-newpath -4.23272528e-16 34.825 moveto
- -2.00961894 27.325 lineto
- 2.00961894 27.325 lineto
- 1.33974596 29.825 0.669872981 32.325 -1.50053581e-16 34.825 curveto
- -1.87503637e-16 34.825 -3.85822471e-16 34.825 -4.23272528e-16 34.825 curveto
-closepath
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-newpath -49.75 -49.75 moveto
- -49.4183333 -49.4183333 -49.0866667 -49.0866667 -48.755 -48.755 curveto
- -48.4233333 -48.4233333 -48.0916667 -48.0916667 -47.76 -47.76 curveto
- -47.4283333 -47.4283333 -47.0966667 -47.0966667 -46.765 -46.765 curveto
- -46.4333333 -46.4333333 -46.1016667 -46.1016667 -45.77 -45.77 curveto
- -45.4383333 -45.4383333 -45.1066667 -45.1066667 -44.775 -44.775 curveto
- -44.4433333 -44.4433333 -44.1116667 -44.1116667 -43.78 -43.78 curveto
- -43.4483333 -43.4483333 -43.1166667 -43.1166667 -42.785 -42.785 curveto
- -42.4533333 -42.4533333 -42.1216667 -42.1216667 -41.79 -41.79 curveto
- -41.4583333 -41.4583333 -41.1266667 -41.1266667 -40.795 -40.795 curveto
- -40.4633333 -40.4633333 -40.1316667 -40.1316667 -39.8 -39.8 curveto
- -39.4683333 -39.4683333 -39.1366667 -39.1366667 -38.805 -38.805 curveto
- -38.4733333 -38.4733333 -38.1416667 -38.1416667 -37.81 -37.81 curveto
- -37.4783333 -37.4783333 -37.1466667 -37.1466667 -36.815 -36.815 curveto
- -36.4833333 -36.4833333 -36.1516667 -36.1516667 -35.82 -35.82 curveto
- -35.4883333 -35.4883333 -35.1566667 -35.1566667 -34.825 -34.825 curveto
- -34.4933333 -34.4933333 -34.1616667 -34.1616667 -33.83 -33.83 curveto
- -33.4983333 -33.4983333 -33.1666667 -33.1666667 -32.835 -32.835 curveto
- -32.5033333 -32.5033333 -32.1716667 -32.1716667 -31.84 -31.84 curveto
- -31.5083333 -31.5083333 -31.1766667 -31.1766667 -30.845 -30.845 curveto
- -30.5133333 -30.5133333 -30.1816667 -30.1816667 -29.85 -29.85 curveto
- -29.5183333 -29.5183333 -29.1866667 -29.1866667 -28.855 -28.855 curveto
- -28.5233333 -28.5233333 -28.1916667 -28.1916667 -27.86 -27.86 curveto
- -27.5283333 -27.5283333 -27.1966667 -27.1966667 -26.865 -26.865 curveto
- -26.5333333 -26.5333333 -26.2016667 -26.2016667 -25.87 -25.87 curveto
- -25.5383333 -25.5383333 -25.2066667 -25.2066667 -24.875 -24.875 curveto
- -24.5433333 -24.5433333 -24.2116667 -24.2116667 -23.88 -23.88 curveto
- -23.5483333 -23.5483333 -23.2166667 -23.2166667 -22.885 -22.885 curveto
- -22.5533333 -22.5533333 -22.2216667 -22.2216667 -21.89 -21.89 curveto
- -21.5583333 -21.5583333 -21.2266667 -21.2266667 -20.895 -20.895 curveto
- -20.5633333 -20.5633333 -20.2316667 -20.2316667 -19.9 -19.9 curveto
- -19.5683333 -19.5683333 -19.2366667 -19.2366667 -18.905 -18.905 curveto
- -18.5733333 -18.5733333 -18.2416667 -18.2416667 -17.91 -17.91 curveto
- -17.5783333 -17.5783333 -17.2466667 -17.2466667 -16.915 -16.915 curveto
- -16.5833333 -16.5833333 -16.2516667 -16.2516667 -15.92 -15.92 curveto
- -15.5883333 -15.5883333 -15.2566667 -15.2566667 -14.925 -14.925 curveto
- -14.5933333 -14.5933333 -14.2616667 -14.2616667 -13.93 -13.93 curveto
- -13.5983333 -13.5983333 -13.2666667 -13.2666667 -12.935 -12.935 curveto
- -12.6033333 -12.6033333 -12.2716667 -12.2716667 -11.94 -11.94 curveto
- -11.6083333 -11.6083333 -11.2766667 -11.2766667 -10.945 -10.945 curveto
- -10.6133333 -10.6133333 -10.2816667 -10.2816667 -9.95 -9.95 curveto
- -9.61833333 -9.61833333 -9.28666667 -9.28666667 -8.955 -8.955 curveto
- -8.62333333 -8.62333333 -8.29166667 -8.29166667 -7.96 -7.96 curveto
- -7.62833333 -7.62833333 -7.29666667 -7.29666667 -6.965 -6.965 curveto
- -6.63333333 -6.63333333 -6.30166667 -6.30166667 -5.97 -5.97 curveto
- -5.63833333 -5.63833333 -5.30666667 -5.30666667 -4.975 -4.975 curveto
- -4.64333333 -4.64333333 -4.31166667 -4.31166667 -3.98 -3.98 curveto
- -3.64833333 -3.64833333 -3.31666667 -3.31666667 -2.985 -2.985 curveto
- -2.65333333 -2.65333333 -2.32166667 -2.32166667 -1.99 -1.99 curveto
- -1.65833333 -1.65833333 -1.32666667 -1.32666667 -0.995 -0.995 curveto
- -0.663333333 -0.663333333 -0.331666667 -0.331666667 0 0 curveto
- 0.331666667 0.331666667 0.663333333 0.663333333 0.995 0.995 curveto
- 1.32666667 1.32666667 1.65833333 1.65833333 1.99 1.99 curveto
- 2.32166667 2.32166667 2.65333333 2.65333333 2.985 2.985 curveto
- 3.31666667 3.31666667 3.64833333 3.64833333 3.98 3.98 curveto
- 4.31166667 4.31166667 4.64333333 4.64333333 4.975 4.975 curveto
- 5.30666667 5.30666667 5.63833333 5.63833333 5.97 5.97 curveto
- 6.30166667 6.30166667 6.63333333 6.63333333 6.965 6.965 curveto
- 7.29666667 7.29666667 7.62833333 7.62833333 7.96 7.96 curveto
- 8.29166667 8.29166667 8.62333333 8.62333333 8.955 8.955 curveto
- 9.28666667 9.28666667 9.61833333 9.61833333 9.95 9.95 curveto
- 10.2816667 10.2816667 10.6133333 10.6133333 10.945 10.945 curveto
- 11.2766667 11.2766667 11.6083333 11.6083333 11.94 11.94 curveto
- 12.2716667 12.2716667 12.6033333 12.6033333 12.935 12.935 curveto
- 13.2666667 13.2666667 13.5983333 13.5983333 13.93 13.93 curveto
- 14.2616667 14.2616667 14.5933333 14.5933333 14.925 14.925 curveto
- 15.2566667 15.2566667 15.5883333 15.5883333 15.92 15.92 curveto
- 16.2516667 16.2516667 16.5833333 16.5833333 16.915 16.915 curveto
- 17.2466667 17.2466667 17.5783333 17.5783333 17.91 17.91 curveto
- 18.2416667 18.2416667 18.5733333 18.5733333 18.905 18.905 curveto
- 19.2366667 19.2366667 19.5683333 19.5683333 19.9 19.9 curveto
- 20.2316667 20.2316667 20.5633333 20.5633333 20.895 20.895 curveto
- 21.2266667 21.2266667 21.5583333 21.5583333 21.89 21.89 curveto
- 22.2216667 22.2216667 22.5533333 22.5533333 22.885 22.885 curveto
- 23.2166667 23.2166667 23.5483333 23.5483333 23.88 23.88 curveto
- 24.2116667 24.2116667 24.5433333 24.5433333 24.875 24.875 curveto
- 25.2066667 25.2066667 25.5383333 25.5383333 25.87 25.87 curveto
- 26.2016667 26.2016667 26.5333333 26.5333333 26.865 26.865 curveto
- 27.1966667 27.1966667 27.5283333 27.5283333 27.86 27.86 curveto
- 28.1916667 28.1916667 28.5233333 28.5233333 28.855 28.855 curveto
- 29.1866667 29.1866667 29.5183333 29.5183333 29.85 29.85 curveto
- 30.1816667 30.1816667 30.5133333 30.5133333 30.845 30.845 curveto
- 31.1766667 31.1766667 31.5083333 31.5083333 31.84 31.84 curveto
- 32.1716667 32.1716667 32.5033333 32.5033333 32.835 32.835 curveto
- 33.1666667 33.1666667 33.4983333 33.4983333 33.83 33.83 curveto
- 34.1616667 34.1616667 34.4933333 34.4933333 34.825 34.825 curveto
- 35.1566667 35.1566667 35.4883333 35.4883333 35.82 35.82 curveto
- 36.1516667 36.1516667 36.4833333 36.4833333 36.815 36.815 curveto
- 37.1466667 37.1466667 37.4783333 37.4783333 37.81 37.81 curveto
- 38.1416667 38.1416667 38.4733333 38.4733333 38.805 38.805 curveto
- 39.1366667 39.1366667 39.4683333 39.4683333 39.8 39.8 curveto
- 40.1316667 40.1316667 40.4633333 40.4633333 40.795 40.795 curveto
- 41.1266667 41.1266667 41.4583333 41.4583333 41.79 41.79 curveto
- 42.1216667 42.1216667 42.4533333 42.4533333 42.785 42.785 curveto
- 43.1166667 43.1166667 43.4483333 43.4483333 43.78 43.78 curveto
- 44.1116667 44.1116667 44.4433333 44.4433333 44.775 44.775 curveto
- 45.1066667 45.1066667 45.4383333 45.4383333 45.77 45.77 curveto
- 46.1016667 46.1016667 46.4333333 46.4333333 46.765 46.765 curveto
- 47.0966667 47.0966667 47.4283333 47.4283333 47.76 47.76 curveto
- 48.0916667 48.0916667 48.4233333 48.4233333 48.755 48.755 curveto
- 49.0866667 49.0866667 49.4183333 49.4183333 49.75 49.75 curveto
-stroke
-[3.98 3.98 ] 0 setdash
-newpath -49.75 -19.9 moveto
- 49.75 -19.9 lineto
-stroke
-newpath -49.75 19.9 moveto
- 49.75 19.9 lineto
-stroke
-showpage
-%%EOF
-
-%%EndDocument
- @endspecial 0.000000 TeXcolorgray 2324 2734 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 2324 2734
-a 2295 2777 a Fc(n)2353 2734 y
- currentpoint grestore moveto
- 2353 2734 a 1916 2443
-a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 1916 2443 a 1865 2464 a Fc(a)1916 2443 y
- currentpoint grestore moveto
- 1916 2443
-a 1946 3036 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 1946 3036 a 1628 3061 a Fc(a)28 b Fb(=)g
-Fc(pur)s(el)r(in)p Fb(\()p Fc(n)p Fb(\))2264 3036 y
- currentpoint grestore moveto
- 2264
-3036 a 1979 2754 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 1979 2754 a 1955 2786 a Fb(0)2004
-2754 y
- currentpoint grestore moveto
- 2004 2754 a 2045 2928 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 2045 2928 a 1982 2956
-a Fa(\000)p Fb(1)2109 2928 y
- currentpoint grestore moveto
- 2109 2928 a 2070 2480 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
-
-2070 2480 a 2008 2509 a Fb(+1)2133 2480 y
- currentpoint grestore moveto
- 2133 2480 a
-Black 0.000000 TeXcolorgray eop end
-%%Trailer
-
-userdict /end-hook known{end-hook}if
-%%EOF
Binary file main/nnet/doc/latex/users/octave/neuroPackage/graphics/purelin.pdf has changed
--- a/main/nnet/doc/latex/users/octave/neuroPackage/graphics/purelinlogo.eps	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,129 +0,0 @@
-%!PS-Adobe-3.0 EPSF-3.0
-%%BoundingBox: 291 381 320 410
-%%HiResBoundingBox: 291.0768 381.0768 319.9232 409.9232
-%%Creator: Asymptote 1.25
-%%CreationDate: 2007.07.02 22:11:43
-%%Pages: 1
-%%EndProlog
-%%Page: 1 1
-gsave
- 291.3268 381.3268 translate
-0 setgray
- 0 0.5 dtransform truncate idtransform setlinewidth pop
-1 setlinecap
-1 setlinejoin
-newpath 0 0 moveto
- 28.3464009 0 lineto
- 28.3464009 28.3464009 lineto
- 0 28.3464009 lineto
- 0 0 lineto
-closepath
-stroke
-newpath 2.83464009 14.1732004 moveto
- 25.5117608 14.1732004 lineto
-stroke
-newpath 2.83464009 2.83464009 moveto
- 2.91023049 2.91023049 2.98582089 2.98582089 3.0614113 3.0614113 curveto
- 3.1370017 3.1370017 3.2125921 3.2125921 3.2881825 3.2881825 curveto
- 3.36377291 3.36377291 3.43936331 3.43936331 3.51495371 3.51495371 curveto
- 3.59054411 3.59054411 3.66613452 3.66613452 3.74172492 3.74172492 curveto
- 3.81731532 3.81731532 3.89290572 3.89290572 3.96849613 3.96849613 curveto
- 4.04408653 4.04408653 4.11967693 4.11967693 4.19526733 4.19526733 curveto
- 4.27085773 4.27085773 4.34644814 4.34644814 4.42203854 4.42203854 curveto
- 4.49762894 4.49762894 4.57321934 4.57321934 4.64880975 4.64880975 curveto
- 4.72440015 4.72440015 4.79999055 4.79999055 4.87558095 4.87558095 curveto
- 4.95117136 4.95117136 5.02676176 5.02676176 5.10235216 5.10235216 curveto
- 5.17794256 5.17794256 5.25353297 5.25353297 5.32912337 5.32912337 curveto
- 5.40471377 5.40471377 5.48030417 5.48030417 5.55589458 5.55589458 curveto
- 5.63148498 5.63148498 5.70707538 5.70707538 5.78266578 5.78266578 curveto
- 5.85825618 5.85825618 5.93384659 5.93384659 6.00943699 6.00943699 curveto
- 6.08502739 6.08502739 6.16061779 6.16061779 6.2362082 6.2362082 curveto
- 6.3117986 6.3117986 6.387389 6.387389 6.4629794 6.4629794 curveto
- 6.53856981 6.53856981 6.61416021 6.61416021 6.68975061 6.68975061 curveto
- 6.76534101 6.76534101 6.84093142 6.84093142 6.91652182 6.91652182 curveto
- 6.99211222 6.99211222 7.06770262 7.06770262 7.14329303 7.14329303 curveto
- 7.21888343 7.21888343 7.29447383 7.29447383 7.37006423 7.37006423 curveto
- 7.44565463 7.44565463 7.52124504 7.52124504 7.59683544 7.59683544 curveto
- 7.67242584 7.67242584 7.74801624 7.74801624 7.82360665 7.82360665 curveto
- 7.89919705 7.89919705 7.97478745 7.97478745 8.05037785 8.05037785 curveto
- 8.12596826 8.12596826 8.20155866 8.20155866 8.27714906 8.27714906 curveto
- 8.35273946 8.35273946 8.42832987 8.42832987 8.50392027 8.50392027 curveto
- 8.57951067 8.57951067 8.65510107 8.65510107 8.73069148 8.73069148 curveto
- 8.80628188 8.80628188 8.88187228 8.88187228 8.95746268 8.95746268 curveto
- 9.03305308 9.03305308 9.10864349 9.10864349 9.18423389 9.18423389 curveto
- 9.25982429 9.25982429 9.33541469 9.33541469 9.4110051 9.4110051 curveto
- 9.4865955 9.4865955 9.5621859 9.5621859 9.6377763 9.6377763 curveto
- 9.71336671 9.71336671 9.78895711 9.78895711 9.86454751 9.86454751 curveto
- 9.94013791 9.94013791 10.0157283 10.0157283 10.0913187 10.0913187 curveto
- 10.1669091 10.1669091 10.2424995 10.2424995 10.3180899 10.3180899 curveto
- 10.3936803 10.3936803 10.4692707 10.4692707 10.5448611 10.5448611 curveto
- 10.6204515 10.6204515 10.6960419 10.6960419 10.7716323 10.7716323 curveto
- 10.8472227 10.8472227 10.9228131 10.9228131 10.9984035 10.9984035 curveto
- 11.0739939 11.0739939 11.1495844 11.1495844 11.2251748 11.2251748 curveto
- 11.3007652 11.3007652 11.3763556 11.3763556 11.451946 11.451946 curveto
- 11.5275364 11.5275364 11.6031268 11.6031268 11.6787172 11.6787172 curveto
- 11.7543076 11.7543076 11.829898 11.829898 11.9054884 11.9054884 curveto
- 11.9810788 11.9810788 12.0566692 12.0566692 12.1322596 12.1322596 curveto
- 12.20785 12.20785 12.2834404 12.2834404 12.3590308 12.3590308 curveto
- 12.4346212 12.4346212 12.5102116 12.5102116 12.585802 12.585802 curveto
- 12.6613924 12.6613924 12.7369828 12.7369828 12.8125732 12.8125732 curveto
- 12.8881636 12.8881636 12.963754 12.963754 13.0393444 13.0393444 curveto
- 13.1149348 13.1149348 13.1905252 13.1905252 13.2661156 13.2661156 curveto
- 13.341706 13.341706 13.4172964 13.4172964 13.4928868 13.4928868 curveto
- 13.5684772 13.5684772 13.6440676 13.6440676 13.719658 13.719658 curveto
- 13.7952484 13.7952484 13.8708388 13.8708388 13.9464292 13.9464292 curveto
- 14.0220196 14.0220196 14.09761 14.09761 14.1732004 14.1732004 curveto
- 14.2487908 14.2487908 14.3243813 14.3243813 14.3999717 14.3999717 curveto
- 14.4755621 14.4755621 14.5511525 14.5511525 14.6267429 14.6267429 curveto
- 14.7023333 14.7023333 14.7779237 14.7779237 14.8535141 14.8535141 curveto
- 14.9291045 14.9291045 15.0046949 15.0046949 15.0802853 15.0802853 curveto
- 15.1558757 15.1558757 15.2314661 15.2314661 15.3070565 15.3070565 curveto
- 15.3826469 15.3826469 15.4582373 15.4582373 15.5338277 15.5338277 curveto
- 15.6094181 15.6094181 15.6850085 15.6850085 15.7605989 15.7605989 curveto
- 15.8361893 15.8361893 15.9117797 15.9117797 15.9873701 15.9873701 curveto
- 16.0629605 16.0629605 16.1385509 16.1385509 16.2141413 16.2141413 curveto
- 16.2897317 16.2897317 16.3653221 16.3653221 16.4409125 16.4409125 curveto
- 16.5165029 16.5165029 16.5920933 16.5920933 16.6676837 16.6676837 curveto
- 16.7432741 16.7432741 16.8188645 16.8188645 16.8944549 16.8944549 curveto
- 16.9700453 16.9700453 17.0456357 17.0456357 17.1212261 17.1212261 curveto
- 17.1968165 17.1968165 17.2724069 17.2724069 17.3479973 17.3479973 curveto
- 17.4235877 17.4235877 17.4991782 17.4991782 17.5747686 17.5747686 curveto
- 17.650359 17.650359 17.7259494 17.7259494 17.8015398 17.8015398 curveto
- 17.8771302 17.8771302 17.9527206 17.9527206 18.028311 18.028311 curveto
- 18.1039014 18.1039014 18.1794918 18.1794918 18.2550822 18.2550822 curveto
- 18.3306726 18.3306726 18.406263 18.406263 18.4818534 18.4818534 curveto
- 18.5574438 18.5574438 18.6330342 18.6330342 18.7086246 18.7086246 curveto
- 18.784215 18.784215 18.8598054 18.8598054 18.9353958 18.9353958 curveto
- 19.0109862 19.0109862 19.0865766 19.0865766 19.162167 19.162167 curveto
- 19.2377574 19.2377574 19.3133478 19.3133478 19.3889382 19.3889382 curveto
- 19.4645286 19.4645286 19.540119 19.540119 19.6157094 19.6157094 curveto
- 19.6912998 19.6912998 19.7668902 19.7668902 19.8424806 19.8424806 curveto
- 19.918071 19.918071 19.9936614 19.9936614 20.0692518 20.0692518 curveto
- 20.1448422 20.1448422 20.2204326 20.2204326 20.296023 20.296023 curveto
- 20.3716134 20.3716134 20.4472038 20.4472038 20.5227942 20.5227942 curveto
- 20.5983846 20.5983846 20.6739751 20.6739751 20.7495655 20.7495655 curveto
- 20.8251559 20.8251559 20.9007463 20.9007463 20.9763367 20.9763367 curveto
- 21.0519271 21.0519271 21.1275175 21.1275175 21.2031079 21.2031079 curveto
- 21.2786983 21.2786983 21.3542887 21.3542887 21.4298791 21.4298791 curveto
- 21.5054695 21.5054695 21.5810599 21.5810599 21.6566503 21.6566503 curveto
- 21.7322407 21.7322407 21.8078311 21.8078311 21.8834215 21.8834215 curveto
- 21.9590119 21.9590119 22.0346023 22.0346023 22.1101927 22.1101927 curveto
- 22.1857831 22.1857831 22.2613735 22.2613735 22.3369639 22.3369639 curveto
- 22.4125543 22.4125543 22.4881447 22.4881447 22.5637351 22.5637351 curveto
- 22.6393255 22.6393255 22.7149159 22.7149159 22.7905063 22.7905063 curveto
- 22.8660967 22.8660967 22.9416871 22.9416871 23.0172775 23.0172775 curveto
- 23.0928679 23.0928679 23.1684583 23.1684583 23.2440487 23.2440487 curveto
- 23.3196391 23.3196391 23.3952295 23.3952295 23.4708199 23.4708199 curveto
- 23.5464103 23.5464103 23.6220007 23.6220007 23.6975911 23.6975911 curveto
- 23.7731815 23.7731815 23.848772 23.848772 23.9243624 23.9243624 curveto
- 23.9999528 23.9999528 24.0755432 24.0755432 24.1511336 24.1511336 curveto
- 24.226724 24.226724 24.3023144 24.3023144 24.3779048 24.3779048 curveto
- 24.4534952 24.4534952 24.5290856 24.5290856 24.604676 24.604676 curveto
- 24.6802664 24.6802664 24.7558568 24.7558568 24.8314472 24.8314472 curveto
- 24.9070376 24.9070376 24.982628 24.982628 25.0582184 25.0582184 curveto
- 25.1338088 25.1338088 25.2093992 25.2093992 25.2849896 25.2849896 curveto
- 25.36058 25.36058 25.4361704 25.4361704 25.5117608 25.5117608 curveto
-stroke
-grestore
-showpage
-%%EOF
Binary file main/nnet/doc/latex/users/octave/neuroPackage/graphics/purelinlogo.pdf has changed
--- a/main/nnet/doc/latex/users/octave/neuroPackage/graphics/tansig.eps	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,698 +0,0 @@
-%!PS-Adobe-3.0 EPSF-3.0
-%%Creator: dvips(k) 5.94b Copyright 2004 Radical Eye Software
-%%Title: tansig_.dvi
-%%CreationDate: Fri May 11 09:24:33 2007
-%%Pages: 1
-%%PageOrder: Ascend
-%%BoundingBox: 255 354 356 437
-%%HiResBoundingBox: 255.5 354.623708 355.5 436.376292
-%%DocumentFonts: CMMI12 CMR12 CMSY10
-%%EndComments
-%DVIPSWebPage: (www.radicaleye.com)
-%DVIPSCommandLine: C:\texmf\miktex\bin\dvips.exe -R -O 127.1bp,229.824bp
-%+ -T 612bp,792bp -q -o tansig_.ps tansig_.dvi
-%DVIPSParameters: dpi=600
-%DVIPSSource:  TeX output 2007.05.11:0924
-%%BeginProcSet: tex.pro 0 0
-%!
-/TeXDict 300 dict def TeXDict begin/N{def}def/B{bind def}N/S{exch}N/X{S
-N}B/A{dup}B/TR{translate}N/isls false N/vsize 11 72 mul N/hsize 8.5 72
-mul N/landplus90{false}def/@rigin{isls{[0 landplus90{1 -1}{-1 1}ifelse 0
-0 0]concat}if 72 Resolution div 72 VResolution div neg scale isls{
-landplus90{VResolution 72 div vsize mul 0 exch}{Resolution -72 div hsize
-mul 0}ifelse TR}if Resolution VResolution vsize -72 div 1 add mul TR[
-matrix currentmatrix{A A round sub abs 0.00001 lt{round}if}forall round
-exch round exch]setmatrix}N/@landscape{/isls true N}B/@manualfeed{
-statusdict/manualfeed true put}B/@copies{/#copies X}B/FMat[1 0 0 -1 0 0]
-N/FBB[0 0 0 0]N/nn 0 N/IEn 0 N/ctr 0 N/df-tail{/nn 8 dict N nn begin
-/FontType 3 N/FontMatrix fntrx N/FontBBox FBB N string/base X array
-/BitMaps X/BuildChar{CharBuilder}N/Encoding IEn N end A{/foo setfont}2
-array copy cvx N load 0 nn put/ctr 0 N[}B/sf 0 N/df{/sf 1 N/fntrx FMat N
-df-tail}B/dfs{div/sf X/fntrx[sf 0 0 sf neg 0 0]N df-tail}B/E{pop nn A
-definefont setfont}B/Cw{Cd A length 5 sub get}B/Ch{Cd A length 4 sub get
-}B/Cx{128 Cd A length 3 sub get sub}B/Cy{Cd A length 2 sub get 127 sub}
-B/Cdx{Cd A length 1 sub get}B/Ci{Cd A type/stringtype ne{ctr get/ctr ctr
-1 add N}if}B/CharBuilder{save 3 1 roll S A/base get 2 index get S
-/BitMaps get S get/Cd X pop/ctr 0 N Cdx 0 Cx Cy Ch sub Cx Cw add Cy
-setcachedevice Cw Ch true[1 0 0 -1 -.1 Cx sub Cy .1 sub]{Ci}imagemask
-restore}B/D{/cc X A type/stringtype ne{]}if nn/base get cc ctr put nn
-/BitMaps get S ctr S sf 1 ne{A A length 1 sub A 2 index S get sf div put
-}if put/ctr ctr 1 add N}B/I{cc 1 add D}B/bop{userdict/bop-hook known{
-bop-hook}if/SI save N @rigin 0 0 moveto/V matrix currentmatrix A 1 get A
-mul exch 0 get A mul add .99 lt{/QV}{/RV}ifelse load def pop pop}N/eop{
-SI restore userdict/eop-hook known{eop-hook}if showpage}N/@start{
-userdict/start-hook known{start-hook}if pop/VResolution X/Resolution X
-1000 div/DVImag X/IEn 256 array N 2 string 0 1 255{IEn S A 360 add 36 4
-index cvrs cvn put}for pop 65781.76 div/vsize X 65781.76 div/hsize X}N
-/p{show}N/RMat[1 0 0 -1 0 0]N/BDot 260 string N/Rx 0 N/Ry 0 N/V{}B/RV/v{
-/Ry X/Rx X V}B statusdict begin/product where{pop false[(Display)(NeXT)
-(LaserWriter 16/600)]{A length product length le{A length product exch 0
-exch getinterval eq{pop true exit}if}{pop}ifelse}forall}{false}ifelse
-end{{gsave TR -.1 .1 TR 1 1 scale Rx Ry false RMat{BDot}imagemask
-grestore}}{{gsave TR -.1 .1 TR Rx Ry scale 1 1 false RMat{BDot}
-imagemask grestore}}ifelse B/QV{gsave newpath transform round exch round
-exch itransform moveto Rx 0 rlineto 0 Ry neg rlineto Rx neg 0 rlineto
-fill grestore}B/a{moveto}B/delta 0 N/tail{A/delta X 0 rmoveto}B/M{S p
-delta add tail}B/b{S p tail}B/c{-4 M}B/d{-3 M}B/e{-2 M}B/f{-1 M}B/g{0 M}
-B/h{1 M}B/i{2 M}B/j{3 M}B/k{4 M}B/w{0 rmoveto}B/l{p -4 w}B/m{p -3 w}B/n{
-p -2 w}B/o{p -1 w}B/q{p 1 w}B/r{p 2 w}B/s{p 3 w}B/t{p 4 w}B/x{0 S
-rmoveto}B/y{3 2 roll p a}B/bos{/SS save N}B/eos{SS restore}B end
-
-%%EndProcSet
-%%BeginProcSet: texps.pro 0 0
-%!
-TeXDict begin/rf{findfont dup length 1 add dict begin{1 index/FID ne 2
-index/UniqueID ne and{def}{pop pop}ifelse}forall[1 index 0 6 -1 roll
-exec 0 exch 5 -1 roll VResolution Resolution div mul neg 0 0]/Metrics
-exch def dict begin Encoding{exch dup type/integertype ne{pop pop 1 sub
-dup 0 le{pop}{[}ifelse}{FontMatrix 0 get div Metrics 0 get div def}
-ifelse}forall Metrics/Metrics currentdict end def[2 index currentdict
-end definefont 3 -1 roll makefont/setfont cvx]cvx def}def/ObliqueSlant{
-dup sin S cos div neg}B/SlantFont{4 index mul add}def/ExtendFont{3 -1
-roll mul exch}def/ReEncodeFont{CharStrings rcheck{/Encoding false def
-dup[exch{dup CharStrings exch known not{pop/.notdef/Encoding true def}
-if}forall Encoding{]exch pop}{cleartomark}ifelse}if/Encoding exch def}
-def end
-
-%%EndProcSet
-%%BeginProcSet: special.pro 0 0
-%!
-TeXDict begin/SDict 200 dict N SDict begin/@SpecialDefaults{/hs 612 N
-/vs 792 N/ho 0 N/vo 0 N/hsc 1 N/vsc 1 N/ang 0 N/CLIP 0 N/rwiSeen false N
-/rhiSeen false N/letter{}N/note{}N/a4{}N/legal{}N}B/@scaleunit 100 N
-/@hscale{@scaleunit div/hsc X}B/@vscale{@scaleunit div/vsc X}B/@hsize{
-/hs X/CLIP 1 N}B/@vsize{/vs X/CLIP 1 N}B/@clip{/CLIP 2 N}B/@hoffset{/ho
-X}B/@voffset{/vo X}B/@angle{/ang X}B/@rwi{10 div/rwi X/rwiSeen true N}B
-/@rhi{10 div/rhi X/rhiSeen true N}B/@llx{/llx X}B/@lly{/lly X}B/@urx{
-/urx X}B/@ury{/ury X}B/magscale true def end/@MacSetUp{userdict/md known
-{userdict/md get type/dicttype eq{userdict begin md length 10 add md
-maxlength ge{/md md dup length 20 add dict copy def}if end md begin
-/letter{}N/note{}N/legal{}N/od{txpose 1 0 mtx defaultmatrix dtransform S
-atan/pa X newpath clippath mark{transform{itransform moveto}}{transform{
-itransform lineto}}{6 -2 roll transform 6 -2 roll transform 6 -2 roll
-transform{itransform 6 2 roll itransform 6 2 roll itransform 6 2 roll
-curveto}}{{closepath}}pathforall newpath counttomark array astore/gc xdf
-pop ct 39 0 put 10 fz 0 fs 2 F/|______Courier fnt invertflag{PaintBlack}
-if}N/txpose{pxs pys scale ppr aload pop por{noflips{pop S neg S TR pop 1
--1 scale}if xflip yflip and{pop S neg S TR 180 rotate 1 -1 scale ppr 3
-get ppr 1 get neg sub neg ppr 2 get ppr 0 get neg sub neg TR}if xflip
-yflip not and{pop S neg S TR pop 180 rotate ppr 3 get ppr 1 get neg sub
-neg 0 TR}if yflip xflip not and{ppr 1 get neg ppr 0 get neg TR}if}{
-noflips{TR pop pop 270 rotate 1 -1 scale}if xflip yflip and{TR pop pop
-90 rotate 1 -1 scale ppr 3 get ppr 1 get neg sub neg ppr 2 get ppr 0 get
-neg sub neg TR}if xflip yflip not and{TR pop pop 90 rotate ppr 3 get ppr
-1 get neg sub neg 0 TR}if yflip xflip not and{TR pop pop 270 rotate ppr
-2 get ppr 0 get neg sub neg 0 S TR}if}ifelse scaleby96{ppr aload pop 4
--1 roll add 2 div 3 1 roll add 2 div 2 copy TR .96 dup scale neg S neg S
-TR}if}N/cp{pop pop showpage pm restore}N end}if}if}N/normalscale{
-Resolution 72 div VResolution 72 div neg scale magscale{DVImag dup scale
-}if 0 setgray}N/psfts{S 65781.76 div N}N/startTexFig{/psf$SavedState
-save N userdict maxlength dict begin/magscale true def normalscale
-currentpoint TR/psf$ury psfts/psf$urx psfts/psf$lly psfts/psf$llx psfts
-/psf$y psfts/psf$x psfts currentpoint/psf$cy X/psf$cx X/psf$sx psf$x
-psf$urx psf$llx sub div N/psf$sy psf$y psf$ury psf$lly sub div N psf$sx
-psf$sy scale psf$cx psf$sx div psf$llx sub psf$cy psf$sy div psf$ury sub
-TR/showpage{}N/erasepage{}N/copypage{}N/p 3 def @MacSetUp}N/doclip{
-psf$llx psf$lly psf$urx psf$ury currentpoint 6 2 roll newpath 4 copy 4 2
-roll moveto 6 -1 roll S lineto S lineto S lineto closepath clip newpath
-moveto}N/endTexFig{end psf$SavedState restore}N/@beginspecial{SDict
-begin/SpecialSave save N gsave normalscale currentpoint TR
-@SpecialDefaults count/ocount X/dcount countdictstack N}N/@setspecial{
-CLIP 1 eq{newpath 0 0 moveto hs 0 rlineto 0 vs rlineto hs neg 0 rlineto
-closepath clip}if ho vo TR hsc vsc scale ang rotate rwiSeen{rwi urx llx
-sub div rhiSeen{rhi ury lly sub div}{dup}ifelse scale llx neg lly neg TR
-}{rhiSeen{rhi ury lly sub div dup scale llx neg lly neg TR}if}ifelse
-CLIP 2 eq{newpath llx lly moveto urx lly lineto urx ury lineto llx ury
-lineto closepath clip}if/showpage{}N/erasepage{}N/copypage{}N newpath}N
-/@endspecial{count ocount sub{pop}repeat countdictstack dcount sub{end}
-repeat grestore SpecialSave restore end}N/@defspecial{SDict begin}N
-/@fedspecial{end}B/li{lineto}B/rl{rlineto}B/rc{rcurveto}B/np{/SaveX
-currentpoint/SaveY X N 1 setlinecap newpath}N/st{stroke SaveX SaveY
-moveto}N/fil{fill SaveX SaveY moveto}N/ellipse{/endangle X/startangle X
-/yrad X/xrad X/savematrix matrix currentmatrix N TR xrad yrad scale 0 0
-1 startangle endangle arc savematrix setmatrix}N end
-
-%%EndProcSet
-%%BeginProcSet: color.pro 0 0
-%!
-TeXDict begin/setcmykcolor where{pop}{/setcmykcolor{dup 10 eq{pop
-setrgbcolor}{1 sub 4 1 roll 3{3 index add neg dup 0 lt{pop 0}if 3 1 roll
-}repeat setrgbcolor pop}ifelse}B}ifelse/TeXcolorcmyk{setcmykcolor}def
-/TeXcolorrgb{setrgbcolor}def/TeXcolorgrey{setgray}def/TeXcolorgray{
-setgray}def/TeXcolorhsb{sethsbcolor}def/currentcmykcolor where{pop}{
-/currentcmykcolor{currentrgbcolor 10}B}ifelse/DC{exch dup userdict exch
-known{pop pop}{X}ifelse}B/GreenYellow{0.15 0 0.69 0 setcmykcolor}DC
-/Yellow{0 0 1 0 setcmykcolor}DC/Goldenrod{0 0.10 0.84 0 setcmykcolor}DC
-/Dandelion{0 0.29 0.84 0 setcmykcolor}DC/Apricot{0 0.32 0.52 0
-setcmykcolor}DC/Peach{0 0.50 0.70 0 setcmykcolor}DC/Melon{0 0.46 0.50 0
-setcmykcolor}DC/YellowOrange{0 0.42 1 0 setcmykcolor}DC/Orange{0 0.61
-0.87 0 setcmykcolor}DC/BurntOrange{0 0.51 1 0 setcmykcolor}DC
-/Bittersweet{0 0.75 1 0.24 setcmykcolor}DC/RedOrange{0 0.77 0.87 0
-setcmykcolor}DC/Mahogany{0 0.85 0.87 0.35 setcmykcolor}DC/Maroon{0 0.87
-0.68 0.32 setcmykcolor}DC/BrickRed{0 0.89 0.94 0.28 setcmykcolor}DC/Red{
-0 1 1 0 setcmykcolor}DC/OrangeRed{0 1 0.50 0 setcmykcolor}DC/RubineRed{
-0 1 0.13 0 setcmykcolor}DC/WildStrawberry{0 0.96 0.39 0 setcmykcolor}DC
-/Salmon{0 0.53 0.38 0 setcmykcolor}DC/CarnationPink{0 0.63 0 0
-setcmykcolor}DC/Magenta{0 1 0 0 setcmykcolor}DC/VioletRed{0 0.81 0 0
-setcmykcolor}DC/Rhodamine{0 0.82 0 0 setcmykcolor}DC/Mulberry{0.34 0.90
-0 0.02 setcmykcolor}DC/RedViolet{0.07 0.90 0 0.34 setcmykcolor}DC
-/Fuchsia{0.47 0.91 0 0.08 setcmykcolor}DC/Lavender{0 0.48 0 0
-setcmykcolor}DC/Thistle{0.12 0.59 0 0 setcmykcolor}DC/Orchid{0.32 0.64 0
-0 setcmykcolor}DC/DarkOrchid{0.40 0.80 0.20 0 setcmykcolor}DC/Purple{
-0.45 0.86 0 0 setcmykcolor}DC/Plum{0.50 1 0 0 setcmykcolor}DC/Violet{
-0.79 0.88 0 0 setcmykcolor}DC/RoyalPurple{0.75 0.90 0 0 setcmykcolor}DC
-/BlueViolet{0.86 0.91 0 0.04 setcmykcolor}DC/Periwinkle{0.57 0.55 0 0
-setcmykcolor}DC/CadetBlue{0.62 0.57 0.23 0 setcmykcolor}DC
-/CornflowerBlue{0.65 0.13 0 0 setcmykcolor}DC/MidnightBlue{0.98 0.13 0
-0.43 setcmykcolor}DC/NavyBlue{0.94 0.54 0 0 setcmykcolor}DC/RoyalBlue{1
-0.50 0 0 setcmykcolor}DC/Blue{1 1 0 0 setcmykcolor}DC/Cerulean{0.94 0.11
-0 0 setcmykcolor}DC/Cyan{1 0 0 0 setcmykcolor}DC/ProcessBlue{0.96 0 0 0
-setcmykcolor}DC/SkyBlue{0.62 0 0.12 0 setcmykcolor}DC/Turquoise{0.85 0
-0.20 0 setcmykcolor}DC/TealBlue{0.86 0 0.34 0.02 setcmykcolor}DC
-/Aquamarine{0.82 0 0.30 0 setcmykcolor}DC/BlueGreen{0.85 0 0.33 0
-setcmykcolor}DC/Emerald{1 0 0.50 0 setcmykcolor}DC/JungleGreen{0.99 0
-0.52 0 setcmykcolor}DC/SeaGreen{0.69 0 0.50 0 setcmykcolor}DC/Green{1 0
-1 0 setcmykcolor}DC/ForestGreen{0.91 0 0.88 0.12 setcmykcolor}DC
-/PineGreen{0.92 0 0.59 0.25 setcmykcolor}DC/LimeGreen{0.50 0 1 0
-setcmykcolor}DC/YellowGreen{0.44 0 0.74 0 setcmykcolor}DC/SpringGreen{
-0.26 0 0.76 0 setcmykcolor}DC/OliveGreen{0.64 0 0.95 0.40 setcmykcolor}
-DC/RawSienna{0 0.72 1 0.45 setcmykcolor}DC/Sepia{0 0.83 1 0.70
-setcmykcolor}DC/Brown{0 0.81 1 0.60 setcmykcolor}DC/Tan{0.14 0.42 0.56 0
-setcmykcolor}DC/Gray{0 0 0 0.50 setcmykcolor}DC/Black{0 0 0 1
-setcmykcolor}DC/White{0 0 0 0 setcmykcolor}DC end
-
-%%EndProcSet
-%%BeginFont: CMSY10
-%!PS-AdobeFont-1.1: CMSY10 1.0
-%%CreationDate: 1991 Aug 15 07:20:57
-% Copyright (C) 1997 American Mathematical Society. All Rights Reserved.
-11 dict begin
-/FontInfo 7 dict dup begin
-/version (1.0) readonly def
-/Notice (Copyright (C) 1997 American Mathematical Society. All Rights Reserved) readonly def
-/FullName (CMSY10) readonly def
-/FamilyName (Computer Modern) readonly def
-/Weight (Medium) readonly def
-/ItalicAngle -14.035 def
-/isFixedPitch false def
-end readonly def
-/FontName /CMSY10 def
-/PaintType 0 def
-/FontType 1 def
-/FontMatrix [0.001 0 0 0.001 0 0] readonly def
-/Encoding 256 array
-0 1 255 {1 index exch /.notdef put} for
-dup 0 /minus put
-readonly def
-/FontBBox{-29 -960 1116 775}readonly def
-/UniqueID 5000820 def
-currentdict end
-currentfile eexec
-D9D66F633B846A97B686A97E45A3D0AA052F09F9C8ADE9D907C058B87E9B6964
-7D53359E51216774A4EAA1E2B58EC3176BD1184A633B951372B4198D4E8C5EF4
-A213ACB58AA0A658908035BF2ED8531779838A960DFE2B27EA49C37156989C85
-E21B3ABF72E39A89232CD9F4237FC80C9E64E8425AA3BEF7DED60B122A52922A
-221A37D9A807DD01161779DDE7D31FF2B87F97C73D63EECDDA4C49501773468A
-27D1663E0B62F461F6E40A5D6676D1D12B51E641C1D4E8E2771864FC104F8CBF
-5B78EC1D88228725F1C453A678F58A7E1B7BD7CA700717D288EB8DA1F57C4F09
-0ABF1D42C5DDD0C384C7E22F8F8047BE1D4C1CC8E33368FB1AC82B4E96146730
-DE3302B2E6B819CB6AE455B1AF3187FFE8071AA57EF8A6616B9CB7941D44EC7A
-71A7BB3DF755178D7D2E4BB69859EFA4BBC30BD6BB1531133FD4D9438FF99F09
-4ECC068A324D75B5F696B8688EEB2F17E5ED34CCD6D047A4E3806D000C199D7C
-515DB70A8D4F6146FE068DC1E5DE8BC5703711DA090312BA3FC00A08C453C609
-C627A8B1550654AD5E22C5F3F3CC8C1C0A6C7ADDAB55016A76EC46213FD9BAAF
-03F7A5FD261BF647FCA5049118033F809370A84AC3ADA3D5BE032CBB494D7851
-A6242E785CCC20D81FC5EE7871F1E588DA3E31BD321C67142C5D76BC6AC708DF
-C21616B4CC92F0F8B92BD37A4AB83E066D1245FAD89B480CB0AC192D4CAFA6AD
-241BD8DF7AD566A2022FBC67364AB89F33608554113D210FE5D27F8FB1B2B78A
-F22EC999DBAAFC9C60017101D5FB2A3B6E2BF4BE47B8E5E4662B8C41AB471DFC
-A31EE1
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-cleartomark
-%%EndFont 
-%%BeginFont: CMR12
-%!PS-AdobeFont-1.1: CMR12 1.0
-%%CreationDate: 1991 Aug 20 16:38:05
-% Copyright (C) 1997 American Mathematical Society. All Rights Reserved.
-11 dict begin
-/FontInfo 7 dict dup begin
-/version (1.0) readonly def
-/Notice (Copyright (C) 1997 American Mathematical Society. All Rights Reserved) readonly def
-/FullName (CMR12) readonly def
-/FamilyName (Computer Modern) readonly def
-/Weight (Medium) readonly def
-/ItalicAngle 0 def
-/isFixedPitch false def
-end readonly def
-/FontName /CMR12 def
-/PaintType 0 def
-/FontType 1 def
-/FontMatrix [0.001 0 0 0.001 0 0] readonly def
-/Encoding 256 array
-0 1 255 {1 index exch /.notdef put} for
-dup 40 /parenleft put
-dup 41 /parenright put
-dup 43 /plus put
-dup 48 /zero put
-dup 49 /one put
-dup 61 /equal put
-readonly def
-/FontBBox{-34 -251 988 750}readonly def
-/UniqueID 5000794 def
-currentdict end
-currentfile eexec
-D9D66F633B846A97B686A97E45A3D0AA052A014267B7904EB3C0D3BD0B83D891
-016CA6CA4B712ADEB258FAAB9A130EE605E61F77FC1B738ABC7C51CD46EF8171
-9098D5FEE67660E69A7AB91B58F29A4D79E57022F783EB0FBBB6D4F4EC35014F
-D2DECBA99459A4C59DF0C6EBA150284454E707DC2100C15B76B4C19B84363758
-469A6C558785B226332152109871A9883487DD7710949204DDCF837E6A8708B8
-2BDBF16FBC7512FAA308A093FE5CF4E9D2405B169CD5365D6ECED5D768D66D6C
-68618B8C482B341F8CA38E9BB9BAFCFAAD9C2F3FD033B62690986ED43D9C9361
-3645B82392D5CAE11A7CB49D7E2E82DCD485CBA04C77322EB2E6A79D73DC194E
-59C120A2DABB9BF72E2CF256DD6EB54EECBA588101ABD933B57CE8A3A0D16B28
-51D7494F73096DF53BDC66BBF896B587DF9643317D5F610CD9088F9849126F23
-DDE030F7B277DD99055C8B119CAE9C99158AC4E150CDFC2C66ED92EBB4CC092A
-AA078CE16247A1335AD332DAA950D20395A7384C33FF72EAA31A5B89766E635F
-45C4C068AD7EE867398F0381B07CB94D29FF097D59FF9961D195A948E3D87C31
-821E9295A56D21875B41988F7A16A1587050C3C71B4E4355BB37F255D6B237CE
-96F25467F70FA19E0F85785FF49068949CCC79F2F8AE57D5F79BB9C5CF5EED5D
-9857B9967D9B96CDCF73D5D65FF75AFABB66734018BAE264597220C89FD17379
-26764A9302D078B4EB0E29178C878FD61007EEA2DDB119AE88C57ECFEF4B71E4
-140A34951DDC3568A84CC92371A789021A103A1A347050FDA6ECF7903F67D213
-1D0C7C474A9053866E9C88E65E6932BA87A73686EAB0019389F84D159809C498
-1E7A30ED942EB211B00DBFF5BCC720F4E276C3339B31B6EABBB078430E6A09BB
-377D3061A20B1EB98796B8607EECBC699445EAA866C38E02DF59F5EDD378303A
-0733B90E7835C0AAF32BA04F1566D8161EA89CD4D14DDB953F8B910BFC8A7F03
-5020F55EF8FC2640ADADA156F6CF8F2EB6610F7EE8874A26CBE7CD154469B9F4
-ED76886B3FB679FFDEB59BB6C55AF7087BA48B75EE2FB374B19BCC421A963E15
-FE05ECAAF9EECDF4B2715010A320102E6F8CCAA342FA11532671C8926C9ED415
-D9C320876265E289F7FA41B4BB6252B17463EF2AC4A92D616D39E58816A6F8F2
-367DBF4EC567A70AF0E7BD49173056591769FB20BD5048CA92C6B1994457323B
-9950B5F84037A826CC226EE233EF4D0E893CEE5C1F652F4F3E71E7CEA4A01879
-EA41FAB023FC06B7ABCF70C48E5F934B765298142FF142EBCEB4A96DD478F51E
-C4923850A838B1A21DAA720558EA0B46AA90175AC1413FC2AE9729C8D0A0AE60
-8308EF0474B68ECC49D2BDD08E003D38DD06EB2B4BFF2D670CB67075B26D39CD
-2E06571D410CAFEB8D5A5CD85316AC3480FFD6F13332CB610F821594247A8160
-A75CE2C3B81601604174C634417F1F8214BA467438F6A1AA72DF3D30195BA544
-B7EBE7B387D15C9135A3DFC67392964E192909B8F78DC39D458A5E8B6EB9EB97
-2946FE6D7A91BCED70DF5CC879A0D3386BD4A0446ACE5500A45F3976C8AE60C5
-4B18CE7283C9763C179A02BD59631825B95740BAB616858ED5FEC11D6590D4C5
-B1EBC7E78DD271A45AB212BD86297B706DDFACEE146F388A20EE30F1378F1E5C
-C4F5743EDECCF4C256A1FE53A655553DF1783C2BC6768C3A24F5D691C962691C
-2E1870D8BB49455851A5CFFFAD7D5B4045D66236FEB0F318D83788BC166A4C9E
-4EE1636CDFBB59BD8A1A6520D9F31AE3DD129D3F81B4786A82CA43B9E6BAFB55
-EED33714E2CBADEE7BF01BD2B560A3A70577D6BD9B5F05B9DA70FB0CA5676C53
-A2385727BFD5F471D5570F40FBE5F1A6BF76C0A17EBE6D468BFDB2FCE1BF1EC5
-3351B5EA44A54BF405AC94DED3DE28EFE253678550056DDEA892DB08E90541EE
-935DE706E8D1CB155DD4EB762A3E18CC7D3E7DEE85C1775A082DCA68BC4FA433
-B81F7E608FB86D6D8F30A67003DF771ACE5DA00293F1FF4137CD87ECC5713309
-E4FD2DCF054A7301462C5AB3C024CD16E8311BE610034610B13911C14A457E0E
-528345ECED9B063EF7D5C69A73CE9799CCC9A23DAC7C90C4FF29DC70025EC2D0
-736EB59000F02F27F3AD6F645B28C5C69A43EF1537E4FA44EDDE536AF5C6C5B5
-763111E88F29B86B9783623ED39EA704B38B193F6DCDF202A1AF04FCFFFDA2DC
-DF887BEA50F5800C3C821388EF3E3189067FE0541BE609FCF6E5A0DAD8C4FC1B
-EB51267D02E3CEC620AB85D8D624DB85FC04005C1AE9DCE7A209A3CD3BCF89C5
-5B3CA84ADA7CA6E3DAFB07C5E46DF7AF29F31346B395E839F074D8B889C60837
-842024F7E6A7A5C50A54AD97D89F5DCBD671B6735D6D1D4E9AA95111449EA839
-4A642ACA
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-cleartomark
-%%EndFont 
-%%BeginFont: CMMI12
-%!PS-AdobeFont-1.1: CMMI12 1.100
-%%CreationDate: 1996 Jul 27 08:57:55
-% Copyright (C) 1997 American Mathematical Society. All Rights Reserved.
-11 dict begin
-/FontInfo 7 dict dup begin
-/version (1.100) readonly def
-/Notice (Copyright (C) 1997 American Mathematical Society. All Rights Reserved) readonly def
-/FullName (CMMI12) readonly def
-/FamilyName (Computer Modern) readonly def
-/Weight (Medium) readonly def
-/ItalicAngle -14.04 def
-/isFixedPitch false def
-end readonly def
-/FontName /CMMI12 def
-/PaintType 0 def
-/FontType 1 def
-/FontMatrix [0.001 0 0 0.001 0 0] readonly def
-/Encoding 256 array
-0 1 255 {1 index exch /.notdef put} for
-dup 97 /a put
-dup 103 /g put
-dup 105 /i put
-dup 110 /n put
-dup 115 /s put
-dup 116 /t put
-readonly def
-/FontBBox{-30 -250 1026 750}readonly def
-/UniqueID 5087386 def
-currentdict end
-currentfile eexec
-D9D66F633B846A97B686A97E45A3D0AA0529731C99A784CCBE85B4993B2EEBDE
-3B12D472B7CF54651EF21185116A69AB1096ED4BAD2F646635E019B6417CC77B
-532F85D811C70D1429A19A5307EF63EB5C5E02C89FC6C20F6D9D89E7D91FE470
-B72BEFDA23F5DF76BE05AF4CE93137A219ED8A04A9D7D6FDF37E6B7FCDE0D90B
-986423E5960A5D9FBB4C956556E8DF90CBFAEC476FA36FD9A5C8175C9AF513FE
-D919C2DDD26BDC0D99398B9F4D03D6A8F05B47AF95EF28A9C561DBDC98C47CF5
-5250011D19E9366EB6FD153D3A100CAA6212E3D5D93990737F8D326D347B7EDC
-4391C9DF440285B8FC159D0E98D4258FC57892DCC57F7903449E07914FBE9E67
-3C15C2153C061EB541F66C11E7EE77D5D77C0B11E1AC55101DA976CCACAB6993
-EED1406FBB7FF30EAC9E90B90B2AF4EC7C273CA32F11A5C1426FF641B4A2FB2F
-4E68635C93DB835737567FAF8471CBC05078DCD4E40E25A2F4E5AF46C234CF59
-2A1CE8F39E1BA1B2A594355637E474167EAD4D97D51AF0A899B44387E1FD933A
-323AFDA6BA740534A510B4705C0A15647AFBF3E53A82BF320DD96753639BE49C
-2F79A1988863EF977B800C9DB5B42039C23EB86953713F730E03EA22FF7BB2C1
-D97D33FD77B1BDCC2A60B12CF7805CFC90C5B914C0F30A673DF9587F93E47CEA
-5932DD1930560C4F0D97547BCD805D6D854455B13A4D7382A22F562D7C55041F
-0FD294BDAA1834820F894265A667E5C97D95FF152531EF97258F56374502865D
-A1E7C0C5FB7C6FB7D3C43FEB3431095A59FBF6F61CEC6D6DEE09F4EB0FD70D77
-2A8B0A4984C6120293F6B947944BE23259F6EB64303D627353163B6505FC8A60
-00681F7A3968B6CBB49E0420A691258F5E7B07B417157803FCBE9B9FB1F80FD8
-CA0BD2E774E4D04F1F0CB9AD88152DF9799FB90EC43955871EB7F0338141CF69
-3A94F81431168EFFF7462ABF70F1AAD9909E0183601E417073F4EC7DF0180A48
-73C309956ED2BC852965D7D4EF3F2A3F2A798CD61AE418D9573497D3911F5323
-ED3496F6AEBE685EE322F58EA7402EF6A7B6EB9E433EB7D0F6E3C3BDAD24F983
-AC4415A43C9687642E3BF1E4F4A99F03FA39177E5FFF4A9205E20954906ACE66
-1BF1C9E2E43707530FF446F58B37C73CF2857A7ABB3355DC42F2E66AAA8E40FB
-4F9A575B9C83CF9529A2AF30DA023468630AF059A7DC07EFF8041298B7AAEE9F
-010E4C93C08FCDA085657E92D98E9B33E1A28D3DA18FCBCBC7839C0744DD5CE0
-17FCC070EFE545CB2387F92A4B74262D7729B2DD458248397176142195B59718
-AA5429ED39CDE4F9CD1F92837B1EDAC168765EDD6395239B7C1CC552A6EC2A8A
-76E87AE3D015F874FECEF9406C030BE3732916C975F583FC660BE945F1A3EEFA
-A3B4E315BC32CF5EC239A9CC1B8ACB2C09540B1A42B6D057F6EC11DC7BD2F474
-72592808C08B7725B4F629671C96961BEA8F3C44C56A09C74FEE732584F36B00
-27977D6B37B2827E64FF0CA96215E62E3A5C325700D9B26E3550CFE92EB1ADB8
-E291B92E4BDEB32E539CD690B41B639E21B547FCF698B77B18417E645C6DCD63
-3FD68D26835FB59036B3EC45D58EB879F03FD8DF16CB785948643D059790CE79
-3BA847D6F75BE113B64E703A059B090ED349D382B2A73506C004B8A6D183AE18
-5AD305146A6DA14E3A7A16E3C5F095B249A8BE5CD1CC5BE1E0FADEDE5FB469A3
-CF8DE193CD5E42769D1F86F927B9752A982E8E42365FAAA3E3C33421D78CE39F
-F56E3C711136B926D7ADD91A6CA8BD527B0F0A28C1D16720B0E2F4FEB2BA12D8
-81BE8B788A6D8C42A8EB37E0E58C7D92BD698585C537B402C8C33DF60D4AA2BB
-D31AA8174BFB18C4D95B7579A8D6D525D46B14D2B541184054B4A853A2E8F2EA
-B3F9C7FA449195AA22FEB1E93C3ACCD94AAA3421D74F5EDCAF13A90E10489859
-A285DEC29A8A966A49EEE994099A6396F3AF8FBB65CCFFE23291C1E8C43F5AD6
-65CC83011AA050F064B1437935B50FE7D961E4D38271C48156F3362E21A06235
-8670F7AF90F121E7E63CF21CD91B3CDE2ABB83B9AC06F93D821E977EEBDC3FC7
-64CE5E26F25E9BF4FAEB921F085B86BC79437116B0A380BB524EA9C916EA9631
-617F47AAF7F48ECF6EA23004030845ACFAADCC5246D7DD91E27AB945716EF42A
-E405E74653815E6B8A12A54DFF961F5DE8DB985E09EFEC30D1757073E9D801CA
-4D2A2C13372C0324D2DBB801C46485260F4EB54F1E51C026FB612004E336ED5E
-8216D0A7C35B2876EEFEC947427D40CF90FB18DA489CAD6D49249967BE0B0701
-72445726CE409D2CB6181B41CD9BAFDD752180CF72A89496F1DACF41B9449A15
-2940EDF1A10F4D7AE497A1280508FABF06B7BB2CDF145FE5E727B1AA8F829E60
-69A753D4A1D15C589C0B5A11FDFD506E41E688A3DF2AB73E2FC9CAAE638112F3
-CEABFE56FD9B42A8BB9AA991FC1373094DF254D9F0FB6F91C22D4C1DD381EF16
-FDE4B1F581E21622F25DD0B334404580A62CBC5987C0D9FD6FEC592EB69AABD8
-07AE68B63C55DB6649959F318D24B86C72954F820D821204C961EDE197954D06
-A7ED11C028F909E36E1F1A1604A8A73C93C4D1EEB13B9804B833258EE39AF2CB
-B1CF8E7B9EB7A35FB9EF41D0EE075F7E639E3327078D57A1086C1B6E8D139E26
-3CE96110C8147DAC5BB94263329081F0E82A21965C3FC43C43E961AF2A287891
-8118F15ACEB297762FC13A254768642CC998ABB2BB6E568ECEE5E49992FD1BC0
-0F11B4C6DAE121F828F10DF054994AC389906AC50F7D44DAD766CC7EF8A29978
-24637C817BC8EB7C1D63F0375F0D1E9827A5885A27C382985312A698F3F9EA26
-E0F822CC35EDA3832C45B0F59A8C711A6198894611A1BA1642A410DFA092E495
-496F3F5CEC63B04D49DA668B8A1A3F717BF22253E3CCBF6647B3FF7A54B1848E
-8652BB8EDAA1A1C8F2E049CF76AA2ED57134014D2A244AA80443FC4430C04654
-7B20F968DAD82171539CCE1C66BE32BE273612225D21B9F98A18D3563BD8B6DF
-F09A71BC4AEF36144CE700ED605541D0C7191DF3DF28803624AE6E9BC0C18149
-F6B3A05D344C98AF1B5906D91697E316E9899853715F54F69B88ACA663D44D76
-CC716AD9CD82F0812993695D14A41961FF2CD7F2F38E5B0D53EC503B07378D5B
-5F4D06B197C90E71DC9A517C5B75D95F4673F7248C3654BE4A392F3C26E12381
-9E83C9ED3213C4864A7F700D2B2A9B7A4DCEE15A02217D1F4F2CBEE3A5A110DB
-343E32ACC22DA8683E53B92FE0D5C7FE9B64D337BE040A32FCA7D2A920379637
-2CBE8478396420A6797C094DAC929D7446F1B401F0DFDDD6F754D81E5A035640
-00F05CC6ACC7AF085F6D587C30318B42E4F3025A609E02643BC2DC1FDEDAED6F
-7E736050941318927622F70697E03BC2E75F43B20BF2251F1D4905B2CB44155B
-E8F06EECA3F05DF6152520AFB32A33D71F50EDE81C64911EACE19C3F1BC5442E
-FB86AD068568F91BC1BA2EC3A23F493071ED53B44094893C2ABD85BBB1A4E4E0
-6878F4B3C2F42A75E6AF88F2EDB307B420E9440F23BE50B620CA7FAFB7CFE743
-74D7CB1A6C5245803F4DB2D1593B8A5631977604DAE2ABA7522CF4FC5F00CC75
-7737CD661B69D3219F93BA851D25001E64C08EF828207498B514BB44C598F7D7
-82F73BE97C58F607BF236C7EE1A4C818B0EF18A0FEA2CC148F60A9C0095524F5
-44B2544055BA21525C241D8EA2CA
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-0000000000000000000000000000000000000000000000000000000000000000
-cleartomark
-%%EndFont 
-TeXDict begin 40258584 52099344 1000 600 600 (tansig_.dvi)
-@start /Fa 255[77{}1 99.6264 /CMSY10 rf /Fb 194[76 11[49
-49 4[76 1[38 38 40[{}6 99.6264 /CMR12 rf /Fc 139[35 46
-4[58 4[33 1[47 5[51 97[{}6 99.6264 /CMMI12 rf end
-%%EndProlog
-%%BeginSetup
-%%Feature: *Resolution 600dpi
-TeXDict begin
- end
-%%EndSetup
-%%Page: 1 1
-TeXDict begin 1 0 bop Black Black 1529 3045 a @beginspecial
--50 @llx -46.677586 @lly 50 @urx 35.075001 @ury 1000
-@rwi @setspecial
-%%BeginDocument: tansig_0.eps
-%!PS-Adobe-3.0 EPSF-3.0
-%%BoundingBox: -51 -47 51 36
-%%HiResBoundingBox: -50 -46.6775841 50 35.075
-%%Creator: Asymptote 1.25
-%%CreationDate: 2007.05.11 09:24:32
-%%Pages: 1
-%%EndProlog
-%%Page: 1 1
-0 setgray
- 0 0.5 dtransform truncate idtransform setlinewidth pop
-1 setlinecap
-1 setlinejoin
-gsave
- 0 0 translate
-newpath 42.25 0 moveto
- 11.5833333 0 -19.0833333 0 -49.75 0 curveto
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-gsave
- 0 0 translate
-newpath 49.75 1.5959456e-16 moveto
- 42.25 2.00961894 lineto
- 42.25 -2.00961894 lineto
- 44.75 -1.33974596 47.25 -0.669872981 49.75 -1.5959456e-16 curveto
- 49.75 -1.15843382e-16 49.75 1.15843382e-16 49.75 1.5959456e-16 curveto
-closepath
-fill
-grestore
-gsave
- 0 0 translate
-newpath 49.75 1.5959456e-16 moveto
- 42.25 2.00961894 lineto
- 42.25 -2.00961894 lineto
- 44.75 -1.33974596 47.25 -0.669872981 49.75 -1.5959456e-16 curveto
- 49.75 -1.15843382e-16 49.75 1.15843382e-16 49.75 1.5959456e-16 curveto
-closepath
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-gsave
- 0 0 translate
-newpath 0 27.325 moveto
- 0 6.60833333 0 -14.1083333 0 -34.825 curveto
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-gsave
- 0 0 translate
-newpath -4.23272528e-16 34.825 moveto
- -2.00961894 27.325 lineto
- 2.00961894 27.325 lineto
- 1.33974596 29.825 0.669872981 32.325 -1.50053581e-16 34.825 curveto
- -1.87503637e-16 34.825 -3.85822471e-16 34.825 -4.23272528e-16 34.825 curveto
-closepath
-fill
-grestore
-gsave
- 0 0 translate
-newpath -4.23272528e-16 34.825 moveto
- -2.00961894 27.325 lineto
- 2.00961894 27.325 lineto
- 1.33974596 29.825 0.669872981 32.325 -1.50053581e-16 34.825 curveto
- -1.87503637e-16 34.825 -3.85822471e-16 34.825 -4.23272528e-16 34.825 curveto
-closepath
-[ 1 0 0 1 0 0] concat
-stroke
-grestore
-newpath -49.75 -19.6336245 moveto
- -49.4183204 -19.6248246 -49.0866534 -19.6155553 -48.755 -19.6058167 curveto
- -48.4233184 -19.5960772 -48.0916507 -19.5858683 -47.76 -19.5751297 curveto
- -47.428315 -19.5643899 -47.0966473 -19.5531203 -46.765 -19.5412707 curveto
- -46.4333111 -19.5294197 -46.1016431 -19.5169886 -45.77 -19.5039183 curveto
- -45.4383063 -19.490846 -45.1066381 -19.4771343 -44.775 -19.4627197 curveto
- -44.4433006 -19.4483024 -44.111632 -19.4331819 -43.78 -19.4172883 curveto
- -43.4482936 -19.4013912 -43.1166247 -19.3847207 -42.785 -19.3672007 curveto
- -42.4532852 -19.3496759 -42.1216158 -19.3313012 -41.79 -19.3119935 curveto
- -41.4582751 -19.2926795 -41.1266051 -19.272432 -40.795 -19.2511605 curveto
- -40.463263 -19.2298806 -40.1315923 -19.2075761 -39.8 -19.1841488 curveto
- -39.4682484 -19.1607103 -39.136577 -19.1361481 -38.805 -19.1103558 curveto
- -38.4732309 -19.0845486 -38.1415586 -19.0575101 -37.81 -19.0291254 curveto
- -37.47821 -19.0007209 -37.1465366 -18.9709685 -36.815 -18.9397445 curveto
- -36.4831851 -18.9084943 -36.1515104 -18.8757703 -35.82 -18.8414397 curveto
- -35.4881554 -18.8070744 -35.1564793 -18.7710996 -34.825 -18.7333732 curveto
- -34.4931203 -18.6956013 -34.1614425 -18.6560741 -33.83 -18.6146405 curveto
- -33.4980788 -18.5731471 -33.1663991 -18.5297424 -32.835 -18.4842667 curveto
- -32.50303 -18.4387126 -32.1713481 -18.3910811 -31.84 -18.3412042 curveto
- -31.5079729 -18.2912251 -31.1762885 -18.2389926 -30.845 -18.1843313 curveto
- -30.5129063 -18.1295371 -30.1812193 -18.0723038 -29.85 -18.0124502 curveto
- -29.5178292 -17.9524248 -29.1861393 -17.8897658 -28.855 -17.8242882 curveto
- -28.5227405 -17.7585891 -28.1910475 -17.6900546 -27.86 -17.6184978 curveto
- -27.5276392 -17.5466572 -27.1959429 -17.4717732 -26.865 -17.3936604 curveto
- -26.5325244 -17.3151859 -26.200825 -17.2334563 -25.87 -17.1482909 curveto
- -25.5373956 -17.0626674 -25.2056932 -16.9735758 -24.875 -16.8808444 curveto
- -24.5422527 -16.7875371 -24.2105475 -16.6905504 -23.88 -16.5897267 curveto
- -23.5470959 -16.4881842 -23.2153886 -16.3827568 -22.885 -16.2733062 curveto
- -22.5519262 -16.162966 -22.2202177 -16.0485452 -21.89 -15.9299305 curveto
- -21.5567455 -15.810225 -21.2250367 -15.6862579 -20.895 -15.5579465 curveto
- -20.5615563 -15.4283106 -20.2298487 -15.2942518 -19.9 -15.1557237 curveto
- -19.5663623 -15.0156044 -19.2346576 -14.8709257 -18.905 -14.7216827 curveto
- -18.571168 -14.5705499 -18.239468 -14.4147513 -17.91 -14.2543276 curveto
- -17.5759789 -14.0916869 -17.2442856 -13.9243088 -16.915 -13.7522824 curveto
- -16.5808012 -13.5776893 -16.2491165 -13.398326 -15.92 -13.2143317 curveto
- -15.5856415 -13.0274069 -15.2539674 -12.8357214 -14.925 -12.6394642 curveto
- -14.5905067 -12.4399102 -14.2588448 -12.2356495 -13.93 -12.0269188 curveto
- -13.5954035 -11.8145372 -13.2637552 -11.5975481 -12.935 -11.3762323 curveto
- -12.6003382 -11.1509403 -12.2687042 -10.9211848 -11.94 -10.6872864 curveto
- -11.605316 -10.4491328 -11.2736966 -10.2067035 -10.945 -9.9603522 curveto
- -10.610341 -9.70953226 -10.2787361 -9.45466512 -9.95 -9.19613143 curveto
- -9.61541607 -8.93299877 -9.28382506 -8.66608471 -8.955 -8.3957902 curveto
- -8.62054233 -8.12086569 -8.28896416 -7.84245876 -7.96 -7.56098435 curveto
- -7.62571941 -7.274961 -7.29415275 -6.98578287 -6.965 -6.69387333 curveto
- -6.63094537 -6.39761656 -6.29938871 -6.09855672 -5.97 -5.79712099 curveto
- -5.63621683 -5.49166371 -5.30466855 -5.1837747 -4.975 -4.87388138 curveto
- -4.64152907 -4.56041378 -4.30998759 -4.24490114 -3.98 -3.92776887 curveto
- -3.6468763 -3.60762268 -3.31534008 -3.2858297 -2.985 -2.96281217 curveto
- -2.65225179 -2.63743988 -2.32071938 -2.31082724 -1.99 -1.98339309 curveto
- -1.65764811 -1.65434266 -1.32611818 -1.32446347 -0.995 -0.994171662 curveto
- -0.663057363 -0.663057456 -0.331528647 -0.331528693 0 0 curveto
- 0.331528647 0.331528693 0.663057363 0.663057456 0.995 0.994171662 curveto
- 1.32611818 1.32446347 1.65764811 1.65434266 1.99 1.98339309 curveto
- 2.32071938 2.31082724 2.65225179 2.63743988 2.985 2.96281217 curveto
- 3.31534008 3.2858297 3.6468763 3.60762268 3.98 3.92776887 curveto
- 4.30998759 4.24490114 4.64152907 4.56041378 4.975 4.87388138 curveto
- 5.30466855 5.1837747 5.63621683 5.49166371 5.97 5.79712099 curveto
- 6.29938871 6.09855672 6.63094537 6.39761656 6.965 6.69387333 curveto
- 7.29415275 6.98578287 7.62571941 7.274961 7.96 7.56098435 curveto
- 8.28896416 7.84245876 8.62054233 8.12086569 8.955 8.3957902 curveto
- 9.28382506 8.66608471 9.61541607 8.93299877 9.95 9.19613143 curveto
- 10.2787361 9.45466512 10.610341 9.70953226 10.945 9.9603522 curveto
- 11.2736966 10.2067035 11.605316 10.4491328 11.94 10.6872864 curveto
- 12.2687042 10.9211848 12.6003382 11.1509403 12.935 11.3762323 curveto
- 13.2637552 11.5975481 13.5954035 11.8145372 13.93 12.0269188 curveto
- 14.2588448 12.2356495 14.5905067 12.4399102 14.925 12.6394642 curveto
- 15.2539674 12.8357214 15.5856415 13.0274069 15.92 13.2143317 curveto
- 16.2491165 13.398326 16.5808012 13.5776893 16.915 13.7522824 curveto
- 17.2442856 13.9243088 17.5759789 14.0916869 17.91 14.2543276 curveto
- 18.239468 14.4147513 18.571168 14.5705499 18.905 14.7216827 curveto
- 19.2346576 14.8709257 19.5663623 15.0156044 19.9 15.1557237 curveto
- 20.2298487 15.2942518 20.5615563 15.4283106 20.895 15.5579465 curveto
- 21.2250367 15.6862579 21.5567455 15.810225 21.89 15.9299305 curveto
- 22.2202177 16.0485452 22.5519262 16.162966 22.885 16.2733062 curveto
- 23.2153886 16.3827568 23.5470959 16.4881842 23.88 16.5897267 curveto
- 24.2105475 16.6905504 24.5422527 16.7875371 24.875 16.8808444 curveto
- 25.2056932 16.9735758 25.5373956 17.0626674 25.87 17.1482909 curveto
- 26.200825 17.2334563 26.5325244 17.3151859 26.865 17.3936604 curveto
- 27.1959429 17.4717732 27.5276392 17.5466572 27.86 17.6184978 curveto
- 28.1910475 17.6900546 28.5227405 17.7585891 28.855 17.8242882 curveto
- 29.1861393 17.8897658 29.5178292 17.9524248 29.85 18.0124502 curveto
- 30.1812193 18.0723038 30.5129063 18.1295371 30.845 18.1843313 curveto
- 31.1762885 18.2389926 31.5079729 18.2912251 31.84 18.3412042 curveto
- 32.1713481 18.3910811 32.50303 18.4387126 32.835 18.4842667 curveto
- 33.1663991 18.5297424 33.4980788 18.5731471 33.83 18.6146405 curveto
- 34.1614425 18.6560741 34.4931203 18.6956013 34.825 18.7333732 curveto
- 35.1564793 18.7710996 35.4881554 18.8070744 35.82 18.8414397 curveto
- 36.1515104 18.8757703 36.4831851 18.9084943 36.815 18.9397445 curveto
- 37.1465366 18.9709685 37.47821 19.0007209 37.81 19.0291254 curveto
- 38.1415586 19.0575101 38.4732309 19.0845486 38.805 19.1103558 curveto
- 39.136577 19.1361481 39.4682484 19.1607103 39.8 19.1841488 curveto
- 40.1315923 19.2075761 40.463263 19.2298806 40.795 19.2511605 curveto
- 41.1266051 19.272432 41.4582751 19.2926795 41.79 19.3119935 curveto
- 42.1216158 19.3313012 42.4532852 19.3496759 42.785 19.3672007 curveto
- 43.1166247 19.3847207 43.4482936 19.4013912 43.78 19.4172883 curveto
- 44.111632 19.4331819 44.4433006 19.4483024 44.775 19.4627197 curveto
- 45.1066381 19.4771343 45.4383063 19.490846 45.77 19.5039183 curveto
- 46.1016431 19.5169886 46.4333111 19.5294197 46.765 19.5412707 curveto
- 47.0966473 19.5531203 47.428315 19.5643899 47.76 19.5751297 curveto
- 48.0916507 19.5858683 48.4233184 19.5960772 48.755 19.6058167 curveto
- 49.0866534 19.6155553 49.4183204 19.6248246 49.75 19.6336245 curveto
-stroke
-[3.98 3.98 ] 0 setdash
-newpath -49.75 -19.9 moveto
- 49.75 -19.9 lineto
-stroke
-newpath -49.75 19.9 moveto
- 49.75 19.9 lineto
-stroke
-showpage
-%%EOF
-
-%%EndDocument
- @endspecial 0.000000 TeXcolorgray 2324 2686 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 2324 2686
-a 2295 2729 a Fc(n)2353 2686 y
- currentpoint grestore moveto
- 2353 2686 a 1916 2395
-a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 1916 2395 a 1865 2416 a Fc(a)1916 2395 y
- currentpoint grestore moveto
- 1916 2395
-a 1946 2987 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 1946 2987 a 1651 3012 a Fc(a)27 b Fb(=)h
-Fc(tansig)t Fb(\()p Fc(n)p Fb(\))2241 2987 y
- currentpoint grestore moveto
- 2241 2987
-a 1979 2706 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 1979 2706 a 1955 2738 a Fb(0)2004 2706
-y
- currentpoint grestore moveto
- 2004 2706 a 2045 2880 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 2045 2880 a 1982 2908 a Fa(\000)p
-Fb(1)2109 2880 y
- currentpoint grestore moveto
- 2109 2880 a 2070 2432 a
- gsave currentpoint currentpoint translate [1.000000 -0.000000 -0.000000
-1.000000 0 0] concat neg exch neg exch translate
- 2070 2432 a
-2008 2460 a Fb(+1)2133 2432 y
- currentpoint grestore moveto
- 2133 2432 a Black 0.000000
-TeXcolorgray eop end
-%%Trailer
-
-userdict /end-hook known{end-hook}if
-%%EOF
Binary file main/nnet/doc/latex/users/octave/neuroPackage/graphics/tansig.pdf has changed
--- a/main/nnet/doc/latex/users/octave/neuroPackage/graphics/tansiglogo.eps	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,129 +0,0 @@
-%!PS-Adobe-3.0 EPSF-3.0
-%%BoundingBox: 291 381 320 410
-%%HiResBoundingBox: 291.0768 381.0768 319.9232 409.9232
-%%Creator: Asymptote 1.25
-%%CreationDate: 2007.05.22 23:21:01
-%%Pages: 1
-%%EndProlog
-%%Page: 1 1
-gsave
- 291.3268 381.3268 translate
-0 setgray
- 0 0.5 dtransform truncate idtransform setlinewidth pop
-1 setlinecap
-1 setlinejoin
-newpath 0 0 moveto
- 28.3464009 0 lineto
- 28.3464009 28.3464009 lineto
- 0 28.3464009 lineto
- 0 0 lineto
-closepath
-stroke
-newpath 2.83464009 14.1732004 moveto
- 25.5117608 14.1732004 lineto
-stroke
-newpath 2.83464009 8.707858 moveto
- 2.91024628 8.71319256 2.98583699 8.71874422 3.0614113 8.72451292 curveto
- 3.13701931 8.73028419 3.21261059 8.73627266 3.2881825 8.74249885 curveto
- 3.36379355 8.74872827 3.43938482 8.75519563 3.51495371 8.76191723 curveto
- 3.59056799 8.76864286 3.66615943 8.775623 3.74172492 8.78287639 curveto
- 3.81734299 8.79013482 3.89293458 8.79766684 3.96849613 8.80549188 curveto
- 4.04411854 8.81332323 4.11971029 8.82144801 4.19526733 8.82988682 curveto
- 4.27089472 8.8383335 4.34648667 8.84709473 4.42203854 8.85619215 curveto
- 4.49767161 8.86529935 4.57326377 8.87474338 4.64880975 8.88454693 curveto
- 4.72444931 8.89436263 4.8000417 8.90453862 4.87558095 8.9150986 curveto
- 4.95122789 8.92567364 5.02682055 8.93663363 5.10235216 8.94800318 curveto
- 5.17800746 8.95939135 5.25360041 8.97119027 5.32912337 8.9834254 curveto
- 5.40478813 8.9956835 5.48038138 9.00837924 5.55589458 9.02153882 curveto
- 5.63156998 9.03472667 5.70716356 9.0483801 5.78266578 9.06252585 curveto
- 5.85835313 9.07670629 5.93394706 9.09138118 6.00943699 9.10657764 curveto
- 6.08513768 9.12181652 6.16073197 9.13757952 6.2362082 9.1538939 curveto
- 6.31192371 9.17025999 6.38751838 9.18718053 6.4629794 9.2046826 curveto
- 6.5387113 9.2222475 6.61430634 9.24039759 6.68975061 9.2591595 curveto
- 6.76550051 9.27799742 6.84109592 9.29745149 6.91652182 9.31754749 curveto
- 6.99229137 9.33773505 7.06788712 9.35856966 7.14329303 9.3800758 curveto
- 7.21908388 9.40169172 7.29467993 9.42398518 7.37006423 9.44697891 curveto
- 7.445878 9.47010363 7.5214743 9.49393561 7.59683544 9.51849532 curveto
- 7.67267362 9.5432105 7.7482701 9.56866149 7.82360665 9.59486599 curveto
- 7.89947061 9.62125395 7.97506716 9.64840467 8.05037785 9.67633253 curveto
- 8.12626874 9.70447554 8.20186525 9.73340621 8.27714906 9.76313505 curveto
- 8.35306772 9.79311459 8.42866406 9.82390411 8.50392027 9.85550979 curveto
- 8.5798672 9.88740556 8.65546321 9.92013064 8.73069148 9.95368635 curveto
- 8.80666676 9.98757528 8.88226226 10.0223093 8.95746268 10.0578846 curveto
- 9.0334659 10.0938397 9.10906071 10.1306518 9.18423389 10.1683114 curveto
- 9.26026405 10.2064004 9.33585799 10.2453539 9.4110051 10.2851567 curveto
- 9.48706063 10.3254408 9.56265352 10.3665922 9.6377763 10.4085899 curveto
- 9.71385501 10.4511221 9.78944666 10.4945196 9.86454751 10.5387555 curveto
- 9.94064654 10.5835794 10.0162368 10.6292613 10.0913187 10.6757688 curveto
- 10.1674346 10.7229167 10.2430233 10.7709101 10.3180899 10.8197113 curveto
- 10.3942186 10.8692028 10.4698057 10.9195222 10.5448611 10.9706266 curveto
- 10.6209979 11.0224672 10.6965833 11.0751128 10.7716323 11.128516 curveto
- 10.8477721 11.1826954 10.9233559 11.2376518 10.9984035 11.2933341 curveto
- 11.0745409 11.3498248 11.1501229 11.40706 11.2251748 11.4649852 curveto
- 11.3013039 11.5237419 11.3768843 11.583206 11.451946 11.6433204 curveto
- 11.528061 11.7042784 11.6036398 11.7659027 11.6787172 11.8281342 curveto
- 11.7548121 11.8912092 11.8303894 11.9549058 11.9054884 12.0191633 curveto
- 11.9815573 12.0842508 12.0571333 12.1499121 12.1322596 12.2160854 curveto
- 12.2082968 12.283061 12.2838716 12.3505596 12.3590308 12.418519 curveto
- 12.4350308 12.4872386 12.5106045 12.5564281 12.585802 12.6260251 curveto
- 12.6617598 12.6963258 12.7373326 12.7670412 12.8125732 12.8381091 curveto
- 12.8884843 12.9098102 12.9640563 12.9818694 13.0393444 13.0542245 curveto
- 13.1152047 13.1271295 13.1907762 13.2003344 13.2661156 13.2737776 curveto
- 13.3419218 13.3476759 13.4174928 13.4218149 13.4928868 13.4961336 curveto
- 13.5686362 13.5708026 13.6442069 13.6456528 13.719658 13.7206231 curveto
- 13.7953487 13.7958314 13.8709191 13.8711606 13.9464292 13.9465501 curveto
- 14.0220599 14.0220599 14.0976302 14.0976302 14.1732004 14.1732004 curveto
- 14.2487707 14.2487707 14.324341 14.324341 14.3999717 14.3998508 curveto
- 14.4754818 14.4752403 14.5510522 14.5505695 14.6267429 14.6257778 curveto
- 14.702194 14.7007481 14.7777647 14.7755982 14.8535141 14.8502673 curveto
- 14.9289081 14.924586 15.0044791 14.998725 15.0802853 15.0726233 curveto
- 15.1556247 15.1460665 15.2311961 15.2192714 15.3070565 15.2921764 curveto
- 15.3823446 15.3645315 15.4579166 15.4365906 15.5338277 15.5082918 curveto
- 15.6090682 15.5793597 15.6846411 15.6500751 15.7605989 15.7203758 curveto
- 15.8357963 15.7899728 15.9113701 15.8591623 15.9873701 15.9278819 curveto
- 16.0625293 15.9958413 16.1381041 16.0633399 16.2141413 16.1303155 curveto
- 16.2892676 16.1964888 16.3648436 16.26215 16.4409125 16.3272376 curveto
- 16.5160115 16.3914951 16.5915888 16.4551917 16.6676837 16.5182667 curveto
- 16.7427611 16.5804982 16.8183399 16.6421225 16.8944549 16.7030805 curveto
- 16.9695166 16.7631949 17.045097 16.822659 17.1212261 16.8814157 curveto
- 17.196278 16.9393408 17.27186 16.9965761 17.3479973 17.0530668 curveto
- 17.423045 17.1087491 17.4986287 17.1637055 17.5747686 17.2178849 curveto
- 17.6498176 17.2712881 17.725403 17.3239336 17.8015398 17.3757743 curveto
- 17.8765952 17.4268787 17.9521823 17.4771981 18.028311 17.5266896 curveto
- 18.1033776 17.5754908 18.1789663 17.6234842 18.2550822 17.6706321 curveto
- 18.3301641 17.7171396 18.4057544 17.7628215 18.4818534 17.8076454 curveto
- 18.5569542 17.8518813 18.6325459 17.8952788 18.7086246 17.9378109 curveto
- 18.7837474 17.9798087 18.8593403 18.0209601 18.9353958 18.0612442 curveto
- 19.0105429 18.101047 19.0861368 18.1400005 19.162167 18.1780895 curveto
- 19.2373402 18.2157491 19.312935 18.2525612 19.3889382 18.2885163 curveto
- 19.4641386 18.3240916 19.5397341 18.3588256 19.6157094 18.3927145 curveto
- 19.6909377 18.4262703 19.7665337 18.4589953 19.8424806 18.4908911 curveto
- 19.9177368 18.5224968 19.9933332 18.5532863 20.0692518 18.5832658 curveto
- 20.1445356 18.6129947 20.2201322 18.6419253 20.296023 18.6700684 curveto
- 20.3713337 18.6979962 20.4469303 18.7251469 20.5227942 18.7515349 curveto
- 20.5981308 18.7777394 20.6737273 18.8031904 20.7495655 18.8279056 curveto
- 20.8249266 18.8524653 20.9005229 18.8762973 20.9763367 18.899422 curveto
- 21.051721 18.9224157 21.127317 18.9447092 21.2031079 18.9663251 curveto
- 21.2785138 18.9878312 21.3541095 19.0086658 21.4298791 19.0288534 curveto
- 21.505305 19.0489494 21.5809004 19.0684035 21.6566503 19.0872414 curveto
- 21.7320946 19.1060033 21.8076896 19.1241534 21.8834215 19.1417183 curveto
- 21.9588825 19.1592204 22.0344772 19.1761409 22.1101927 19.192507 curveto
- 22.1856689 19.2088214 22.2612632 19.2245844 22.3369639 19.2398233 curveto
- 22.4124538 19.2550197 22.4880478 19.2696946 22.5637351 19.283875 curveto
- 22.6392373 19.2980208 22.7148309 19.3116742 22.7905063 19.3248621 curveto
- 22.8660195 19.3380216 22.9416128 19.3507174 23.0172775 19.3629755 curveto
- 23.0928005 19.3752106 23.1683934 19.3870095 23.2440487 19.3983977 curveto
- 23.3195803 19.4097673 23.395173 19.4207273 23.4708199 19.4313023 curveto
- 23.5463592 19.4418623 23.6219516 19.4520383 23.6975911 19.461854 curveto
- 23.7731371 19.4716575 23.8487293 19.4811015 23.9243624 19.4902087 curveto
- 23.9999142 19.4993062 24.0755062 19.5080674 24.1511336 19.5165141 curveto
- 24.2266906 19.5249529 24.3022824 19.5330777 24.3779048 19.540909 curveto
- 24.4534663 19.548734 24.5290579 19.5562661 24.604676 19.5635245 curveto
- 24.6802415 19.5707779 24.7558329 19.577758 24.8314472 19.5844837 curveto
- 24.9070161 19.5912053 24.9826073 19.5976726 25.0582184 19.603902 curveto
- 25.1337903 19.6101282 25.2093816 19.6161167 25.2849896 19.621888 curveto
- 25.3605639 19.6276567 25.4361546 19.6332083 25.5117608 19.6385429 curveto
-stroke
-grestore
-showpage
-%%EOF
Binary file main/nnet/doc/latex/users/octave/neuroPackage/graphics/tansiglogo.pdf has changed
--- a/main/nnet/doc/latex/users/octave/neuroPackage/logsig.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,15 +0,0 @@
-\subsection{logsig}
-
-\begin{figure}[htb]
-\centering
-  \includegraphics{octave/neuroPackage/graphics/logsig}
-\caption{Log-Sigmoid transfer function}
-\label{fig:logsigTransferFunction}
-\end{figure}
-
-\begin{figure}[htb]
-\centering
-  \includegraphics{octave/neuroPackage/graphics/logsiglogo}
-\caption{Log-Sigmoid transfer function logo}
-\label{fig:logsigTransferFunctionLogo}
-\end{figure}
\ No newline at end of file
--- a/main/nnet/doc/latex/users/octave/neuroPackage/min_max.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,27 +0,0 @@
-\subsection{min\_max}
-\textit{min\_max} get the minimal and maximal values of an training input matrix. So the dimension of this matrix must be an RxN matrix where R is the number of input neurons and N depends on the number of training sets.\\
-
-\noindent \textbf{\textcolor{brown}{Syntax:}}\\
-
-\noindent mMinMaxElements = min\_max(RxN);\\
-
-\noindent \textbf{\textcolor{brown}{Description:}}\\
-
-\noindent RxN: R x N matrix of min and max values for R input elements with N columns\\ 
-
-\noindent \textbf{\textcolor{brown}{Example:}}\\
-
-\begin{equation}
-	\left[
-		\begin{array}{cc}
-     	1 &  11 \\
-    	0  & 21
-   \end{array} 
-	 \right]            = min\_max\left[ 
-	 															   \begin{array}{ccccc}
-	 															   3 & 1 & 3 & 5 & 11 \\
-	 															   12& 0 & 21& 8 & 6  \\
-	 															   \end{array}
-	 															   	 \right]
-\end{equation}
-
--- a/main/nnet/doc/latex/users/octave/neuroPackage/neuroPackage.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,22 +0,0 @@
-\chapter{Neural Network Package for Octave}
-This chapter describes all functions available in the neural network package of Octave.
-
-Eventhough it will be as compatible as possible to the one of MATLAB(TM).
-
-\section{Available Functions}
-\input{octave/neuroPackage/min_max}
-\input{octave/neuroPackage/newff}
-\input{octave/neuroPackage/prestd}
-\input{octave/neuroPackage/poststd}
-\input{octave/neuroPackage/saveMLPStruct}
-\input{octave/neuroPackage/sim}
-\input{octave/neuroPackage/subset}
-\input{octave/neuroPackage/train}
-\input{octave/neuroPackage/trastd}
-
-
-\section{Transfer functions}
-\input{octave/neuroPackage/logsig}
-\input{octave/neuroPackage/purelin}
-\input{octave/neuroPackage/tansig}
-
--- a/main/nnet/doc/latex/users/octave/neuroPackage/newff.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,27 +0,0 @@
-\subsection{newff}
-\textit{newff} is the short form for \textit{\textbf{new f}eed \textbf{f}orward network}. This command creates a feed-forward backpropagation network structure.\\
-
-\noindent \textbf{\textcolor{brown}{Syntax:}}\\
-
-\noindent net = newff(Rx2,[S1 S2 ... SN],\{TF1 TF2 ... TFN\},BTF,BLF,PF)\\
-
-\noindent \textbf{\textcolor{brown}{Description:}}\\
-
-\noindent Rx2: R x 2 matrix of min and max values for R input elements\\ 
-\noindent Si: Size of ith layer, for N layers\\ 
-\noindent TFi: Transfer function of ith layer, default = "tansig"\\
-\noindent BTF: Backpropagation network training function, default = "trainlm" \\
-\noindent BLF: Backpropagation weight/bias learning function, NOT USED, is only for MATLAB(TM) compatibility\\
-\noindent PF: Performance function, default = "mse"\\
-
-\noindent \textbf{\textcolor{brown}{Examples:}}\\
-
-\noindent net = newff(Rx2,[2 1])\\
-\noindent net = newff(Rx2,[2 1],\{"tansig","purelin"\});\\
-\noindent net = newff(Rx2,[2 1],\{"tansig","purelin"\},"trainlm");\\
-\noindent net = newff(Rx2,[2 1],\{"tansig","purelin"\},"trainlm","notUsed","mse");\\
-
-\noindent \textbf{\textcolor{brown}{Comments:}}\\
-In this version, you can have as much output neurons as you want. The same with the number of hidden layers. This means you can have one input layer, unrestricted number of hidden layers and one output layer. That's it.\\
-
-
--- a/main/nnet/doc/latex/users/octave/neuroPackage/poststd.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,1 +0,0 @@
-\subsection{poststd}
\ No newline at end of file
--- a/main/nnet/doc/latex/users/octave/neuroPackage/prestd.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,1 +0,0 @@
-\subsection{prestd}
\ No newline at end of file
--- a/main/nnet/doc/latex/users/octave/neuroPackage/purelin.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,16 +0,0 @@
-\subsection{purelin}
-
-\begin{figure}[htb]
-\centering
-  \includegraphics{octave/neuroPackage/graphics/purelin}
-\caption{Linear transfer function}
-\label{fig:purelinTransferFunction}
-\end{figure}
-
-\begin{figure}[htb]
-\centering
-  \includegraphics{octave/neuroPackage/graphics/purelinlogo}
-\caption{Linear transfer function logo}
-\label{fig:purelinTransferFunctionLogo}
-\end{figure}
-
--- a/main/nnet/doc/latex/users/octave/neuroPackage/saveMLPStruct.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,10 +0,0 @@
-\subsection{saveMLPStruct}
-This is an additional function which doesn't exist in the neural network toolbox of MathWorks (TM). To see the network structure, you can use this command and save the complete structure to a file. Open this file and you have the same view like you would open the \textit{network type} of MATLAB(TM).\\
-
-\noindent \textbf{\textcolor{brown}{Syntax:}}\\
-
-\noindent saveMLPStruct(net,"initNetwork.txt");\\
-
-
-
-
--- a/main/nnet/doc/latex/users/octave/neuroPackage/sim.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,18 +0,0 @@
-\subsection{sim}
-
-\noindent \textbf{\textcolor{brown}{Syntax:}}\\
-
-\noindent simout = sim(net,P);\\
-
-\noindent \textbf{\textcolor{brown}{Description:}}\\
-
-\noindent  \textbf{Left-Hand-Side:}\\
-\noindent simout: Output values of the simulated network.\\
-
-\noindent  \textbf{Right-Hand-Side:}\\
-\noindent net: Network, created and trained with newff(...) and train(...)\\
-\noindent P: Input data\\ 
-
-
-
-
--- a/main/nnet/doc/latex/users/octave/neuroPackage/subset.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,47 +0,0 @@
-\subsection{subset}
-
-\textit{subset} can be used to optimize the data sets for train, test and validation of a neural
-network.\\
-
-\noindent \textbf{\textcolor{brown}{Syntax:}}\\
-
-\noindent [mTrain, mTest, mVali] = subset(mData,nTargets,iOpti,fTest,fVali);\\
-
-\noindent \textbf{\textcolor{brown}{Description:}}\\
-
-\noindent  \textbf{Left-Hand-Side:}\\
-\noindent mTrain: (R+T) x M matrix with R input rows, T output rows and M columns
-				where M <= N.\\
-\noindent mTest:  (R+T) x S matrix with R input rows, T output rows and S columns
-				where S <= N.\\
-\noindent mVali:  (R+T) x U matrix with R input rows, T output rows and U columns
-				where U <= N. And U can only exist, if S also exist.\\
-
-\noindent  \textbf{Right-Hand-Side:}\\
-\noindent mData: (R+T) x N matrix with R input rows, T output rows and N columns\\ 
-\noindent nTargets: Number of T output rows\\ 
-\noindent iOpti: Integer value to define level of optimization.\\
-\noindent fTest: Fraction to define the percentage of data sets which should be used for testing. \\
-\noindent fVali: Fraction to define the percentage of data sets which should be used for testing.\\
-
-\noindent iOpti can have following values:\\
-0	: no optimization\\
-1	: will randomise the column order and rerange the columns containing min and max values to be in the train set\\
-2	:	will NOT randomise the column order, but rerange the columns containing min and max values to be in the train set\\
-
-\noindent fTest or fValie have following meaning:\\
-Each of this arguments can be a fraction or zero. The value 1 is not allowed! The sum of both values
-must also be smaller than 1!\\
-Example: fTest = 1/3\\
-
-\noindent \textbf{Default values}\\
-\noindent iOpti		= 1\\
-\noindent fTest		= 1/3\\
-\noindent fVali		= 1/6\\
-
-\noindent \textbf{\textcolor{brown}{Examples:}}\\
-
-\noindent mTrain = subset(mData,2,1,0,0)\\
-\noindent [mTrain,mTest] = subset(mData,2,1,1/3,0);\\
-\noindent [mTrain,mTest,mVali] = subset(mData,1);\\
-\noindent [mTrain,mTest,mVali] = subset(mData,1,1,1/3,1/6);\\
\ No newline at end of file
--- a/main/nnet/doc/latex/users/octave/neuroPackage/tansig.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,29 +0,0 @@
-\subsection{tansig}
-
-\noindent
-I solved all of my real life problems with this transfer function if a non-linear function was used. In [4] page 2-6 the tansig is defined as in equation \eqref{equ:tansigTransferFunction}. A look on the MathWorks homepage with the keyword tansig will show that tansig is programed as in equation \eqref{equ:tansigTransferFunctionNnet}.
-
-\begin{equation}
-	a = \frac{e^n - e^{-n}}{e^n + e^{-n}}
-	\label{equ:tansigTransferFunction}
-\end{equation}
-
-\begin{equation}
-	a = \frac{2}{(1 + e^{-2*n})-1}
-	\label{equ:tansigTransferFunctionNnet}
-\end{equation}
-
-
-\begin{figure}[htb]
-\centering
-  \includegraphics{octave/neuroPackage/graphics/tansig}
-\caption{Tansig transfer function}
-\label{fig:tansigTransferFunction}
-\end{figure}
-
-\begin{figure}[htb]
-\centering
-  \includegraphics{octave/neuroPackage/graphics/tansiglogo}
-\caption{Tansig transfer function logo}
-\label{fig:tansigTransferFunctionLogo}
-\end{figure}
--- a/main/nnet/doc/latex/users/octave/neuroPackage/train.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,34 +0,0 @@
-\subsection{train}
-
-\noindent \textbf{\textcolor{brown}{Syntax:}}\\
-
-\noindent net = train(MLPnet,P,T,[],[],VV);\\
-
-\noindent \textbf{\textcolor{brown}{Description:}}\\
-
-\noindent  \textbf{Left-Hand-Side:}\\
-\noindent net: Trained multi-layer network.\\
-
-\noindent  \textbf{Right-Hand-Side:}\\
-\noindent MLPnet: Multi-layer network, created with newff(...)\\ 
-\noindent P: Input data for training\\ 
-\noindent T: Target data for training\\
-\noindent []: Not used right now, only for compatibility with Matlab\\
-\noindent []: Not used right now, only for compatibility with Matlab\\
-\noindent VV: Validation data. Contains input and target values\\
-
-\noindent \textbf{\textcolor{brown}{Examples:}}\\
-
-\noindent net = train(MLPnet,P,T)\\
-\noindent net = train(MLPnet,P,T,[],[],VV)\\
-
-\noindent P,T must have the same number of rows as in newff(...).\\
-\noindent VV.P, VV.T must have the same number of rows as in newff(...)\\
-
-\noindent \textbf{\textcolor{brown}{Comments:}}\\
-\noindent Please be sure to put the validation values in a structure named
-VV.P and VV.T.\\
-VV can be changed, but not .P and .T!
-
-
-
--- a/main/nnet/doc/latex/users/octave/neuroPackage/trastd.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,1 +0,0 @@
-\subsection{trastd}
\ No newline at end of file
--- a/main/nnet/doc/latex/users/octave/octave.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,6 +0,0 @@
-\chapter{Octave}
-This chapter describes all functions available in the neural network toolbox in Octave.
-
-
-
-\input{octave/functions/isposint_netw_privOct}
\ No newline at end of file
--- a/main/nnet/doc/latex/users/title2.tex	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,8 +0,0 @@
-% start text here!!
-
-\title{A neural network package for Octave\\
-		User's Guide \\
-				\input{../common/version}}
-
-\author{Michel D. Schmid}
-\maketitle
Binary file main/nnet/doc/pdf/neuralNetworkPackageForOctaveDevelopersGuide.pdf has changed
Binary file main/nnet/doc/pdf/neuralNetworkPackageForOctaveUsersGuide.pdf has changed
--- a/main/nnet/inst/dhardlim.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,28 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{a} = dhardlim (@var{n})
-##
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function a = dhardlim(n)
-
-  a = 0;
-
-endfunction
--- a/main/nnet/inst/dividerand.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,72 +0,0 @@
-## Copyright (C) 2009 Luiz Angelo Daros de Luca <luizluca@gmail.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} [@var{trainVectors},@var{validationVectors},@var{testVectors},@var{indexOfTrain},@var{indexOfValidation},@var{indexOfTest}] = dividerand (@var{allCases},@var{trainRatio},@var{valRatio},@var{testRatio})
-## Divide the vectors in training, validation and test group according to
-## the informed ratios
-##
-##
-## @example
-##
-## [trainVectors,validationVectors,testVectors,indexOfTrain,indexOfValidatio
-## n,indexOfTest] = dividerand(allCases,trainRatio,valRatio,testRatio)
-##
-## The ratios are normalized. This way:
-##
-## dividerand(xx,1,2,3) == dividerand(xx,10,20,30)
-##
-## @end example
-##
-## @end deftypefn
-
-
-function [trainVectors,validationVectors,testVectors,indexOfTrain,indexOfValidation,indexOfTest] = dividerand(allCases,trainRatio,valRatio,testRatio)
-  #
-  # Divide the vectors in training, validation and test group according to
-  # the informed ratios
-  #
-  # [trainVectors,validationVectors,testVectors,indexOfTrain,indexOfValidatio
-  # n,indexOfTest] = dividerand(allCases,trainRatio,valRatio,testRatio)
-  #
-  # The ratios are normalized. This way:
-  #
-  # dividerand(xx,1,2,3) == dividerand(xx,10,20,30)
-  #
-
-  ## Normalize ratios
-  total = trainRatio + valRatio + testRatio;
-  #trainRatio = trainRatio / total; not used
-  validationRatio = valRatio / total;
-  testRatio = testRatio / total;
-
-  ## Calculate the number of cases for each type
-  numerOfCases = size(allCases,2);
-  numberOfValidation = floor(validationRatio*numerOfCases);
-  numberOfTest = floor(testRatio*numerOfCases);
-  numberOfTrain = numerOfCases - numberOfValidation - numberOfTest;
-
-  ## Find their indexes
-  indexOfAll=randperm(numerOfCases);
-  indexOfValidation=sort(indexOfAll(1:numberOfValidation));
-  indexOfTest=sort(indexOfAll((1:numberOfTest)+numberOfValidation));
-  indexOfTrain=sort(indexOfAll((1:numberOfTrain)+numberOfTest+numberOfValidation));
-
-  ## Return vectors
-  trainVectors = allCases(:,indexOfTrain);
-  testVectors = allCases(:,indexOfTest);
-  validationVectors = allCases(:,indexOfValidation);
-endfunction
-
--- a/main/nnet/inst/dposlin.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,32 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}@var{a}= poslin (@var{n})
-## @code{poslin} is a positive linear transfer function used
-## by neural networks
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function a = dposlin(n)
-
-   if (n<0)
-     a = 0;
-   else
-     a = 1;
-   endif
-
-endfunction
--- a/main/nnet/inst/dsatlin.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,38 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{a} = dsatlin (@var{n})
-##
-## @end deftypefn
-
-## @seealso{dpurelin,dtansig,dlogsig}
-
-## Author: Michel D. Schmid
-
-
-function a = dsatlin(n)
-
-
-  # the derivative of satlin is easy:
-  # where satlin is constant, the derivative is 0
-  # else, because without variable n, the derivative is 1
-  if (n>=0 && n<=1)
-    a = 1;
-  else
-    a = 0;
-  endif
-
-endfunction
--- a/main/nnet/inst/dsatlins.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,38 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{a} = satlins (@var{n})
-## A neural feed-forward network will be trained with @code{trainlm}
-##
-## @end deftypefn
-
-## @seealso{purelin,tansig,logsig,satlin,hardlim,hardlims}
-
-## Author: Michel D. Schmid
-
-
-function a = dsatlins(n)
-
-  # comment see dsatlin
-  # a = 1 if (n>=-1 && n<=1),
-  # 0 otherwise
-  if (n>=-1 && n<=1)
-    a = 1;
-  else
-    a = 0;
-  endif
-
-endfunction
--- a/main/nnet/inst/hardlim.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,31 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{a} = hardlim (@var{n})
-##
-## @end deftypefn
-
-## @seealso{purelin,tansig}
-
-## Author: Michel D. Schmid <michael.schmid@plexso.com>
-
-
-function a = hardlim(n)
-
-  # a=1 if n>=0, a=0 otherwise
-  a = (n>=0);
-
-endfunction
--- a/main/nnet/inst/hardlims.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,35 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{a} = hardlims (@var{n})
-##
-## @end deftypefn
-
-## @seealso{purelin,tansig,hardlim}
-
-## Author: Michel D. Schmid <michael.schmid@plexso.com>
-
-
-function a = hardlims(n)
-
-  # a=1 if n>0, a=-1 otherwise
-  if n>=0
-    a=1;
-  else
-    a=-1;
-  endif
-
-endfunction
--- a/main/nnet/inst/ind2vec.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,41 +0,0 @@
-## Copyright (C) 2009 Luiz Angelo Daros de Luca <luizluca@gmail.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {@var{vec}} = ind2vec (@var{ind})
-## @code{vec2ind} convert indices to vector
-##
-##
-## @example
-## EXAMPLE 1
-## vec = [1 2 3; 4 5 6; 7 8 9];
-##
-## ind = vec2ind(vec)
-## The prompt output will be:
-## ans = 
-##    1 2 3 1 2 3 1 2 3
-## @end example
-##
-## @end deftypefn
-
-## @seealso{vec2ind}
-
-function [vector]=ind2vec(ind)
-  # Converts indices to vectors
-  #
-  #
-  vectors = length(ind);
-  vector = sparse(ind,1:vectors,ones(1,vectors));
-endfunction
--- a/main/nnet/inst/isposint.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,59 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{f} = isposint(@var{n})
-## @code{isposint} returns true for positive integer values.
-## 
-## @example
-##   isposint(1)   # this returns TRUE
-##   isposint(0.5) # this returns FALSE
-##   isposint(0)   # this also return FALSE
-##   isposint(-1)  # this also returns FALSE
-## @end example
-##
-## 
-## @end deftypefn
-
-## Author: Michel D. Schmid 
-
-function f = isposint(n)
-
-  ## check range of input arguments
-  error(nargchk(1,1,nargin))
-  
-  ## check input arg
-  if (length(n)>1)
-    error("Input argument must not be a vector, only scalars are allowed!")
-  endif
-
-  f = 1;
-  if ( (!isreal(n)) || (n<=0) || (round(n) != n) )
-    f = 0;
-  endif
-
-
-endfunction
-
-%!shared
-%! disp("testing isposint")
-%!assert(isposint(1)) # this should pass
-%!assert(isposint(0.5),0) # should return zero
-%!assert(isposint(-1),0) # should return zero
-%!assert(isposint(-1.5),0) # should return zero
-%!assert(isposint(0),0) # should return zero
-%!fail("isposint([0 0])","Input argument must not be a vector, only scalars are allowed!")
-%!fail("isposint('testString')","Input argument must not be a vector, only scalars are allowed!")
-
--- a/main/nnet/inst/logsig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,39 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}@var{a} = logsig (@var{n})
-## @code{logsig} is a non-linear transfer function used to train
-## neural networks.
-## This function can be used in newff(...) to create a new feed forward
-## multi-layer neural network.
-##
-## @end deftypefn
-
-## @seealso{purelin,tansig}
-
-## Author: Michel D. Schmid
-
-
-function a = logsig(n)
-
-
-  a = 1 ./ (1 + exp(-n));
-  ## attention with critical values ==> infinite values
-  ## must be set to 1! Still the same problem as in "tansig"
-  i = find(!finite(a));
-  a(i) = sign(n(i));
-
-endfunction
--- a/main/nnet/inst/mapstd.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,327 +0,0 @@
-## Copyright (C) 2009 Luiz Angelo Daros de Luca <luizluca@gmail.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} [@var{YY},@var{PS}] = mapstd (@var{XX},@var{ymean},@var{ystd})
-## Map values to mean 0 and standard derivation to 1.
-##
-## @example
-## [YY,PS] = mapstd(XX,ymean,ystd)
-##
-##    Apply the conversion and returns YY as (YY-ymean)/ystd.
-##
-## [YY,PS] = mapstd(XX,FP)
-##
-##    Apply the conversion but using an struct to inform target mean/stddev.
-##    This is the same of [YY,PS]=mapstd(XX,FP.ymean, FP.ystd).
-##
-## YY = mapstd('apply',XX,PS)
-##
-##    Reapply the conversion based on a previous operation data.
-##    PS stores the mean and stddev of the first XX used.
-##
-## XX = mapstd('reverse',YY,PS)
-##
-##    Reverse a conversion of a previous applied operation.
-##
-## dx_dy = mapstd('dx',XX,YY,PS)
-##
-##    Returns the derivative of Y with respect to X.
-##
-## dx_dy = mapstd('dx',XX,[],PS)
-##
-##    Returns the derivative (less efficient).
-##
-## name = mapstd('name');
-##
-##    Returns the name of this convesion process.
-##
-## FP = mapstd('pdefaults');
-##
-##    Returns the default process parameters.
-##
-## names = mapstd('pnames');
-##
-##    Returns the description of the process parameters.
-##
-## mapstd('pcheck',FP);
-##
-##    Raises an error if FP has some inconsistent.
-## @end example
-##
-## @end deftypefn
-
-function [out1,out2]=mapstd(in1,in2,in3,in4)
-  #
-  # Map values to mean 0 and standard derivation to 1.
-  #
-  # [YY,PS] = mapstd(XX,ymean,ystd)
-  #
-  #    Apply the conversion and returns YY as (YY-ymean)/ystd.
-  #
-  # [YY,PS] = mapstd(XX,FP)
-  #
-  #    Apply the conversion but using an struct to inform target mean/stddev.
-  #    This is the same of [YY,PS]=mapstd(XX,FP.ymean, FP.ystd).
-  #
-  # YY = mapstd('apply',XX,PS)
-  #
-  #    Reapply the conversion based on a previous operation data.
-  #    PS stores the mean and stddev of the first XX used.
-  #
-  # XX = mapstd('reverse',YY,PS)
-  #
-  #    Reverse a conversion of a previous applied operation.
-  #
-  # dx_dy = mapstd('dx',XX,YY,PS)
-  #
-  #    Returns the derivative of Y with respect to X.
-  #
-  # dx_dy = mapstd('dx',XX,[],PS)
-  #
-  #    Returns the derivative (less efficient).
-  #
-  # name = mapstd('name');
-  #
-  #    Returns the name of this convesion process.
-  #
-  # FP = mapstd('pdefaults');
-  #
-  #    Returns the default process parameters.
-  #
-  # names = mapstd('pnames');
-  #
-  #    Returns the description of the process parameters.
-  #
-  # mapstd('pcheck',FP);
-  #
-  #    Raises an error if FP has some inconsistent.
-  #
-
-  if nargin==0
-    error("Not enough arguments.")
-  endif
-
-  # Defaults
-  ps.name="mapstd";
-  ps.ymean=0;
-  ps.ystd=1;
-
-  if ischar(in1)
-    switch in1
-        case "name"
-            if nargout>1
-                error("Too many output arguments");
-            endif
-            if nargin>1
-                error("Too many input arguments");
-            endif
-            out1="Map Mean and Standard Deviation";
-            return;
-        case "pdefaults"
-            if nargout>1
-                error("Too many output arguments");
-            endif
-            if nargin>1
-                error("Too many input arguments");
-            endif
-            out1=ps;
-        case "pcheck"
-            if nargout>1
-                error("Too many output arguments");
-            endif
-            if nargin<2
-                error("Not enough input arguments");
-            endif
-            if nargin>2
-                error("Too many input arguments");
-            endif
-            
-            fp=in2;           
-            if ~isstruct(fp)
-                error("FP must be a struct")                
-            elseif ~isfield(fp,"ymean")
-                error("FP must include ymean field")
-            elseif ~isfield(fp,"ystd")
-                error("FP must include ystd field")
-            elseif isdouble(fp.ymean)
-                error("FP.ymean must be a real scalar value")
-            elseif isdouble(fp.ystd)
-                error("FP.ystd must be a real scalar value")
-            else
-                out1='';
-            endif
-            return;
-        # MATLAB uses pnames but documents as pdesc (that does not work)
-        case "pnames"
-            if nargout>1
-                error("Too many output arguments");
-            endif
-            if nargin>1
-                error("Too many input arguments");
-            endif
-            # MATLAB seems to be buggy in the second element
-            #out1={'Mean value for each row of Y.','Maximum value for each
-            #row of Y.'};            
-            out1={"Mean value for each row of Y.","Standart deviation value for each row of Y."};                        
-        case "apply"
-            if nargin<3
-                error("Not enough input arguments");
-            endif
-            if nargin>3
-                error("Too many input arguments");
-            endif
-            if nargout>1
-                error("Too many output arguments");
-            endif
-            xx=in2;
-            ps=in3;
-            yy=apply(xx,ps);
-            out1=yy;
-            out2=ps;
-            return;
-        case "reverse"
-            if nargin<3
-                error("Not enough input arguments");
-            endif
-            if nargin>3
-                error("Too many input arguments");
-            endif
-            if nargout>1
-                error("Too many output arguments");
-            endif
-            yy=in2;
-            ps=in3;
-            xx=reverse(yy,ps);
-            out1=xx;
-            out2=ps;
-            return;
-        case "dx"
-            if nargin<3
-                error("Not enough input arguments");
-            endif
-            if nargin>3
-                error("Too many input arguments");
-            endif
-            if nargout>1
-                error("Too many output arguments");
-            endif
-            xx=in2;
-            yy=in3;
-            ps=in4;
-            xx_yy=derivate(xx,yy,ps);
-            out1=xx_yy;
-            return;
-    endswitch
-  else
-    xx=in1;
-    ps.xrows=size(xx,1);
-    ps.yrows=size(xx,1);
-    ps.xmean=mean(xx,2);
-    ps.xstd=std(xx,0,2);      
-    
-    if nargin==1
-        # All correct
-    elseif nargin==2
-        if isstruct(in2)
-            ps.ymean=in2.ymean;
-            ps.ystd=in2.ystd;
-        else
-            ps.ymean=in2;
-        endif
-    elseif nargin == 3
-        ps.ymean=in2;
-        ps.ystd=in3;
-    else
-        error("Too many input arguments");
-    endif
-    
-    out1=apply(xx,ps);   
-    out2=ps;
-  endif
-
-  # Verify args
-  function checkargs(values,ps)
-    # check xx is matrix
-    if ~isnumeric(values)
-        error("Just numeric values are accepted")
-    endif
-    # check ps is struct
-    if ~isstruct(ps)
-        error("PS should be a struct")
-    endif
-    # check ymean,ystd
-    if ~isa(ps.ymean,"double")
-        error("PS.ymean should be a double")
-    endif
-    if ~isa(ps.ystd,"double")
-        error("PS.ystd should be a double")
-    endif
-    if ~all(size(ps.ymean)==[1 1])
-        error("PS.ymean should be a scalar")
-    endif
-    if ~all(size(ps.ystd)==[1 1])
-        error("PS.ystd should be a scalar")
-    endif
-    # check xmean,ystd
-    if ~isnumeric(ps.xmean)
-        error("PS.xmean should be a numeric")
-    endif
-    if ~isnumeric(ps.xstd)
-        error("PS.xstd should be a numeric")
-    endif
-    if ~all(size(ps.xmean)==size(ps.xstd))
-        error("Size of PS.xmean and PS.xstd must match")
-    endif
-  endfunction
-
-  # Apply the mapping operation
-  function [yy]=apply(xx,ps)
-    checkargs(xx,ps)
-
-    if ~all(size(xx,1)==size(ps.xmean,1))
-        error("Size of XX rows should match PS.xmean and PS.xstd")
-    endif
-    # Avoid multiply/division by zero
-    ps.xstd(ps.xstd == 0) = 1;
-    yy=(xx - (ps.xmean*ones(1,size(xx,2)))) ./ (ps.xstd*ones(1,size(xx,2)));
-    yy=(yy + ps.ymean) .* ps.ystd;
-  endfunction
-
-  # Reverse the mapping operation
-  function [xx]=reverse(yy,ps)
-    checkargs(yy,ps)
-    if ~all(size(yy,1)==size(ps.xmean,1))
-        error("Size of YY rows should match PS.xmean and PS.xstd")
-    endif
-    # Avoid multiply/division by zero
-    ps.xstd(ps.xstd == 0) = 1;
-    yy=(yy ./ ps.ystd) - ps.ymean;
-    xx=(yy .* (ps.xstd*ones(1,size(yy,2)))) + (ps.xmean*ones(1,size(yy,2)));
-  endfunction
-
-  # I don't know why this exists but matlab implements it
-  function [dy_dx]=derivate(xx,yy,ps)
-    checkargs(yy,ps)
-    checkargs(xx,ps)
-
-    cols = size(xx,2);
-    diagonal = diag(ps.ystd ./ ps.xstd);
-    dy_dx = diagonal(:,:,ones(1,cols));
-  endfunction
-
-#end
-
-endfunction
--- a/main/nnet/inst/min_max.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,70 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{Pr} = min_max (@var{Pp})
-## @code{min_max} returns variable Pr with range of matrix rows
-##
-## @emph{warning}: @code{min_max} has been deprecated in favor of
-## @code{minmax}. This function will be removed from future versions of the
-## 'nnet' package".
-##
-## @example
-## PR - R x 2 matrix of min and max values for R input elements
-## @end example
-##
-## @example
-## Pp = [1 2 3; -1 -0.5 -3]
-## pr = min_max(Pp);
-## pr = [1 3; -0.5 -3];
-## @end example
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function Pr = min_max(Pp)
-  persistent warned = false;
-  if (! warned)
-    warned = true;
-    warning ("Octave:deprecated-function",
-             "`min_max' has been deprecated in favor of `minmax'. This function will be removed from future versions of the `nnet' package");
-  endif
-
-  ## check number of input args
-  error(nargchk(1,1,nargin))
-
-  Pr = []; # returns an empty matrix
-  #if ismatrix(Pp)
-  if (!(size(Pp,1)==1) && !(size(Pp,2)==1)) # ismatrix(1) will return 1!!!
-    if isreal(Pp) # be sure, this is no complex matrix
-      Pr = [min(Pp,[],2) max(Pp,[],2)];
-    else
-      error("Argument has illegal type.")
-    endif
-  else
-    error("Argument must be a matrix.")
-  endif
-
-endfunction
-
-%!shared
-%! disp("testing min_max")
-%!test fail("min_max(1)","Argument must be a matrix.")
-%!test fail("min_max('testString')","Argument must be a matrix.")
-%!test fail("min_max(cellA{1}=1)","Argument must be a matrix.")
-%!test fail("min_max([1+1i, 2+2i])","Argument must be a matrix.")
-%!test fail("min_max([1+1i, 2+2i; 3+1i, 4+2i])","Argument has illegal type.")
-
-
--- a/main/nnet/inst/minmax.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,116 +0,0 @@
-## Copyright (C) 2012 Carnë Draug <carandraug+dev@gmail.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {@var{Pr} =} minmax (@var{Pp})
-## Calculate maximum and mininum of rows.
-##
-## For each row of the matrix @var{Pp}, outputs its minimum and maximum on the
-## first and second column of @var{Pr} respectively.  @var{Pr} will have the
-## same number of rows as @var{Pp} and 2 columns.
-##
-## @group
-## @example
-## Pp = [5 7 9 2 5 0 6
-##       5 3 6 2 7 9 3
-##       8 3 2 3 5 6 8]
-## minmax (Pp)
-##   @result{} 0 9
-##      2 9
-##      0 8
-## @end example
-## @end group
-##
-## @var{Pp} can also be a cell array of matrices in wich case they all must have
-## the same number of columns, and all matrices on each row of cells must have
-## the same number of rows.  In this case, matrices of each row of @var{Pp} are
-## concatenated horizontally for calculating the minimum ad maximum values.
-## @var{Pr} will be a single column cell array with same number of rows as
-## @var{Pp}. For example:
-##
-## @group
-## @example
-## Pp = @{[0 1; 1 2; 4 6] [2 3; 8 0; 3 1] [9 1; 5 2; 4 8];
-##        [1 2; 9 7] [5 2; 3 1] [7 6; 0 3]@}
-## minmax (Pp)
-##   @result{} @{
-##        [1,1] =
-##           0   9
-##           0   8
-##           1   8
-##        [2,1] =
-##           1   7
-##           0   9
-##      @}
-## @end example
-## @end group
-##
-## If drawn on a table, it would look like:
-##
-## @verbatim
-##   2x3 cell array      2x1 cell array
-##
-##  0 1   2 3   9 1   >      0 9
-##  1 2   8 0   5 2   >      0 8
-##  4 6   3 1   4 8   >      1 8
-##
-##  1 2   5 2   7 6   >      1 7
-##  9 7   3 1   0 3   >      0 9
-## @end verbatim
-##
-## Note how on this example: the number of columns (3) in the cell array is
-## irrelevant but the output has the same number of rows (2); all matrices have
-## the same number of columns (2).
-##
-## @seealso {cell2mat, max, min}
-## @end deftypefn
-
-function Pr = minmax (Pp)
-
-  if (nargin != 1)
-    print_usage;
-  elseif (minmax_check (Pp))
-    Pr = single_minmax (Pp);
-  elseif (iscell (Pp) && ndims (Pp) == 2 && all (cellfun (@minmax_check, Pp(:))))
-    Pr_rows = cellfun (@rows, Pp(:,1));
-    if (!all (cellfun (@columns, Pp(:)) == columns (Pp{1})))
-      error ("minmax: all matrices must have the same number of columns.");
-    elseif (!all (bsxfun (@eq, cellfun (@rows, Pp), Pr_rows)(:)))
-      error ("minmax: all matrices in a row of cells must have same number of rows.");
-    endif
-    Pr = mat2cell (single_minmax (cell2mat (Pp)), Pr_rows, 2);
-  else
-    error ("minmax: input must be one, or a 2D cell array of, 2D non-complex matrix.");
-  endif
-
-endfunction
-
-function retval = minmax_check (val)
-  retval = isnumeric (val) && !iscomplex (val) && ndims (val) == 2;
-endfunction
-
-function Pr = single_minmax (Pp)
-  Pr = [min(Pp, [], 2) max(Pp, [], 2)];
-endfunction
-
-%!assert (minmax ([2 5 4; -2 6 5]), [2 5; -2 6]);                         # basic usage
-%!assert (minmax ([2 5 4]), [2 5]);                                       # single row, basic usage
-%!assert (minmax ({[0 1; -1 -2; 34 56] [2 3; 8 0; 21 23]; [1 -2; 9 7] [12 5; 13 11]}), ...
-%!        {[0 3; -2 8; 21 56]; [-2 12; 7 13]});                           # basic usage with cell arrays
-%!assert (minmax (1), [1 1]);                                             # matlab compatibility
-%!fail ("minmax ([i 2; 3 4])");                                           # do not accept complex values
-%!fail ("minmax (rand (2, 2, 2))");                                       # only 2D matrix
-%!fail ("minmax ({[0 1; 1 2] [2 3 2; 8 0 2]; [1 2] [9 7 3]})");           # number of columns must be the same everywhere
-%!fail ("minmax ({[0 1; 1 2] [2 3; 8 0; 5 5]; [1 2; 9 7] [1 5; 1 1]})");  # each row of cells must have matrices with same number of rows
--- a/main/nnet/inst/newff.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,277 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {@var{net}} = newff (@var{Pr},@var{ss},@var{trf},@var{btf},@var{blf},@var{pf})
-## @code{newff} create a feed-forward backpropagation network
-##
-## @example
-## Pr - R x 2 matrix of min and max values for R input elements
-## Ss - 1 x Ni row vector with size of ith layer, for N layers
-## trf - 1 x Ni list with transfer function of ith layer,
-##       default = "tansig"
-## btf - Batch network training function,
-##       default = "trainlm"
-## blf - Batch weight/bias learning function,
-##       default = "learngdm"
-## pf  - Performance function,
-##       default = "mse".
-## @end example
-##
-## @example
-## EXAMPLE 1
-## Pr = [0.1 0.8; 0.1 0.75; 0.01 0.8];
-##      it's a 3 x 2 matrix, this means 3 input neurons
-##
-## net = newff(Pr, [4 1], @{"tansig","purelin"@}, "trainlm", "learngdm", "mse");
-## @end example
-##
-## @end deftypefn
-
-## @seealso{sim, init, train}
-
-## Author: Michel D. Schmid
-
-function net = newff(Pr,ss,transFunc,trainFunc,notUsed,performFunc)
-
-  ## initial descriptipn
-  ##  newff(Pr,ss,transfunc,trainFunc,notUsed,performFunc)
-  ##  * Pr is a nx2 matrix with min and max values of standardized inputs
-  ##    Pr means: p-range
-  ##  * ss is a row vector, the first element describes the number
-  ##    of hidden neurons, the second element describes the number
-  ##    of output neurons
-  ##  * transFunc is a cell array of transfer function, standard is "tansig"
-  ##  * trainFunc is the training algorithm
-  ##  * notUsed exist only because we have only one train algorithm which doesn't
-  ##    need a weight learning function
-  ##  * performFunc is written for the performance function, standard is "mse"
-
-  ## check range of input arguments
-  error(nargchk(2,6,nargin))
-
-  ## get number of layers (without input layer)
-  nLayers = length(ss);
-
-  ## set defaults
-  if (nargin <3)
-    # the number of transfer functions depends on the number of
-    # hidden layers, so we have to create a loop here 30.09.09 (dd.mm.yy)
-	for i=1:nLayers
-	  if (i==nLayers)
-	    transFunc{i,1} = "purelin";
-	  else
-        transFunc{i,1}= "tansig";
-      endif
-    endfor
-  endif
-  if (nargin <4)
-    trainFunc = "trainlm";
-  endif
-  if (nargin <5)
-    notUsed = "noSense";
-  endif
-  if (nargin==5)
-    ## it doesn't matter what nargin 5 is ...!
-    ## it won't be used ... it's only for matlab compatibility
-    notUsed = "noSense"
-  endif
-  if (nargin <6)
-    performFunc = "mse";
-  endif
-
-  ## check input args
-  checkInputArgs(Pr,ss);
-
-  ## Standard architecture of neural network
-  net = __newnetwork(1,nLayers,1,"newff");
-  ## description:
-  ##	first argument: number of inputs, nothing else allowed till now
-  ## it's not the same like the number of neurons in this input
-  ## second argument: number of layers, including output layer
-  ## third argument: number of outputs, nothing else allowed till now
-  ## it's not the same like the number of neurons in this output
-
-  ## set inputs with limit of only ONE input
-  net.inputs{1}.range = Pr;
-  [nRows, nColumns] = size(Pr);
-  net.inputs{1}.size = nRows;
-
-  ## set size of IW
-  net.IW{1,1} = zeros(1,nRows);
-  ## set more needed empty cells
-  for iLayers = 2:nLayers
-    net.IW{iLayers,1} = [];
-    #  net.IW{2:nLayers,1} = [];    # old code
-  endfor
-  ## set number of bias, one per layer
-  for iBiases = 1:nLayers
-    net.b{iBiases,1} = 0;
-  endfor
-
-  ## set rest of layers
-
-  ## set size of LayerWeights LW
-  ## the numbers of rows and columns depends on the
-  ## number of hidden neurons and output neurons...
-  ## 2 hidden neurons match 2 columns ...
-  ## 2 output neurons match 2 rows ...
-  for i=2:nLayers
-    net.LW{i,i-1} = zeros(ss(i),ss(i-1));
-  endfor
-  for iLayers = 1:nLayers
-    net.layers{iLayers}.size = ss(iLayers);
-    net.layers{iLayers}.transferFcn = transFunc{iLayers};
-  endfor
-
-  ## define everything with "targets"
-  net.numTargets = ss(end);
-  net.targets = cell(1,nLayers);
-  for i=1:nLayers
-    if (i==nLayers)
-      net.targets{i}.size = ss(end);
-      ## next row of code is only for MATLAB(TM) compatibility
-      ## I never used this the last 4 years ...
-      net.targets{i}.userdata = "Put your custom informations here!";
-    else
-      net.targets{i} = [];
-    endif
-  endfor
-
-  ## Performance
-  net.performFcn = performFunc;
-
-  ## Adaption
-  for i=1:nLayers
-#    net.biases{i}.learnFcn = blf;
-#    net.layerWeights{i,:}.learnFcn = blf;
-    net.biases{i}.size = ss(i);
-  endfor
-
-  ## Training
-  net.trainFcn = trainFunc; # actually, only trainlm will exist
-  net = setTrainParam(net);
-  ## Initialization
-  net = __init(net);
-
-# ======================================================
-#
-# additional check functions...
-#
-# ======================================================
-  function checkInputArgs(Pr,ss)
-    
-    ## check if Pr has correct format
-    if !isreal(Pr) || (size(Pr,2)!=2)
-      error("Input ranges must be a two column matrix!")
-    endif
-    if any(Pr(:,1) > Pr(:,2)) # check if numbers in the second column are larger as in the first one
-      error("Input ranges has values in the second column larger as in the same row of the first column.")
-    endif
-
-    ## check if ss has correct format, must be 1xR row vector
-    if (size(ss,1)!=1)
-      error("Layer sizes is not a row vector.")
-    endif
-    if (size(ss,2)<2)
-      error("There must be at least one hidden layer and one output layer!")
-    endif
-    for k=1:length(ss)
-      sk = ss(k);
-      if !isreal(sk) || any(sk<1) || any(round(sk)!=sk)
-        error("Layer sizes is not a row vector of positive integers.")
-      endif
-    endfor
-
-  endfunction
-# ======================================================
-#
-# additional set functions...
-#
-# ======================================================
-  function net = setTrainParam(net)
-
-    trainFunc = net.trainFcn;
-    switch(trainFunc)
-
-    case "trainlm"
-      net.trainParam.epochs = 100;
-      net.trainParam.goal = 0;
-      net.trainParam.max_fail = 5;
-      net.trainParam.mem_reduc = 1;
-      net.trainParam.min_grad = 1.0000e-010;
-      net.trainParam.mu = 0.0010;
-      net.trainParam.mu_dec = 0.1;
-      net.trainParam.mu_inc = 10;
-      net.trainParam.mu_max = 1.0000e+010;
-      net.trainParam.show = 50;
-      net.trainParam.time = Inf;
-    otherwise
-      error("newff:setTrainParam: this train algorithm isn't available till now!")
-    endswitch
-
-  endfunction
-# ========================================================  
-
-
-endfunction
-
-%!shared
-%! disp("testing newff")
-
-# if input range Pr has only one column
-%!test
-%! Pr = [1;2];
-%! fail("newff(Pr,[1 1],{'tansig','purelin'},'trainlm','unused','mse')","Input ranges must be a two column matrix!")
-
-# if input range Pr has two columns
-%!test
-%! Pr = [1 2 ; 4  6];
-%! assert(__checknetstruct(newff(Pr,[1 1],{'tansig','purelin'},'trainlm','unused','mse')))
-  ## __checknetstruct returns TRUE is input arg is a network structure ...
-
-# if input range Pr has three columns
-%!test
-%! Pr = [1 2 3; 4 5 6];
-%! fail("newff(Pr,[1 1],{'tansig','purelin'},'trainlm','unused','mse')","Input ranges must be a two column matrix!")
-
-# if input range has in the second col greater values as in the first col ...
-%!test
-%! Pr = [5 3; 4 5];
-%! fail("newff(Pr,[1 1],{'tansig','purelin'},'trainlm','unused','mse')",\
-%!  "Input ranges has values in the second column larger as in the same row of the first column.")
-
-# check if ss has correct format
-%!test
-%! Pr = [1 2 ; 4 6];
-%! fail("newff(Pr,[1 1; 2 3],{'tansig','purelin'},'trainlm','unused','mse')",\
-%!  "Layer sizes is not a row vector.")
-
-# check if ss has correct format
-%!test
-%! Pr = [1 2 ; 4 6];
-%! assert(__checknetstruct(newff(Pr,[ 2 3],{'tansig','purelin'},'trainlm','unused','mse')))
-
-# check if ss has correct format
-%!test
-%! Pr = [1 2 ; 4 6];
-%! fail("newff(Pr,[1],{'tansig','purelin'},'trainlm','unused','mse')",\
-%!  "There must be at least one hidden layer and one output layer!")
-
-# check if ss has correct format
-%!test
-%! Pr = [1 2 ; 4 6];
-%! fail("newff(Pr,[-1 1],{'tansig','purelin'},'trainlm','unused','mse')",\
-%!  "Layer sizes is not a row vector of positive integers.")
--- a/main/nnet/inst/newp.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,140 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {@var{net}} = newp (@var{Pr},@var{ss},@var{transFunc},@var{learnFunc})
-## @code{newp} create a perceptron
-##
-## @example
-## PLEASE DON'T USE THIS FUNCTIONS, IT'S STILL NOT FINISHED!
-## =========================================================
-## @end example
-## @example
-## Pr - R x 2 matrix of min and max values for R input elements
-## ss - a scalar value with the number of neurons
-## transFunc - a string with the transfer function
-##       default = "hardlim"
-## learnFunc - a string with the learning function
-##       default = "learnp"
-## @end example
-##
-##
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function net = newp(Pr,ss,transFunc,learnFunc)
-
-  ## initial descriptipn
-  ##  newp(Pr,ss,transFunc,learnFunc)
-  ##  * Pr is a nx2 matrix with min and max values of standardized inputs
-  ##    Pr means: p-range
-  ##  * ss is a scalar value which describes the number of neurons
-  ##    of output neurons
-  ##  * transFunc is the transfer function, standard is "hardlim"
-  ##  * learnFunc is the learning function, standard is "learnp"
-
-  ## check range of input arguments
-  error(nargchk(1,4,nargin))
-
-  ## set defaults
-  if (nargin <2)
-    ss = 1; # means one neuron
-  endif
-  if (nargin <3)
-    transFunc = "hardlim";
-  endif
-  if (nargin <4)
-    learnFunc = "learnp";
-  endif
-
-  ## check input args
-  checkInputArgs(Pr,ss);
-
-#   ## get number of layers (without input layer)
-#   nLayers = length(ss);
-
-  ## Standard architecture of neural network
-  net = __newnetwork(1,1,1,"newp");
-  ## description:
-  ##	first argument: number of inputs, nothing else allowed till now
-  ## it's not the same like the number of neurons in this input
-  ## second argument: number of layers, including output layer
-  ## third argument: number of outputs, nothing else allowed till now
-  ## it's not the same like the number of neurons in this output
-  ## fourth argument: network type
-
-
-  ## set inputs with limit of only ONE input
-  net.inputs{1}.range = Pr;
-  [nRows, nColumns] = size(Pr);
-  net.inputs{1}.size = nRows;
-
-  ## set size of IW
-  net.IW{1,1} = zeros(1,nRows);
-  ## set number of bias, one per layer
-  net.b{iBiases,1} = 0;
-
-  ## define everything with "layers"
-  net.numLayers = ss(end);
-  net.layers = cell(1,1);
-  net.layers{1}.size = ss(end);
-  net.layers{1}.transFcn = transFunc;
-  ## next row of code is only for MATLAB(TM) compatibility
-  ## I never used this the last 4 years ...
-  net.targets{i}.userdata = "Put your custom informations here!";
-
-  ## performance function
-  net.performFnc = "mae";
-
-  ## learning
-  net.biases{1}.learnFcn = learnFunc;
-  net.inputWeights{1,1}.learnFcn = learnFunc;
-
-  ## adaption
-  net.adaptFcn = "trains";
-
-  ## Training
-  net.trainFcn = "trainc";
-
-  ## Initialization
-  net = __init(net);
-
-# ======================================================
-#
-# additional check functions...
-#
-# ======================================================
-  function checkInputArgs(Pr,ss)
-    
-    ## check if Pr has correct format
-    if !isreal(Pr) | (size(Pr,2)!=2)
-      error("Input ranges must be a two column matrix!")
-    endif
-    if any(Pr(:,1) > Pr(:,2)) # check if numbers in the second column are larger as in the first one
-      error("Input ranges has values in the second column larger as in the same row of the first column.")
-    endif
-
-    ## check if ss has correct format, must be a scalar value
-    if ( (size(ss,1)!=1) || (size(ss,2)!=1))
-      error("Layer sizes is not a scalar value.")
-    endif
-
-  endfunction
-
-# ========================================================  
-
-
-endfunction
--- a/main/nnet/inst/poslin.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,32 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}@var{a}= poslin (@var{n})
-## @code{poslin} is a positive linear transfer function used
-## by neural networks
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function a = poslin(n)
-
-   if (n<0)
-     a = 0;
-   else
-     a = n;
-   endif
-
-endfunction
--- a/main/nnet/inst/poststd.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,78 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{Pp},@var{Tt}] = poststd(@var{Pn},@var{meanp},,@var{stdP},@var{Tn},@var{meanT},@var{stdT})
-## @code{poststd} postprocesses the data which has been preprocessed by @code{prestd}.
-## @end deftypefn
-
-## @seealso{prestd,trastd}
-
-## Author: Michel D. Schmid
-
-function [Pp,Tt] = poststd(Pn,meanp,stdp,Tn,meant,stdt)
-
-  ## check range of input arguments
-  error(nargchk(3,6,nargin))
-  if (nargin==4)
-    error("4 input arguments are not allowed!");
-  endif
-  if (nargin==5)
-    error("5 input arguments are not allowed!");
-  endif
-
-  ## do first inputs
-  ## set all standard deviations which are zero to 1
-  [nRowsII, nColumnsII] = size(Pn);
-  rowZeros = zeros(nRowsII,1);
-  findZeros = find(stdp==0);
-  rowZeros(findZeros)=1;
-  nequal = !rowZeros;
-  if (sum(rowZeros) != 0)
-    warning("Some standard deviations are zero. Those inputs won't be transformed.");
-    meanpZero = meanp.*nequal;
-    stdpZero = stdp.*nequal + 1*rowZeros;
-  else
-    meanpZero = meanp;
-    stdpZero = stdp;
-  endif
-  
-  ## calculate the postprocessed inputs
-  nColumnsIIone = ones(1,nColumnsII);
-  Pp = (stdpZero*nColumnsIIone).*Pn + meanpZero*nColumnsIIone;
-
-  ## do also targets
-  if ( nargin==6 )
-    # now set all standard deviations which are zero to 1
-    [nRowsIII, nColumnsIII] = size(stdt);
-    rowZeros = zeros(nRowsIII,1);
-    findZeros = find(stdt==0);
-    rowZeros(findZeros)=1;
-    nequal = !rowZeros;
-    if (sum(rowZeros) != 0)
-      warning("Some standard deviations are zero. Those targets won't be transformed.");
-      meantZero = meant.*nequal;
-      stdtZero = stdt.*nequal + 1*rowZeros;
-    else
-      meantZero = meant;
-      stdtZero = stdt;
-    endif
-
-    ## calculate the postprocessed targets
-    nColumnsIIIone = ones(1,nColumnsIII);
-    Tt = (stdtZero*nColumnsIIIone).*Tn + meantZero*nColumnsIIIone;
-  endif
-
-endfunction
--- a/main/nnet/inst/prestd.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,98 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{pn},@var{meanp},@var{stdp},@var{tn},@var{meant},@var{stdt}] =prestd(@var{p},@var{t})
-## @code{prestd} preprocesses the data so that the mean is 0 and the standard deviation is 1.
-## @end deftypefn
-
-## @seealso{trastd}
-
-## Author: Michel D. Schmid
-
-function [pn,meanp,stdp,tn,meant,stdt] = prestd(Pp,Tt)
-
-  ## inital description
-  ## prestd(p,t)
-  ##  * p are the general descriptions for the inputs of
-  ##    neural networks
-  ##  * t is written for "targets" and these are the outputs
-  ##    of a neural network
-  
-  ## some more detailed description:
-  ## for more informations about this
-  ## formula programmed in this file, see:
-  ## 1. http://en.wikipedia.org/wiki/Standard_score
-  ## 2. http://www.statsoft.com/textbook/stathome.html
-  ##    choose "statistical glossary", choose "standardization"
-
-  ## check range of input arguments
-  error(nargchk(1,2,nargin))
-
-  ## do first inputs
-  meanp = mean(Pp')';
-  stdp = std(Pp')';
-  [nRows,nColumns]=size(Pp);
-  rowOnes = ones(1,nColumns);
-
-  ## now set all standard deviations which are zero to 1
-  [nRowsII, nColumnsII] = size(stdp);
-  rowZeros = zeros(nRowsII,1); # returning a row containing only zeros
-  findZeros = find(stdp==0); # returning a vector containing index where zeros are
-  rowZeros(findZeros)=1; #
-  nequal = !rowZeros;
-  if (sum(rowZeros) != 0)
-    warning("Some standard deviations are zero. Those inputs won't be transformed.");
-    meanpZero = meanp.*nequal;
-    stdpZero = stdp.*nequal + 1*rowZeros;
-  else
-    meanpZero = meanp;
-    stdpZero = stdp;
-  endif
-
-  ## calculate the standardized inputs
-  pn = (Pp-meanpZero*rowOnes)./(stdpZero*rowOnes);
-
-  ## do also targets
-  if ( nargin==2 )
-    meant = mean(Tt')';
-    stdt = std(Tt')';
-
-    ## now set all standard deviations which are zero to 1
-    [nRowsIII, nColumnsIII] = size(stdt);
-    rowZeros = zeros(nRowsIII,1);
-    findZeros = find(stdt==0);
-    rowZeros(findZeros)=1;
-    nequal = !rowZeros;
-    if (sum(rowZeros) != 0)
-      warning("Some standard deviations are zero. Those targets won't be transformed.");
-      meantZero = meant.*nequal;
-      stdtZero = stdt.*nequal + 1*rowZeros;
-    else
-      meantZero = meant;
-      stdtZero = stdt;
-    endif
-
-    ## calculate the standardized targets
-    tn = (Tt-meantZero*rowOnes)./(stdtZero*rowOnes);
-  endif
-endfunction
-
-
-%!shared Pp, Tt, pn
-%!  Pp = [1 2 3 4; -1 3 2 -1];
-%!  Tt = [3 4 5 6];
-%!  [pn,meanp,stdp] = prestd(Pp);
-%!assert(pn,[-1.16190 -0.38730 0.38730 1.16190; -0.84887 1.09141 0.60634 -0.84887],0.00001);
--- a/main/nnet/inst/private/__analyzerows.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,115 +0,0 @@
-## Copyright (C) 2008 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{retmatrix} = __analyzerows(@var{matrix})
-## @code{__analyzerows} takes a matrix as input argument and checks what kind of
-## data are contained in the rows.
-##   a.) binary values? Means the row contains only 0 and 1
-##   b.) unique values?
-##   c.) Min values are several times contained in the row
-##   d.) Max values are several times contained in the row
-## @end deftypefn
-
-## Author: mds
-
-function retmatrix = __analyzerows(matrix)
-
-  ## check number of inputs
-  error(nargchk(1,1,nargin));
-
-  nRows = size(matrix,1);   # get number or rows
-  retmatrix = zeros(nRows,4);
-  doneVec = zeros(nRows,1);
-
-  ## now let's check which rows are binary
-  i = 1;
-  while (i <= nRows)
-    vec = matrix(i,:);
-    n1 = find(vec==1);
-    n0 = find(vec==0);
-    if (length(n1)==0 || length(n0)==0)
-      #do nothing
-    else
-      if (length(vec)==(length(n1)+length(n0)))
-        # in this case, the vector contains only ones and zeros
-        retmatrix(i,1) = 1;
-        doneVec(i) = 1;
-      endif
-    endif
-    i += 1;
-  endwhile
-
-  ## now let's check which rows are unique
-  i = 1;
-  while (i <= nRows)
-    if (doneVec(i)==0)
-      vec = matrix(i,:);
-      n1 = find(vec==vec(1));
-      if (length(vec)==(length(n1)))
-        # in this case, the vector contains only unique data
-        retmatrix(i,2) = 1;
-        doneVec(i) = 1;
-      endif
-    endif
-  i += 1;
-  endwhile
-
-  
-  ## now let's check how often we can find the min value
-  i = 1;
-  while (i <= nRows)
-	if (doneVec(i)==0)
-      vec = matrix(i,:);
-      n1 = min(vec);
-	  retmatrix(i,3) = length(find(n1==vec));
-	endif
-  i += 1;
-  endwhile
-  
-  ## now let's check how often we can find the max value
-  i = 1;
-  while (i <= nRows)
-	if (doneVec(i)==0)
-      vec = matrix(i,:);
-      n1 = max(vec);
-	  retmatrix(i,4) = length(find(n1==vec));
-	endif
-  i += 1;
-  endwhile
-
-endfunction
-
-%!shared b, retmat
-%! disp("testing __analyzerows")
-%! b = [1 0 0 1; 1 0 0 0; 1 2 0 1];
-%! retmat = __analyzerows(b);
-%!assert(retmat(1,1)==1);#%!assert(retmat(1,1)==1);
-%!assert(retmat(2,1)==1);
-%!assert(retmat(3,1)==0);
-%! b = [1 0 0 2; 1 0 0 0; 1 1 1 1];
-%! retmat = __analyzerows(b);
-%!assert(retmat(1,2)==0);
-%!assert(retmat(2,2)==0);
-%!assert(retmat(3,2)==1);
-%! b = [1 0 0 2; 1 0 0 0; 1 1 1 1];
-%! retmat = __analyzerows(b);
-%!assert(retmat(1,3)==2);
-%!assert(retmat(2,3)==0);
-%!assert(retmat(3,3)==0);
-%! retmat = __analyzerows(b);
-%!assert(retmat(1,4)==1);
-%!assert(retmat(2,4)==0);
-%!assert(retmat(3,4)==0);
--- a/main/nnet/inst/private/__calcjacobian.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,281 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid   <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}@var{Jj} = __calcjacobian (@var{net},@var{Im},@var{Nn},@var{Aa},@var{vE})
-## This function calculates the jacobian matrix. It's used inside the
-## Levenberg-Marquardt algorithm of the neural network toolbox.
-## PLEASE DO NOT USE IT ELSEWEHRE, it proparly will not work!
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function [Jj] = __calcjacobian(net,Im,Nn,Aa,vE)
-
-  ## comment:
-  ## - return value Jj is jacobi matrix
-  ##   for this calculation, see "Neural Network Design; Hagan, Demuth & Beale page 12-45"
-
-
-  ## check range of input arguments
-  error(nargchk(5,5,nargin))
-
-  ## get signals from inside the network
-  bias  = net.b;
-
-  ## calculate some help matrices
-  mInputWeight = net.IW{1} * Im;
-  nLayers = net.numLayers;
-  for i=2:nLayers
-    mLayerWeight{i,1} = net.LW{i,i-1} * Aa{i-1,1};
-  endfor
-
-  ## calculate number of columns and rows in jacobi matrix
-  ## firstly, number of columns
-  a = ones(nLayers+1,1); # +1 is for the input
-  a(1) = net.inputs{1}.size;
-  for iLayers = 1:nLayers
-    a(iLayers+1) = net.layers{iLayers}.size;
-  endfor
-  nColumnsJacobiMatrix = 0;
-  for iLayers = 1:nLayers
-    nColumnsJacobiMatrix = (a(iLayers)+1)*a(iLayers+1) + nColumnsJacobiMatrix;
-  endfor
-  ## secondly, number of rows
-  ve = vE{nLayers,1};
-  nRowsJacobiMatrix = length(ve(:));
-
-
-  ## FIRST STEP -----------------------------------------------------
-  ## calculate the neuron outputs without the transfer function
-  ## - n1_1 = W^1*a_1^0+b^1: the ^x factor defined the xth train data set
-  ##   the _x factor defines the layer
-  ## **********  this datas should be hold in Nn
-  ## **********  should be calculated in "__calcperf"
-  ## **********  so Nn{1} means hidden layer
-  ## **********  so Nn{2} means second hidden layer or output layer
-  ## **********  and so on ...
-  ## END FIRST STEP -------------------------------------------------
-
-  ## now we can rerange the signals ... this will be done only for
-  ## matrix calculation ...
-  [nRowsError nColumnsError] = size(ve);
-  errorSize = size(ve(:),1); # this will calculate, if only one row
-  # of errors exist... in other words... two rows will be reranged to
-  # one row with the same number of elements.
-  rerangeIndex = floor([0:(errorSize-1)]/nRowsError)+1;
-  nLayers = net.numLayers;
-
-  for i = 1:nLayers
-    Nn{i,1} = Nn{i,1}(:,rerangeIndex);
-    Aa{i,1} = Aa{i,1}(:,rerangeIndex);
-    [nRows nColumns] = size(Nn{i,1});
-    bTemp = bias{i,1};
-    bias{i,1} = repmat(bTemp,1,nColumns);
-    bias{i,1} = bias{i,1}(:,rerangeIndex);
-  endfor
-  mInputWeight = mInputWeight(:,rerangeIndex);
-  for i=2:nLayers
-    mLayerWeight{i,1} = mLayerWeight{i,1}(:,rerangeIndex);
-  endfor
-  Im = Im(:,rerangeIndex);
-
-  ## define how the errors are connected
-  ## ATTENTION! this happens in row order...
-  numTargets = net.numTargets;
-  mIdentity = -eye(numTargets);
-  cols = size(mIdentity,2);
-  mIdentity = mIdentity(:,rem(0:(cols*nColumnsError-1),cols)+1);
-  errorConnect = cell(net.numLayers,1);
-  startPos = 0;
-  for i=net.numLayers
-    targSize = net.layers{i}.size;
-    errorConnect{i} = mIdentity(startPos+[1:targSize],:);
-    startPos = startPos + targSize;
-  endfor
-
-  ## SECOND STEP ----------------------------------------------
-  ## define and calculate the derivative matrix dF
-  ## - this is "done" by the two first derivative functions
-  ##   of the transfer functions
-  ##   e.g. __dpureline, __dtansig, __dlogsig and so on ...
-
-  ## calculate the sensitivity matrix tildeS
-  ## start at the end layer, this means of course the output layer,
-  ## the transfer function is selectable
-  
-  ## for calculating the last layer
-  ## this should happen like following:
-  ## tildeSx = -dFx(n_x^x)
-  ## use mIdentity to calculate the number of targets correctly
-  ## for all other layers, use instead:
-  ## tildeSx(-1) = dF1(n_x^(x-1))(W^x)' * tildeSx;
-
-  for iLayers = nLayers:-1:1 # this will count from the last
-                             # layer to the first layer ...
-    n = Nn{iLayers}; # nLayers holds the value of the last layer...
-    ## which transfer function should be used?
-    if (iLayers==nLayers)
-      switch(net.layers{iLayers}.transferFcn)
-        case "radbas"
-          tildeSxTemp = __dradbas(n);
-        case "purelin"
-          tildeSxTemp = __dpurelin(n);
-        case "tansig"
-          n = tansig(n);
-          tildeSxTemp = __dtansig(n);
-        case "logsig"
-          n = logsig(n);
-          tildeSxTemp = __dlogsig(n);
-        otherwise	
-          error(["transfer function argument: " net.layers{iLayers}.transferFcn  " is not valid!"])
-      endswitch
-      tildeSx{iLayers,1} = tildeSxTemp .* mIdentity;
-      n = bias{nLayers,1};
-      switch(net.layers{iLayers}.transferFcn)
-        case "radbas"
-          tildeSbxTemp = __dradbas(n);
-        case "purelin"
-          tildeSbxTemp = __dpurelin(n);
-        case "tansig"
-          n = tansig(n);
-          tildeSbxTemp = __dtansig(n);
-        case "logsig"
-          n = logsig(n);
-          tildeSbxTemp = __dlogsig(n);
-        otherwise
-          error(["transfer function argument: " net.layers{iLayers}.transferFcn  " is not valid!"])
-      endswitch
-      tildeSbx{iLayers,1} = tildeSbxTemp .* mIdentity;
-    endif
-
-    if (iLayers<nLayers)
-      dFx = ones(size(n));
-      switch(net.layers{iLayers}.transferFcn) ######## new lines ...
-        case "radbas"
-          nx = radbas(n);
-          dFx = __dradbas(nx);
-        case "purelin"
-	  nx = purelin(n);
-	  dFx = __dpurelin(nx);
-        case "tansig"         ######## new lines ...
-	  nx = tansig(n);
-	  dFx = __dtansig(nx);
-	case "logsig"    ######## new lines ...
-          nx = logsig(n);  ######## new lines ...
-	  dFx = __dlogsig(nx); ######## new lines ...
-	otherwise     ######## new lines ...
-	  error(["transfer function argument: " net.layers{iLayers}.transferFcn  " is not valid!"])######## new lines ...
-       endswitch ############# new lines ....
-	  LWtranspose = net.LW{iLayers+1,iLayers};
-      if iLayers<(nLayers-1)
-        mIdentity = -ones(net.layers{iLayers}.size,size(mIdentity,2));
-      endif
-
-      mTest = tildeSx{iLayers+1,1};
-      LWtranspose = LWtranspose' * mTest;
-      tildeSx{iLayers,1} = dFx .* LWtranspose;
-      tildeSxTemp = dFx .* LWtranspose;
-      tildeSbx{iLayers,1} = ones(size(nx)).*tildeSxTemp;
-    endif
-
-  endfor #  if iLayers = nLayers:-1:1
-  ## END SECOND STEP -------------------------------------------------
-
-  ## THIRD STEP ------------------------------------------------------
-  ## some problems occur if we have more than only one target... so how
-  ## does the jacobi matrix looks like?
-
-  ## each target will cause an extra row in the jacobi matrix, for
-  ## each training set..  this means, 2 targets --> double of rows in the
-  ## jacobi matrix ... 3 targets --> three times the number of rows like
-  ## with one target and so on.
-
-  ## now calculate jacobi matrix
-  ## to do this, define first the transposed of it
-  ## this makes it easier to calculate on the "batch" way, means all inputs
-  ## at the same time...
-  ## and it makes it easier to use the matrix calculation way..
-
-  JjTrans = zeros(nRowsJacobiMatrix,nColumnsJacobiMatrix)'; # transposed jacobi matrix
-
-  ## Weight Gradients
-  for i=1:net.numLayers
-    if i==1
-      newInputs = Im;
-      newTemps =  tildeSx{i,1};
-      gIW{i,1} = copyRows(newTemps,net.inputs{i}.size) .* copyRowsInt(newInputs,net.layers{i}.size);
-    endif
-    if i>1
-      Ad = cell2mat(Aa(i-1,1)');
-      newInputs = Ad;
-      newTemps = tildeSx{i,1};
-      gLW{i,1} = copyRows(newTemps,net.layers{i-1}.size) .* copyRowsInt(newInputs,net.layers{i}.size);
-    endif
-  endfor
-
-  for i=1:net.numLayers
-    [nRows, nColumns] = size(Im);
-    if (i==1)
-      nWeightElements = a(i)*a(i+1); # n inputs * n hidden neurons
-      JjTrans(1:nWeightElements,:) =  gIW{i}(1:nWeightElements,:);
-      nWeightBias = a(i+1);
-      start = nWeightElements;
-      JjTrans(start+1:start+nWeightBias,:) = tildeSbx{i,1};
-      start = start+nWeightBias;
-    endif
-    if (i>1)
-      nLayerElements = a(i)*a(i+1); # n hidden neurons * n output neurons
-      JjTrans(start+1:start+nLayerElements,:)=gLW{i}(1:nLayerElements,:);
-      start = start +  nLayerElements;
-      nLayerBias = a(i+1);
-      JjTrans(start+1:start+nLayerBias,:) = tildeSbx{i,1};
-      start = start + nLayerBias;
-    endif
-  endfor
-  Jj = JjTrans';
-  ## END THIRD STEP -------------------------------------------------
-
-
-#=======================================================
-#
-# additional functions
-#
-#=======================================================
-
-  function k = copyRows(k,m)
-    # make copies of the ROWS of Aa matrix
-
-    mRows = size(k,1);
-    k = k(rem(0:(mRows*m-1),mRows)+1,:);
-  endfunction
-
-# -------------------------------------------------------
-
-  function k = copyRowsInt(k,m)
-    # make copies of the ROWS of matrix with elements INTERLEAVED
-
-    mRows = size(k,1);
-    k = k(floor([0:(mRows*m-1)]/m)+1,:);
-  endfunction
-
-# =====================================================================
-#
-# END additional functions
-#
-# =====================================================================
-
-endfunction
--- a/main/nnet/inst/private/__calcperf.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,108 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid    <email: michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{perf}, @var{Ee}, @var{Aa}, @var{Nn}] = __calcperf (@var{net},@var{xx},@var{Im},@var{Tt})
-## @code{__calcperf} calculates the performance of a multi-layer neural network.
-## PLEASE DON'T USE IT ELSEWHERE, it proparly won't work.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function [perf,Ee,Aa,Nn] = __calcperf(net,xx,Im,Tt)
-
-  ## comment:
-  ## perf, net performance.. from input to output through the hidden layers
-  ## Aa, output values of the hidden and last layer (output layer)
-  ## is used for NEWFF network types
-
-  ## calculate bias terms
-  ## must have the same number of columns like the input matrix Im
-  [nRows, nColumns] = size(Im);
-  Btemp = cell(net.numLayers,1); # Btemp: bias matrix
-  ones1xQ = ones(1,nColumns);
-  for i= 1:net.numLayers
-    Btemp{i} = net.b{i}(:,ones1xQ);
-  endfor
-
-  ## shortcuts
-  IWtemp = cell(net.numLayers,net.numInputs,1);# IW: input weights ...
-  LWtemp = cell(net.numLayers,net.numLayers,1);# LW: layer weights ...
-  Aa = cell(net.numLayers,1);# Outputs hidden and output layer
-  Nn = cell(net.numLayers,1);# outputs before the transfer function
-  IW = net.IW; # input weights
-  LW = net.LW; # layer weights
-
-  ## calculate the whole network till outputs are reached...
-  for iLayers = 1:net.numLayers
-
-    ## calculate first input weights to weighted inputs..
-    ## this can be done with matrix calculation...
-    ## called "dotprod"
-    ## to do this, there must be a special matrix ...
-    ## e.g.  IW = [1 2 3 4 5; 6 7 8 9 10] * [ 1 2 3; 4 5 6; 7 8 9; 10 11 12; 1 2 3];
-    if (iLayers==1)
-      IWtemp{iLayers,1} = IW{iLayers,1} * Im;
-      onlyTempVar = [IWtemp(iLayers,1) Btemp(iLayers)];
-    else
-      IWtemp{iLayers,1} = [];
-    endif
-
-    ## now calculate layer weights to weighted layer outputs
-    if (iLayers>1)
-      Ad = Aa{iLayers-1,1};
-      LWtemp{iLayers,1} = LW{iLayers,iLayers-1} * Ad;
-      onlyTempVar = [LWtemp(iLayers,1) Btemp(iLayers)];
-    else
-      LWtemp{iLayers,1} = [];
-    endif
-
-    Nn{iLayers,1} = onlyTempVar{1};
-    for k=2:length(onlyTempVar)
-      Nn{iLayers,1} = Nn{iLayers,1} + onlyTempVar{k};
-    endfor
-
-    ## now calculate with the transfer functions the layer output
-    switch net.layers{iLayers}.transferFcn
-    case "purelin"
-      Aa{iLayers,1} = purelin(Nn{iLayers,1});
-    case "tansig"
-      Aa{iLayers,1} = tansig(Nn{iLayers,1});
-    case "logsig"
-      Aa{iLayers,1} = logsig(Nn{iLayers,1});
-    otherwise
-      error(["Transfer function: " net.layers{iLayers}.transferFcn " doesn't exist!"])
-    endswitch
-
-  endfor  # iLayers = 1:net.numLayers
-
-  ## now calc network error
-  Ee = cell(net.numLayers,1);
-
-  for i=net.numLayers
-    Ee{i,1} = Tt{i,1} - Aa{i,1};# Tt: target
-    # Ee will be the error vector cell array
-  endfor
-
-  ## now calc network performance
-  switch(net.performFcn)
-  case "mse"
-    perf = __mse(Ee);
-  otherwise
-    error("for performance functions, only mse is currently valid!")
-  endswitch
-
-endfunction
--- a/main/nnet/inst/private/__checknetstruct.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,47 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{isTrue}] = __checknetstruct (@var{net})
-## This function will check if a valid structure seems to be a neural network
-## structure
-##
-## @noindent
-##
-## left side arguments:
-## @noindent
-##
-## right side arguments:
-## @noindent
-##
-##
-## @noindent
-## are equivalent.
-## @end deftypefn
-
-## @seealso{newff,prestd,trastd}
-
-## Author: Michel D. Schmid
-
-
-function isTrue = __checknetstruct(net)
-
-  isTrue = 0;
-  ## first check, if it's a structure
-  if (isstruct(net) && isfield(net,"networkType"))
-    isTrue = 1;
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__copycoltopos1.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,43 +0,0 @@
-## Copyright (C) 2008 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{retmatrix} = __copycoltopos1(@var{matrix},@var{colIndex})
-## @code{__copycoltopos1} copies the column of position colIndex to the first position.
-## Moves the rest of the matrix one position to the right.
-## @end deftypefn
-
-## Author: mds
-
-function retmatrix = __copycoltopos1(matrix,colIndex)
-
-  ## check number of inputs
-  error(nargchk(2,2,nargin));
-
-  temp = matrix(:,colIndex);
-  matrix(:,colIndex) = []; # delete col
-  retmatrix = [temp matrix ];
-
-endfunction
-
-%!shared a, retmat
-%! disp("testing __copycoltopos1")
-%! a = [0 1 2 3 4; 5 6 7 8 9];
-%! retmat = __copycoltopos1(a,3);
-%!assert(retmat(1,1)==2);
-%!assert(retmat(2,1)==7);
-%! retmat = __copycoltopos1(a,5);
-%!assert(retmat(1,1)==4);
-%!assert(retmat(2,1)==9);
--- a/main/nnet/inst/private/__dlogsig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,30 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{a} = __dlogsig (@var{n})
-##
-## @end deftypefn
-
-## @seealso{__dpurelin,__dtansig}
-
-## Author: Michel D. Schmid
-
-
-function a = __dlogsig(n)
-  
-  a = n.*(1-n);
-
-endfunction
--- a/main/nnet/inst/private/__dpurelin.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,33 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{a} = __dpurelin (@var{n})
-## @code{dpurelin}, first derivative of purelin
-## @example
-##
-## purelin is a linear transfer function used by neural networks
-## @end example
-##
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function a = __dpurelin(n)
-
-   [nRows, nColumns] = size(n);
-   a = ones(nRows,nColumns);
-
-endfunction
--- a/main/nnet/inst/private/__dradbas.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,42 +0,0 @@
-## Copyright (C) 2010 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __dradbas (@var{n})
-## First derivative of the radial basis transfer function.
-##
-## @code{__dradbas(n) = exp(-n^2)*-2*x}
-##
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function retval = __dradbas (n)
-
-  if (nargin != 1)
-    print_usage ();
-  else
-    retval = exp (-n^2)*(-2)*x;
-    # the derivative of exp(-n^2) must be calculated
-    # with help of the chain-rule!
-    # d/dx of e^x = e^x
-    # d/dx of -x^2 = -2x
-    # now calculate the product of both
-  endif
-endfunction
-
-
-#%!assert (radbas (3), exp (-3^2));
--- a/main/nnet/inst/private/__dtansig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,36 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{n} = __dtansig (@var{n})
-## first derivative of @code{tansig}
-##
-## @example
-##
-## tansig is a symmetric non linear transfer function
-## used by neural networks.
-## Input n must be calculated with "n = tansig(n)".
-## @end example
-##
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function a = __dtansig(n)
-
-  a = 1-(n.*n);
-
-endfunction
--- a/main/nnet/inst/private/__getx.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,52 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid     <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{x} = __getx (@var{net})
-## @code{__getx} will rerange the weights in one columns vector.
-##
-##
-## @noindent
-## @end deftypefn
-
-
-## Author: Michel D. Schmid
-
-function x = __getx(net)
-
-  ## check number of inputs
-  error(nargchk(1,1,nargin));
-
-  ## check input args
-  ## check "net", must be a net structure
-  if !__checknetstruct(net)
-    error("Structure doesn't seem to be a neural network")
-  endif
-
-  ## inputs
-  x = net.IW{1,1}(:);
-  x = [x; net.b{1}(:)];
-
-  nNumLayers = net.numLayers;
-  for iLayers = 2:nNumLayers # 1 would be the input layer
-
-    ## layers
-    x = [x; net.LW{iLayers,iLayers-1}(:)];
-    x = [x; net.b{iLayers}(:)];
-
-  endfor
-
-
-endfunction
--- a/main/nnet/inst/private/__init.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,82 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{net} = __init (@var{net})
-## @code{__init} initializes a neural network. This will be done
-## with the function @code{rand} from octave.
-##
-## @example
-## net = __init(net);
-## @end example
-##
-## This function takes the octave function "rand" to init the 
-## neural network weights.
-##
-## @noindent
-## @end deftypefn
-
-
-## Author: Michel D. Schmid
-
-function net=__init(net)
-
-  ## check number of inputs
-  error(nargchk(1,1,nargin));
-
-  ## check input
-  if ( !__checknetstruct(net) )
-    error("__init: wrong argument type, must be a structure!");
-  endif
-
-
-  if (strcmp(net.networkType,"newff"))
-
-    ## init with random numbers between +-1
-    ## input weight layer
-    mRand = rand(net.layers{1}.size,net.inputs{1}.size);
-    net.IW{1} = mRand*2-1;
-
-    ## hidden layers
-    nLayers = net.numLayers;
-    for i=2:nLayers
-      mRand = rand(net.layers{i}.size,net.layers{i-1}.size);
-      net.LW{i,i-1} = mRand*2-1;
-    endfor
-    for i=1:nLayers
-      mRand = rand(net.biases{i}.size,1);
-      net.b{i} = mRand*2-1;
-    endfor
-  elseif (strcmp(net.networkType,"newp"))
-
-    ## init with zeros
-    inputRows = size(net.inputs{1,1}.range,1);
-    net.IW{1} = zeros(inputRows,1);
-    net.b{1} = zeros(1,1);
-  endif
-
-  ## warn user of constant inputs
-  for i=1:net.numInputs
-    prange = net.inputs{i}.range;
-    if (any(prange(:,1) == prange(:,2)))
-      fprintf("\n")
-      fprintf("** Warning in INIT\n")
-      fprintf("** Network net.inputs{%g}.range has a row with equal min and max values.\n",i)
-      fprintf("** Constant inputs do not provide useful information.\n")
-      fprintf("\n")
-    end
-  end
-
-endfunction
--- a/main/nnet/inst/private/__mae.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,51 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}@var{perf} = __mae (@var{E})
-## @code{__mse} returns the Mean-Square-Error of a vector E
-##
-## @example
-##
-## This function is used to calculate the perceptron performance
-## @end example
-##
-## @end deftypefn
-
-## @seealso{__mse}
-
-## Author: Michel D. Schmid
-
-function perf = __mae(E)
-
-  ## check number of inputs
-  error(nargchk(1,1,nargin));
-
-  if iscell(E)
-    perf = 0;
-    elements = 0;
-    for i=1:size(E,1)
-      for j=1:size(E,2)
-        perf = perf + sum(sum(E{i,j}.^2));
-        elements = elements + prod(size(E{i,j}));
-      endfor
-    endfor
-    perf = perf / elements;
-  else
-    error("Error vector should be a cell array!")
-  endif
-
-
-endfunction
--- a/main/nnet/inst/private/__mse.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,51 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}@var{perf} = __mse (@var{E})
-## @code{__mse} returns the Mean-Square-Error of a vector E
-##
-## @example
-##
-## This function is used to calculate the network performance
-## @end example
-##
-## @end deftypefn
-
-## @seealso{__mae}
-
-## Author: Michel D. Schmid
-
-function perf = __mse(E)
-
-  ## check number of inputs
-  error(nargchk(1,1,nargin));
-
-  if iscell(E)
-    perf = 0;
-    elements = 0;
-    for i=1:size(E,1)
-      for j=1:size(E,2)
-        perf = perf + sum(sum(E{i,j}.^2));
-        elements = elements + prod(size(E{i,j}));
-      endfor
-    endfor
-    perf = perf / elements;
-  else
-    error("Error vector should be a cell array!")
-  endif
-
-
-endfunction
--- a/main/nnet/inst/private/__newnetwork.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,193 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}{@var{net}} = __newnetwork(@var{numInputs},@var{numLayers},@var{numOutputs},@var{networkType})
-## @code{__newnetwork} create a custom 'zero'-network
-##
-##
-## @example
-## net = __newnetwork(numInputs,numLayers,numOutputs,networkType)
-##
-## numInputs : number of input vectors, actually only 1 allowed
-## numLayers : number of layers
-## numOutputs: number of output vectors, actually only 1 allowed
-## networkType: e.g. feed-forward-network "newff"
-## @end example
-##
-## @example
-## net = __newnetwork(1,2,1,"newff")
-##       1 input layer, two hidden layers, one output layer
-##       and the network type
-## @end example
-##
-## @noindent
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function net = __newnetwork(numInputs,numLayers,numOutputs,networkType)
-
-  ## check range of input arguments
-  error(nargchk(4,4,nargin))
-
-  ## check input args
-  if ( !isposint(numInputs) )
-    error("network: at least 1 input must be defined! ")
-    # this can't happen actually, only one is allowed and this
-    # one is hard coded
-  elseif ( !isposint(numLayers) )
-    error("network: at least 1 hidden- and one output layer must be defined! ")
-  endif
-  ## second check for numLayers... must be at least "2" for the
-  ## newff, this means at least 1 hidden and 1 output layer
-  if (strcmp(networkType,"newff")  && (numLayers<2))
-    error("network: not enough layers are defined! ")
-  endif
-
-  ## define network type
-  net.networkType = networkType;
-
-  ## ZERO NETWORK
-  net.numInputs = 0;
-  net.numLayers = 0;
-  net.numInputDelays = 0;
-  net.numLayerDelays = 0;
-  # the next five parameters aren't used till now, they are used
-  # only for matlab nnet type compatibility ==> saveMLPStruct
-  net.biasConnect = [];   # not used parameter till now
-  net.inputConnect = [];  # not used parameter till now
-  net.layerConnect = [];  # not used parameter till now
-  net.outputConnect = []; # not used parameter till now
-  net.targetConnect = []; # not used parameter till now
-  net.numOutputs = 0;
-  net.numTargets = 0;
-  net.inputs = cell(0,1);
-  net.layers = cell(0,1);
-  net.biases = cell(0,1);
-  net.inputWeights = cell(0,0);
-  net.layerWeights = cell(0,0);
-  net.outputs = cell(1,0);
-  net.targets = cell(1,0);
-  net.performFcn = "";
-  net.performParam = [];
-  net.trainFcn = "";
-  net.trainParam = [];
-  net.IW = {};
-  net.LW = {};
-  net.b = cell(0,1);
-  net.userdata.note = "Put your custom network information here.";
-
-
-  ## ARCHITECTURE
-  
-  ## define everything with "inputs"
-  net.numInputs = numInputs;
-  ## actually, it's only possible to have "one" input vector
-  net.inputs{1,1}.range = [0 0];
-  net.inputs{1,1}.size = 0;
-  net.inputs{1,1}.userdata = "Put your custom informations here!";
-  
-  ## define everything with "layers"
-  net.numLayers = numLayers;
-  net = newLayers(net,numLayers);
-
-  ## define unused variables, must be defined for saveMLPStruct
-  net.biasConnect = [0; 0];
-  net.inputConnect = [0; 0];
-  net.layerConnect = [0 0; 0 0];
-  net.outputConnect = [0 0];
-  net.targetConnect = [0 0];
-  net.numInputDelays = 0;
-  net.numLayerDelays = 0;
-
-  ## define everything with "outputs"
-  net.numOutputs = numOutputs;
-  net.outputs = cell(1,numLayers);
-  for i=1:numLayers
-    if (i==numLayers)
-      net.outputs{i}.size = 1; # nothing else allowed till now
-      net.outputs{i}.userdata = "Put your custom informations here!";
-    else
-      net.outputs{i} = [];
-    endif
-  endfor
-
-  ## define everything with "biases"
-  net = newBiases(net,numLayers);
-
-
-
-#=====================================================
-#
-# Additional ARCHITECTURE Functions
-#
-#=====================================================
-  function net = newLayers(net,numLayers)
-
-    ## check range of input arguments
-    error(nargchk(2,2,nargin))
-
-    ## check type of arguments
-    if ( !isscalar(numLayers) || !isposint(numLayers) )
-      error("second argument must be a positive integer scalar value!")
-    endif
-    if ( !__checknetstruct(net) )
-      error("first argument must be a network structure!")
-    endif
-
-    for iRuns=1:numLayers
-      net.layers{iRuns,1}.dimension = 0;
-      net.layers{iRuns,1}.netInputFcn = "";
-      net.layers{iRuns,1}.size = 0;
-### TODO: test with newff      net.layers{iRuns,1}.transferFcn = "tansig";
-      net.layers{iRuns,1}.transferFcn = "";
-      net.layers{iRuns,1}.userdata = "Put your custom informations here!";
-    endfor
-
-  endfunction
-
-#-----------------------------------------------------
-
-  function net = newBiases(net,numLayers)
-
-    ## check range of input arguments
-    error(nargchk(2,2,nargin))
-
-    ## check type of arguments
-    if ( !isscalar(numLayers) || !isposint(numLayers) )
-      error("second argument must be a positive integer scalar value!")
-    endif
-    if ( !isstruct(net) )
-      error("first argument must be a network structure!")
-    endif
-
-    for iRuns=1:numLayers
-      net.biases{iRuns,1}.learn = 1;
-      net.biases{iRuns,1}.learnFcn = "";
-      net.biases{iRuns,1}.learnParam = "undefined...";
-      net.biases{iRuns,1}.size = 0;
-      net.biases{iRuns,1}.userdata = "Put your custom informations here!";
-    endfor
-
-  endfunction
-
-# ================================================================
-#
-#             END Additional ARCHITECTURE Functions
-#
-# ================================================================
-
-endfunction
--- a/main/nnet/inst/private/__optimizedatasets.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,87 +0,0 @@
-## Copyright (C) 2008 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{retmatrix} = __optimizedatasets (@var{matrix},@var{nTrainSets},@var{nTargets},@var{bRand})
-## @code{__optimizedatasets} reranges the data sets depending on the input arguments.
-## @code{matrix} is the data set matrix containing inputs and outputs (targets) in row order.
-## This means for example: the first three rows are inputs and the fourth row is an output row.
-## The second argument is used in the optimizing algorithm. All cols with min and max values must
-## be in the range of the train data sets. The third argument defines how much rows are equal to the
-## neural network targets. These rows must be at the end of the data set!
-## The fourth arguemnt is optional and defines if the data sets have to be randomised before
-## optimizing.
-## Default value for bRand is 1, means randomise the columns.
-## @end deftypefn
-
-## Author: mds
-
-function retmatrix = __optimizedatasets(matrix,nTrainSets,nTargets,bRand)
-
-  ## check number of inputs
-  error(nargchk(3,4,nargin));
-
-  # set default values
-  bRandomise = 1;
-  
-  if (nargin==4)
-    bRandomise = bRand;
-  endif
-  
-  # if needed, randomise the cols
-  if (bRandomise)
-    matrix = __randomisecols(matrix);
-  endif
-  
-  # analyze matrix, which row contains what kind of data?
-  # a.) binary values? Means the row contains only 0 and 1
-  # b.) unique values?
-  # c.) Min values are several times contained in the row
-  # d.) Max values are several times contained in the row
-  matrix1 = matrix(1:end-nTargets,:);
-  analyzeMatrix = __analyzerows(matrix1);
-  
-  # now sort "matrix" with help of analyzeMatrix
-  # following conditions must be kept:
-  # a.) rows containing unique values aren't sorted!
-  # b.) sort first rows which contains min AND max values only once
-  # c.) sort secondly rows which contains min OR max values only once
-  # d.) at last, sort binary data if still needed!
-  retmatrix = __rerangecolumns(matrix,analyzeMatrix,nTrainSets);
-
-
-endfunction
-
-%!shared retmatrix, matrix
-%! disp("testing __optimizedatasets")
-%! matrix = [1 2 3 2 1 2 3 0 5 4 3 2 2 2 2 2 2; \
-%!			 0 1 1 0 0 0 0 0 0 0 0 0 0 0 1 1 0; \
-%!			-1 3 2 4 9 1 1 1 1 1 9 1 1 1 9 9 0; \
-%!			 2 3 2 3 2 2 2 2 3 3 3 3 1 1 1 1 1];
-%! ## The last row is equal to the neural network targets
-%! retmatrix = __optimizedatasets(matrix,9,1);
-%! ## the above statement can't be tested with assert!
-%! ## it contains random values! So pass a "success" message
-%!assert(1==1);
-%! matrix = [1 2 3 2 1 2 3 0 5 4 3 2 2 2 2 2 2; \
-%!			 0 1 1 0 0 0 0 0 0 0 0 0 0 0 1 1 0; \
-%!			-1 3 2 4 9 1 1 1 1 1 9 1 1 1 9 9 0; \
-%!			 2 3 2 3 2 2 2 2 3 3 3 3 1 1 1 1 1];
-%! ## The last row is equal to the neural network targets
-%! retmatrix = __optimizedatasets(matrix,9,1,0);
-%!assert(retmatrix(1,1)==5);
-%!assert(retmatrix(2,1)==0);
-%!assert(retmatrix(3,1)==1);
-%!assert(retmatrix(4,1)==3);
--- a/main/nnet/inst/private/__printAdaptFcn.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,34 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printAdaptFcn (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printAdaptFcn(fid,net)
-
-  if isfield(net,"adaptFcn")
-    if isempty(net.adaptFcn)
-      fprintf(fid,"            adaptFcn:  '%s'\n","empty");
-    else
-      fprintf(fid,"            adaptFcn:  '%s'\n",net.adaptFcn);
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printAdaptParam.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,34 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printAdaptParam (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printAdaptParam(fid,net)
-
-  if isfield(net,"adaptParam")
-    if isempty(net.adaptParam)
-      fprintf(fid,"          adaptParam:  '%s'\n","not yet used item");
-    else
-      fprintf(fid,"          adaptParam:  '%s'\n",net.adaptParam);
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printB.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,46 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printB (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printB(fid,net)
-
-  if isfield(net,"b")
-    nBiases = 0;
-    # check if it's cell array
-    if iscell(net.b)
-      [nRows, nColumns] = size(net.b);
-      for i=1:nRows
-        for k=1:nColumns
-          if !isempty(net.b{i,k})
-            nBiases = nBiases+1;
-          endif
-        endfor
-      endfor
-      # insert enough spaces to put ":" to position 20
-      # insert 2 spaces for distance between ":" and "%"
-      fprintf(fid,"                   b: {%dx%d cell} containing %d bias vectors\n",nRows,nColumns,nBiases);
-    else
-      fprintf(fid,"unsure if this is possible\n")
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printBiasConnect.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,58 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printBiasConnect (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function __printBiasConnect(fid,net)
-
-  if isfield(net,"biasConnect")
-    # net.biasConnect can be a matrix..!
-    # check if it's a matrix
-    if isscalar(net.biasConnect)
-      error("unsure if this is possible..")
-    elseif isnumeric(net.biasConnect)
-      if ismatrix(net.biasConnect)
-        if issquare(net.biasConnect)
-             # nothing prgrammed till now
-        elseif isvector(net.biasConnect)
-          # insert enough spaces to put ":" to position 20
-          # insert 2 spaces for distance between ":" and "%"
-          # print bracket for open
-          fprintf(fid,"         biasConnect:  [");
-          [nRows nColumns] = size(net.biasConnect);
-          for k = 1:1:nRows
-            for i = 1:1:nColumns
-              fprintf(fid,"%d",net.biasConnect(i*k));
-            endfor
-            if k!=nRows
-              #print ; for newline in matrix
-              fprintf(fid,";");
-            endif
-          endfor
-          # print last bracket
-          fprintf(fid,"] not yet used item\n");
-        endif  # if issquare..
-      endif #if ismatrix
-    endif # isscalar(net.biasConnect)
-  endif  # if isfield(...)
-
-endfunction
--- a/main/nnet/inst/private/__printBiases.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,38 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printBiases (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printBiases(fid,net)
-
-  if isfield(net,"biases")
-    # check if it's cell array
-    if iscell(net.biases)
-      [nRows, nColumns] = size(net.biases);
-      # insert enough spaces to put ":" to position 20
-      # insert 2 spaces for distance between ":" and "%"
-      fprintf(fid,"              biases: {%dx%d cell} containing %d biases\n",nRows,nColumns,length(net.biases));
-    else
-      fprintf(fid,"unsure if this is possible\n");
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printIW.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,46 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printIW (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printIW(fid,net)
-
-  if isfield(net,"IW")
-    nInputs = 0;
-    # check if it's cell array
-    if iscell(net.IW)
-      [nRows, nColumns] = size(net.IW);
-      for i=1:nRows
-        for k=1:nColumns
-          if !isempty(net.IW{i,k})
-            nInputs = nInputs+1;
-          endif
-        endfor
-      endfor
-      # insert enough spaces to put ":" to position 20
-      # insert 2 spaces for distance between ":" and "%"
-      fprintf(fid,"                  IW: {%dx%d cell} containing %d input weight matrix\n",nRows,nColumns,nInputs);
-    else
-      fprintf(fid,"unsure if this is possible\n");
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printInitFcn.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,35 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printInitFcn (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function __printInitFcn(fid,net)
-
-  if isfield(net,"initFcn")
-    if isempty(net.initFcn)
-      fprintf(fid,"             initFcn:  '%s'\n","empty");
-    else
-      fprintf(fid,"             initFcn:  '%s'\n",net.initFcn);
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printInitParam.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,34 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printInitParam (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printInitParam(fid,net)
-
-  if isfield(net,"initParam")
-    if isempty(net.initParam)
-      fprintf(fid,"           initParam:  '%s'\n","not yet used item");
-    else
-      fprintf(fid,"           initParam:  '%s'\n",net.initParam);
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printInputConnect.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,57 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printInputConnect (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printInputConnect(fid,net)
-
-  if isfield(net,"inputConnect")
-    # net.inputConnect can be a matrix..!
-    # check if it's a matrix
-    if isscalar(net.inputConnect)
-      error("unsure if this is possible..")
-    elseif isnumeric(net.inputConnect)
-      if ismatrix(net.inputConnect)
-        if issquare(net.inputConnect)
-          # nothing prgrammed till now
-        elseif isvector(net.inputConnect)
-          # insert enough spaces to put ":" to position 20
-          # insert 2 spaces for distance between ":" and "%"
-          # print bracket for open
-          fprintf(fid,"        inputConnect:  [");
-          [nRows nColumns] = size(net.inputConnect);
-          for k = 1:1:nRows
-            for i = 1:1:nColumns
-              fprintf(fid,"%d",net.inputConnect(i*k));
-            endfor
-            if k!=nRows
-              #print ; for newline in matrix
-              fprintf(fid,";");
-            endif
-          endfor
-          # print last bracket
-          fprintf(fid,"] not yet used item\n");
-        endif  # if issquare..
-      endif #if ismatrix
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printInputWeights.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,40 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printInputsWeights (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printInputWeights(fid,net)
-
-  if isfield(net,"inputweights")
-    # check if it's cell array
-    if iscell(net.inputweights)
-      [nRows, nColumns] = size(net.inputweights);
-      # insert enough spaces to put ":" to position 20
-      # insert 2 spaces for distance between ":" and "%"
-      fprintf(fid,"        inputweights: {%dx%d cell} containing xx input weight\n",nRows,nColumns);
-    else
-      fprintf(fid,"unsure if this is possible\n");
-    endif
-  else
-    fprintf(fid,"field inputweights not found & not yet used item\n");
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printInputs.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,49 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printInputs (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printInputs(fid,net)
-
-  if isfield(net,"inputs")
-    # check if it's cell array
-    if iscell(net.inputs)
-      [nRows, nColumns] = size(net.inputs);
-      # insert enough spaces to put ":" to position 20
-      # insert 2 spaces for distance between ":" and "%"
-      fprintf(fid,"              inputs: {%dx%d cell} of inputs\n",nRows,nColumns);
-    else
-      fprintf(fid,"unsure if this is possible\n");
-    endif
-
-  endif
-
-endfunction
-
-
-
-
-
-
-
-
-
-
--- a/main/nnet/inst/private/__printLW.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,46 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printLW (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printLW(fid,net)
-
-  if isfield(net,"LW")
-    nLayers = 0;
-    # check if it's cell array
-    if iscell(net.LW)
-      [nRows, nColumns] = size(net.LW);
-      for i=1:nRows
-        for k=1:nColumns
-          if !isempty(net.LW{i,k})
-            nLayers = nLayers+1;
-          endif
-        endfor
-      endfor
-      # insert enough spaces to put ":" to position 20
-      # insert 2 spaces for distance between ":" and "%"
-      fprintf(fid,"                  LW: {%dx%d cell} containing %d layer weight matrix\n",nRows,nColumns,nLayers);
-    else
-      fprintf(fid,"unsure if this is possible\n");
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printLayerConnect.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,77 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printLayerConnect (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printLayerConnect(fid,net)
-
-  if isfield(net,"layerConnect")
-    # net.layerConnect can be a matrix..!
-    # check if it's a matrix
-    if isscalar(net.layerConnect)
-      error("unsure if this is possible..")
-    elseif isnumeric(net.layerConnect)
-      if ismatrix(net.layerConnect)
-        if issquare(net.layerConnect)
-          # insert enough spaces to put ":" to position 20
-          # insert 2 spaces for distance between ":" and "%"
-          fprintf(fid,"        layerConnect:  [");
-          [nRows nColumns] = size(net.layerConnect);
-          for k = 1:1:nRows
-            for i = 1:1:nColumns
-              if i<nColumns
-                fprintf(fid,"%d ",net.layerConnect(i*k));
-              else
-                fprintf(fid,"%d",net.layerConnect(i*k));
-              endif
-            endfor
-            if k!=nRows
-              #print ; for newline in matrix
-              fprintf(fid,";");
-            endif
-          endfor
-          # print last bracket
-          fprintf(fid,"] not yet used item\n");
-        elseif isvector(net.layerConnect)
-        # insert enough spaces to put ":" to position 20
-        # insert 2 spaces for distance between ":" and "%"
-        # print bracket for open
-          fprintf(fid,"        layerConnect:  [");
-          [nRows nColumns] = size(net.layerConnect);
-             for k = 1:1:nRows
-               for i = 1:1:nColumns
-                 fprintf(fid,"%d",net.layerConnect(i*k));
-               endfor
-               if k!=nRows
-                 #print ; for newline in matrix
-                 fprintf(fid,";");
-               endif
-             endfor
-             # print last bracket
-             fprintf(fid,"] not yet used item\n");
-           endif  # if issquare..
-         endif #if ismatrix
-      endif
-    else
-      fprintf(fid," ERROR...");
-    endif
-
-endfunction
--- a/main/nnet/inst/private/__printLayerWeights.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,40 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printLayerWeights (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printLayerWeights(fid,net)
-
-  if isfield(net,"layerweights")
-    # check if it's cell array
-    if iscell(net.layerweights)
-      [nRows, nColumns] = size(net.layerweights);
-      # insert enough spaces to put ":" to position 20
-      # insert 2 spaces for distance between ":" and "%"
-      fprintf(fid,"        layerweights: {%dx%d cell} containing xx layer weight\n",nRows,nColumns);
-    else
-      fprintf(fid,"layerweights:unsure if this is possible\n");
-    endif
-  else
-     fprintf(fid,"field layerweights not found & not yet used item\n");
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printLayers.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,38 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printLayers (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printLayers(fid,net)
-
-  if isfield(net,"layers")
-   # check if it's cell array
-    if iscell(net.layers)
-      [nRows, nColumns] = size(net.layers);
-      # insert enough spaces to put ":" to position 20
-      # insert 2 spaces for distance between ":" and "%"
-      fprintf(fid,"              layers: {%dx%d cell} of layers\n",nRows,nColumns);
-    else
-      fprintf(fid,"unsure if this is possible\n");
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printMLPHeader.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,41 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid   <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printMLPHeader (@var{fid})
-## @code{__printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printMLPHeader(fid)
-
-     # one empty row
-     fprintf(fid,"\n");
-     # write "net="
-     fprintf(fid,"net=\n");
-     # next empty row
-     fprintf(fid,"\n");
-     # write "Neural Network object:", insert two spaces..
-     fprintf(fid,"  Neural Network object:\n");
-     # next empty row
-     fprintf(fid,"\n");
-     # write "architecture:", insert two spaces..
-     fprintf(fid,"  architecture:\n");
-     # one more time an empty row
-     fprintf(fid,"\n");
-     
-endfunction
--- a/main/nnet/inst/private/__printNetworkType.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,34 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid   <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printMLPHeader (@var{fid})
-## @code{__printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printNetworkType(fid,net)
-
-  if isfield(net,"networkType")
-    if strcmp(net.networkType,"newff")
-      fprintf(fid,"          Network type:  '%s'\n","Feed forward multi-layer network");
-    else
-      fprintf(fid,"          Network type:  '%s'\n","error: undefined network type");
-    endif
-  endif
-     
-endfunction
--- a/main/nnet/inst/private/__printNumInputDelays.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,40 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printNumInputDelays (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function __printNumInputDelays(fid,net)
-
-     ## now check the structure fields..
-     cellNetFields = fieldnames(net);
-     # search for numInputDelays
-     if isfield(net,"numInputDelays")
-        # test on scalar
-        if isscalar(net.numInputDelays)
-           fprintf(fid,"      numInputDelays:  %d  (read-only)\n",net.numInputDelays);
-        # net.numInputDelays must be an integer... till now, 11-01-2006
-        else
-            error("numInputDelays must be a scalar value!");
-        endif
-     endif
-     
-endfunction
--- a/main/nnet/inst/private/__printNumInputs.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,42 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printNumInputs (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function __printNumInputs(fid,net)
-
-     ## now check the structure fields..
-     cellNetFields = fieldnames(net);
-     # search for numInputs
-     if isfield(net,"numInputs")
-        # test on scalar
-        if isscalar(net.numInputs)
-           # insert enough spaces to put ":" to position 20
-           # insert 2 spaces for distance between ":" and "%"
-           fprintf(fid,"           numInputs:  %d\n",net.numInputs);
-        # net.numInputs must be an integer... till now, 11-01-2006
-        else
-            error("numInputs must be a scalar value!");
-        endif
-     endif
-     
-endfunction
--- a/main/nnet/inst/private/__printNumLayerDelays.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,41 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printNumLayerDelays (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printNumLayerDelays(fid,net)
-
-     ## now check the structure fields..
-     cellNetFields = fieldnames(net);
-     # search for numLayerDelays
-     if isfield(net,"numLayerDelays")
-        # test on scalar
-        if isscalar(net.numLayerDelays)
-          # insert enough spaces to put ":" to position 20
-          # insert 2 spaces for distance between ":" and "%"
-          fprintf(fid,"      numLayerDelays:  %d  (read-only)\n",net.numLayerDelays);
-        # net.numLayerDelays must be an integer... till now, 11-01-2006
-        else
-            error("numLayerDelays must be a scalar value!");
-        endif
-     endif
-     
-endfunction
--- a/main/nnet/inst/private/__printNumLayers.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,37 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printNumLayers (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printNumLayers(fid,net)
-
-     if isfield(net,"numLayers")
-        if isscalar(net.numLayers)
-					 # insert enough spaces to put ":" to position 20
-           # insert 2 spaces for distance between ":" and "%"
-           fprintf(fid,"           numLayers:  %d\n",net.numLayers);
-        # net.numLayers must be an integer... till now, 11-01-2006
-        else
-            error("numLayers must be a scalar value!");
-        endif
-     endif
-     
-endfunction
--- a/main/nnet/inst/private/__printNumOutputs.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,41 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printNumOutputs (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printNumOutputs(fid,net)
-
-     ## now check the structure fields..
-     cellNetFields = fieldnames(net);
-     # search for numOutputs
-     if isfield(net,"numOutputs")
-        # test on scalar
-        if isscalar(net.numOutputs)
-          # insert enough spaces to put ":" to position 20
-          # insert 2 spaces for distance between ":" and "%"
-          fprintf(fid,"          numOutputs:  %d  (read-only)\n",net.numOutputs);
-        # net.numOutputs must be an integer... till now, 11-01-2006
-        else
-            error("numOutputs must be a scalar value!");
-        endif
-     endif
-     
-endfunction
--- a/main/nnet/inst/private/__printNumTargets.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,41 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printNumTargets (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printNumTargets(fid,net)
-
-     ## now check the structure fields..
-     cellNetFields = fieldnames(net);
-     # search for numTargets
-     if isfield(net,"numTargets")
-        # test on scalar
-        if isscalar(net.numTargets)
-          # insert enough spaces to put ":" to position 20
-          # insert 2 spaces for distance between ":" and "%"
-          fprintf(fid,"          numTargets:  %d  (read-only)\n",net.numTargets);
-        # net.numTargets must be an integer... till now, 11-01-2006
-        else
-            error("numTargets must be a scalar value!");
-        endif
-     endif
-     
-endfunction
--- a/main/nnet/inst/private/__printOutputConnect.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,79 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printOutputConnect (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printOutputConnect(fid,net)
-
-  if isfield(net,"outputConnect")
-    # net.outputConnect can be a matrix..!
-    # check if it's a matrix
-    if isscalar(net.outputConnect)
-      error("unsure if this is possible..")
-    elseif isnumeric(net.outputConnect)
-      if ismatrix(net.outputConnect)
-        if issquare(net.outputConnect)
-          fprintf(fid,"       outputConnect:  [");
-          [nRows nColumns] = size(net.outputConnect);
-          for k = 1:1:nRows
-            for i = 1:1:nColumns
-              if i<nColumns
-                fprintf(fid,"%d ",net.outputConnect(i*k));
-              else
-                fprintf(fid,"%d",net.outputConnect(i*k));
-              endif
-            endfor
-            if k!=nRows
-              #print ; for newline in matrix
-              fprintf(fid,";");
-            endif
-          endfor
-          # print last bracket
-          fprintf(fid,"]\n");
-        elseif isvector(net.outputConnect)
-          # insert enough spaces to put ":" to position 20
-          # insert 2 spaces for distance between ":" and "%"
-          # print bracket for open
-          fprintf(fid,"       outputConnect:  [");
-          [nRows nColumns] = size(net.outputConnect);
-             for k = 1:1:nRows
-               for i = 1:1:nColumns
-                 if (i<nColumns)
-                   fprintf(fid,"%d ",net.outputConnect(i*k));
-                 else
-                   fprintf(fid,"%d",net.outputConnect(i*k));
-                 endif
-               endfor
-               if k!=nRows
-                 #print ; for newline in matrix
-                 fprintf(fid,";");
-               endif
-             endfor
-             # print last bracket
-             fprintf(fid,"] not yet used item\n");
-           endif  # if issquare..
-         endif #if ismatrix
-      endif
-    else
-      fprintf(fid," ERROR...");
-    endif
-
-endfunction
--- a/main/nnet/inst/private/__printOutputs.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,43 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printOutputs (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function __printOutputs(fid,net)
-
-  if isfield(net,"outputs")
-    # check if it's cell array
-    if iscell(net.outputs)
-      [nRows, nColumns] = size(net.outputs);
-      # insert enough spaces to put ":" to position 20
-      # insert 2 spaces for distance between ":" and "%"
-      if (net.numOutputs>1)
-        fprintf(fid,"             outputs: {%dx%d cell} containing %d output\n",nRows,nColumns,net.numOutputs);      
-      else
-        fprintf(fid,"             outputs: {%dx%d cell} containing %d output\n",nRows,nColumns,net.numOutputs);
-      endif
-    else
-      fprintf(fid,"unsure if this is possible\n");
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printPerformFcn.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,34 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printPerformFcn (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printPerformFcn(fid,net)
-
-  if isfield(net,"performFcn")
-    if isempty(net.performFcn)
-      fprintf(fid,"          performFcn:  '%s'\n","empty");
-    else
-      fprintf(fid,"          performFcn:  '%s'\n",net.performFcn);
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printPerformParam.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,34 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} printInputs (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid 
-
-function __printPerformParam(fid,net)
-
-  if isfield(net,"performParam")
-    if isempty(net.performParam)
-      fprintf(fid,"        performParam:  '%s'\n","not yet used item");
-    else
-      fprintf(fid,"        performParam:  '%s'\n",net.performParam);
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printTargetConnect.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,82 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printTargetConnect (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printTargetConnect(fid,net)
-
-  if isfield(net,"targetConnect")
-    # net.targetConnect can be a matrix..!
-    # check if it's a matrix
-    if isscalar(net.targetConnect)
-      error("unsure if this is possible..")
-    elseif isnumeric(net.targetConnect)
-      if ismatrix(net.targetConnect)
-        if issquare(net.targetConnect)
-          # insert enough spaces to put ":" to position 20
-          # insert 2 spaces for distance between ":" and "%"
-          fprintf(fid,"       targetConnect:  [");
-          [nRows nColumns] = size(net.targetConnect);
-          for k = 1:1:nRows
-            for i = 1:1:nColumns
-              if i<nColumns
-                fprintf(fid,"%d ",net.targetConnect(i*k));
-              else
-                fprintf(fid,"%d",net.targetConnect(i*k));
-              endif
-            endfor
-            if k!=nRows
-              #print ; for newline in matrix
-              fprintf(fid,";");
-            endif
-          endfor
-          # print last bracket
-          fprintf(fid,"]\n");
-        elseif isvector(net.targetConnect)
-          # insert enough spaces to put ":" to position 20
-          # insert 2 spaces for distance between ":" and "%"
-          # print bracket for open
-          fprintf(fid,"       targetConnect:  [");
-          [nRows nColumns] = size(net.targetConnect);
-             for k = 1:1:nRows
-               for i = 1:1:nColumns
-                 if (i<nColumns)
-                   fprintf(fid,"%d ",net.targetConnect(i*k));
-                 else
-                   fprintf(fid,"%d",net.targetConnect(i*k));
-                 endif
-               endfor
-               if k!=nRows
-                 #print ; for newline in matrix
-                 fprintf(fid,";");
-               endif
-             endfor
-             # print last bracket
-             fprintf(fid,"] not yet used item\n");
-           endif  # if issquare..
-         endif #if ismatrix
-      endif
-    else
-      fprintf(fid," ERROR...");
-    endif
-
-
-endfunction
--- a/main/nnet/inst/private/__printTargets.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,38 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printTargets (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid 
-
-function __printTargets(fid,net)
-
-  if isfield(net,"targets")
-    # check if it's cell array
-    if iscell(net.targets)
-      [nRows, nColumns] = size(net.targets);
-      # insert enough spaces to put ":" to position 20
-      # insert 2 spaces for distance between ":" and "%"
-      fprintf(fid,"             targets: {%dx%d cell} containing %d targets\n",nRows,nColumns,net.numTargets);
-    else
-      fprintf(fid,"unsure if this is possible\n");
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printTrainFcn.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,35 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printTrainFcn (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function __printTrainFcn(fid,net)
-
-  if isfield(net,"trainFcn")
-    if isempty(net.trainFcn)
-      fprintf(fid,"            trainFcn:  '%s'\n","empty");
-    else
-      fprintf(fid,"            trainFcn:  '%s'\n",net.trainFcn);
-    endif
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__printTrainParam.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,59 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} __printTrainParam (@var{fid})
-## @code{printMLPHeader} saves the header of a  neural network structure
-## to a *.txt file with identification @code{fid}.
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function __printTrainParam(fid,net)
-
-  if isfield(net,"trainParam")
-    str2 = "";
-    str3 = "";
-    if isempty(net.trainParam)
-      fprintf(fid,"          trainParam:  '%s'\n","not yet used item");
-    else
-      cellFieldNames = fieldnames(net.trainParam);
-      [nRows, nColumns] = size(cellFieldNames);
-      if (nRows<4)
-      else
-        for iRuns = 1:nRows
-          if (iRuns==1)
-            str1 =  ["." char(cellFieldNames(iRuns,1)) ", "];
-          endif
-          if (iRuns<=4 & iRuns>1)
-            str1 = [str1 "." char(cellFieldNames(iRuns,1)) ", "];
-          endif
-          if (iRuns>4 & iRuns<=8)
-            str2 = [str2 "." char(cellFieldNames(iRuns,1)) ", "];
-          endif
-          if (iRuns>8)
-            str3 = [str3 "." char(cellFieldNames(iRuns,1)) ", "];
-          endif
-        endfor
-        fprintf(fid,"          trainParam:  %s\n",str1);
-        fprintf(fid,"                       %s\n",str2);
-        fprintf(fid,"                       %s\n",str3);
-      endif
-    endif
-  else
-    fprintf(fid,"field trainparam not found\n");
-  endif
-
-endfunction
--- a/main/nnet/inst/private/__randomisecols.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,42 +0,0 @@
-## Copyright (C) 2008 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{retmatrix} = __randomisecols (@var{matrix})
-## @code{__randomisecols} takes a matrix as input argument and changes the order
-## of the columns. The rows aren't affected.
-## @end deftypefn
-
-## Author: mds
-
-function [retmatrix] = __randomisecols(matrix)
-
-  ## check number of inputs
-  error(nargchk(1,1,nargin));
-
-  # get number of cols
-  nCols = size(matrix,2);
-  
-  # now create random column order
-  colOrder = randperm(nCols);
-  
-  # now sort the matrix new
-  retmatrix = matrix(:,[colOrder]);
-
-
-endfunction
-
-%!# no test possible, contains randperm which is using
-%!# some randome functions
--- a/main/nnet/inst/private/__rerangecolumns.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,166 +0,0 @@
-## Copyright (C) 2008 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{retmatrix} = __rerangecolumns (@var{matrix},@var{analyzeMatrix},@var{nTrainSets})
-## @code{__rerangecolumns} reranges the data sets depending on the input arguments.
-## @code{matrix} is the data set matrix containing inputs and outputs (targets) in row order.
-## This means for example: the first three rows are inputs and the fourth row is an output row.
-## The second argument is used in the optimizing algorithm. This matrix contains informations about
-## the description of the rows data of matrix.
-## The third argument is used to be sure, rerange all the columns to the correct position.
-## @end deftypefn
-
-## Author: mds
-
-function retmatrix = __rerangecolumns(matrix,analyzeMatrix,nTrainSets)
-
-  ## check number of inputs
-  error(nargchk(3,3,nargin));
-
-  # set default values
-
-  # now sort "matrix" with help of analyzeMatrix
-  # following conditions must be kept:
-  # a.) rows containing unique values aren't sorted!
-  # b.) sort first rows which contains min AND max values only once
-  # c.) sort secondly rows which contains min OR max values only once
-  # d.) at last, sort binary data if still needed!
-
-  nRows = size(analyzeMatrix,1);   # get number of rows
-
-  ## create i-vector
-  i = 1;
-  iVec = [];
-  while (i <= nRows)
-    if ( (analyzeMatrix(i,3)==1) && (analyzeMatrix(i,4)==1) )
-      iVec = [iVec i];
-    endif
-    i += 1;
-  endwhile
-  i = 1;
-  while (i <= nRows)
-	if ( (analyzeMatrix(i,3)>1) || (analyzeMatrix(i,4)>1) )
-	  iVec = [iVec i];
-	endif
-	i += 1;
-  endwhile
-  i = 1;
-  while (i <= nRows)
-    if (analyzeMatrix(i,1)==1)
-      iVec = [iVec i];
-    endif
-  i += 1;
-  endwhile
-
-
-  ## now do main loop
-  j = 1;
-  i = iVec(j);
-  nRows = length(iVec);
-  while (j < nRows)
-    if (analyzeMatrix(i,2)==1)
-      # easiest case, nothing to do
-    else
-
-      # now let's see if min AND max values are only once in the row
-      if ( (analyzeMatrix(i,3)==1) && (analyzeMatrix(i,4)==1) )
-		# search at which index the min value is
-		minVal = min(matrix(i,:));
-        [rowInd, colInd] = find(matrix(i,:)==minVal);# colInd is searched
-        if (colInd >= nTrainSets ) # move column
-		  matrix = __copycoltopos1(matrix,colInd);
-        endif
-        # search at which index the max value is
-        maxVal = max(matrix(i,:));
-        [rowInd, colInd] = find(matrix(i,:)==maxVal);# colInd is searched
-        if (colInd >= nTrainSets ) # move column
-		  matrix = __copycoltopos1(matrix,colInd);
-        endif
-
-      else
-        
-		# now here, we have to copy the rows, if min OR max values are more than once in a row
-        if ( (analyzeMatrix(i,3)>=1) || (analyzeMatrix(i,4)>=1) )
-
-		  # search at which index the min value is
-		  minVal = min(matrix(i,:));
-          [rowInd, colInd] = find(matrix(i,:)==minVal);# colInd is searched
-          if (colInd(1) >= nTrainSets ) # move column
-		    matrix = __copycoltopos1(matrix,colInd(1));
-          endif
-          
-          # search at which index the max value is
-          maxVal = max(matrix(i,:));
-          [rowInd, colInd] = find(matrix(i,:) == maxVal);# colInd is searched
-          if (colInd(1) >= nTrainSets ) # move column
-		    matrix = __copycoltopos1(matrix,colInd(1));
-          endif
-
-		else
-		  # now sort binary data, if needed
-		  
-          # search at which index the 0-value is
-		  [rowInd, colInd] = find(matrix(i,:)==0);# colInd is searched
-          if (colInd(1) >= nTrainSets ) # move column
-		    matrix = __copycoltopos1(matrix,colInd(1));
-          endif
-          # search at which index the 1-value is
-          [rowInd, colInd] = find(matrix(i,:)==1);# colInd is searched
-          if (colInd(1) >= nTrainSets ) # move column
-		    matrix = __copycoltopos1(matrix,colInd(1));
-          endif
-
-        endif# END OF if ( (analyzeMatrix(i,3)>=1) || (analyzeMatrix(i,4)>=1) )
-
-      endif # END OF if ( (analyzeMatrix(i,3)==1) AND (analyzeMatrix(i,4)==1) )
-
-    endif # END OF if (analyzeMatrix(i,2)==1)
-    j += 1;
-    i = iVec(j);
-  endwhile
-  retmatrix = matrix;
-endfunction
-
-%!shared matrix,analyzeMatrix,nTrainSets, returnmatrix
-%! disp("testing __rerangecolumns")
-%! matrix = [0 1 0 0 0 0 1 0 1 1;  \
-%!			 4 4 4 4 4 4 4 4 4 4;  \
-%!        -1.1 -1.1 2 3 4 3.2 1 8 9 10; \
-%!           0 1.1 3 4 5 2 10 10 2 3; \
-%!          -1 1 1 1 1 2 3 4 1 5];
-%! analyzeMatrix = [1 0 0 0; 0 1 0 0; 0 0 2 1; 0 0 1 2; 0 0 1 1];
-%! nTrainSets = 8;
-%! returnmatrix = __rerangecolumns(matrix,analyzeMatrix,nTrainSets);
-%!assert(returnmatrix(1,1)==1);
-%!assert(returnmatrix(2,1)==4);
-%!assert(returnmatrix(3,1)==1);
-%!assert(returnmatrix(4,1)==10);
-%!assert(returnmatrix(5,1)==3);
-%! matrix = [0 1 0 0 0 0 1 0 1 1; 			\
-%!			 4 4 4 4 4 4 4 4 4 4; 			\
-%!          -1.1 -1.1 2 3 4 3.2 1 8 9 10; 	\
-%!           0 1.1 3 4 5 2 10 10 2 3; 		\
-%!          -1 1 1 1 1 2 3 4 1 5;     		\
-%!			 0 1 2 1 2 1 2 3 4 5;];  # the last row is euqal to the nnet targets
-%! analyzeMatrix = [1 0 0 0; 0 1 0 0; 0 0 2 1; 0 0 1 2; 0 0 1 1];
-%! nTrainSets = 8;
-%! returnmatrix = __rerangecolumns(matrix,analyzeMatrix,nTrainSets);
-%!assert(returnmatrix(1,1)==1);
-%!assert(returnmatrix(2,1)==4);
-%!assert(returnmatrix(3,1)==1);
-%!assert(returnmatrix(4,1)==10);
-%!assert(returnmatrix(5,1)==3);
-%!assert(returnmatrix(6,1)==2);
--- a/main/nnet/inst/private/__setx.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,59 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} @var{net} = __setx (@var{net},@var{X2})
-## @code{__setx} sets the new weights to the neural network structure
-## @end deftypefn
-
-## @seealso{getx}
-
-## Author: Michel D. Schmid
-
-function net = __setx(net,xx)
-
-  ## check number of inputs
-  error(nargchk(2,2,nargin));
-
-  ## check input args
-  ## check "net", must be a net structure
-  if !__checknetstruct(net)
-    error("Structure doesn't seem to be a neural network")
-  endif
-
-  ## inputs
-  [nRows, nColumns] = size(net.IW{1,1});
-  nElementsIW = nRows*nColumns;
-  net.IW{1,1}(:) = xx(1:nElementsIW);
-
-  [nRows, nColumns] = size(net.b{1,1});
-  nElementsB1 = nRows*nColumns;
-  net.b{1,1}(:) = xx(1+nElementsIW:nElementsIW+nElementsB1);
-  start = nElementsIW + nElementsB1;
-
-  ## layers
-  nLayers = net.numLayers;
-  for i = 2:nLayers
-    [nRows, nColumns] = size(net.LW{i,i-1});
-    nElementsLW = nRows*nColumns;
-    net.LW{i,i-1}(:) = xx(1+start:start+nElementsLW);
-
-    [nRows, nColumns] = size(net.b{i,1});
-    nElementsBx = nRows*nColumns;
-    net.b{i,1}(:) = xx(1+start+nElementsLW:start+nElementsLW+nElementsBx);
-    start = start + nElementsLW + nElementsBx;
-  endfor
-
-endfunction
--- a/main/nnet/inst/private/__trainlm.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,359 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{netOut}] = __trainlm (@var{net},@var{mInputN},@var{mOutput},@var{[]},@var{[]},@var{VV})
-## A neural feed-forward network will be trained with @code{__trainlm}
-##
-## @example
-## [netOut,tr,out,E] = __trainlm(net,mInputN,mOutput,[],[],VV);
-## @end example
-## @noindent
-##
-## left side arguments:
-## @example
-## netOut: the trained network of the net structure @code{MLPnet}
-## tr :
-## out:
-## E  : Error
-## @end example
-## @noindent
-##
-## right side arguments:
-## @example
-## net    : the untrained network, created with @code{newff}
-## mInputN: normalized input matrix
-## mOutput: output matrix
-## []     : unused parameter
-## []     : unused parameter
-## VV     : validize structure
-## out:
-## E  : Error
-## @end example
-## @noindent
-##
-##
-## @noindent
-## are equivalent.
-## @end deftypefn
-
-## @seealso{newff,prestd,trastd}
-
-## Author: Michel D. Schmid
-
-## Comments: see in "A neural network toolbox for Octave User's Guide" [4]
-##  for variable naming... there have inputs or targets only one letter,
-## e.g. for inputs is P written. To write a program, this is stupid, you can't
-## search for 1 letter variable... that's why it is written here like Pp, or Tt
-## instead only P or T.
-
-function [net] = __trainlm(net,Im,Pp,Tt,VV)
-
-  ## check range of input arguments
-  error(nargchk(5,5,nargin))
-
-  ## Initialize
-  ##------------
-
-  ## get parameters for training
-  epochs   = net.trainParam.epochs;
-  goal     = net.trainParam.goal;
-  maxFail  = net.trainParam.max_fail;
-  minGrad  = net.trainParam.min_grad;
-  mu       = net.trainParam.mu;
-  muInc    = net.trainParam.mu_inc;
-  muDec    = net.trainParam.mu_dec;
-  muMax    = net.trainParam.mu_max;
-  show     = net.trainParam.show;
-  time     = net.trainParam.time;
-
-  ## parameter checking
-  checkParameter(epochs,goal,maxFail,minGrad,mu,...
-	               muInc,muDec,muMax,show,time);
-
-  ## Constants
-  shortStr = "TRAINLM";    # TODO: shortStr is longer as TRAINLM !!!!!!!!!!!
-  doValidation = !isempty(VV);
-  stop = "";
-
-
-  #startTime = clock(); # TODO: maybe this row can be placed
-                       # some rows later
-
-  ## the weights are used in column vector format
-  xx = __getx(net); # x is the variable with respect to, but no
-                    # variables with only one letter!!
-  ## define identity matrix
-  muI = eye(length(xx));                  
-
-  startTime = clock();  # if the next some tests are OK, I can delete
-                        # startTime = clock(); 9 rows above..
-
-  ## calc performance of the actual net
-  [perf,vE,Aa,Nn] = __calcperf(net,xx,Im,Tt);
-  if (doValidation)
-    ## calc performance if validation is used
-    VV.net = net; # save the actual net in the validate
-    # structure... if no train loop will show better validate
-    # performance, this will be the returned net
-    vperf = __calcperf(net,xx,VV.Im,VV.Tt);
-    VV.perf = vperf;
-    VV.numFail = 0; # one of the stop criterias
-  endif
-
-  nLayers = net.numLayers;
-  for iEpochs = 0:epochs # longest loop & one of the stop criterias
-    ve = vE{nLayers,1};
-    ## calc jacobian
-    ## Jj is jacobian matrix
-    [Jj] = __calcjacobian(net,Im,Nn,Aa,vE);
-
-    ## rerange error vector for jacobi matrix
-    ve = ve(:);
-
-    Jjve = (Jj' * ve); # will be used to calculate the gradient
-
-    normGradX = sqrt(Jjve'*Jjve);
-
-    ## record training progress for later plotting
-    ## if requested
-    trainRec.perf(iEpochs+1) = perf;
-    trainRec.mu(iEpochs+1) = mu;
-    if (doValidation)
-      trainRec.vperf(iEpochs+1) = VV.perf;
-    endif
-
-    ## stoping criteria
-    [stop,currentTime] = stopifnecessary(stop,startTime,perf,goal,...
-                           iEpochs,epochs,time,normGradX,minGrad,mu,muMax,...
-                           doValidation,VV,maxFail);
-
-    ## show train progress
-    showtrainprogress(show,stop,iEpochs,epochs,time,currentTime,...
-		  goal,perf,minGrad,normGradX,shortStr,net);
-
-    ## show performance plot, if needed
-    if !isnan(show) # if no performance plot is needed
-      ## now make it possible to define after how much loops the
-      ## performance plot should be updated
-      if (mod(iEpochs,show)==0)
-        plot(1:length(trainRec.perf),trainRec.perf);
-	if (doValidation)
-	  hold on;
-	  plot(1:length(trainRec.vperf),trainRec.vperf,"--g");
-	endif
-      endif
-    endif # if !(strcmp(show,"NaN"))
-#    legend("Training","Validation");
-
-    ## stop if one of the criterias is reached.
-    if length(stop)
-      if (doValidation)
-        net = VV.net;
-      endif
-      break
-    endif
-
-    ## calculate DeltaX
-    while (mu <= muMax)
-      ## calculate change in x
-      ## see [4], page 12-21
-      dx = -((Jj' * Jj) + (muI*mu)) \ Jjve;
-
-      ## add changes in x to actual x values (xx)
-      x1 = xx + dx;
-      ## now add x1 to a new network to see if performance will be better
-      net1 = __setx(net,x1);
-      ## calc now new performance with the new net
-      [perf1,vE1,Aa1,N1] = __calcperf(net1,x1,Im,Tt);
-
-      if (perf1 < perf)
-        ## this means, net performance with new weight values is better...
-        ## so save the new values
-        xx = x1;
-        net = net1;
-        Nn = N1;
-        Aa = Aa1;
-        vE = vE1;
-        perf = perf1;
-
-        mu = mu * muDec;
-        if (mu < 1e-20)   # 1e-20 is properly the hard coded parameter in MATLAB(TM)
-          mu = 1e-20;
-        endif
-        break
-      endif
-      mu = mu * muInc;
-    endwhile
-
-    ## validate with DeltaX
-    if (doValidation)
-      vperf = __calcperf(net,xx,VV.Im,VV.Tt);
-      if (vperf < VV.perf)
-        VV.perf = vperf;
-    	VV.net = net;
-    	## if actual validation performance is better,
-        ## set numFail to zero again
-    	VV.numFail = 0;
-      elseif (vperf > VV.perf)
-        VV.numFail = VV.numFail + 1;
-      endif
-    endif
-
-  endfor #for iEpochs = 0:epochs
-
-#=======================================================
-#
-# additional functions
-#
-#=======================================================
-  function checkParameter(epochs,goal,maxFail,minGrad,mu,...
-	               muInc, muDec, muMax, show, time)
-    ## Parameter Checking
-
-    ## epochs must be a positive integer
-    if ( !isposint(epochs) )
-      error("Epochs is not a positive integer.")
-    endif
-
-    ## goal can be zero or a positive double
-    if ( (goal<0) || !(isa(goal,"double")) )
-      error("Goal is not zero or a positive real value.")
-    endif
-
-    ## maxFail must be also a positive integer
-    if ( !isposint(maxFail) ) # this will be used, to see if validation can
-      # break the training
-      error("maxFail is not a positive integer.")
-    endif
-
-    if (!isa(minGrad,"double")) || (!isreal(minGrad)) || (!isscalar(minGrad)) || ...
-      (minGrad < 0)
-      error("minGrad is not zero or a positive real value.")
-    end
-
-    ## mu must be a positive real value. this parameter is responsible
-    ## for moving from stepest descent to quasi newton
-    if ((!isa(mu,"double")) || (!isreal(mu)) || (any(size(mu)) != 1) || (mu <= 0))
-      error("mu is not a positive real value.")
-    endif
-
-    ## muDec defines the decrement factor
-    if ((!isa(muDec,"double")) || (!isreal(muDec)) || (any(size(muDec)) != 1) || ...
-  		 (muDec < 0) || (muDec > 1))
-      error("muDec is not a real value between 0 and 1.")
-    endif
-
-    ## muInc defines the increment factor
-    if (~isa(muInc,"double")) || (!isreal(muInc)) || (any(size(muInc)) != 1) || ...
-      (muInc < 1)
-      error("muInc is not a real value greater than 1.")
-    endif
-
-    ## muMax is the upper boundary for the mu value
-    if (!isa(muMax,"double")) || (!isreal(muMax)) || (any(size(muMax)) != 1) || ...
-      (muMax <= 0)
-      error("muMax is not a positive real value.")
-    endif
-
-    ## check for actual mu value
-    if (mu > muMax)
-      error("mu is greater than muMax.")
-    end
-
-    ## check if show is activated
-    if (!isnan(show))
-	  if (!isposint(show))
-        error(["Show is not " "NaN" " or a positive integer."])
-      endif
-    endif
-
-    ## check at last the time argument, must be zero or a positive real value
-    if (!isa(time,"double")) || (!isreal(time)) || (any(size(time)) != 1) || ...
-      (time < 0)
-      error("Time is not zero or a positive real value.")
-    end
-
-  endfunction # parameter checking
-
-#
-# -----------------------------------------------------------------------------
-#
-
-  function showtrainprogress(show,stop,iEpochs,epochs,time,currentTime, ...
-          goal,perf,minGrad,normGradX,shortStr,net)
-
-    ## check number of inputs
-    error(nargchk(12,12,nargin));
-
-    ## show progress
-    if isfinite(show) && (!rem(iEpochs,show) || length(stop))
-      fprintf(shortStr);   # outputs the training algorithm
-      if isfinite(epochs)
-        fprintf(", Epoch %g/%g",iEpochs, epochs);
-      endif
-      if isfinite(time)
-        fprintf(", Time %4.1f%%",currentTime/time*100);   # \todo: Time wird nicht ausgegeben
-      endif
-      if isfinite(goal)
-        fprintf(", %s %g/%g",upper(net.performFcn),perf,goal); # outputs the performance function
-      endif
-      if isfinite(minGrad)
-        fprintf(", Gradient %g/%g",normGradX,minGrad);
-      endif
-      fprintf("\n")
-      if length(stop)
-        fprintf("%s, %s\n\n",shortStr,stop);
-      endif
-      fflush(stdout); # writes output to stdout as soon as output messages are available
-    endif
-  endfunction
-  
-#
-# -----------------------------------------------------------------------------
-#
-
-  function [stop,currentTime] = stopifnecessary(stop,startTime,perf,goal,...
-                        iEpochs,epochs,time,normGradX,minGrad,mu,muMax,...
-						doValidation,VV,maxFail)
-
-    ## check number of inputs
-    error(nargchk(14,14,nargin));
-
-    currentTime = etime(clock(),startTime);
-    if (perf <= goal)
-      stop = "Performance goal met.";
-    elseif (iEpochs == epochs)
-      stop = "Maximum epoch reached, performance goal was not met.";
-    elseif (currentTime > time)
-      stop = "Maximum time elapsed, performance goal was not met.";
-    elseif (normGradX < minGrad)
-      stop = "Minimum gradient reached, performance goal was not met.";
-    elseif (mu > muMax)
-      stop = "Maximum MU reached, performance goal was not met.";
-    elseif (doValidation) 
-	  if (VV.numFail > maxFail)
-        stop = "Validation stop.";
-      endif
-    endif
-  endfunction
-
-# =====================================================================
-#
-# END additional functions
-#
-# =====================================================================
-
-endfunction
--- a/main/nnet/inst/purelin.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,35 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}@var{a}= purelin (@var{n})
-## @code{purelin} is a linear transfer function used
-## by neural networks
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function a = purelin(n)
-
-   a = n;
-
-endfunction
-
-%!assert(purelin(2),2);
-%!assert(purelin(-2),-2);
-%!assert(purelin(0),0);
-
-%!error  # this test must throw an error!
-%! assert(purelin(2),1);
--- a/main/nnet/inst/radbas.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,37 +0,0 @@
-## Copyright (C) 2009 Luca Favatella <slackydeb@gmail.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} radbas (@var{n})
-## Radial basis transfer function.
-##
-## @code{radbas(n) = exp(-n^2)}
-##
-## @end deftypefn
-
-## Author: Luca Favatella <slackydeb@gmail.com>
-## Version: 0.1
-
-function retval = radbas (n)
-
-  if (nargin != 1)
-    print_usage ();
-  else
-    retval = exp (-n^2);
-  endif
-endfunction
-
-
-%!assert (radbas (3), exp (-3^2));
--- a/main/nnet/inst/satlin.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,37 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{a} = satlin (@var{n})
-## A neural feed-forward network will be trained with @code{trainlm}
-##
-## @end deftypefn
-
-## @seealso{purelin,tansig,logsig}
-
-## Author: Michel D. Schmid
-
-
-function a = satlin(n)
-
-  if (n<0)
-    a = 0;
-  elseif (n>=0 && n<=1)
-    a = n;
-  else
-    a = 1; # if n>1
-  endif
-
-endfunction
--- a/main/nnet/inst/satlins.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,37 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{a} = satlins (@var{n})
-## A neural feed-forward network will be trained with @code{trainlm}
-##
-## @end deftypefn
-
-## @seealso{purelin,tansig,logsig,satlin,hardlim,hardlims}
-
-## Author: Michel D. Schmid
-
-
-function a = satlins(n)
-
-  if (n<-1)
-    a = -1;
-  elseif (n>=-1 && n<=1)
-    a = n;
-  else
-    a = 1; # if n>1
-  endif
-
-endfunction
--- a/main/nnet/inst/saveMLPStruct.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,185 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {} saveMLPStruct (@var{net},@var{strFileName})
-## @code{saveStruct} saves a neural network structure to *.txt files
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-function saveMLPStruct(net,strFileName)
-
-  ## the variable net holds the neural network structure..
-  # check if "net" is a structure type
-  if !__checknetstruct(net)
-    error("Structure doesn't seem to be a neural network")
-  endif
-
-  # open the first level file
-  fid1 = fopen(strFileName,"w+t","ieee-le");
-
-  if (fid1 < 0)
-    error ("Can not open %s", strFileName);
-  endif
-
-  ## print header
-#   try            ## wird nicht mehr benötigt..
-    __printMLPHeader(fid1);
-#   catch
-#     ## Add saveMLPStructure directory to the path and try again
-#     addpath ([fileparts(mfilename()),"/saveMLPStructure"]);
-#     __printMLPHeader(fid1);
-#   end_try_catch
-  
-  ## check for field "networkType"
-  __printNetworkType(fid1,net);
-
-  ## check for field "numInputs"
-  __printNumInputs(fid1,net);
-
-  ## check for field "numLayers"
-  __printNumLayers(fid1,net)
-
-  ## check for field "biasConnect"
-  __printBiasConnect(fid1,net)
-
-  ## check for field "inputConnect"
-  __printInputConnect(fid1,net)
-
-  ## check for field "layerConnect"
-  __printLayerConnect(fid1,net)
-
-  ## check for field "outputConnect"
-  __printOutputConnect(fid1,net)
-
-  ## check for field "targetConnect"
-  __printTargetConnect(fid1,net)
-
-  ## print one empty line
-  fprintf(fid1,"\n");
-
-  ## check for numOutputs
-  __printNumOutputs(fid1,net);
-
-  ## check for numTargets
-  __printNumTargets(fid1,net);
-
-  ## check for numInputDelays
-  __printNumInputDelays(fid1,net);
-
-  ## check for numLayerDelays
-  __printNumLayerDelays(fid1,net);
-
-  ## print one empty line
-  fprintf(fid1,"\n");
-
-  ## print subobject structures:
-  fprintf(fid1,"  subobject structures:\n");
-
-  ## print one empty line
-  fprintf(fid1,"\n");
-
-  ## print inputs
-  __printInputs(fid1,net);
-
-  ## print layers
-  __printLayers(fid1,net);
-
-  ## print outputs
-  __printOutputs(fid1,net);
-
-  ## print targets
-  __printTargets(fid1,net);
-
-  ## print biases
-  __printBiases(fid1,net);
-
-  ## print inputweights
-  __printInputWeights(fid1,net);
-
-  ## print layerweights
-  __printLayerWeights(fid1,net);
-
-  ## print one empty line
-  fprintf(fid1,"\n");
-
-  ## print subobject structures:
-  fprintf(fid1,"  functions:\n");
-
-  ## print one empty line
-  fprintf(fid1,"\n");
-
-  ## print adaptFcn
-  __printAdaptFcn(fid1,net);
-
-  ## print initFcn
-  __printInitFcn(fid1,net);
-
-  ## print performFcn
-  __printPerformFcn(fid1,net);
-
-  ## print performFcn
-  __printTrainFcn(fid1,net);
-
-  ## print one empty line
-  fprintf(fid1,"\n");
-
-  ## print subobject structures:
-  fprintf(fid1,"  parameters:\n");
-
-  ## print one empty line
-  fprintf(fid1,"\n");
-
-  ## print adaptParam
-  __printAdaptParam(fid1,net);
-
-  ## print initParam
-  __printInitParam(fid1,net);
-
-  ## print performParam
-  __printPerformParam(fid1,net);
-
-  ## print trainParam
-  __printTrainParam(fid1,net);
-
-  ## print one empty line
-  fprintf(fid1,"\n");
-
-  ## print subobject structures:
-  fprintf(fid1,"  weight & bias values:\n");
-
-  ## print one empty line
-  fprintf(fid1,"\n");
-
-  ## print IW
-  __printIW(fid1,net);
-
-  ## print LW
-  __printLW(fid1,net);
-
-  ## print b
-  __printB(fid1,net);
-
-  ## print one empty line
-  fprintf(fid1,"\n");
-
-  ## print subobject structures:
-  fprintf(fid1,"  other:\n");
-
-
-  fclose(fid1);
-
-endfunction
--- a/main/nnet/inst/sim.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,82 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {@var{netoutput} =} sim (@var{net}, @var{mInput})
-## @code{sim} is usuable to simulate a before defined neural network.
-## @code{net} is created with newff(@dots{}) and @var{mInput} should be the
-## corresponding input data set!
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-## Comments: see in "A neural network toolbox for Octave User's Guide" [4]
-##  for variable naming... there have inputs or targets only one letter,
-## e.g. for inputs is written P. To write a program, this is stupid, you can't
-##  search for 1 letter variable... that's why it is written here like Pp, or Tt
-## instead only P or T.
-
-function [netoutput] = sim(net,mInput)
-
-  ## check range of input arguments
-  error(nargchk(2,2,nargin))
-
-  ## check input args
-  ## check "net", must be a net structure
-  if !__checknetstruct(net)
-    error("Structure doesn't seem to be a neural network")
-  endif
-  ## check "mInput", must have defined size
-  [nRows, nColumns] = size(mInput);
-  if (nRows != net.inputs{1}.size)
-    error(["Simulation input data must have: " num2str(net.inputs{1}.size) " rows."])
-  endif
-
-  ## first get weights...
-  IW = net.IW{1};
-  b1 = net.b{1};
-  b1 = repmat(b1,1,size(mInput,2));
-  nLoops = net.numLayers;
-  for i=1:nLoops
-
-    trf = net.layers{i}.transferFcn;
-    ## calculate the outputs for each layer from input to output
-
-    if i==1
-      Nn{i,1} = IW*mInput + b1;
-    else
-      LWx = net.LW{i,i-1};
-      bx = net.b{i};
-      bx = repmat(bx,1,size(Aa{i-1,1},2));
-      Nn{i,1} = LWx*Aa{i-1,1} + bx;
-    endif
-
-    switch(trf)
-      case "tansig"
-        Aa{i,1} = tansig(Nn{i,1});
-      case "purelin"
-        Aa{i,1} = purelin(Nn{i,1});
-      case "logsig"
-        Aa{i,1} = logsig(Nn{i,1});
-      otherwise
-        error(["sim:Unknown transfer fucntion: " trf "!"]);
-    endswitch
-  endfor
-
-  netoutput = Aa{i,1};
-
-endfunction
-
--- a/main/nnet/inst/subset.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,210 +0,0 @@
-## Copyright (C) 2008 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{mTrain}, @var{mTest}, @var{mVali}] = subset (@var{mData},@var{nTargets},@var{iOpti},@var{fTest},@var{fVali})
-## @code{subset} splits the main data matrix which contains inputs and targets into 2 or 3 subsets
-## depending on the parameters. 
-##
-## The first parameter @var{mData} must be in row order. This means if the network
-## contains three inputs, the matrix must be have 3 rows and x columns to define the
-## data for the inputs. And some more rows for the outputs (targets), e.g. a neural network
-## with three inputs and two outputs must have 5 rows with x columns!
-## The second parameter @var{nTargets} defines the number or rows which contains the target values!
-## The third argument @code{iOpti} is optional and can have three status:
-## 	   0: no optimization
-##     1: will randomise the column order and order the columns containing min and max values to be in the train set
-##     2: will NOT randomise the column order, but order the columns containing min and max values to be in the train set
-##	   default value is @code{1}
-## The fourth argument @code{fTest} is also optional and defines how 
-## much data sets will be in the test set. Default value is @code{1/3}
-## The fifth parameter @code{fTrain} is also optional and defines how
-## much data sets will be in the train set. Default value is @code{1/6}
-## So we have 50% of all data sets which are for training with the default values.
-##
-## @example
-##   [mTrain, mTest] = subset(mData,1)
-##   returns three subsets of the complete matrix
-##   with randomized and optimized columns!
-## @end example
-## @example
-##   [mTrain, mTest] = subset(mData,1,)
-##   returns two subsets
-## @end example
-##
-## @end deftypefn
-
-## Author: Michel D. Schmid
-
-
-function [mTrain, mTest, mVali] = subset(mData,nTargets,iOpti,fTest,fVali)
-
-  ## check range of input arguments
-  error(nargchk(2,5,nargin))
-  
-  ## check the input arguments ...!
-  if (nTargets==0)
-    error("No TARGETS defined! This doesn't make any sense for feed-forward neural networks! Please define at least one row of targets")
-  endif
-
-  ## set default values
-  if (nargin==2)
-    iOpti = 1;
-    fTest = 1/3;
-    fVali = 1/6;
-  elseif (nargin==3)
-    fTest = 1/3;
-    fVali = 1/6;
-  elseif (nargin==4)
-    ## if fTest is set and nothing is set
-    ## for fVali I assume that fVali is not used!
-    fVali = 0;
-  endif
-  
-  ## calculate the number of train, test and validation sets
-  fTrain = 1-fTest-fVali;
-  nTrainSets = floor(size(mData,2)*fTrain);
-  diffRestSets = size(mData,2)-nTrainSets;
-  nTestSets = floor(size(mData,2)*fTest);
-  nValiSets = size(mData,2)-nTrainSets-nTestSets;
-
-
-  ## now let's see if matrix must be optimized!
-  bOptiAgain = 1;
-  while (bOptiAgain)
-    if (iOpti == 1)
-    # check that only one optimizing run is enough!!
-    # maybe it's necessary to do it twice ..!
-    # check that all min and max values are in the train set ...!
-      mData = __optimizedatasets(mData,nTrainSets,nTargets,iOpti);
-      mTrain = mData(:,1:nTrainSets);
-      iRuns = size(mTrain,1);
-      i = 1;
-      j = 1;
-      while (i < iRuns)
-    	  if ( max(mTrain(i,:)) == max(mData(i,:)) )
-    	    j += 1;
-    	  endif
-    	  i +=1;
-      endwhile
-      if (i==j)
-        bOptiAgain = 0;
-      endif
-    elseif (iOpti == 2)
-      # check that only one optimizing run is enough!!
-      # maybe it's necessary to do it twice ..!
-      # check that all min and max values are in the train set ...!
-      mData = __optimizedatasets(mData,nTrainSets,nTargets,iOpti);
-      mTrain = mData(:,1:nTrainSets);
-      iRuns = size(mTrain,1);
-      j = 1;
-      i = 1;
-      while (i < iRuns)
-    	  if (max(mTrain(i,:))==max(mData(i,:)))
-			j += 1;
-    	  endif
-    	  i += 1;
-      endwhile
-      if (i==j)
-        bOptiAgain = 0;
-      endif
-    else
-      ## in this case, iOpti must be 0 ==> nothing todo
-      bOptiAgain = 0;
-    endif
-  endwhile #END OF while(bOptiAgain)
-
-  ## now split up 
-  if (nargout==1)
-    mTrain = mData;
-  elseif (nargout==2);
-    mTrain = mData(:,1:nTrainSets);
-    mTest = mData(:,nTrainSets+1:nTrainSets+nTestSets);
-  elseif (nargout==3)
-    mTrain = mData(:,1:nTrainSets);
-    mTest = mData(:,nTrainSets+1:nTrainSets+nTestSets);
-    mVali = mData(:,nTrainSets+nTestSets+1:end);
-  endif
-
-endfunction
-
-%!shared matrix, nTargets, mTrain, mTest, mVali
-%! disp("testing subset")
-%! matrix = [1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 18 20; \
-%!			 0 2 4 1 3 5 3 4 1 -1 -2 -9 -1 10 12 20 11 11 11 11; \
-%!			-2 2 2 2 2 0 0 0 0  0 10 12 13 12 13 44 33 32 98 11; \
-%!			 0 0 0 0 1 1 1 1 0  0  1  1  1  0  0  1  1  1  0  0; \
-%!           4 4 4 4 4 4 4 4 4  4  4  4  4  4  4  4  4  4  4  4; \
-%!           1 2 3 4 5 6 7 8 9 10 11 12 13 33 44 55 66 77 88 99];
-%! nTargets = 1; # the last row is equivalent to the target values.
-%! [mTrain, mTest, mVali] = subset(matrix,nTargets);  ############################
-%!assert(size(mTrain,2)==10);# 50% of 20
-%!assert(size(mTest,2)==6);# 1/3 of 20 = 6 (floor)
-%!assert(size(mVali,2)==4);# 1/6 of 20 = 4 (floor)
-%! # It's not possible to test the column order with this call!
-%! # randomizing is used! But all max and min values should be
-%! # in the training set
-%!assert(max(mTrain(1,:))==max(matrix(1,:)));
-%!assert(min(mTrain(1,:))==min(matrix(1,:)));
-%!assert(max(mTrain(2,:))==max(matrix(2,:)));
-%!assert(min(mTrain(2,:))==min(matrix(2,:)));
-%!assert(max(mTrain(3,:))==max(matrix(3,:)));
-%!assert(min(mTrain(3,:))==min(matrix(3,:)));
-%!assert(max(mTrain(4,:))==max(matrix(4,:)));
-%!assert(min(mTrain(4,:))==min(matrix(4,:)));
-%!
-%!
-%! [mTrain, mTest, mVali] = subset(matrix,nTargets,0);  ############################
-%!assert(size(mTrain,2)==10);# 50% of 20
-%!assert(size(mTest,2)==6);# 1/3 of 20 = 6 (floor)
-%!assert(size(mVali,2)==4);# 1/6 of 20 = 4 (floor)
-%!assert(mTrain==matrix(:,1:10));
-%!assert(mTest==matrix(:,11:16));
-%!assert(mVali==matrix(:,17:20));
-%!
-%!
-%! [mTrain, mTest, mVali] = subset(matrix,nTargets,2);  ############################
-%!assert(size(mTrain,2)==10);# 50% of 20
-%!assert(size(mTest,2)==6);# 1/3 of 20 = 6 (floor)
-%!assert(size(mVali,2)==4);# 1/6 of 20 = 4 (floor)
-%!assert(max(mTrain(1,:))==max(matrix(1,:)));
-%!assert(min(mTrain(1,:))==min(matrix(1,:)));
-%!assert(max(mTrain(2,:))==max(matrix(2,:)));
-%!assert(min(mTrain(2,:))==min(matrix(2,:)));
-%!assert(max(mTrain(3,:))==max(matrix(3,:)));
-%!assert(min(mTrain(3,:))==min(matrix(3,:)));
-%!assert(max(mTrain(4,:))==max(matrix(4,:)));
-%!assert(min(mTrain(4,:))==min(matrix(4,:)));
-%!
-%!
-%! ## next test ... optimize twice
-%! matrix = [1 2 3 4 5 6 7 20 8 10 11 12 13 14 15 16 17 18 18 9; \
-%!			 0 2 4 1 3 5 3 4 1 -1 -2 -9 -1 10 12 20 11 11 11 11; \
-%!			-2 2 2 2 2 0 0 0 0  0 10 12 13 12 13 44 33 32 98 11; \
-%!			 0 0 0 0 1 1 1 1 0  0  1  1  1  0  0  1  1  1  0  0; \
-%!           4 4 4 4 4 4 4 4 4  4  4  4  4  4  4  4  4  4  4  4; \
-%!           1 2 3 4 5 6 7 8 9 10 11 12 13 33 44 55 66 77 88 99];
-%! [mTrain, mTest, mVali] = subset(matrix,nTargets,2);  ############################
-%!assert(max(mTrain(1,:))==max(matrix(1,:)));
-%!assert(min(mTrain(1,:))==min(matrix(1,:)));
-%!assert(max(mTrain(2,:))==max(matrix(2,:)));
-%!assert(min(mTrain(2,:))==min(matrix(2,:)));
-%!assert(max(mTrain(3,:))==max(matrix(3,:)));
-%!assert(min(mTrain(3,:))==min(matrix(3,:)));
-%!assert(max(mTrain(4,:))==max(matrix(4,:)));
-%!assert(min(mTrain(4,:))==min(matrix(4,:)));
-
-## \todo, a lot of tests to be sure, everything is working OK!!
-## all combinations of arguments must be testet!
--- a/main/nnet/inst/tansig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,39 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}@var{a} = tansig (@var{n})
-## @code{tansig} is a non-linear transfer function used to train
-## neural networks.
-## This function can be used in newff(...) to create a new feed forward
-## multi-layer neural network.
-##
-## @end deftypefn
-
-## @seealso{purelin, logsig}
-
-## Author: Michel D. Schmid
-
-
-function a = tansig(n)
-
-  ## see MATLAB(TM) online help
-  a = 2 ./ (1 + exp(-2*n)) - 1;
-  ## attention with critical values ==> infinite values
-  ## must be set to 1
-  i = find(!finite(a));
-  a(i) = sign(n(i));
-
-endfunction
--- a/main/nnet/inst/train.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,215 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}[@var{net}] = train (@var{MLPnet},@var{mInputN},@var{mOutput},@var{[]},@var{[]},@var{VV})
-## A neural feed-forward network will be trained with @code{train}
-##
-## @example
-## [net,tr,out,E] = train(MLPnet,mInputN,mOutput,[],[],VV);
-## @end example
-## @noindent
-##
-## @example
-## left side arguments:
-##   net: the trained network of the net structure @code{MLPnet}
-## @end example
-## @noindent
-##
-## @example
-## right side arguments:
-##   MLPnet : the untrained network, created with @code{newff}
-##   mInputN: normalized input matrix
-##   mOutput: output matrix (normalized or not)
-##   []     : unused parameter
-##   []     : unused parameter
-##   VV     : validize structure
-## @end example
-## @end deftypefn
-
-## @seealso{newff,prestd,trastd}
-
-## Author: Michel D. Schmid
-
-## Comments: see in "A neural network toolbox for Octave User's Guide" [4]
-## for variable naming... there have inputs or targets only one letter,
-## e.g. for inputs is P written. To write a program, this is stupid, you can't
-## search for 1 letter variable... that's why it is written here like Pp, or Tt
-## instead only P or T.
-
-function [net] = train(net,Pp,Tt,notUsed1,notUsed2,VV)
-
-  ## check range of input arguments
-  error(nargchk(3,6,nargin))
-
-  ## set defaults
-  doValidation = 0;
-  if nargin==6
-    # doValidation=1;
-    ## check if VV is in MATLAB(TM) notation
-    [VV, doValidation] = checkVV(VV);
-  endif
-
-  ## check input args
-  checkInputArgs(net,Pp,Tt)
-
-  ## nargin ...
-  switch(nargin)
-  case 3
-    [Pp,Tt] = trainArgs(net,Pp,Tt);
-    VV = [];
-  case 6
-    [Pp,Tt] = trainArgs(net,Pp,Tt);
-    if isempty(VV)
-      VV = [];
-    else
-      if !isfield(VV,"Pp")
-        error("VV.Pp must be defined or VV must be [].")
-      endif
-      if !isfield(VV,"Tt")
-        error("VV.Tt must be defined or VV must be [].")
-      endif
-      [VV.Pp,VV.Tt] = trainArgs(net,VV.Pp,VV.Tt);
-    endif
-  otherwise
-    error("train: impossible code execution in switch(nargin)")
-  endswitch
-
-
-  ## so now, let's start training the network
-  ##===========================================
-
-  ## first let's check if a train function is defined ...
-  if isempty(net.trainFcn)
-    error("train:net.trainFcn not defined")
-  endif
-
-  ## calculate input matrix Im
-  [nRowsInputs, nColumnsInputs] = size(Pp);
-  Im = ones(nRowsInputs,nColumnsInputs).*Pp{1,1};
-
-  if (doValidation)
-    [nRowsVal, nColumnsVal] = size(VV.Pp);
-    VV.Im = ones(nRowsVal,nColumnsVal).*VV.Pp{1,1};
-  endif
-
-  ## make it MATLAB(TM) compatible
-  nLayers = net.numLayers;
-  Tt{nLayers,1} = Tt{1,1};
-  Tt{1,1} = [];
-  if (!isempty(VV))
-    VV.Tt{nLayers,1} = VV.Tt{1,1};
-    VV.Tt{1,1} = [];
-  endif
-
-  ## which training algorithm should be used
-  switch(net.trainFcn)
-    case "trainlm"
-      if !strcmp(net.performFcn,"mse")
-        error("Levenberg-Marquardt algorithm is defined with the MSE performance function, so please set MSE in NEWFF!")
-      endif
-      net = __trainlm(net,Im,Pp,Tt,VV);
-    otherwise
-      error("train algorithm argument is not valid!")
-  endswitch
-
-
-# =======================================================
-#
-# additional check functions...
-#
-# =======================================================
-
-  function checkInputArgs(net,Pp,Tt)
-      
-    ## check "net", must be a net structure
-    if !__checknetstruct(net)
-      error("Structure doesn't seem to be a neural network!")
-    endif
-
-    ## check Pp (inputs)
-    nInputSize = net.inputs{1}.size; #only one exists
-    [nRowsPp, nColumnsPp] = size(Pp);
-    if ( (nColumnsPp>0) )
-      if ( nInputSize==nRowsPp )
-      # seems to be everything i.o.
-      else
-        error("Number of rows must be the same, like in net.inputs.size defined!")
-      endif
-    else
-      error("At least one column must exist")
-    endif
-    
-    ## check Tt (targets)
-    [nRowsTt, nColumnsTt] = size(Tt);
-    if ( (nRowsTt | nColumnsTt)==0 )
-      error("No targets defined!")
-    elseif ( nColumnsTt!=nColumnsPp )
-      error("Inputs and targets must have the same number of data sets (columns).")
-    elseif ( net.layers{net.numLayers}.size!=nRowsTt )
-      error("Defined number of output neurons are not identically to targets data sets!")
-    endif
-
-  endfunction
-# -------------------------------------------------------
-  function [Pp,Tt] = trainArgs(net,Pp,Tt);
-
-    ## check number of arguments
-    error(nargchk(3,3,nargin));
-
-    [PpRows, PpColumns] = size(Pp);
-    Pp = mat2cell(Pp,PpRows,PpColumns);    # mat2cell is the reason
-    									   # why octave-2.9.5 doesn't work
-										   # octave-2.9.x with x>=6 should be
-										   # ok
-    [TtRows, TtColumns] = size(Tt);
-    Tt = mat2cell(Tt,TtRows,TtColumns);
-
-  endfunction
-
-# -------------------------------------------------------
-
-  function [VV, doValidation] = checkVV(VV)
-
-    ## check number of arguments
-    error(nargchk(1,1,nargin));
-
-	if (isempty(VV))	
-	  doValidation = 0;	
-	else
-	  doValidation = 1;
-      ## check if MATLAB(TM) naming convention is used
-      if isfield(VV,"P")
-        VV.Pp = VV.P;
-        VV.P = [];
-      elseif !isfield(VV,"Pp")
-        error("VV is defined but inside exist no VV.P or VV.Pp")
-      endif
-
-      if isfield(VV,"T")
-        VV.Tt = VV.T;
-        VV.T = [];
-      elseif !isfield(VV,"Tt")
-        error("VV is defined but inside exist no VV.TP or VV.Tt")
-      endif
-	
-	endif
-
-
-  endfunction
-
-# ============================================================
-
-endfunction
--- a/main/nnet/inst/trastd.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,89 +0,0 @@
-## Copyright (C) 2005 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {}@var{pn} = trastd (@var{p},@var{meanp},@var{stdp})
-## @code{trastd} preprocess additional data for neural network simulation.
-##
-## @example
-##   @code{p}    : test input data
-##   @code{meanp}: vector with standardization parameters of prestd(...)
-##   @code{stdp} : vector with standardization parameters of prestd(...)
-##
-##   meanp = [2.5; 6.5];
-##   stdp = [1.2910; 1.2910];
-##   p = [1 4; 2 5];
-##
-##   pn = trastd(p,meanp,stdp);
-## @end example
-## @noindent
-## @end deftypefn
-
-## @seealso{prestd, poststd}
-
-## Author: Michel D. Schmid
-
-function [Pn] = trastd(Pp,meanp,stdp)
-
-  ## check number of inputs
-  error(nargchk(3,3,nargin));
-
-  
-  [nRows,nColumns]=size(Pp);
-  rowOnes = ones(1,nColumns);
-  
-  ## now set all standard deviations which are zero to 1
-  [nRowsII, nColumnsII] = size(stdp);
-  rowZeros = zeros(nRowsII,1);
-  findZeros = find(stdp==0);
-  rowZeros(findZeros)=1;
-  equal = rowZeros;
-  nequal = !equal;
-  if ( sum(equal) != 0 )
-    warning("Some standard deviations are zero. Those inputs won't be transformed.");
-    meanp = meanp.*nequal;
-    stdp = stdp.*nequal + 1*equal;
-  end
-
-  Pn = (Pp-meanp*rowOnes)./(stdp*rowOnes);
-
-endfunction
-
-##
-## >> mInput = [1 2 3 4; 5 6 7 8]
-##
-## mInput =
-##
-##      1     2     3     4
-##      5     6     7     8
-##
-## >> [pn,meanp,stdp] = prestd(mInput)
-##
-## pn =
-##
-##    -1.1619   -0.3873    0.3873    1.1619
-##    -1.1619   -0.3873    0.3873    1.1619
-##
-##
-## meanp =
-##
-##     2.5000
-##     6.5000
-##
-##
-## stdp =
-##
-##     1.2910
-##     1.2910
--- a/main/nnet/inst/vec2ind.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,42 +0,0 @@
-## Copyright (C) 2009 Luiz Angelo Daros de Luca <luizluca@gmail.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## -*- texinfo -*-
-## @deftypefn {Function File} {@var{ind}} = vec2ind (@var{vector})
-## @code{vec2ind} convert vectors to indices
-##
-##
-## @example
-## EXAMPLE 1
-## vec = [1 2 3; 4 5 6; 7 8 9];
-##
-## ind = vec2ind(vec)
-## The prompt output will be:
-## ans = 
-##    1 2 3 1 2 3 1 2 3
-## @end example
-##
-## @end deftypefn
-
-## @seealso{ind2vec}
-
-
-function [ind]=vec2ind(vector)
-  # Convert vectors to indices
-  #
-  #
-  [ind,col,vals] = find(vector);
-  ind=ind';
-endfunction
--- a/main/nnet/tests/MLP/MLPScripts.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,55 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## This file is used to call all MLP tests, but without
-## assert and so on, means if no error occure, everything
-## seems to be ok.
-## This is only a very basic test which should be run
-## from this directory
-
-tic
-cd example1;
-
-
-mlp9_1_1_tansig
-mlp9_2_1_tansig
-mlp9_2_2_1_tansig
-mlp9_2_2_tansig
-mlp9_2_3_tansig
-mlp9_5_3_tansig
-
-cd ..
-
-
-cd example2
-
-mlp9_1_1_logsig
-mlp9_2_1_logsig
-mlp9_2_2_1_logsig
-mlp9_2_2_logsig
-mlp9_2_3_logsig
-mlp9_5_3_logsig
-
-cd ..
-elapsed_time = toc;
-
-disp("Running 12 very basic tests successfully!")
-disp("Secondes needed to running all the tests: ");
-disp(elapsed_time);
-
-
--- a/main/nnet/tests/MLP/example1/mData.txt	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,94 +0,0 @@
-# Created by Octave 2.9.5, Wed May 24 10:33:36 2006     <sim@TL3124>
-# name: mData
-# type: matrix
-# rows: 89
-# columns: 13
- 306 286 2 0 12 61 2 0 3 28 4 0 2
- 368 188 1 0 6 49 0 0 3 37 0 0 1
- 511 73 0 0 40 21 0 0 16 22 0 0 1
- 418 43 0 0 34 30 1 0 9 27 5 0 1
- 299 173 1 0 8 63 1 0 1 37 3 0 1
- 312 253 0 0 2 63 2 0 0 35 3 0 1
- 492 98 0 0 7 23 0 0 13 36 0 0 2
- 506 64 0 0 32 23 0 0 13 29 0 0 3
- 476 41 0 0 32 5 0 0 19 26 0 0 3
- 483 66 0 0 17 16 0 0 10 28 2 0 3
- 429 44 0 0 37 19 0 0 23 19 0 0 1
- 521 137 0 0 16 17 0 0 7 24 0 0 2
- 340 163 1 0 16 72 3 0 3 31 1 0 1
- 323 177 0 0 8 68 4 0 0 37 2 0 1
- 344 240 2 0 4 43 1 0 4 39 6 0 2
- 459 22 0 0 46 14 0 0 27 17 0 0 1
- 487 36 0 0 34 10 0 0 19 27 0 0 3
- 331 169 0 0 19 66 1 0 0 34 1 0 1
- 541 269 1 0 12 5 0 0 6 20 0 0 3
- 475 23 0 0 37 5 0 0 26 14 0 0 3
- 475 186 1 0 15 28 0 0 5 35 0 0 2
- 496 319 0 0 2 4 0 0 2 14 0 0 3
- 525 41 0 0 38 9 0 0 37 18 0 0 3
- 484 37 0 0 50 13 0 0 29 18 0 0 2
- 511 55 0 0 32 15 0 0 23 21 0 0 2
- 515 44 0 0 29 16 0 0 19 31 0 0 2
- 478 101 0 0 15 34 0 0 11 40 0 0 2
- 429 433 3 0 8 11 0 0 5 15 0 0 3
- 471 8 0 0 58 2 0 0 25 13 0 0 1
- 303 269 3 0 10 66 2 0 1 32 7 0 2
- 445 74 0 0 19 30 0 0 4 43 0 0 1
- 488 83 0 0 34 18 0 0 15 34 0 0 3
- 264 298 4 0 7 68 10 0 1 30 7 0 2
- 489 44 0 0 21 12 0 0 29 21 0 0 1
- 475 34 0 0 38 13 0 0 28 18 0 0 3
- 492 14 0 0 44 4 0 0 40 7 0 0 1
- 454 173 1 0 14 21 0 0 2 42 0 0 2
- 306 285 5 0 5 68 3 0 1 33 1 0 2
- 508 64 0 0 31 14 0 0 17 33 0 0 2
- 477 61 0 0 36 13 0 0 22 15 0 0 2
- 349 224 1 0 3 66 3 0 2 41 1 0 1
- 447 189 0 0 11 34 0 0 5 45 0 0 2
- 496 34 0 0 49 6 0 0 31 11 0 0 3
- 484 222 0 0 16 12 0 0 9 24 0 0 2
- 412 50 0 0 25 47 0 0 11 29 0 0 1
- 316 184 0 0 13 65 3 0 5 34 3 0 1
- 345 163 1 0 17 57 2 0 2 32 1 0 1
- 285 273 2 0 7 50 13 0 1 28 10 0 1
- 317 179 0 0 11 64 0 0 4 35 0 0 1
- 500 68 0 0 23 15 0 0 18 30 0 0 1
- 495 34 0 0 39 4 0 0 21 21 0 0 3
- 387 294 1 0 10 37 0 0 4 30 5 0 2
- 258 236 2 0 2 72 7 0 0 26 8 0 1
- 423 25 0 0 26 16 0 0 16 29 0 0 3
- 501 32 0 0 40 11 0 0 24 20 0 0 3
- 459 37 0 0 46 4 0 0 32 16 0 0 3
- 511 48 0 0 35 7 0 0 27 15 0 0 2
- 295 271 3 0 6 71 5 0 3 29 4 0 1
- 502 34 0 0 25 9 0 0 23 11 0 0 3
- 458 36 0 0 12 7 0 0 24 19 0 0 2
- 470 273 1 0 9 17 0 0 2 32 0 0 3
- 477 30 0 0 24 14 0 0 18 26 0 0 3
- 406 77 1 0 15 24 2 0 6 37 1 0 1
- 291 251 2 0 8 53 2 0 0 43 2 0 1
- 407 51 0 0 28 47 0 0 5 35 0 0 1
- 390 47 0 0 17 45 1 0 3 33 2 0 1
- 347 123 1 0 7 52 7 0 4 38 3 0 1
- 300 175 0 0 8 71 2 0 0 35 0 0 1
- 473 59 0 0 44 15 0 0 22 23 0 0 1
- 487 35 0 0 34 12 0 0 29 24 0 0 3
- 532 57 0 0 19 6 0 0 28 15 0 0 2
- 286 207 1 0 5 74 2 0 1 38 4 0 1
- 405 199 0 0 13 37 0 0 6 45 0 0 2
- 337 177 0 0 11 47 1 0 4 39 3 0 1
- 527 20 0 0 32 6 0 0 23 16 0 0 1
- 326 218 1 0 4 71 2 0 1 33 1 0 1
- 486 14 0 0 41 4 0 0 40 8 0 0 1
- 420 43 0 0 26 30 9 0 9 24 4 0 1
- 286 293 3 0 7 83 9 0 1 22 5 0 1
- 395 59 0 0 30 33 0 0 4 36 0 0 1
- 506 24 0 0 34 12 0 0 26 17 0 0 3
- 396 217 0 0 16 43 2 0 6 30 0 0 2
- 457 15 0 0 44 1 0 0 30 13 0 0 1
- 472 139 0 0 16 25 0 0 11 38 0 0 2
- 493 21 0 0 33 16 0 0 22 19 0 0 3
- 311 236 0 0 6 66 4 0 2 24 9 0 1
- 490 23 0 0 34 6 0 0 30 14 0 0 1
- 485 29 0 0 33 7 0 0 14 20 0 0 3
- 481 43 0 0 38 11 0 0 21 24 0 0 2
--- a/main/nnet/tests/MLP/example1/mlp9_1_1_tansig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,5 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 2.27367/0, Gradient 54.9248/1e-10
-TRAINLM, Epoch 8/100, MSE 0.570748/0, Gradient 0.0154084/1e-10
-TRAINLM, Validation stop.
-
-1.8392 1.8391 1.5589 1.5589 1.5589 1.5589 1.5589 1.8392 1.8392 1.8392 1.8392 1.8392 1.5589 1.8392 1.8392 1.5589 1.8392 1.8392 1.5589 1.8384 1.8392 1.5589 1.5589 1.8392 1.5589 1.5589 1.5589 1.8392 1.5589 1.5589
\ No newline at end of file
--- a/main/nnet/tests/MLP/example1/mlp9_1_1_tansig.dat_orig	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,16 +0,0 @@
->> TRAINLM, Epoch 0/100, MSE 2.27367/0, Gradient 54.9248/1e-010
-TRAINLM, Epoch 8/100, MSE 0.570748/0, Gradient 0.0154084/1e-010
-TRAINLM, Validation stop.
-
-
-simOut =
-
-  Columns 1 through 17 
-
-    1.8392    1.8391    1.5589    1.5589    1.5589    1.5589    1.5589    1.8392    1.8392    1.8392    1.8392    1.8392    1.5589    1.8392    1.8392    1.5589    1.8392
-
-  Columns 18 through 30 
-
-    1.8392    1.5589    1.8384    1.8392    1.5589    1.5589    1.8392    1.5589    1.5589    1.5589    1.8392    1.5589    1.5589
-
->> 
\ No newline at end of file
--- a/main/nnet/tests/MLP/example1/mlp9_1_1_tansig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,103 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-##
-## This is a test to train a 9-1-1 MLP (was a real project).
-##
-
-## load data
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## this neural network example isn't a mirror of "the real life work"
-## it will be used to compare the results with MATLAB(TM) results.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-
-# now split the data matrix in 3 pieces,
-# train data, test data and validate data
-# the proportion should be about 1/2 train,
-# 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights,
-# for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-# now divide mInput & mOutput in the three sets
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = []; # delete validation set
-mOutput(:,1:nValiSets) = []; # delete validation set
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = []; # delete test set
-mOutput(:,1:nTestSets) = []; # delete test set
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); # input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 1;
-nOutputNeurons = 1;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"tansig","purelin"},"trainlm","learngdm","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(:) = 1.5;
-MLPnet.LW{2,1}(:) = 0.5;
-MLPnet.b{1,1}(:) = 1.5;
-MLPnet.b{2,1}(:) = 0.5;
-
-## define validation data new, for MATLAB(TM) compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-## now train the network structure
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-
-## make preparations for net test and test MLPnet
- # standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-# will output the network results
-[simOut] = sim(net,mTestInputN);
--- a/main/nnet/tests/MLP/example1/mlp9_2_1_tansig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,5 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 2.5516/0, Gradient 69.8821/1e-010
-TRAINLM, Epoch 16/100, MSE 0.45224/0, Gradient 0.0212947/1e-010
-TRAINLM, Validation stop.
-
-2.0386 2.0386 1.3267 2.2917 2.0386 2.0386 2.2917 2.0386 2.0386 2.0386 2.0386 1.3592 2.2917 2.0386 1.3267 1.3267 2.0386 1.3267 1.3405 2.0386 2.0386 1.3267 1.5778 2.0386 2.0386 1.3267 1.3292 2.0386 2.2917 1.3267
\ No newline at end of file
--- a/main/nnet/tests/MLP/example1/mlp9_2_1_tansig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,102 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## load data
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## this neural network example isn't a mirror of "the real life work"
-## it will be used to compare the results with MATLAB(TM) results.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-
-# now split the data matrix in 3 pieces,
-# train data, test data and validate data
-# the proportion should be about 1/2 train,
-# 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights,
-# for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-# now divide mInput & mOutput in the three sets
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = []; # delete validation set
-mOutput(:,1:nValiSets) = []; # delete validation set
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = []; # delete test set
-mOutput(:,1:nTestSets) = []; # delete test set
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); # input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 2;
-nOutputNeurons = 1;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"tansig","purelin"},"trainlm","learngdm","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(1,:) = 0.5;
-MLPnet.IW{1,1}(2,:) = 1.5;
-MLPnet.LW{2,1}(:) = 0.5;
-MLPnet.b{1,1}(1,:) = 0.5;
-MLPnet.b{1,1}(2,:) = 1.5;
-MLPnet.b{2,1}(:) = 0.5;
-
-## define validation data new, for MATLAB(TM) compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-## now train the network structure
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-
-## make preparations for net test and test MLPnet
- # standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-# will output the network results
-[simOut] = sim(net,mTestInputN);
--- a/main/nnet/tests/MLP/example1/mlp9_2_2_1_tansig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,5 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 2.78009/0, Gradient 79.2812/1e-010
-TRAINLM, Epoch 6/100, MSE 0.202812/0, Gradient 2.34665/1e-010
-TRAINLM, Validation stop.
-
-1.4051 0.5235 1.3323 1.4051 1.4051 1.4051 1.4051 1.4051 0.1939 1.0805 1.4051 0.9002 1.4051 1.3975 1.1845 1.3363 0.5385 1.4051 1.3051 1.4051 1.4051 1.4051 1.3527 1.4021 1.4051 0.6403 1.4051 1.4029 1.4051 1.4051
\ No newline at end of file
--- a/main/nnet/tests/MLP/example1/mlp9_2_2_1_tansig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,101 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## load data
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## this neural network example isn't a mirror of "the real life work"
-## it will be used to compare the results with MATLAB(TM) results.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-
-
-# now split the data matrix in 3 pieces,
-# train data, test data and validate data
-# the proportion should be about 1/2 train,
-# 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights,
-# for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-# now divide mInput & mOutput in the three sets
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = []; # delete validation set
-mOutput(:,1:nValiSets) = []; # delete validation set
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = []; # delete test set
-mOutput(:,1:nTestSets) = []; # delete test set
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); # input matrix with (R x 2)...
-
-%% define network
-nHiddenNeurons = [2 2];
-nOutputNeurons = 1;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{'tansig','tansig','tansig','purelin'},'trainlm','learngdm','mse');
-%% for test purpose, define weights by hand
-# MLPnet.IW{1,1}(:) = 1.5;
-# MLPnet.LW{2,1}(:) = 0.5;
-# MLPnet.b{1,1}(:) = 1.5;
-# MLPnet.b{2,1}(:) = 0.5;
-
-## define validation data new, for MATLAB(TM) compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-## now train the network structure
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-
-## make preparations for net test and test MLPnet
- # standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-# will output the network results
-[simOut] = sim(net,mTestInputN);
--- a/main/nnet/tests/MLP/example1/mlp9_2_2_tansig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,6 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 26.3503/0, Gradient 416.299/1e-010
-TRAINLM, Epoch 9/100, MSE 3.26824/0, Gradient 0.969955/1e-010
-TRAINLM, Validation stop.
-
-2.0431 2.0431 1.0154 2.0035 2.0431 1.0154 2.0035 2.0431 1.0154 2.0431 1.0154 1.0154 0.9758 2.0431 1.0154 1.0154 1.0154 0.9758 2.0431 2.0431 2.0431 1.0154 1.0154 1.0154 2.0431 1.0154 1.0154 2.0431 2.0037 1.0154
-8.1802 8.1802 4.0490 8.0207 8.1802 4.0490 8.0207 8.1802 4.0490 8.1802 4.0490 4.0490 3.8895 8.1802 4.0490 4.0490 4.0490 3.8895 8.1802 8.1802 8.1802 4.0490 4.0490 4.0490 8.1802 4.0490 4.0490 8.1802 8.0213 4.0490
\ No newline at end of file
--- a/main/nnet/tests/MLP/example1/mlp9_2_2_tansig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,104 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## load data
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## this neural network example isn't a mirror of "the real life work"
-## it will be used to compare the results with MATLAB(TM) results.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-mOutput = [mOutput; mOutput*4];
-
-# now split the data matrix in 3 pieces,
-# train data, test data and validate data
-# the proportion should be about 1/2 train,
-# 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights,
-# for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-# now divide mInput & mOutput in the three sets
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = []; # delete validation set
-mOutput(:,1:nValiSets) = []; # delete validation set
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = []; # delete test set
-mOutput(:,1:nTestSets) = []; # delete test set
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); # input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 2;
-nOutputNeurons = 2;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"tansig","purelin"},"trainlm","learngdm","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(1,:) = 1.5;
-MLPnet.IW{1,1}(2,:) = 0.5;
-MLPnet.LW{2,1}(1,:) = 1.5;
-MLPnet.LW{2,1}(2,:) = 0.5;
-MLPnet.b{1,1}(1,:) = 0.5;
-MLPnet.b{1,1}(2,:) = 1.5;
-MLPnet.b{2,1}(:) = 0.5;
-
-## define validation data new, for MATLAB(TM) compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-## now train the network structure
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-
-## make preparations for net test and test MLPnet
- # standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-# will output the network results
-[simOut] = sim(net,mTestInputN);
--- a/main/nnet/tests/MLP/example1/mlp9_2_3_tansig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,7 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 134.203/0, Gradient 1028.61/1e-010
-TRAINLM, Epoch 14/100, MSE 14.8413/0, Gradient 0.138881/1e-010
-TRAINLM, Validation stop.
-
-2.1874 2.0916 1.1052 2.1874 2.1874 1.1052 1.1052 2.1874 2.1874 2.1874 1.1052 1.1931 2.1874 2.1874 1.1052 2.3418 1.1052 1.1052 2.1874 2.1874 2.1874 2.3418 1.1052 1.1052 2.1874 1.1052 2.3412 2.1874 1.1052 1.1052
-8.7496 8.3662 4.4206 8.7496 8.7496 4.4208 4.4206 8.7496 8.7496 8.7496 4.4206 4.7724 8.7496 8.7496 4.4206 9.3672 4.4206 4.4206 8.7496 8.7496 8.7496 9.3672 4.4206 4.4206 8.7496 4.4206 9.3648 8.7496 4.4206 4.4206
-21.8740 20.9156 11.0515 21.8740 21.8740 11.0520 11.0515 21.8740 21.8740 21.8740 11.0515 11.9311 21.8740 21.8740 11.0515 23.4180 11.0515 11.0515 21.8740 21.8740 21.8740 23.4180 11.0515 11.0515 21.8740 11.0515 23.4119 21.8740 11.0515 11.0515
\ No newline at end of file
--- a/main/nnet/tests/MLP/example1/mlp9_2_3_tansig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,109 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## load data
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## this neural network example isn't a mirror of "the real life work"
-## it will be used to compare the results with MATLAB(TM) results.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-mOutput = [mOutput; mOutput*4; mOutput*10];
-
-
-# now split the data matrix in 3 pieces,
-# train data, test data and validate data
-# the proportion should be about 1/2 train,
-# 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights,
-# for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-# now divide mInput & mOutput in the three sets
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = []; # delete validation set
-mOutput(:,1:nValiSets) = []; # delete validation set
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = []; # delete test set
-mOutput(:,1:nTestSets) = []; # delete test set
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); # input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 2;
-nOutputNeurons = 3;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"tansig","purelin"},"trainlm","learngdm","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(1,:) = 1.5;
-MLPnet.IW{1,1}(2,:) = 0.5;
-MLPnet.LW{2,1}(1,:) = 1.5;
-MLPnet.LW{2,1}(2,:) = 0.5;
-MLPnet.LW{2,1}(3,:) = 0.1;
-MLPnet.b{1,1}(1,:) = 0.5;
-MLPnet.b{1,1}(2,:) = 1.5;
-MLPnet.b{2,1}(1,:) = 0.5;
-MLPnet.b{2,1}(2,:) = 0.5;
-MLPnet.b{2,1}(3,:) = 0.1;
-
-
-## define validation data new, for MATLAB(TM) compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-## now train the network structure
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-
-## make preparations for net test and test MLPnet
- # standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-# will output the network results
-[simOut] = sim(net,mTestInputN);
--- a/main/nnet/tests/MLP/example1/mlp9_5_3_tansig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,7 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 141.889/0, Gradient 1253.56/1e-010
-TRAINLM, Epoch 8/100, MSE 13.6739/0, Gradient 137.397/1e-010
-TRAINLM, Validation stop.
-
-0.8081 2.5413 2.5667 1.1749 1.3791 1.1761 1.2381 0.6851 0.6435 1.4331 2.2694 2.5965 1.0768 0.7442 0.9235 2.5681 1.8238  0.4295 1.0288 0.9430 0.4290 1.1847 1.1398 1.8846 1.7315 0.9500 2.4366 0.4475 1.2249 2.5670
-4.7533 8.7954 8.3791 6.1910 5.8888 6.1933 6.3116 4.5980 4.4590 6.6755 8.2791 8.5445 5.8721 4.4152 5.3732 8.3835 7.4296 3.7666 5.4819 5.3490 3.7647 6.2098 6.0769 7.5455 7.0122 5.3092 8.5951 3.8224 6.2864 8.3796
-12.1298 21.4848 20.4710 15.8321 14.6769 15.8372 16.0940 11.8536 11.5048 16.8779 20.3651 20.8561 15.0359 11.2229 13.7898 20.4812 18.5210 9.7772 14.1764 13.6925 9.7724 15.8731 15.5474 18.7726 17.4322 13.8107 21.0503 9.9154 16.0392 20.4721
\ No newline at end of file
--- a/main/nnet/tests/MLP/example1/mlp9_5_3_tansig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,111 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## load data
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## this neural network example isn't a mirror of "the real life work"
-## it will be used to compare the results with MATLAB(TM) results.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-mOutput = [mOutput; mOutput*4; mOutput*10];
-
-
-# now split the data matrix in 3 pieces,
-# train data, test data and validate data
-# the proportion should be about 1/2 train,
-# 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights,
-# for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-# now divide mInput & mOutput in the three sets
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = []; # delete validation set
-mOutput(:,1:nValiSets) = []; # delete validation set
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = []; # delete test set
-mOutput(:,1:nTestSets) = []; # delete test set
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); # input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 5;
-nOutputNeurons = 3;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"tansig","purelin"},"trainlm","learngdm","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(1,:) = 1.5;
-MLPnet.IW{1,1}(2,:) = 0.5;
-MLPnet.IW{1,1}(3:5,:) = 1;
-MLPnet.LW{2,1}(1,:) = 1.5;
-MLPnet.LW{2,1}(2,:) = 0.5;
-MLPnet.LW{2,1}(3,:) = 0.1;
-MLPnet.b{1,1}(1,:) = 0.5;
-MLPnet.b{1,1}(2,:) = 1.5;
-MLPnet.b{1,1}(3:5,:) = 1;
-MLPnet.b{2,1}(1,:) = 0.5;
-MLPnet.b{2,1}(2,:) = 0.5;
-MLPnet.b{2,1}(3,:) = 0.1;
-
-
-## define validation data new, for MATLAB(TM) compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-## now train the network structure
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-
-## make preparations for net test and test MLPnet
- # standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-# will output the network results
-[simOut] = sim(net,mTestInputN);
--- a/main/nnet/tests/MLP/example1/orig/mData.txt	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,89 +0,0 @@
-390	47	0	0	17	45	1	0	3	33	2	0	1
-395	59	0	0	30	33	0	0	4	36	0	0	1
-420	43	0	0	26	30	9	0	9	24	4	0	1
-412	50	0	0	25	47	0	0	11	29	0	0	1
-418	43	0	0	34	30	1	0	9	27	5	0	1
-406	77	1	0	15	24	2	0	6	37	1	0	1
-407	51	0	0	28	47	0	0	5	35	0	0	1
-445	74	0	0	19	30	0	0	4	43	0	0	1
-317	179	0	0	11	64	0	0	4	35	0	0	1
-300	175	0	0	8	71	2	0	0	35	0	0	1
-331	169	0	0	19	66	1	0	0	34	1	0	1
-347	123	1	0	7	52	7	0	4	38	3	0	1
-323	177	0	0	8	68	4	0	0	37	2	0	1
-337	177	0	0	11	47	1	0	4	39	3	0	1
-299	173	1	0	8	63	1	0	1	37	3	0	1
-286	207	1	0	5	74	2	0	1	38	4	0	1
-429	44	0	0	37	19	0	0	23	19	0	0	1
-459	22	0	0	46	14	0	0	27	17	0	0	1
-340	163	1	0	16	72	3	0	3	31	1	0	1
-311	236	0	0	6	66	4	0	2	24	9	0	1
-326	218	1	0	4	71	2	0	1	33	1	0	1
-316	184	0	0	13	65	3	0	5	34	3	0	1
-258	236	2	0	2	72	7	0	0	26	8	0	1
-457	15	0	0	44	1	0	0	30	13	0	0	1
-471	8	0	0	58	2	0	0	25	13	0	0	1
-285	273	2	0	7	50	13	0	1	28	10	0	1
-489	44	0	0	21	12	0	0	29	21	0	0	1
-345	163	1	0	17	57	2	0	2	32	1	0	1
-473	59	0	0	44	15	0	0	22	23	0	0	1
-486	14	0	0	41	4	0	0	40	8	0	0	1
-368	188	1	0	6	49	0	0	3	37	0	0	1
-490	23	0	0	34	6	0	0	30	14	0	0	1
-291	251	2	0	8	53	2	0	0	43	2	0	1
-492	14	0	0	44	4	0	0	40	7	0	0	1
-286	293	3	0	7	83	9	0	1	22	5	0	1
-349	224	1	0	3	66	3	0	2	41	1	0	1
-295	271	3	0	6	71	5	0	3	29	4	0	1
-500	68	0	0	23	15	0	0	18	30	0	0	1
-511	73	0	0	40	21	0	0	16	22	0	0	1
-312	253	0	0	2	63	2	0	0	35	3	0	1
-527	20	0	0	32	6	0	0	23	16	0	0	1
-264	298	4	0	7	68	10	0	1	30	7	0	2
-478	101	0	0	15	34	0	0	11	40	0	0	2
-344	240	2	0	4	43	1	0	4	39	6	0	2
-303	269	3	0	10	66	2	0	1	32	7	0	2
-306	286	2	0	12	61	2	0	3	28	4	0	2
-515	44	0	0	29	16	0	0	19	31	0	0	2
-306	285	5	0	5	68	3	0	1	33	1	0	2
-472	139	0	0	16	25	0	0	11	38	0	0	2
-454	173	1	0	14	21	0	0	2	42	0	0	2
-458	36	0	0	12	7	0	0	24	19	0	0	2
-405	199	0	0	13	37	0	0	6	45	0	0	2
-475	186	1	0	15	28	0	0	5	35	0	0	2
-447	189	0	0	11	34	0	0	5	45	0	0	2
-387	294	1	0	10	37	0	0	4	30	5	0	2
-396	217	0	0	16	43	2	0	6	30	0	0	2
-492	98	0	0	7	23	0	0	13	36	0	0	2
-521	137	0	0	16	17	0	0	7	24	0	0	2
-511	48	0	0	35	7	0	0	27	15	0	0	2
-532	57	0	0	19	6	0	0	28	15	0	0	2
-477	61	0	0	36	13	0	0	22	15	0	0	2
-481	43	0	0	38	11	0	0	21	24	0	0	2
-484	222	0	0	16	12	0	0	9	24	0	0	2
-484	37	0	0	50	13	0	0	29	18	0	0	2
-508	64	0	0	31	14	0	0	17	33	0	0	2
-511	55	0	0	32	15	0	0	23	21	0	0	2
-541	269	1	0	12	5	0	0	6	20	0	0	3
-477	30	0	0	24	14	0	0	18	26	0	0	3
-496	319	0	0	2	4	0	0	2	14	0	0	3
-470	273	1	0	9	17	0	0	2	32	0	0	3
-429	433	3	0	8	11	0	0	5	15	0	0	3
-506	64	0	0	32	23	0	0	13	29	0	0	3
-487	36	0	0	34	10	0	0	19	27	0	0	3
-475	34	0	0	38	13	0	0	28	18	0	0	3
-423	25	0	0	26	16	0	0	16	29	0	0	3
-488	83	0	0	34	18	0	0	15	34	0	0	3
-487	35	0	0	34	12	0	0	29	24	0	0	3
-475	23	0	0	37	5	0	0	26	14	0	0	3
-496	34	0	0	49	6	0	0	31	11	0	0	3
-493	21	0	0	33	16	0	0	22	19	0	0	3
-459	37	0	0	46	4	0	0	32	16	0	0	3
-525	41	0	0	38	9	0	0	37	18	0	0	3
-501	32	0	0	40	11	0	0	24	20	0	0	3
-506	24	0	0	34	12	0	0	26	17	0	0	3
-502	34	0	0	25	9	0	0	23	11	0	0	3
-485	29	0	0	33	7	0	0	14	20	0	0	3
-495	34	0	0	39	4	0	0	21	21	0	0	3
-476	41	0	0	32	5	0	0	19	26	0	0	3
-483	66	0	0	17	16	0	0	10	28	2	0	3
\ No newline at end of file
--- a/main/nnet/tests/MLP/example2/mData.txt	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,94 +0,0 @@
-# Created by Octave 2.9.5, Wed May 24 10:33:36 2006     <sim@TL3124>
-# name: mData
-# type: matrix
-# rows: 89
-# columns: 13
- 306 286 2 0 12 61 2 0 3 28 4 0 2
- 368 188 1 0 6 49 0 0 3 37 0 0 1
- 511 73 0 0 40 21 0 0 16 22 0 0 1
- 418 43 0 0 34 30 1 0 9 27 5 0 1
- 299 173 1 0 8 63 1 0 1 37 3 0 1
- 312 253 0 0 2 63 2 0 0 35 3 0 1
- 492 98 0 0 7 23 0 0 13 36 0 0 2
- 506 64 0 0 32 23 0 0 13 29 0 0 3
- 476 41 0 0 32 5 0 0 19 26 0 0 3
- 483 66 0 0 17 16 0 0 10 28 2 0 3
- 429 44 0 0 37 19 0 0 23 19 0 0 1
- 521 137 0 0 16 17 0 0 7 24 0 0 2
- 340 163 1 0 16 72 3 0 3 31 1 0 1
- 323 177 0 0 8 68 4 0 0 37 2 0 1
- 344 240 2 0 4 43 1 0 4 39 6 0 2
- 459 22 0 0 46 14 0 0 27 17 0 0 1
- 487 36 0 0 34 10 0 0 19 27 0 0 3
- 331 169 0 0 19 66 1 0 0 34 1 0 1
- 541 269 1 0 12 5 0 0 6 20 0 0 3
- 475 23 0 0 37 5 0 0 26 14 0 0 3
- 475 186 1 0 15 28 0 0 5 35 0 0 2
- 496 319 0 0 2 4 0 0 2 14 0 0 3
- 525 41 0 0 38 9 0 0 37 18 0 0 3
- 484 37 0 0 50 13 0 0 29 18 0 0 2
- 511 55 0 0 32 15 0 0 23 21 0 0 2
- 515 44 0 0 29 16 0 0 19 31 0 0 2
- 478 101 0 0 15 34 0 0 11 40 0 0 2
- 429 433 3 0 8 11 0 0 5 15 0 0 3
- 471 8 0 0 58 2 0 0 25 13 0 0 1
- 303 269 3 0 10 66 2 0 1 32 7 0 2
- 445 74 0 0 19 30 0 0 4 43 0 0 1
- 488 83 0 0 34 18 0 0 15 34 0 0 3
- 264 298 4 0 7 68 10 0 1 30 7 0 2
- 489 44 0 0 21 12 0 0 29 21 0 0 1
- 475 34 0 0 38 13 0 0 28 18 0 0 3
- 492 14 0 0 44 4 0 0 40 7 0 0 1
- 454 173 1 0 14 21 0 0 2 42 0 0 2
- 306 285 5 0 5 68 3 0 1 33 1 0 2
- 508 64 0 0 31 14 0 0 17 33 0 0 2
- 477 61 0 0 36 13 0 0 22 15 0 0 2
- 349 224 1 0 3 66 3 0 2 41 1 0 1
- 447 189 0 0 11 34 0 0 5 45 0 0 2
- 496 34 0 0 49 6 0 0 31 11 0 0 3
- 484 222 0 0 16 12 0 0 9 24 0 0 2
- 412 50 0 0 25 47 0 0 11 29 0 0 1
- 316 184 0 0 13 65 3 0 5 34 3 0 1
- 345 163 1 0 17 57 2 0 2 32 1 0 1
- 285 273 2 0 7 50 13 0 1 28 10 0 1
- 317 179 0 0 11 64 0 0 4 35 0 0 1
- 500 68 0 0 23 15 0 0 18 30 0 0 1
- 495 34 0 0 39 4 0 0 21 21 0 0 3
- 387 294 1 0 10 37 0 0 4 30 5 0 2
- 258 236 2 0 2 72 7 0 0 26 8 0 1
- 423 25 0 0 26 16 0 0 16 29 0 0 3
- 501 32 0 0 40 11 0 0 24 20 0 0 3
- 459 37 0 0 46 4 0 0 32 16 0 0 3
- 511 48 0 0 35 7 0 0 27 15 0 0 2
- 295 271 3 0 6 71 5 0 3 29 4 0 1
- 502 34 0 0 25 9 0 0 23 11 0 0 3
- 458 36 0 0 12 7 0 0 24 19 0 0 2
- 470 273 1 0 9 17 0 0 2 32 0 0 3
- 477 30 0 0 24 14 0 0 18 26 0 0 3
- 406 77 1 0 15 24 2 0 6 37 1 0 1
- 291 251 2 0 8 53 2 0 0 43 2 0 1
- 407 51 0 0 28 47 0 0 5 35 0 0 1
- 390 47 0 0 17 45 1 0 3 33 2 0 1
- 347 123 1 0 7 52 7 0 4 38 3 0 1
- 300 175 0 0 8 71 2 0 0 35 0 0 1
- 473 59 0 0 44 15 0 0 22 23 0 0 1
- 487 35 0 0 34 12 0 0 29 24 0 0 3
- 532 57 0 0 19 6 0 0 28 15 0 0 2
- 286 207 1 0 5 74 2 0 1 38 4 0 1
- 405 199 0 0 13 37 0 0 6 45 0 0 2
- 337 177 0 0 11 47 1 0 4 39 3 0 1
- 527 20 0 0 32 6 0 0 23 16 0 0 1
- 326 218 1 0 4 71 2 0 1 33 1 0 1
- 486 14 0 0 41 4 0 0 40 8 0 0 1
- 420 43 0 0 26 30 9 0 9 24 4 0 1
- 286 293 3 0 7 83 9 0 1 22 5 0 1
- 395 59 0 0 30 33 0 0 4 36 0 0 1
- 506 24 0 0 34 12 0 0 26 17 0 0 3
- 396 217 0 0 16 43 2 0 6 30 0 0 2
- 457 15 0 0 44 1 0 0 30 13 0 0 1
- 472 139 0 0 16 25 0 0 11 38 0 0 2
- 493 21 0 0 33 16 0 0 22 19 0 0 3
- 311 236 0 0 6 66 4 0 2 24 9 0 1
- 490 23 0 0 34 6 0 0 30 14 0 0 1
- 485 29 0 0 33 7 0 0 14 20 0 0 3
- 481 43 0 0 38 11 0 0 21 24 0 0 2
--- a/main/nnet/tests/MLP/example2/mlp9_1_1_logsig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,21 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 3.19597/0, Gradient 87.4172/1e-010
-TRAINLM, Epoch 6/100, MSE 0.219207/0, Gradient 0.455596/1e-010
-TRAINLM, Validation stop.
-
-
-simOut =
-
-  Columns 1 through 12 
-
-    1.0000    0.6866    0.8512    1.0000    1.0000    0.9998    1.0000    1.0000    0.6207    0.7705    1.0000    0.7405
-
-  Columns 13 through 24 
-
-    0.9997    0.9371    0.7935    0.8539    0.6889    1.0000    0.8356    1.0000    1.0000    1.0000    0.8666    0.9578
-
-  Columns 25 through 30 
-
-    0.9999    0.7036    1.0000    0.9635    1.0000    1.0000
-
->>
->>
\ No newline at end of file
--- a/main/nnet/tests/MLP/example2/mlp9_1_1_logsig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,106 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd 
-
-
-
-## load data
-
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## first permute the whole matrix in row wise
-## this won't be used right now for debug and test purpose
-# order = randperm(nRows);
-# mData(order,:) = mData;
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-%mOutput = [mOutput; mOutput*4];
-# now split the data matrix in 3 pieces, train data, test data and validate data
-# the proportion should be about 1/2 train, 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights, for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = [];
-mOutput(:,1:nValiSets) = [];
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = [];
-mOutput(:,1:nTestSets) = [];
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); % input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 1;
-nOutputNeurons = 1;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"logsig","purelin"},"trainlm","learngdm","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(:) = 1.5;
-MLPnet.LW{2,1}(:) = 0.5;
-MLPnet.b{1,1}(:) = 1.5;
-MLPnet.b{2,1}(:) = 0.5;
-
-#saveMLPStruct(MLPnet,"MLP3test.txt");
-#disp("network structure saved, press any key to continue...")
-#pause
-
-## define validation data new, for matlab compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-#[net,tr,out,E] = train(MLPnet,mInputN,mOutput,[],[],VV);
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-# saveMLPStruct(net,"MLP3testNachTraining.txt");
-# disp("network structure saved, press any key to continue...")
-# pause
-
-# % make preparations for net test and test MLPnet
-#     % standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-[simOut] = sim(net,mTestInputN);#%,Pi,Ai,mTestOutput);
--- a/main/nnet/tests/MLP/example2/mlp9_2_1_logsig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,20 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 2.35298/0, Gradient 79.3291/1e-010
-TRAINLM, Epoch 6/100, MSE 0.154324/0, Gradient 0.341107/1e-010
-TRAINLM, Validation stop.
-
-
-simOut =
-
-  Columns 1 through 12 
-
-    1.4912    0.9151    1.1368    1.4883    1.4996    1.4632    1.4907    1.5000    0.8235    1.0274    1.5000    0.9873
-
-  Columns 13 through 24 
-
-    1.4608    1.2652    1.0581    1.1405    0.9181    1.4839    1.1152    1.5000    1.4901    1.4997    1.1583    1.3023
-
-  Columns 25 through 30 
-
-    1.4694    0.9380    1.5000    1.3135    1.4947    1.4993
-
->>
\ No newline at end of file
--- a/main/nnet/tests/MLP/example2/mlp9_2_1_logsig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,113 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd 
-
-
-## load data
-
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## first permute the whole matrix in row wise
-## this won't be used right now for debug and test purpose
-# order = randperm(nRows);
-# mData(order,:) = mData;
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-%mOutput = [mOutput; mOutput*4];
-# now split the data matrix in 3 pieces, train data, test data and validate data
-# the proportion should be about 1/2 train, 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights, for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = [];
-mOutput(:,1:nValiSets) = [];
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = [];
-mOutput(:,1:nTestSets) = [];
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); % input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 2;
-nOutputNeurons = 1;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"logsig","purelin"},"trainlm","learngdm","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(1,:) = 0.5;
-MLPnet.IW{1,1}(2,:) = 1.5;
-MLPnet.LW{2,1}(:) = 0.5;
-MLPnet.b{1,1}(1,:) = 0.5;
-MLPnet.b{1,1}(2,:) = 1.5;
-MLPnet.b{2,1}(:) = 0.5;
-
-saveMLPStruct(MLPnet,"MLP3test.txt");
-#disp("network structure saved, press any key to continue...")
-#pause
-
-## define validation data new, for matlab compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-#[net,tr,out,E] = train(MLPnet,mInputN,mOutput,[],[],VV);
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-# saveMLPStruct(net,"MLP3testNachTraining.txt");
-# disp("network structure saved, press any key to continue...")
-# pause
-# % the names in matlab help, see "train" in help for more informations
-# tr.perf(max(tr.epoch)+1);
-# 
-# % make preparations for net test and test MLPnet
-#     % standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-# % [mTestOutputN] = trastd(mTestOutput,cMeanOutput,cStdOutput);
-#     % define unused parameters to get E-variable of simulation
-# Pi = zeros(6,0);
-# Ai = zeros(6,0);
-#     % simulate net
-[simOut] = sim(net,mTestInputN);#%,Pi,Ai,mTestOutput);
--- a/main/nnet/tests/MLP/example2/mlp9_2_2_1_logsig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,103 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-##
-## This program is free software; you can redistribute it and/or modify it
-## under the terms of the GNU General Public License as published by
-## the Free Software Foundation; either version 2, or (at your option)
-## any later version.
-##
-## This program is distributed in the hope that it will be useful, but
-## WITHOUT ANY WARRANTY; without even the implied warranty of
-## MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
-## General Public License for more details.
-##
-## You should have received a copy of the GNU General Public License
-## along with this program; see the file COPYING.  If not, see
-## <http://www.gnu.org/licenses/>.
-
-
-## author: msd
-
-mData = load("mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## first permute the whole matrix in row wise
-## this won't be used right now for debug and test purpose
-# order = randperm(nRows);
-# mData(order,:) = mData;
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-% now split the data matrix in 3 pieces, train data, test data and validate data
-% the proportion should be about 1/2 train, 1/3 test and 1/6 validate data
-% in this neural network we have 12 weights, for each weight at least 3 train sets..
-% (that's a rule of thumb like 1/2, 1/3 and 1/6)
-% 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-% now the rest of the sets are again 100%
-% ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = [];
-mOutput(:,1:nValiSets) = [];
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = [];
-mOutput(:,1:nTestSets) = [];
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);% standardize inputs
-
-%% comments: there is no reason to standardize the outputs because we have only
-% one output ...
-
-% define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); % input matrix with (R x 2)...
-
-%% define network
-nHiddenNeurons = [2 2];
-nOutputNeurons = 1;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{'tansig','tansig','purelin'},'trainlm','learngdm','mse');
-%% for test purpose, define weights by hand
-# MLPnet.IW{1,1}(:) = 1.5;
-# MLPnet.LW{2,1}(:) = 0.5;
-# MLPnet.b{1,1}(:) = 1.5;
-# MLPnet.b{2,1}(:) = 0.5;
-
- saveMLPStruct(MLPnet,'MLP9-2-2-1.txt');
-# disp('network structure saved, press any key to continue...')
-
-
-## define validation data new, for matlab compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
- %[net,tr,out,E] = train(MLPnet,mInputN,mOutput,[],[],VV);
- MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-# 
-# % % make preparations for net test and test MLPnet
-# %     % standardise input & output test data
-  [mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-# 
-  #[simOut,Pf,Af,simE,simPerf] = sim(net,mTestInputN);
-# 
-[simOut] = sim(net,mTestInputN);
\ No newline at end of file
--- a/main/nnet/tests/MLP/example2/mlp9_2_2_3_logsig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,26 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 227.919/0, Gradient 1662.64/1e-010
-TRAINLM, Epoch 7/100, MSE 9.72986/0, Gradient 2.23561e-009/1e-010
-TRAINLM, Validation stop.
-
-
-simOut =
-
-  Columns 1 through 12 
-
-   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612
-    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161
-   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706
-
-  Columns 13 through 24 
-
-   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612
-    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161
-   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706
-
-  Columns 25 through 30 
-
-   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612   -3.5612
-    2.4161    2.4161    2.4161    2.4161    2.4161    2.4161
-   14.3706   14.3706   14.3706   14.3706   14.3706   14.3706
-
->>
\ No newline at end of file
--- a/main/nnet/tests/MLP/example2/mlp9_2_2_logsig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,23 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 42.6863/0, Gradient 532.547/1e-010
-TRAINLM, Epoch 6/100, MSE 1.62318/0, Gradient 3.12953/1e-010
-TRAINLM, Validation stop.
-
-
-simOut =
-
-  Columns 1 through 12 
-
-    3.4901    1.8132    2.3724    3.4869    3.4995    3.4557    3.4896    3.5000    1.6319    2.0664    3.5000    1.9704
-    1.4967    0.9377    1.1241    1.4956    1.4998    1.4852    1.4965    1.5000    0.8773    1.0221    1.5000    0.9901
-
-  Columns 13 through 24 
-
-    3.4524    2.8360    2.1451    2.3844    1.8195    3.4818    2.3061    3.5000    3.4889    3.4997    2.4421    2.9859
-    1.4841    1.2787    1.0484    1.1281    0.9398    1.4939    1.1020    1.5000    1.4963    1.4999    1.1474    1.3286
-
-  Columns 25 through 30 
-
-    3.4639    1.8612    3.5000    3.0314    3.4941    3.4993
-    1.4880    0.9537    1.5000    1.3438    1.4980    1.4998
-
->>
\ No newline at end of file
--- a/main/nnet/tests/MLP/example2/mlp9_2_2_logsig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,111 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd 
-
-
-## load data
-
-mData = load("mData.txt","mData");
-# mData = mData.mData(1:10,:);
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## first permute the whole matrix in row wise
-## this won't be used right now for debug and test purpose
-# order = randperm(nRows);
-# mData(order,:) = mData;
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-mOutput = [mOutput; mOutput*4];
-# now split the data matrix in 3 pieces, train data, test data and validate data
-# the proportion should be about 1/2 train, 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights, for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = [];
-mOutput(:,1:nValiSets) = [];
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = [];
-mOutput(:,1:nTestSets) = [];
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); % input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 2;
-nOutputNeurons = 2;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"logsig","purelin"},"trainlm","learngdm","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(1,:) = 1.5;
-MLPnet.IW{1,1}(2,:) = 0.5;
-MLPnet.LW{2,1}(1,:) = 1.5;
-MLPnet.LW{2,1}(2,:) = 0.5;
-MLPnet.b{1,1}(1,:) = 0.5;
-MLPnet.b{1,1}(2,:) = 1.5;
-MLPnet.b{2,1}(:) = 0.5;
-
-# saveMLPStruct(MLPnet,"MLP3test.txt");
-#disp("network structure saved, press any key to continue...")
-#pause
-
-## define validation data new, for MATLAB(TM) compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-#[net,tr,out,E] = train(MLPnet,mInputN,mOutput,[],[],VV);
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-# saveMLPStruct(net,"MLP3testNachTraining.txt");
-# disp("network structure saved, press any key to continue...")
-# disp("nach Training")
-# pause
-
-# % make preparations for net test and test MLPnet
-#     % standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-#     % simulate net
-[simOut] = sim(net,mTestInputN);%,Pi,Ai,mTestOutput);
--- a/main/nnet/tests/MLP/example2/mlp9_2_3_logsig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,26 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 245.153/0, Gradient 1576.3/1e-010
-TRAINLM, Epoch 6/100, MSE 8.10503/0, Gradient 15.6612/1e-010
-TRAINLM, Validation stop.
-
-
-simOut =
-
-  Columns 1 through 12 
-
-    3.4901    1.8132    2.3724    3.4869    3.4995    3.4557    3.4896    3.5000    1.6319    2.0664    3.5000    1.9704
-    1.4967    0.9377    1.1241    1.4956    1.4998    1.4852    1.4965    1.5000    0.8773    1.0221    1.5000    0.9901
-    0.2993    0.1875    0.2248    0.2991    0.3000    0.2970    0.2993    0.3000    0.1755    0.2044    0.3000    0.1980
-
-  Columns 13 through 24 
-
-    3.4524    2.8360    2.1451    2.3844    1.8195    3.4818    2.3061    3.5000    3.4889    3.4997    2.4421    2.9859
-    1.4841    1.2787    1.0484    1.1281    0.9398    1.4939    1.1020    1.5000    1.4963    1.4999    1.1474    1.3286
-    0.2968    0.2557    0.2097    0.2256    0.1880    0.2988    0.2204    0.3000    0.2993    0.3000    0.2295    0.2657
-
-  Columns 25 through 30 
-
-    3.4639    1.8612    3.5000    3.0314    3.4941    3.4993
-    1.4880    0.9537    1.5000    1.3438    1.4980    1.4998
-    0.2976    0.1907    0.3000    0.2688    0.2996    0.3000
-
->>
\ No newline at end of file
--- a/main/nnet/tests/MLP/example2/mlp9_2_3_logsig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,115 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd 
-
-
-
-## load data
-
-mData = load("mData.txt","mData");
-# mData = mData.mData(1:10,:);
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## first permute the whole matrix in row wise
-## this won't be used right now for debug and test purpose
-# order = randperm(nRows);
-# mData(order,:) = mData;
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-mOutput = [mOutput; mOutput*4; mOutput*10];
-# now split the data matrix in 3 pieces, train data, test data and validate data
-# the proportion should be about 1/2 train, 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights, for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = [];
-mOutput(:,1:nValiSets) = [];
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = [];
-mOutput(:,1:nTestSets) = [];
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); % input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 2;
-nOutputNeurons = 3;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"logsig","purelin"},"trainlm","learngdm","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(1,:) = 1.5;
-MLPnet.IW{1,1}(2,:) = 0.5;
-MLPnet.LW{2,1}(1,:) = 1.5;
-MLPnet.LW{2,1}(2,:) = 0.5;
-MLPnet.LW{2,1}(3,:) = 0.1;
-MLPnet.b{1,1}(1,:) = 0.5;
-MLPnet.b{1,1}(2,:) = 1.5;
-MLPnet.b{2,1}(1,:) = 0.5;
-MLPnet.b{2,1}(2,:) = 0.5;
-MLPnet.b{2,1}(3,:) = 0.1;
-
-# saveMLPStruct(MLPnet,"MLP3test.txt");
-#disp("network structure saved, press any key to continue...")
-#pause
-
-## define validation data new, for MATLAB(TM) compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-#[net,tr,out,E] = train(MLPnet,mInputN,mOutput,[],[],VV);
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-# saveMLPStruct(net,"MLP3testNachTraining.txt");
-# disp("network structure saved, press any key to continue...")
-# disp("nach Training")
-# pause
-
-# % make preparations for net test and test MLPnet
-#     % standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-#     % simulate net
-[simOut] = sim(net,mTestInputN);%,Pi,Ai,mTestOutput);
--- a/main/nnet/tests/MLP/example2/mlp9_5_3_logsig.dat	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,26 +0,0 @@
-TRAINLM, Epoch 0/100, MSE 240.504/0, Gradient 1971.95/1e-010
-TRAINLM, Epoch 6/100, MSE 7.95197/0, Gradient 203.206/1e-010
-TRAINLM, Validation stop.
-
-
-simOut =
-
-  Columns 1 through 12 
-
-    7.9886    3.6782    5.2496    7.9843    7.9995    7.9278    7.9880    8.0000    3.0623    4.4398    8.0000    4.1632
-    2.9962    1.5594    2.0832    2.9948    2.9998    2.9759    2.9960    3.0000    1.3541    1.8133    3.0000    1.7211
-    0.5992    0.3119    0.4166    0.5990    0.6000    0.5952    0.5992    0.6000    0.2708    0.3627    0.6000    0.3442
-
-  Columns 13 through 24 
-
-    7.9206    6.3668    4.6573    5.2794    3.6986    7.9768    5.0815    8.0000    7.9871    7.9997    5.4228    6.7236
-    2.9735    2.4556    1.8858    2.0931    1.5662    2.9923    2.0272    3.0000    2.9957    2.9999    2.1409    2.5745
-    0.5947    0.4911    0.3772    0.4186    0.3132    0.5985    0.4054    0.6000    0.5991    0.6000    0.4282    0.5149
-
-  Columns 25 through 30 
-
-    7.9450    3.8310    8.0000    6.8332    7.9936    7.9993
-    2.9817    1.6103    3.0000    2.6111    2.9979    2.9998
-    0.5963    0.3221    0.6000    0.5222    0.5996    0.6000
-
->>
\ No newline at end of file
--- a/main/nnet/tests/MLP/example2/mlp9_5_3_logsig.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,116 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## load data
-
-mData = load("mData.txt","mData");
-# mData = mData.mData(1:10,:);
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## first permute the whole matrix in row wise
-## this won't be used right now for debug and test purpose
-# order = randperm(nRows);
-# mData(order,:) = mData;
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-mOutput = [mOutput; mOutput*4; mOutput*10];
-# now split the data matrix in 3 pieces, train data, test data and validate data
-# the proportion should be about 1/2 train, 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights, for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = [];
-mOutput(:,1:nValiSets) = [];
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = [];
-mOutput(:,1:nTestSets) = [];
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
-
-[mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-
-## comments: there is no reason to standardize the outputs because we have only
-# one output ...
-
-# define the max and min inputs for each row
-mMinMaxElements = min_max(mTrainInputN); % input matrix with (R x 2)...
-
-## define network
-nHiddenNeurons = 5;
-nOutputNeurons = 3;
-
-MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"logsig","purelin"},"trainlm","learngdm","mse");
-## for test purpose, define weights by hand
-MLPnet.IW{1,1}(1,:) = 1.5;
-MLPnet.IW{1,1}(2,:) = 0.5;
-MLPnet.IW{1,1}(3:5,:) = 1;
-MLPnet.LW{2,1}(1,:) = 1.5;
-MLPnet.LW{2,1}(2,:) = 0.5;
-MLPnet.LW{2,1}(3,:) = 0.1;
-MLPnet.b{1,1}(1,:) = 0.5;
-MLPnet.b{1,1}(2,:) = 1.5;
-MLPnet.b{1,1}(3:5,:) = 1;
-MLPnet.b{2,1}(1,:) = 0.5;
-MLPnet.b{2,1}(2,:) = 0.5;
-MLPnet.b{2,1}(3,:) = 0.1;
-
-# saveMLPStruct(MLPnet,"MLP3test.txt");
-#disp("network structure saved, press any key to continue...")
-#pause
-
-## define validation data new, for MATLAB(TM) compatibility
-VV.P = mValiInput;
-VV.T = mValliOutput;
-
-## standardize also the validate data
-VV.P = trastd(VV.P,cMeanInput,cStdInput);
-
-#[net,tr,out,E] = train(MLPnet,mInputN,mOutput,[],[],VV);
-MLPnet.trainParam.show = NaN;
-[net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-# saveMLPStruct(net,"MLP3testNachTraining.txt");
-# disp("network structure saved, press any key to continue...")
-# disp("nach Training")
-# pause
-
-# % make preparations for net test and test MLPnet
-#     % standardise input & output test data
-[mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-
-#     % simulate net
-[simOut] = sim(net,mTestInputN);%,Pi,Ai,mTestOutput);
--- a/main/nnet/tests/MLP/example2/orig/mData.txt	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,89 +0,0 @@
-390	47	0	0	17	45	1	0	3	33	2	0	1
-395	59	0	0	30	33	0	0	4	36	0	0	1
-420	43	0	0	26	30	9	0	9	24	4	0	1
-412	50	0	0	25	47	0	0	11	29	0	0	1
-418	43	0	0	34	30	1	0	9	27	5	0	1
-406	77	1	0	15	24	2	0	6	37	1	0	1
-407	51	0	0	28	47	0	0	5	35	0	0	1
-445	74	0	0	19	30	0	0	4	43	0	0	1
-317	179	0	0	11	64	0	0	4	35	0	0	1
-300	175	0	0	8	71	2	0	0	35	0	0	1
-331	169	0	0	19	66	1	0	0	34	1	0	1
-347	123	1	0	7	52	7	0	4	38	3	0	1
-323	177	0	0	8	68	4	0	0	37	2	0	1
-337	177	0	0	11	47	1	0	4	39	3	0	1
-299	173	1	0	8	63	1	0	1	37	3	0	1
-286	207	1	0	5	74	2	0	1	38	4	0	1
-429	44	0	0	37	19	0	0	23	19	0	0	1
-459	22	0	0	46	14	0	0	27	17	0	0	1
-340	163	1	0	16	72	3	0	3	31	1	0	1
-311	236	0	0	6	66	4	0	2	24	9	0	1
-326	218	1	0	4	71	2	0	1	33	1	0	1
-316	184	0	0	13	65	3	0	5	34	3	0	1
-258	236	2	0	2	72	7	0	0	26	8	0	1
-457	15	0	0	44	1	0	0	30	13	0	0	1
-471	8	0	0	58	2	0	0	25	13	0	0	1
-285	273	2	0	7	50	13	0	1	28	10	0	1
-489	44	0	0	21	12	0	0	29	21	0	0	1
-345	163	1	0	17	57	2	0	2	32	1	0	1
-473	59	0	0	44	15	0	0	22	23	0	0	1
-486	14	0	0	41	4	0	0	40	8	0	0	1
-368	188	1	0	6	49	0	0	3	37	0	0	1
-490	23	0	0	34	6	0	0	30	14	0	0	1
-291	251	2	0	8	53	2	0	0	43	2	0	1
-492	14	0	0	44	4	0	0	40	7	0	0	1
-286	293	3	0	7	83	9	0	1	22	5	0	1
-349	224	1	0	3	66	3	0	2	41	1	0	1
-295	271	3	0	6	71	5	0	3	29	4	0	1
-500	68	0	0	23	15	0	0	18	30	0	0	1
-511	73	0	0	40	21	0	0	16	22	0	0	1
-312	253	0	0	2	63	2	0	0	35	3	0	1
-527	20	0	0	32	6	0	0	23	16	0	0	1
-264	298	4	0	7	68	10	0	1	30	7	0	2
-478	101	0	0	15	34	0	0	11	40	0	0	2
-344	240	2	0	4	43	1	0	4	39	6	0	2
-303	269	3	0	10	66	2	0	1	32	7	0	2
-306	286	2	0	12	61	2	0	3	28	4	0	2
-515	44	0	0	29	16	0	0	19	31	0	0	2
-306	285	5	0	5	68	3	0	1	33	1	0	2
-472	139	0	0	16	25	0	0	11	38	0	0	2
-454	173	1	0	14	21	0	0	2	42	0	0	2
-458	36	0	0	12	7	0	0	24	19	0	0	2
-405	199	0	0	13	37	0	0	6	45	0	0	2
-475	186	1	0	15	28	0	0	5	35	0	0	2
-447	189	0	0	11	34	0	0	5	45	0	0	2
-387	294	1	0	10	37	0	0	4	30	5	0	2
-396	217	0	0	16	43	2	0	6	30	0	0	2
-492	98	0	0	7	23	0	0	13	36	0	0	2
-521	137	0	0	16	17	0	0	7	24	0	0	2
-511	48	0	0	35	7	0	0	27	15	0	0	2
-532	57	0	0	19	6	0	0	28	15	0	0	2
-477	61	0	0	36	13	0	0	22	15	0	0	2
-481	43	0	0	38	11	0	0	21	24	0	0	2
-484	222	0	0	16	12	0	0	9	24	0	0	2
-484	37	0	0	50	13	0	0	29	18	0	0	2
-508	64	0	0	31	14	0	0	17	33	0	0	2
-511	55	0	0	32	15	0	0	23	21	0	0	2
-541	269	1	0	12	5	0	0	6	20	0	0	3
-477	30	0	0	24	14	0	0	18	26	0	0	3
-496	319	0	0	2	4	0	0	2	14	0	0	3
-470	273	1	0	9	17	0	0	2	32	0	0	3
-429	433	3	0	8	11	0	0	5	15	0	0	3
-506	64	0	0	32	23	0	0	13	29	0	0	3
-487	36	0	0	34	10	0	0	19	27	0	0	3
-475	34	0	0	38	13	0	0	28	18	0	0	3
-423	25	0	0	26	16	0	0	16	29	0	0	3
-488	83	0	0	34	18	0	0	15	34	0	0	3
-487	35	0	0	34	12	0	0	29	24	0	0	3
-475	23	0	0	37	5	0	0	26	14	0	0	3
-496	34	0	0	49	6	0	0	31	11	0	0	3
-493	21	0	0	33	16	0	0	22	19	0	0	3
-459	37	0	0	46	4	0	0	32	16	0	0	3
-525	41	0	0	38	9	0	0	37	18	0	0	3
-501	32	0	0	40	11	0	0	24	20	0	0	3
-506	24	0	0	34	12	0	0	26	17	0	0	3
-502	34	0	0	25	9	0	0	23	11	0	0	3
-485	29	0	0	33	7	0	0	14	20	0	0	3
-495	34	0	0	39	4	0	0	21	21	0	0	3
-476	41	0	0	32	5	0	0	19	26	0	0	3
-483	66	0	0	17	16	0	0	10	28	2	0	3
\ No newline at end of file
--- a/main/nnet/tests/MLP/loadtestresults.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,54 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## Author: Michel D. Schmid
-
-function [cAr mData] = loadtestresults(strFileName)
-
-  ## check range of input arguments
-  error(nargchk(1,1,nargin))
-
-  i = 1;
-  mData = [];
-  cAr = {};
-
-  fid = fopen(strFileName,"rt"); # open read only
-
-  strLine = fgetl(fid);
-  while (!feof(fid)) # this means, while not eof
-
-    [val, count] = sscanf (strLine, "%f");
-    if (count)
-      mData = [mData; val'];
-    else
-      cAr{i} = strLine;
-    endif
-
-    strLine = fgetl(fid);
-    i += 1;
-  endwhile
-  
-  # here, the strLine contains the last row of a file
-  # so do the complete coding once more
-  [val, count] = sscanf (strLine, "%f");
-  if (count)
-    mData = [mData; val'];
-  else
-    cAr{i} = strLine;
-  endif
-
-  fclose(fid);
-  
-endfunction
--- a/main/nnet/tests/MLP/nnetTestMLP.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,6 +0,0 @@
-
-test testExample1_1 # runs test for complete mlp scripts
-#test testExample1_2 # runs test for complete mlp scripts
-
-#delete('log_test1_1');
-#delete('log_test1_2');
\ No newline at end of file
--- a/main/nnet/tests/MLP/preparedata9_x_1.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,63 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## load data
-mData = load("example1/mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## this neural network example isn't a mirror of "the real life work"
-## it will be used to compare the results with MATLAB(TM) results.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-
-# now split the data matrix in 3 pieces,
-# train data, test data and validate data
-# the proportion should be about 1/2 train,
-# 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights,
-# for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-# now divide mInput & mOutput in the three sets
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = []; # delete validation set
-mOutput(:,1:nValiSets) = []; # delete validation set
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = []; # delete test set
-mOutput(:,1:nTestSets) = []; # delete test set
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
--- a/main/nnet/tests/MLP/preparedata9_x_2.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,63 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## load data
-mData = load("example1/mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## this neural network example isn't a mirror of "the real life work"
-## it will be used to compare the results with MATLAB(TM) results.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-mOutput = [mOutput; mOutput*4];
-# now split the data matrix in 3 pieces,
-# train data, test data and validate data
-# the proportion should be about 1/2 train,
-# 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights,
-# for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-# now divide mInput & mOutput in the three sets
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = []; # delete validation set
-mOutput(:,1:nValiSets) = []; # delete validation set
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = []; # delete test set
-mOutput(:,1:nTestSets) = []; # delete test set
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
--- a/main/nnet/tests/MLP/preparedata9_x_3.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,63 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## load data
-mData = load("example1/mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## this neural network example isn't a mirror of "the real life work"
-## it will be used to compare the results with MATLAB(TM) results.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-mOutput = [mOutput; mOutput*4; mOutput*10];
-# now split the data matrix in 3 pieces,
-# train data, test data and validate data
-# the proportion should be about 1/2 train,
-# 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights,
-# for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-# now divide mInput & mOutput in the three sets
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = []; # delete validation set
-mOutput(:,1:nValiSets) = []; # delete validation set
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = []; # delete test set
-mOutput(:,1:nTestSets) = []; # delete test set
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
--- a/main/nnet/tests/MLP/preparedata9_x_x_1.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,63 +0,0 @@
-## Copyright (C) 2006 Michel D. Schmid  <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## load data
-mData = load("example1/mData.txt","mData");
-mData = mData.mData;
-[nRows, nColumns] = size(mData);
-# this file contains 13 columns. The first 12 columns are the inputs
-# the last column is the output
-# remove column 4, 8 and 12
-# 89 rows
-
-## this neural network example isn't a mirror of "the real life work"
-## it will be used to compare the results with MATLAB(TM) results.
-
-mOutput = mData(:,end);
-mInput = mData(:,1:end-1);
-mInput(:,[4 8 12]) = []; # delete column 4, 8 and 12
-
-## now prepare data
-mInput = mInput';
-mOutput = mOutput';
-
-# now split the data matrix in 3 pieces,
-# train data, test data and validate data
-# the proportion should be about 1/2 train,
-# 1/3 test and 1/6 validate data
-# in this neural network we have 12 weights,
-# for each weight at least 3 train sets..
-# (that's a rule of thumb like 1/2, 1/3 and 1/6)
-# 1/2 of 89 = 44.5; let's take 44 for training
-nTrainSets = floor(nRows/2);
-# now the rest of the sets are again 100%
-# ==> 2/3 for test sets and 1/3 for validate sets
-nTestSets = (nRows-nTrainSets)/3*2;
-nValiSets = nRows-nTrainSets-nTestSets;
-
-# now divide mInput & mOutput in the three sets
-mValiInput = mInput(:,1:nValiSets);
-mValliOutput = mOutput(:,1:nValiSets);
-mInput(:,1:nValiSets) = []; # delete validation set
-mOutput(:,1:nValiSets) = []; # delete validation set
-mTestInput = mInput(:,1:nTestSets);
-mTestOutput = mOutput(:,1:nTestSets);
-mInput(:,1:nTestSets) = []; # delete test set
-mOutput(:,1:nTestSets) = []; # delete test set
-mTrainInput = mInput(:,1:nTrainSets);
-mTrainOutput = mOutput(:,1:nTrainSets);
--- a/main/nnet/tests/MLP/testExample1_1.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,75 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## This file is used to test all the m-files inside
-## the example1 directory.
-
-## it exist for each m-file a corresponding dat-file with
-## the numerical results of matlab
-
-## actually, following m-files will be tested:
-## A. One hidden layer
-## ==================
-## 1. mlp9_1_1_tansig
-## 2. mlp9_2_1_tansig
-## 3. mlp9_2_2_tansig
-## 4. mlp9_2_3_tansig
-## 5. mlp9_5_3_tansig
-##
-## B. Two hidden layer
-## ==================
-
-
-####### mlp9_1_1_tansig ######
-%!shared cAr, mTestResults, simOut, line, fid
-%!  diary log_test1_1
-%!  dir = "example1";
-%!  [cAr, mTestResults] = loadtestresults([dir "/mlp9_1_1_tansig.dat"]);
-%!  preparedata9_x_1;
-%!  [mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-%!  mMinMaxElements = min_max(mTrainInputN); # input matrix with (R x 2)...
-%!  nHiddenNeurons = 1;
-%!  nOutputNeurons = 1;
-%!  MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"tansig","purelin"},"trainlm","learngdm","mse");
-%!  MLPnet.IW{1,1}(:) = 1.5;
-%!  MLPnet.LW{2,1}(:) = 0.5;
-%!  MLPnet.b{1,1}(:) = 1.5;
-%!  MLPnet.b{2,1}(:) = 0.5;
-%!  VV.P = mValiInput;
-%!  VV.T = mValliOutput;
-%!  VV.P = trastd(VV.P,cMeanInput,cStdInput);
-%!  [net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-%!  [mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-%!  [simOut] = sim(net,mTestInputN);
-%!assert(simOut,mTestResults,0.0001)
-%!  diary off ;
-%!  fid = fopen("log_test1_1","r");
-%!  line = fgetl(fid);
-%!	if (line==-1)
-%!    error("no String in Line: Row 67");
-%!  endif
-%!assert(substr(line,16,1),substr(cAr{1,1},16,1))
-%!assert(substr(line,27,7),substr(cAr{1,1},27,7))
-%!assert(substr(line,47,7),substr(cAr{1,1},47,7))
-%!  line = fgetl(fid);
-%!assert(substr(line,16,1),substr(cAr{1,2},16,1))
-%!assert(substr(line,27,7),substr(cAr{1,2},27,7))
-%!assert(substr(line,48,6),substr(cAr{1,2},48,6))
-%!assert(strcmp("TRAINLM, Validation stop.",substr(cAr{1,3},1,25)))
-%!  fclose(fid);
-
--- a/main/nnet/tests/MLP/testExample1_2.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,84 +0,0 @@
-## Copyright (C) 2007 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## author: msd
-
-
-## This file is used to test all the m-files inside
-## the example1 directory.
-
-## it exist for each m-file a corresponding dat-file with
-## the numerical results of matlab
-
-## actually, following m-files will be tested:
-## A. One hidden layer
-## ==================
-## 1. mlp9_1_1_tansig
-## 2. mlp9_2_1_tansig
-## 3. mlp9_2_2_tansig
-## 4. mlp9_2_3_tansig
-## 5. mlp9_5_3_tansig
-##
-## B. Two hidden layer
-## ==================
-
-###### mlp9_2_1_tansig ######
-%!shared cAr, mTestResults, simOut, line, fid
-%!  diary log_test1_2
-%!  dir = "example1";
-%!  [cAr, mTestResults] = loadtestresults([dir "/mlp9_2_1_tansig.dat"]);
-%!  preparedata9_x_1
-%!  [mTrainInputN,cMeanInput,cStdInput] = prestd(mTrainInput);# standardize inputs
-%!  mMinMaxElements = min_max(mTrainInputN); # input matrix with (R x 2)...
-%!  nHiddenNeurons = 2;
-%!  nOutputNeurons = 1;
-%!  MLPnet = newff(mMinMaxElements,[nHiddenNeurons nOutputNeurons],{"tansig","purelin"},"trainlm","learngdm","mse");
-%!  MLPnet.IW{1,1}(1,:) = 0.5;
-%!  MLPnet.IW{1,1}(2,:) = 1.5;
-%!  MLPnet.LW{2,1}(:) = 0.5;
-%!  MLPnet.b{1,1}(1,:) = 0.5;
-%!  MLPnet.b{1,1}(2,:) = 1.5;
-%!  MLPnet.b{2,1}(:) = 0.5;
-%!  VV.P = mValiInput;
-%!  VV.T = mValliOutput;
-%!  VV.P = trastd(VV.P,cMeanInput,cStdInput);
-%!  [net] = train(MLPnet,mTrainInputN,mTrainOutput,[],[],VV);
-%!  [mTestInputN] = trastd(mTestInput,cMeanInput,cStdInput);
-%!  [simOut] = sim(net,mTestInputN);
-%!  diary off
-%!assert(simOut,mTestResults,0.0001)
-%!  fid = fopen("log_test1_2","r");
-%!  line = fgetl(fid);
-%!assert(substr(line,16,1),substr(cAr{1,1},16,1))
-%!assert(substr(line,27,7),substr(cAr{1,1},27,7))
-%!assert(substr(line,48,7),substr(cAr{1,1},48,7))
-%!  line = fgetl(fid);
-%!assert(substr(line,16,2),substr(cAr{1,2},16,2))
-%!assert(substr(line,27,7),substr(cAr{1,2},27,7))
-%!assert(substr(line,48,6),substr(cAr{1,2},48,6))
-%!assert(strcmp("TRAINLM, Validation stop.",substr(cAr{1,3},1,25)))
-%!  fclose(fid);
-
-
-
-
-
-
-
-
-
-
-
- 
--- a/main/nnet/tests/nnetTest.m	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,34 +0,0 @@
-## Copyright (C) 2008 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-## Running this file will test all function files and tests which are
-## available in the nnet package
-
-# change to directory where the function files are
-cd ../inst
-
-# now run the function files in functional order if needed
-test __analyzerows
-test __copycoltopos1
-
-test isposint
-test min_max
-test newff
-
-
-# go back to the test directory
-cd ../tests
-
-
--- a/main/nnet/tests/readme	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,33 +0,0 @@
-tests directory:
-================
-
-nnetTest.m:
------------
-Runs all the m-files containing
-test cases
-
-test_nnet_win32.pl:
--------------------
-Is a perl script. To run this, you must have
-installed a perl interpreter. I tested it with
-the ActiveState Perl for Win32 systems.
-Currently, it does nothing else as nnetTest.m does.
-
-
-
-subdirectory MLP:
-=================
-
-MLPScripts.m:
--------------
-to run some very basic tests please run
-MLPScripts. This will run all scripts
-inside the subdirectories "example1" and
-"example2".
-If no error occur, the first test is passed :-)
-
-nnetTestMLP.m
-----------
-these tests don't run on the windows version of octave-forge.
-Please use a linux/unix version. I tested it on octave-3.0.0
-compiled on ubuntu.
--- a/main/nnet/tests/test_nnet_win32.pl	Wed Feb 24 17:38:45 2016 +0000
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,82 +0,0 @@
-#!C:/perl/bin/perl.exe
-##
-## Copyright (C) 2008 Michel D. Schmid <michael.schmid@plexso.com>
-##
-## This program is free software; you can redistribute it and/or modify it under
-## the terms of the GNU General Public License as published by the Free Software
-## Foundation; either version 3 of the License, or (at your option) any later
-## version.
-##
-## This program is distributed in the hope that it will be useful, but WITHOUT
-## ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-## FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
-## details.
-##
-## You should have received a copy of the GNU General Public License along with
-## this program; if not, see <http://www.gnu.org/licenses/>.
-
-use strict;
-use diagnostics;    # Force verbose warning diagnostics.
-use warnings;
-use English;
-
-#use vars qw( $VERSION );
-#use Getopt::Long;
-#use Win32;
-use Win32::Process;
-use Win32::Console::ANSI;
-use Term::ANSIScreen qw/:color /;
-
-
-#my $command = "D:\\programs\\programming\\octave\\bin\\octave.exe";
-my $args = "D:\\programs\\programming\\octave\\bin\\octave.exe D:\\daten\\octave\\neuroPackage\\0.1.8.1\\nnet\\tests\\nnetTest.m";
-my $numberOfFailedTests = 0;
-my $numberOfSuccessfullTests = 0;
-my $testingFile;
-my @tokens=[];
-#my $process; # Prozess-Objekt
-print "Starting with tests for the neural network package\n";
-# Win32::Process::Create($process,
-# 					   $command,
-# 					   $args,
-# 					   1,
-# 					   CREATE_NEW_CONSOLE,
-# 					   '.');
-#$process->Wait(INFINITE); # script will wait until process is finished
-
-# my $pid = $process->GetProcessID();
-# print "new process pid: $pid\n";
-open(COUNTER, "$args |")
-or die("...: $!\n");
-
-while (<COUNTER>)
-{
-   if (/^testing/)
-   {
-     chomp ($testingFile = $_);
-   }
-   if (/^!!!!!/)
-   {
-     $numberOfFailedTests += 1;
-     print colored("$testingFile $_",'red');
-   }elsif(/^PASSES/)
-   {
-     @tokens = split(/ /, $_);
-     $numberOfSuccessfullTests += $tokens[4];
-     print colored("$testingFile $_",'yellow');
-   }
-
-}
-print "\n\n";
-print colored("Summary:\n",'green');
-print colored("Number of files containing failed tests: $numberOfFailedTests!\n",'red');
-print colored("Number of successfull tests:             $numberOfSuccessfullTests!\n",'yellow');
-my $allTests = $numberOfFailedTests + $numberOfSuccessfullTests;
-print colored("\nRunning complete $allTests tests!\n",'green');
-
-close COUNTER;
-
-
-
-
-