zoukankan      html  css  js  c++  java
  • Spark机器学习(12):神经网络算法

    1. 神经网络基础知识

    1.1 神经元

    神经网络(Neural Net)是由大量的处理单元相互连接形成的网络。神经元是神经网络的最小单元,神经网络由若干个神经元组成。一个神经元的结构如下:

    上面的神经元x1,x2,x3和1是输入,hw,b(x)是输出。

    其中f(x)是激活函数,常用的激活函数有sigmoid函数和tanh(双曲正切)函数。

    sigmoid函数:

    tanh(双曲正切)函数:

    1.2 神经网络

    神经网络由若干个层次,相邻层次之间的神经元存在输入的关系。第一层称为输入层,最后一层称为输出层,中间的层次称为隐含层。

    1.3 信号前向传播和误差反向传播

    设神经网络有n层,第1层为L1,第2层为L2,第n层为Ln,第p(p=1,2,...n)层的神经元节点数量是mp。aj(k)表示第k层第j个节点的输出值。则对于L1(也就是输入层),有

    第(k+1)层第j个神经元的输出

    设一个训练样本的误差为

    整体误差函数

     

    为了防止过拟合,增加了第二项L2正则化。

    目标是求(w,b),使得J(w,b)最小。为此使用梯度下降法,每次迭代按照下面的公式对w和b进行更新

    第n层(也就是输出层)的输出神经元j,其残差为

    第k层第i个节点的残差为

    求解(w,b)的过程如下:

    1) 对于所有的k,令w(k):=0,b(k):=0;

    2) 信号前向传播,根据每个样本的输入值和w(k)、w(k),逐层计算出hw,b(x);

    3) 误差反向传播,逐层计算出每一层每个神经元的残差;

    4) 对w和b的值进行更新。

    反复进行步骤(2)~(4),直到完成指定的迭代次数为止。

    2. MLlib神经网络的实现

    MLlib的神经网络类是NerualNet。主要参数包括:

    Size:Array[Int],神经网络每一层的节点数量;

    Layer:神经网络的层数;

    Activation_function:激活函数,可以是sigm或tanh

    Ouput_function:输出函数,可以是sigm、softmax或linear。

    代码:

    import org.apache.log4j.{ Level, Logger }
    import org.apache.spark.{ SparkConf, SparkContext }
    import breeze.linalg.{
    DenseMatrix => BDM,
    max => Bmax,
    min => Bmin
    }
    import scala.collection.mutable.ArrayBuffer
    
    /**
      * Created by Administrator on 2017/7/27.
      */
    object NNTest {
    
      def main(args: Array[String]) = {
        // 设置运行环境
        val conf = new SparkConf().setAppName("Neural Net")
          .setMaster("spark://master:7077").setJars(Seq("E:\Intellij\Projects\MachineLearning\MachineLearning.jar"))
        val sc = new SparkContext(conf)
        Logger.getRootLogger.setLevel(Level.WARN)
    
        // 随机生成样本数据
        Logger.getRootLogger.setLevel(Level.WARN)
        val sampleRow = 1000
        val sampleColumn = 5
        val randSamp_01 = RandSampleData.RandM(sampleRow, sampleColumn, -10, 10, "sphere")
        // 归一化
        val norMax = Bmax(randSamp_01(::, breeze.linalg.*))
        val norMin = Bmin(randSamp_01(::, breeze.linalg.*))
        val nor1 = randSamp_01 - (BDM.ones[Double](randSamp_01.rows, 1)) * norMin
        val nor2 = nor1 :/ ((BDM.ones[Double](nor1.rows, 1)) * (norMax - norMin))
        // 转换样本
        val randSamp_02 = ArrayBuffer[BDM[Double]]()
        for (i <- 0 to sampleRow - 1) {
          val mi = nor2(i, ::)
          val mi1 = mi.inner
          val mi2 = mi1.toArray
          val mi3 = new BDM(1, mi2.length, mi2)
          randSamp_02 += mi3
        }
        val randSamp_03 = sc.parallelize(randSamp_02, 10)
        sc.setCheckpointDir("hdfs://master:9000/ml/data/checkpoint")
        randSamp_03.checkpoint()
        val trainRDD = randSamp_03.map(f => (new BDM(1, 1, f(::, 0).data), f(::, 1 to -1)))
        // 训练,建立模型
        val opts = Array(100.0, 50.0, 0.0)
        trainRDD.cache
        val numExamples = trainRDD.count()
        println(s"Number of Examples: $numExamples")
        val NNModel = new NeuralNet().
          setSize(Array(5, 10, 10, 10, 10, 10, 1)).
          setLayer(7).
          setActivation_function("tanh_opt").
          setLearningRate(2.0).
          setScaling_learningRate(1.0).
          setWeightPenaltyL2(0.0).
          setNonSparsityPenalty(0.0).
          setSparsityTarget(0.05).
          setInputZeroMaskedFraction(0.0).
          setDropoutFraction(0.0).
          setOutput_function("sigm").
          NNtrain(trainRDD, opts)
    
        // 测试模型
        val NNPrediction = NNModel.predict(trainRDD)
        val NNPredictionError = NNModel.Loss(NNPrediction)
        println(s"NNerror = $NNPredictionError")
        val showPrediction = NNPrediction.map(f => (f.label.data(0), f.predict_label.data(0))).take(100)
        println("Prediction Result")
        println("Value" + "	" + "Prediction" + "	" + "Error")
        for (i <- 0 until showPrediction.length)
          println(showPrediction(i)._1 + "	" + showPrediction(i)._2 + "	" + (showPrediction(i)._2 - showPrediction(i)._1))
    
        var tmpWeight = NNModel.weights(0)
        for (i <-0 to 5) {
          tmpWeight = NNModel.weights(i)
          println(s"Weight of Layer ${i+1}")
          for (j <- 0 to tmpWeight.rows - 1) {
            for (k <- 0 to tmpWeight.cols - 1) {
              print(tmpWeight(j, k) + "	")
            }
            println()
          }
        }
    
      }
    
    }

    以上代码建立了一个7层的神经网络,各层的节点数量为Array(5, 10, 10, 10, 10, 10, 1),对Sphere函数进行了测试。

    运行结果:

    Number of Examples: 1000
    epoch: numepochs = 1 , Took = 17 seconds; Full-batch train mse = 0.066738, val mse = 0.000000.
    epoch: numepochs = 2 , Took = 12 seconds; Full-batch train mse = 0.069649, val mse = 0.000000.
    epoch: numepochs = 3 , Took = 10 seconds; Full-batch train mse = 0.055260, val mse = 0.000000.
    epoch: numepochs = 4 , Took = 10 seconds; Full-batch train mse = 0.016346, val mse = 0.000000.
    epoch: numepochs = 5 , Took = 9 seconds; Full-batch train mse = 0.013802, val mse = 0.000000.
    epoch: numepochs = 6 , Took = 13 seconds; Full-batch train mse = 0.045142, val mse = 0.000000.
    epoch: numepochs = 7 , Took = 7 seconds; Full-batch train mse = 0.031211, val mse = 0.000000.
    epoch: numepochs = 8 , Took = 7 seconds; Full-batch train mse = 0.016334, val mse = 0.000000.
    epoch: numepochs = 9 , Took = 9 seconds; Full-batch train mse = 0.013348, val mse = 0.000000.
    epoch: numepochs = 10 , Took = 7 seconds; Full-batch train mse = 0.017879, val mse = 0.000000.
    epoch: numepochs = 11 , Took = 7 seconds; Full-batch train mse = 0.012627, val mse = 0.000000.
    epoch: numepochs = 12 , Took = 7 seconds; Full-batch train mse = 0.018080, val mse = 0.000000.
    epoch: numepochs = 13 , Took = 7 seconds; Full-batch train mse = 0.016755, val mse = 0.000000.
    epoch: numepochs = 14 , Took = 7 seconds; Full-batch train mse = 0.012250, val mse = 0.000000.
    epoch: numepochs = 15 , Took = 7 seconds; Full-batch train mse = 0.044833, val mse = 0.000000.
    epoch: numepochs = 16 , Took = 7 seconds; Full-batch train mse = 0.024345, val mse = 0.000000.
    epoch: numepochs = 17 , Took = 7 seconds; Full-batch train mse = 0.039005, val mse = 0.000000.
    epoch: numepochs = 18 , Took = 7 seconds; Full-batch train mse = 0.012298, val mse = 0.000000.
    epoch: numepochs = 19 , Took = 7 seconds; Full-batch train mse = 0.012371, val mse = 0.000000.
    epoch: numepochs = 20 , Took = 6 seconds; Full-batch train mse = 0.014077, val mse = 0.000000.
    epoch: numepochs = 21 , Took = 7 seconds; Full-batch train mse = 0.040328, val mse = 0.000000.
    epoch: numepochs = 22 , Took = 6 seconds; Full-batch train mse = 0.036575, val mse = 0.000000.
    epoch: numepochs = 23 , Took = 6 seconds; Full-batch train mse = 0.033986, val mse = 0.000000.
    epoch: numepochs = 24 , Took = 6 seconds; Full-batch train mse = 0.026421, val mse = 0.000000.
    epoch: numepochs = 25 , Took = 6 seconds; Full-batch train mse = 0.036776, val mse = 0.000000.
    epoch: numepochs = 26 , Took = 6 seconds; Full-batch train mse = 0.011838, val mse = 0.000000.
    epoch: numepochs = 27 , Took = 6 seconds; Full-batch train mse = 0.010749, val mse = 0.000000.
    epoch: numepochs = 28 , Took = 6 seconds; Full-batch train mse = 0.012717, val mse = 0.000000.
    epoch: numepochs = 29 , Took = 6 seconds; Full-batch train mse = 0.011883, val mse = 0.000000.
    epoch: numepochs = 30 , Took = 7 seconds; Full-batch train mse = 0.010562, val mse = 0.000000.
    epoch: numepochs = 31 , Took = 6 seconds; Full-batch train mse = 0.010591, val mse = 0.000000.
    epoch: numepochs = 32 , Took = 6 seconds; Full-batch train mse = 0.010389, val mse = 0.000000.
    epoch: numepochs = 33 , Took = 6 seconds; Full-batch train mse = 0.015908, val mse = 0.000000.
    epoch: numepochs = 34 , Took = 6 seconds; Full-batch train mse = 0.012413, val mse = 0.000000.
    epoch: numepochs = 35 , Took = 6 seconds; Full-batch train mse = 0.010442, val mse = 0.000000.
    epoch: numepochs = 36 , Took = 6 seconds; Full-batch train mse = 0.056686, val mse = 0.000000.
    epoch: numepochs = 37 , Took = 6 seconds; Full-batch train mse = 0.054850, val mse = 0.000000.
    epoch: numepochs = 38 , Took = 6 seconds; Full-batch train mse = 0.019422, val mse = 0.000000.
    epoch: numepochs = 39 , Took = 6 seconds; Full-batch train mse = 0.016443, val mse = 0.000000.
    epoch: numepochs = 40 , Took = 6 seconds; Full-batch train mse = 0.010289, val mse = 0.000000.
    epoch: numepochs = 41 , Took = 7 seconds; Full-batch train mse = 0.022615, val mse = 0.000000.
    epoch: numepochs = 42 , Took = 6 seconds; Full-batch train mse = 0.010723, val mse = 0.000000.
    epoch: numepochs = 43 , Took = 6 seconds; Full-batch train mse = 0.010289, val mse = 0.000000.
    epoch: numepochs = 44 , Took = 6 seconds; Full-batch train mse = 0.033933, val mse = 0.000000.
    epoch: numepochs = 45 , Took = 7 seconds; Full-batch train mse = 0.030156, val mse = 0.000000.
    epoch: numepochs = 46 , Took = 7 seconds; Full-batch train mse = 0.022068, val mse = 0.000000.
    epoch: numepochs = 47 , Took = 7 seconds; Full-batch train mse = 0.029382, val mse = 0.000000.
    epoch: numepochs = 48 , Took = 6 seconds; Full-batch train mse = 0.021275, val mse = 0.000000.
    epoch: numepochs = 49 , Took = 6 seconds; Full-batch train mse = 0.039427, val mse = 0.000000.
    epoch: numepochs = 50 , Took = 7 seconds; Full-batch train mse = 0.016674, val mse = 0.000000.
    NNerror = 0.016674267332022572
    Prediction Result
    Value    Prediction    Error
    0.6048934040010798    0.19097551722554007    -0.41391788677553976
    0.5917463309959767    0.35726681238891195    -0.23447951860706479
    0.5798180746808277    0.19232727566724744    -0.38749079901358024
    0.39808885303777447    0.1926440400752866    -0.20544481296248787
    0.4140924247674261    0.19529777426853168    -0.2187946504988944
    0.08847408598189055    0.19110126347316514    0.10262717749127459
    0.3583460134199821    0.21170344602417424    -0.1466425673958079
    0.29635258460747904    0.3549086780038481    0.05855609339636908
    0.21947238532147648    0.19156569159762857    -0.02790669372384791
    0.5357166982629155    0.36018248221537214    -0.17553421604754332
    0.5547810234563126    0.1912501730851674    -0.36353085037114524
    0.40529948654006304    0.21826323152039923    -0.1870362550196638
    0.4765320387665492    0.34409113646061484    -0.13244090230593436
    0.05759629179315594    0.1914373047341408    0.13384101294098488
    0.25415182638221206    0.29169483353745973    0.037543007155247665
    0.2731217394258585    0.19452719525740314    -0.07859454416845535
    0.021103715077802527    0.19131792203441428    0.17021420695661174
    0.24098254783013137    0.334302879677641    0.09332033184750962
    0.6300811731076671    0.3595001582783692    -0.2705810148292979
    0.41827613603130404    0.195477735057971    -0.22279840097333303
    0.2526404805902617    0.1945578268820965    -0.05808265370816518
    0.16619916368077442    0.191265206532793    0.025066042852018577
    0.007724491831775392    0.1909446242319318    0.1832201324001564
    0.08926696720959378    0.19197139383958237    0.10270442662998859
    0.4822857005955674    0.19244393418394434    -0.28984176641162307
    0.12166559083216193    0.19242076231047756    0.07075517147831563
    0.2883494676971952    0.30939742289582284    0.02104795519862762
    0.38817298742061984    0.1909921814285587    -0.19718080599206114
    0.34588396966368695    0.1957690915303307    -0.15011487813335625
    0.19958641570784796    0.19348928314854685    -0.0060971325593011105
    0.31340425691874024    0.19828489007869293    -0.11511936684004731
    0.31775749422734    0.19211592601952254    -0.12564156820781747
    0.48789392695999645    0.19120722177454247    -0.296686705185454
    0.4359840834351843    0.3604340050247724    -0.07555007841041189
    0.17359981155470314    0.1914334455263964    0.01783363397169327
    0.3629355770221922    0.2004476969345776    -0.1624878800876146
    0.4627621372503198    0.2111988404691097    -0.2515632967812101
    0.49652077030838826    0.19101585452942166    -0.3055049157789666
    0.12618599928245963    0.19939585850613975    0.07320985922368012
    0.45276204270081455    0.1924159942977412    -0.26034604840307335
    0.2837721853443281    0.2016124468403725    -0.08215973850395558
    0.34590164213713354    0.3601210376231753    0.014219395486041786
    0.1961497656762427    0.19408639222665872    -0.0020633734495839884
    0.22135763175909048    0.27616370537642354    0.054806073617333056
    0.43356473411523927    0.19150317510575426    -0.242061559009485
    0.09566706862199378    0.19087327269062435    0.09520620406863056
    0.29830626566849494    0.19959705355592236    -0.09870921211257258
    0.3070532379895792    0.34322116560057725    0.036167927610998074
    0.07052673330364767    0.19118739087384276    0.12066065757019509
    0.5501181200918814    0.2024015375945202    -0.34771658249736126
    0.31894277127298554    0.1917670097886867    -0.12717576148429885
    0.08585450906008718    0.20848620726607436    0.12263169820598718
    0.20245657700014166    0.19218060734066275    -0.010275969659478912
    0.1712767340967007    0.1913375355437103    0.020060801447009613
    0.3779192242827297    0.2035011707996587    -0.17441805348307102
    0.241909871430447    0.19089783315658176    -0.051012038273865246
    0.40578032667620945    0.3561807946562045    -0.049599532020004944
    0.20834390560196567    0.19103138812628986    -0.017312517475675804
    0.49675932490421343    0.1915234454414188    -0.3052358794627946
    0.2342257039800733    0.1920213029433058    -0.04220440103676751
    0.18045883957051312    0.20420037376704497    0.023741534196531855
    0.2309607430153665    0.1912620988835584    -0.0396986441318081
    0.40644947116571745    0.19173032204451546    -0.214719149121202
    0.11691561072493983    0.19280159115148832    0.07588598042654848
    0.05696889589626215    0.19083593927270395    0.13386704337644179
    0.47164532559761124    0.2506614607550888    -0.22098386484252242
    0.6208470748110626    0.35822718159638256    -0.26261989321468004
    0.46559040490785325    0.2083058633813562    -0.25728454152649705
    0.5052214973114583    0.19867901868911944    -0.30654247862233885
    0.4127229537166962    0.35982142534462497    -0.05290152837207124
    0.16960925650784137    0.19135483819811452    0.021745581690273158
    0.19722393334464125    0.19080547758699506    -0.006418455757646185
    0.4335762052660574    0.20920751654239156    -0.22436868872366583
    0.1496556423910719    0.19090570957335065    0.04125006718227875
    0.3015215343928844    0.1922591754560472    -0.10926235893683722
    0.0    0.19143943303505245    0.19143943303505245
    0.36555981464056164    0.19189800228180368    -0.17366181235875797
    0.3963164889187304    0.19451555510717428    -0.20180093381155612
    0.313325868748335    0.19168776655589245    -0.12163810219244256
    0.5034713123520999    0.3339326133308013    -0.16953869902129853
    0.4224576693623929    0.3539965263782299    -0.06846114298416295
    0.08523050351506854    0.19132247714662606    0.10609197363155752
    0.26080914691197654    0.19095418139777426    -0.06985496551420228
    0.1324640982588358    0.19304336020222349    0.0605792619433877
    0.13055674031551295    0.19224589387242375    0.06168915355691079
    0.23625412018106318    0.1917614371123628    -0.044492683068700384
    0.5019570376831385    0.3081554524341633    -0.19380158524897517
    0.030390738837917763    0.19083521852879112    0.16044447969087336
    0.34274552561551896    0.19120112478612986    -0.1515444008293891
    0.4514974655646171    0.1916124319598441    -0.259885033604773
    0.531777023034474    0.3515396924867077    -0.18023733054776636
    0.367772718094668    0.3317275143536775    -0.03604520374099052
    0.41600472261866916    0.22278029398575255    -0.1932244286329166
    0.36506543552315474    0.325628070833062    -0.039437364690092735
    0.314008782918081    0.32408907795815034    0.010080295040069354
    0.2925887989109779    0.1921158349811155    -0.10047296392986238
    0.4658619691058588    0.2146831464338164    -0.2511788226720424
    0.2280242270958607    0.19158705334902099    -0.03643717374683972
    0.5003581077100195    0.19149431703681175    -0.3088637906732078
    0.4448442553362914    0.19086228548828346    -0.2539819698480079
    Weight of Layer 1
    1.3741710823232989    1.0997962988037757    -2.3077515870713716    2.0946962013291297    2.24588083756021    0.7186952475525394    
    -1.1885813301306254    0.25046447165487246    -1.253986920617667    1.535570764339994    0.1440090623878452    1.2110656874633237    
    -0.23784821158321864    -0.5133767761738681    0.5355594752965599    -0.9862762256909807    2.234245108441277    -0.5216923380767392    
    2.0496153507146033    -0.9000455162282417    1.3406201642695788    2.1185256789014897    1.038387978643167    -0.011886136436036997    
    2.4017180810229086    0.5342060426581219    2.188686239727936    -0.604587031465719    0.061697537675081446    -0.48429030459304306    
    -1.234468451038262    0.7790398934631602    -0.22067594788975725    -2.0414139797176176    -0.9324514648411226    0.798505045375407    
    0.843464180836847    1.8612698144445792    2.290144904438349    1.2291878648431667    2.3639566790099784    -2.1175568466779437    
    2.1488480623696975    2.253851655104785    -1.879142801282798    -0.23011616258273254    2.4342506675413413    -2.184430097374211    
    1.3446335417651347    0.39411399422872706    -1.4588967794444714    2.6567285233270366    -0.8576819932211762    -1.9914472547514181    
    1.4277752508742856    0.6379599194760166    -0.3783031968398195    1.4158689111045317    1.5318358789872808    -1.201612551759085    
    Weight of Layer 2
    -1.264966034912716    2.045363428231917    0.39087016834115496    1.0930481252787911    -1.571245354275874    -0.9655062462170442    2.1709800176902982    -1.025175316866544    -1.5230797088149843    1.769571487127593    0.2347823358786302    
    -0.7297761283100554    -0.7576138927387723    -0.16523415352888424    -1.8516805610014189    1.3715533800487367    1.95737270209438    0.08246852496952446    1.0190398538209786    0.38679398113645574    0.529334650423367    0.43356573369591683    
    -1.494114816211369    0.47441528091212    -0.4329092188524354    1.8318597886945986    1.7458297556728688    -2.0957897682487707    -0.12195567981037991    0.0378548041005975    -0.6616871154791367    -1.2988919749370011    -2.0146727073983346    
    -2.0369675326219587    -0.6502946004953208    -0.5031500283188425    1.2033687522496184    -0.35900710238718897    -1.213965816985491    -0.6247306519040674    -1.172636102729738    0.1492977187034359    0.7805087252939967    0.42756349073372346    
    1.9444610499195103    -1.684949676049153    -2.214370792742285    1.4529568002779754    -0.817314868362314    -2.0437576009396534    1.414190131329268    0.8002930022447152    0.5103985267247684    1.072725074367445    0.7728306875824079    
    1.3280098832755556    -0.5723175992914568    1.8965322712004982    1.9131475857947766    0.7116308703102919    -1.1337078046228692    1.4624037563264591    -1.2661436485440505    0.2074359557718086    1.7605365810734208    0.5503309132836761    
    -0.12889849968368497    -1.5813673816630978    0.3179522409763945    -1.4093501587164268    0.37747278027064335    2.0973248562810287    1.3796729017078182    -1.32247724141541    0.05176617793309023    0.2797968400006565    0.2649190482622152    
    0.3553074482253876    -1.236624823614366    -1.7608721088406603    1.0024974314351445    -2.174840973945128    -0.6578975450189226    2.301788638015837    1.1479110741419556    -0.3844810208733819    -1.757161551068665    0.005888873948249745    
    1.0889152043326655    -0.4907523692459491    -1.5314215158890179    -2.068119203282905    -0.6966652421295678    1.7172836960101752    -1.030902225947676    1.7894526684520087    1.5213695041770878    -2.0310793395869613    -0.23638299504893515    
    -0.9900330517352804    -1.0052522373970076    -0.25640336055482327    0.5747216493939299    -0.6727702236294272    -0.0968976241335074    1.828632192751469    -1.2027447051840479    1.462449341909922    1.866886724932346    0.337038328583328    
    Weight of Layer 3
    1.836452591890687    -1.1016514469136454    1.9551193591759655    -1.5732459452330396    -1.1800834414335224    -1.1923893040285825    0.06312897816177482    1.457981772832194    1.7288840730921886    1.0295473067747805    -0.6835578234742354    
    -1.281488825782123    1.9855453373804615    -0.21337154884305315    -0.8204219180246067    -0.8260103421573441    -1.3974890573670238    -0.18789338539658415    1.5852967650612315    -0.9475186470859063    -1.0100358806860719    1.0069697324461917    
    0.7336707009234793    -0.8190101361063539    -1.9821987263863545    -2.0789438096783956    -1.1812929830305807    -0.12007384037833814    -1.8876632918570322    -0.8968914717818263    1.3471019678011835    -0.152196929664138    -0.9273968120578239    
    -0.5427506117781439    -1.6379161635634556    1.6603536697042776    -0.2826332939784261    -1.7089295991871511    -0.3086776080327076    -0.838637438570992    0.45911079203796323    0.26754055606009786    -1.8482009521236713    -0.23130366979329475    
    -0.37585662596248814    0.4900765346436331    0.6860109049091877    -0.9572551404136589    0.30380993421860114    -1.73380834254428    -0.9187544507012596    -2.143513804630084    1.9614638862521159    0.513675117374747    1.7364284348955028    
    1.4300160181482728    -0.028644837063659466    0.9434705098905873    -1.256117787020064    -2.286119449246773    0.8392807137479158    1.0463524997789728    0.22524746294136241    -2.065346398428731    -1.3984782782329925    0.8064849904915904    
    -0.020036943079112954    1.3571828713761132    1.8168106691174883    1.1732287314521193    -1.1216050426635968    -0.7992471983374778    1.4737000544885857    -1.2629467521609772    -2.166720398969067    -0.13444189253030867    -0.06990910534870051    
    -1.1475378284991893    0.7939976065553747    -0.6034315465733754    1.2609824176884306    -1.0556544940783124    -1.2600392532679041    0.20515032057156965    -0.9368115470081553    1.8486189613353239    -1.748953620555969    0.5962310435917523    
    -1.6677307685118707    1.6475809798433472    0.8635630357973535    -1.794106867032129    0.23576724690928239    1.9345586718254242    1.8665220736111652    1.715231095424259    1.4448153663136947    1.4585220061056003    -2.097657471713302    
    0.8605510379359859    0.48221398085417627    1.5176826373865975    1.4652552486449435    1.9094578768378816    1.1144707218150902    0.4891850304975143    -1.6217012198757983    -0.26648664353939455    -0.702859436768092    1.0351022938433367    
    Weight of Layer 4
    -0.911096181634767    1.393613461284606    -0.9330396913369214    1.2364172665419595    0.6539246160388238    -0.3497594712315348    -0.5272926339150141    1.0073562856426537    -0.7754136627565675    0.3978221698601255    0.6516969484538829    
    -2.1313935565863322    0.6870222803829747    -0.14063545929924332    -1.4601798330750633    -0.5046337368523596    0.45490880803140493    -0.3469665791558201    -1.1242340473155688    -1.2723993607026778    0.7754297084802044    -0.1971882092105515    
    1.268461391063688    1.433232422744769    -1.9734642121928314    -2.1500226866603094    1.69187327087415    -0.5342565995220732    -1.8236939987229444    -0.9248308295424809    -1.0585137171567485    -0.758573013381397    -1.4786474341672335    
    -1.3004208980147893    -1.4206386974340788    -1.83226719079787    1.853287649173723    -1.5346566916523345    0.18937676571029147    -2.14876104739136    0.19300765463829608    2.2600431704632133    -0.7439149441396848    -0.7944037931661944    
    -1.5998484896496197    -1.6524100951400051    1.5947517659570918    -1.8662327423253926    1.2338013894973758    -0.4884517327249603    -1.933617967710128    -0.16207327398832927    -0.257515806279318    0.8147979296307275    1.3995976816521387    
    2.094656544970861    -0.9941271469702547    1.6445550567163796    -1.8006734384410419    -1.639945317843356    -0.04008156577686328    -1.5983003403214098    -0.5203002571128433    1.7875581773282712    -1.094555900585581    1.2068765635018603    
    2.1254252250165493    -0.32295734337741594    -1.3909813447797978    -1.7675398403329456    1.6196489153696365    -0.11330208718956442    1.2249494342821745    -0.9623282852038381    0.5949871990921731    -0.4589253834864904    -1.4480879152384845    
    0.41124565810833846    1.8266387733496516    -1.3444906678350586    -1.8466017258629779    -1.2710822759712612    -2.3672244544192775    0.07985138355524748    1.1928505759997003    0.1224150977985233    1.581079240096574    -1.800780284328034    
    -1.7595274297258634    1.2703227203108252    -0.01664878904783711    -1.9795940886734285    -0.020392614773247077    -1.4314141516158432    -1.365137154825855    1.7302923882870662    1.9823859978980145    0.35232814451148275    0.7343215278791788    
    1.394419951767305    1.2499547753082167    -1.549641575886935    1.848772251833023    0.7361855983730334    0.6310928126046181    -0.8813463668969193    0.004308048034432659    0.43321638933450207    0.8087966273251945    0.07110982269880414    
    Weight of Layer 5
    0.03585680860332078    0.6094562888567525    1.7968463081634347    -2.362627715032082    2.2247821225164652    -1.0784288939947677    -1.0554498524378448    0.5861411371363056    -1.3689382763820177    -1.6272403702710627    -1.6503549168843508    
    1.6301987233118784    -0.8906914047579433    0.24243168770601023    0.4009515829010596    -1.0758568810272509    -0.9051235220484805    -2.2014662058001826    -0.6016562255004753    -1.54618907737507    0.3068858681819345    -0.1910939351954038    
    0.3949377293208745    -0.7652271919004981    0.68090033725971    1.2348269143656605    2.197797374061151    0.7456474083783481    2.177221715481275    0.577448303135589    0.6713302440403498    -1.422241156275792    0.3923309252775803    
    2.016732930847877    0.9072966194060744    -1.173786606850799    0.762874680679405    0.5785929523331224    -0.2517270024395546    2.0525009570621475    1.3299906196120428    -0.3016789963343337    1.6433222480576049    1.4658027315448252    
    -0.7093708542273217    -0.363654080609567    0.14670807608588535    1.2229445344521144    2.0365363421196743    -0.025435674952313806    1.1537986276254326    -1.2324297702242806    0.7761473466711113    0.8799485068219668    -0.5873930067918902    
    0.3139747558740749    -1.6573144697633098    0.7038102951541025    -1.408818539999227    2.097157257533922    -0.7036366036514073    0.4160486916023455    0.8262172799352652    -1.4795358417499833    0.2852056896956041    -1.542169273346858    
    -1.3692080951184145    -0.18564381183893783    1.4773933766691916    1.566187371091047    2.2303361972935196    1.6867117590655034    -2.2562477027382437    -0.7074433661402227    0.31697668962943404    -0.7387985710568243    -1.1533917617505614    
    -1.2255743400186403    -1.0127411060516087    -1.3756205847124954    1.930937272879095    0.5512007768312437    -1.8860525843110458    1.83432047092595    1.7499230942835498    -0.05470238124314854    0.1415405841710963    0.5347734456158572    
    0.8685622790833061    1.0117880568953437    -1.0680283034993974    -0.6423950104042628    2.0957313900176207    -0.3292735667051877    0.4115339082100468    0.2448817017887727    -0.36690487429065827    1.0946609803320706    -1.2972731428065445    
    0.929738769296319    -0.8483315095032794    0.9886368647914796    -1.1490945738647198    0.48817906098502184    1.201937687948849    -1.8405795878382836    1.6127096671527423    -2.0480015245423417    -0.9757299992342688    0.5211781810863436    
    Weight of Layer 6
    -1.1548299180301753    -1.6001306147116388    -2.387282014077577    -1.082677370520492    -0.013943138965433734    0.10533899958501511    0.5742412321742517    -1.5014155245560539    0.8057997937824102    1.8479652037781695    -0.23508934192649694    
  • 相关阅读:
    Python编程题24--回文数
    Python编程题23--移动零
    Python编程题22--只出现一次的数字
    Python编程题21--每日温度
    Python编程题20--最大子序和
    Python编程题19--比特位计数
    推荐系统中对比实验工具--伯乐bole
    pytorch中反向传播的loss.backward(retain_graph=True)报错
    win10彻底禁止自动更新(家庭版操作)
    linux使用清华源安装工具,切换清华源
  • 原文地址:https://www.cnblogs.com/mstk/p/7245832.html
Copyright © 2011-2022 走看看