zoukankan      html  css  js  c++  java
  • keras+resnet实现车牌识别

    1.使用PIL和opencv生成车牌图像数据

    from PIL import ImageFont,Image,ImageDraw
    import cv2
    import numpy as np
    import os
    from math import *
    #创建 生成车牌图像数据 的类
    index = {"京": 0, "沪": 1, "津": 2, "渝": 3, "冀": 4, "晋": 5, "蒙": 6, "辽": 7, "吉": 8, "黑": 9, "苏": 10, "浙": 11, "皖": 12,
             "闽": 13, "赣": 14, "鲁": 15, "豫": 16, "鄂": 17, "湘": 18, "粤": 19, "桂": 20, "琼": 21, "川": 22, "贵": 23, "云": 24,
             "藏": 25, "陕": 26, "甘": 27, "青": 28, "宁": 29, "新": 30, "0": 31, "1": 32, "2": 33, "3": 34, "4": 35, "5": 36,
             "6": 37, "7": 38, "8": 39, "9": 40, "A": 41, "B": 42, "C": 43, "D": 44, "E": 45, "F": 46, "G": 47, "H": 48,
             "J": 49, "K": 50, "L": 51, "M": 52, "N": 53, "P": 54, "Q": 55, "R": 56, "S": 57, "T": 58, "U": 59, "V": 60,
             "W": 61, "X": 62, "Y": 63, "Z": 64}
     
    chars = ["京", "沪", "津", "渝", "冀", "晋", "蒙", "辽", "吉", "黑", "苏", "浙", "皖", "闽", "赣", "鲁", "豫", "鄂", "湘", "粤", "桂",
                 "琼", "川", "贵", "云", "藏", "陕", "甘", "青", "宁", "新", "0", "1", "2", "3", "4", "5", "6", "7", "8", "9", "A",
                 "B", "C", "D", "E", "F", "G", "H", "J", "K", "L", "M", "N", "P", "Q", "R", "S", "T", "U", "V", "W", "X",
                 "Y", "Z"
                 ]
    def r(val):
        return int(np.random.random()*val)
    def GenCh(f,val):#生成中文字符,f是中文字体对象,
        img=Image.new('RGB',(45,70),(255,255,255))#创建中文字符区域的全白画布,中文一般要方正一些,所以画布设置大一点,等下再resize
        draw=ImageDraw.Draw(img)#对画布创建画画对象
        draw.text((0,3),val,(0,0,0),font=f)#画画对象画出规定字体的黑色的规定val【0】文字,在左上角位置是(0,3)的点开始
        img =  img.resize((23,70))
        A=np.array(img)#图画转换成array格式
        return A
    def GenCh1(f,val):#生成英文和数字字符,f是中文字体对象,
        img=Image.new('RGB',(23,70),(255,255,255))
        draw=ImageDraw.Draw(img)
        draw.text((0,2),val,(0,0,0),font=f)#画画对象在左上角(0,2)出开始画出val,颜色全黑,字体是f
        A=np.array(img)
        return A
    def rot(img,angle,shape,max_angle):#透视畸变 
        size_o=[shape[1],shape[0]]#cv读的(h,w)换成[w,h]
        size=(shape[1]+int(shape[0]*cos((float(max_angle)/180)*3.14)),shape[0])#【变化后的w,h】
        interval=abs(int(sin(float(angle)/180)*3.14)*shape[0])#h变换的绝对值
        pts1=np.float32([[0,0],[0,size_o[1]],[size_o[0],0],[size_o[0],size_o[1]]])#(00,0h,w0,wh)
        if (angle>0):
            pts2=np.float32([[interval,0],[0,size[1]],[size[0],0],[size[0]-interval,size_o[1]]])
        else:
            pts2 = np.float32([[0,0],[interval,size[1]  ],[size[0]-interval,0  ],[size[0],size_o[1]]])
        M=cv2.getPerspectiveTransform(pts1,pts2)#根据两幅图的四个坐标点计算透视矩阵
        dst=cv2.warpPerspective(img,M,size)#img再M矩阵的变化下生成size大小的变化图
        return dst
    def rotRandrom(img,factor,shape):#仿射畸变
        pts1=np.float32([[0,0],[0,shape[0]],[shape[1],0],[shape[1],shape[0]]])#00,0h,w0,wh
        pts2=np.float32([[r(factor),r(factor)],[r(factor),shape[0]-r(factor)],[shape[1]-factor,r(factor)],[shape[1]-r(factor),shape[0]-r(factor)]])
        M=cv2.getPerspectiveTransform(pts1,pts2)
        dst=cv2.warpPerspective(img,M,shape)
        return dst
    def tfactor(img):#添加饱和度光照的噪声
        hsv=cv2.cvtColor(img,cv2.COLOR_BGR2HSV)
        hsv[:,:,0] = hsv[:,:,0]*(0.8+ np.random.random()*0.2)
        hsv[:,:,1] = hsv[:,:,1]*(0.3+ np.random.random()*0.7)
        hsv[:,:,2] = hsv[:,:,2]*(0.2+ np.random.random()*0.8)
        img=cv2.cvtColor(hsv,cv2.COLOR_HSV2BGR)
        return img
    def random_envirment(img,data_set):#添加自然环境噪声
        index=r(len(data_set))
        env=cv2.imread(data_set[index])#随机读出一张环境噪声图片
        env=cv2.resize(env,(img.shape[1],img.shape[0]))#改变环境噪声图片的大下
        bak=(img==0)
        bak=bak.astype(np.uint8)*255#图片的黑色部分变成白色部分
        inv=cv2.bitwise_and(bak,env)#图片的非黑部分和噪声求and,
        img=cv2.bitwise_or(inv,img)
        return img
    def AddGauss(img,level):#添加高斯模糊
        return cv2.blur(img,(level*2+1,level*2+1))
    def AddNoiseSingleChannel(single):#高斯噪声
        diff=255-single.max()
        noise=np.random.normal(0,1+r(6),single.shape)
        noise = (noise - noise.min())/(noise.max()-noise.min())
        noise= diff*noise
        noise= noise.astype(np.uint8)
        dst = single + noise
        return dst
    def addNoise(img,sdev = 0.5,avg=10):#每个通道随机添加高斯噪声
        img[:,:,0] =  AddNoiseSingleChannel(img[:,:,0])
        img[:,:,1] =  AddNoiseSingleChannel(img[:,:,1])
        img[:,:,2] =  AddNoiseSingleChannel(img[:,:,2])
        return img

    def AddSmudginess(img, Smu):#模拟车牌的污渍特征 rows = r(Smu.shape[0] - 50) cols = r(Smu.shape[1] - 50) adder = Smu[rows:rows + 50, cols:cols + 50]; adder = cv2.resize(adder, (50, 50)); #adder = cv2.bitwise_not(adder) img = cv2.resize(img,(50,50)) img = cv2.bitwise_not(img) img = cv2.bitwise_and(adder, img) img = cv2.bitwise_not(img)
    return img class GenPlate:#生成车牌图像数据 的类 def __init__(self,fontCh,fontEng,NoPlates): self.fontC=ImageFont.truetype(fontCh,43,0)#创建中文字体对象,fontCh是字体文件地址,43是字体大小,规定文字字体 self.fontE=ImageFont.truetype(fontEng,60,0) self.img=np.array(Image.new('RGB',(226,70),(255,255,255)))#创建(226,70)大小的全白图像,并转换成array格式 self.bg=cv2.resize(cv2.imread('./input_data/images/template.bmp'),(226,70))#读取出背景模板图 self.smu=cv2.imread('./input_data/images/smu2.jpg')#******************************** self.noplates_path=[] for parent,parent_folder,filenames in os.walk(NoPlates):#NoPlates的上级目录,NoPlates的子目录(没有),NoPlates的子文件 for filename in filenames: path=parent+'/'+filename self.noplates_path.append(path)#环境噪声图片 def draw(self,val): offset=2 self.img[0:70,offset+8:offset+8+23]=GenCh(self.fontC,val[0])#再self.img画布上画出中文字 self.img[0:79,offset+8+23+6:offset+8+23+6+23]=GenCh1(self.fontE,val[1])#英文字 for i in range(5):#画出5个数字 base=offset+8+23+6+23+17+ i*23+ i*6 self.img[0:70,base:base+23]=GenCh1(self.fontE,val[i+2]) return self.img#画出背景白色,文字黑色的文字内容 def generate(self,text): if len(text)==7: fg=self.draw(text)#调用draw函数,画出这7个字符) fg=cv2.bitwise_not(fg)#图像的array按位取反:黑色背景,白色文字 com=cv2.bitwise_or(fg,self.bg)#再与背景图片取或,则生成背景是背景图片,前景是白色文字的图片array格式 com=rot(com,r(60)-30,com.shape,30)#透视畸变效果(img,angle,shape,max_angle) com = rotRandrom(com,10,(com.shape[1],com.shape[0]))#仿射畸变 com = tfactor(com)#添加饱和度光照的噪声 com = random_envirment(com,self.noplates_path)#环境图片的噪声 com = AddGauss(com, 1+r(4))#高斯模糊 com = addNoise(com)#高斯噪声 return com def genPlateString(self,pos,val):#随机生成(中文 英文 数字*5)的字符 #pos!=-1时,读取出val值就是车牌值************************pos=-1时,val随机,主要是生成车牌 plateStr='' if (pos!=-1):# plateStr+=val#读出你想要生成的车牌号 else:#随机生成车牌号 for cpos in range(7): if cpos==0: plateStr+=chars[r(31)] elif cpos==1: plateStr+=chars[41+r(24)] else: plateStr+=chars[31+r(10)] return plateStr def genBatch(self,batchSize,size):#生成batch——size个车牌数据 for i in range(batchSize): plateStr=self.genPlateString(-1,-1) print(plateStr) img=self.generate(plateStr) img=cv2.resize(img,size) filename=str(plateStr)+'.jpg' cv2.imencode('.jpg',img)[1].tofile(filename)#此处解决中文名乱码 if __name__=='__main__': G=GenPlate("./input_data/font/platech.ttf",'./input_data/font/platechar.ttf',"./input_data/NoPlates") G.genBatch(10,(224,224))

    2.使用keras生成resnet34模型

    #2.构建网络模型
    from keras.models import Model
    from keras.layers import Input,Dropout,BatchNormalization,Conv2D,MaxPooling2D,AveragePooling2D,concatenate,Activation,ZeroPadding2D,add,Flatten,Dense
    import numpy as np
    seed=1
    np.random.seed(seed)
    import matplotlib.pyplot as plt
    %matplotlib inline
    characters="京沪津渝冀晋蒙辽吉黑苏浙皖闽赣鲁豫鄂湘粤桂琼川贵云藏陕甘青宁新0123456789ABCDEFGHJKLMNPQRSTUVWXYZ"
    width,height,n_len,n_classes=224,224,7,len(characters)
    def gen(batch_size=32):
        X = np.zeros((batch_size, height, width, 3), dtype=np.uint8)
        y = [np.zeros((batch_size, n_classes), dtype=np.uint8) for i in range(n_len)]
       # generator = ImageCaptcha(width=width, height=height)
        while True:
            for i in range(batch_size):
                a=G.genPlateString(-1,-1)
                b=G.generate(a)
                img = cv2.resize(b,(224,224))
                X[i] = img.astype('float32')
                for j, ch in enumerate(a):
                    y[j][i, :] = 0
                    y[j][i, characters.find(ch)] = 1
            yield X, y
    
    def decode(y):
        y = np.argmax(np.array(y), axis=2)[:,0]
        return ''.join([characters[x] for x in y])
     
    """
    使用者三段代码测试生成结果,以及解码结果
    X, y = next(gen(1))
    plt.imshow(X[0])
    print(decode(y))
    plt.title(decode(y))
    
    """
    def gen(batch_size=32):#生成数据
        x=np.zeros((batch_size,height,width,3),dtype=np.uint8)
        #y=[np.zeros((batch_size,n_classes),dtype=np.uint8) for i in range (n_len)]
        y=np.zeros((batch_size,n_len,n_classes),dtype=np.uint8)
        for i in range(batch_size):
            string=G.genPlateString(-1,1)
            img=G.generate(string)
            img=cv2.resize(img,(224,224))
            x[i]=img.astype('float32')
            for j,ch in enumerate(string):
                y[i][j,:]=0
                y[i][j,characters.find(ch)]=1
        yield x,y
    def decode(y):
    
        y=np.argmax(np.array(y),axis=2)[0]
        return ''.join([characters[x] for x in y])
    #x,y=next(gen(1))
    #plt.imshow(x[0])
    #print(decode(y))
    #plt.title(decode(y))
    #resnet34网络
    def Conv2d_BN(x,nb_filter,kernel_size,strides=(1,1),padding='same',name=None):
        if name is not None:
            bn_name=name+'_bn'
            conv_name=name+'_conv'
        else:
            bn_name=None
            conv_name=None
        x=Conv2D(nb_filter,kernel_size,padding=padding,strides=strides,activation='relu',name=conv_name)(x)
        x=BatchNormalization(name=bn_name)(x)
        return x
    def Conv_Block(input_layer,nb_filter,kernel_size,strides=(1,1),with_conv_shortcut=False):
        x=Conv2d_BN(input_layer,nb_filter=nb_filter,kernel_size=kernel_size,strides=strides,padding='same')
        x=Conv2d_BN(x,nb_filter=nb_filter,kernel_size=kernel_size,padding='same')
        if with_conv_shortcut:#残差边
            shortcut=Conv2d_BN(input_layer,nb_filter=nb_filter,strides=strides,kernel_size=kernel_size)
            x=add([x,shortcut])#残差边卷积后相加,缩小网络输出图像尺寸
            return x
        else:
            x=add([x,input_layer])#残差边直接相加,加深网络卷积深度
            return x
    input_x=Input(shape=(224,224,3))
    x=ZeroPadding2D((3,3))(input_x)
    x=Conv2d_BN(x,nb_filter=64,kernel_size=(7,7),strides=(2,2),padding='valid')
    x=MaxPooling2D(pool_size=(3,3),strides=(2,2),padding='same')(x)#56,56,64
    x = Conv_Block(x,nb_filter=64,kernel_size=(3,3))  
    x = Conv_Block(x,nb_filter=64,kernel_size=(3,3))  
    x = Conv_Block(x,nb_filter=64,kernel_size=(3,3))  #56,56,34
    x = Conv_Block(x,nb_filter=128,kernel_size=(3,3),strides=(2,2),with_conv_shortcut=True)  
    x = Conv_Block(x,nb_filter=128,kernel_size=(3,3))  
    x = Conv_Block(x,nb_filter=128,kernel_size=(3,3))  
    x = Conv_Block(x,nb_filter=128,kernel_size=(3,3)) #28,28,128
    x = Conv_Block(x,nb_filter=256,kernel_size=(3,3),strides=(2,2),with_conv_shortcut=True)  
    x = Conv_Block(x,nb_filter=256,kernel_size=(3,3))  
    x = Conv_Block(x,nb_filter=256,kernel_size=(3,3))  
    x = Conv_Block(x,nb_filter=256,kernel_size=(3,3))  
    x = Conv_Block(x,nb_filter=256,kernel_size=(3,3))  
    x = Conv_Block(x,nb_filter=256,kernel_size=(3,3))  #14,14,256
    x = Conv_Block(x,nb_filter=512,kernel_size=(3,3),strides=(2,2),with_conv_shortcut=True)  
    x = Conv_Block(x,nb_filter=512,kernel_size=(3,3))  
    x = Conv_Block(x,nb_filter=512,kernel_size=(3,3))#7,7,512  
    x = AveragePooling2D(pool_size=(7,7))(x) #1,1,512
    x=Flatten()(x)#512
    #x = Dense(1000,activation='softmax')(x)  
    x = [Dense(n_classes, activation='softmax', name='P%d'%(i+1))(x) for i in range(7)]#有7个输出结果,所以7个Dense层和x分别连接
    model = Model(inputs=input_x,outputs=x)  
    model.compile(loss='categorical_crossentropy',optimizer='sgd',metrics=['accuracy'])  
    model.summary() 
    model.fit_generator(gen(), epochs=1000, steps_per_epoch=20,validation_data=gen(),validation_steps=1)        
    #保存模型图
    from keras.utils import plot_model
    plot_model(model, to_file='model.png',show_shapes='True')
    #测试下结果
    X, y = next(gen(1))
    y_pred = model.predict(X)
    print(X)
    print("ddddd")
    print(X[0])
    print(decode(y))
    print(decode(y_pred))
    plt.title('real: %s
    pred:%s'%(decode(y), decode(y_pred)))
    plt.imshow(X[0], cmap='gray')
    #这里显示是和正常的颜色不一样,这是因为,plt读取的通道顺序和cv2的通道顺序是不同的
    #保存模型和参数
    model.save('resnet34_model.h5')
    score=model.evaluate(X,y,verbose=0)
    print("test score=",score[0])
    print("test accuracy=",score[1])
    

    3.利用模型进行实际预测输出

    #进行实际预测
    from keras.models import load_model
    model=load_model("resnet34_model.h5")
    print("导入模型完成")
    print("读取图片")
    #pic = Image.open("./宁R46974.jpg")
    #pic.show()
    img = cv2.imread('./x.jpg')#地址不能有中文
    img=img[np.newaxis,:,:,:]#图片是三维的但是训练时是转换成4维了所以需要增加一个维度
    predict = model.predict(img)
    print("车牌号为:",decode(predict))
    

    参考:https://github.com/Haveoneriver/License-Plate-Recognition-Items

    https://github.com/szad670401/end-to-end-for-chinese-plate-recognition

  • 相关阅读:
    (转)大数据量高并发的数据库优化与sql优化
    SQL Server查询优化方法参考(转)
    CString和LPCSTR区别(转)
    delphi读写剪贴板的一些参考
    Delphi和VC混合编程总结
    Delphi 一些函数解释
    伪共享与volatile
    happens-before原则
    递归与回溯-排立组合
    二叉树
  • 原文地址:https://www.cnblogs.com/Turing-dz/p/13288055.html
Copyright © 2011-2022 走看看