CNN 배치 처리

IngCoding·2022년 6월 28일
1

머신러닝

목록 보기
32/34

내용출처 : 이수안컴퓨터연구소 Youtube

import cv2
import numpy as np 
import matplotlib.pyplot as plt
import urllib
import requests
from io import BytesIO

1. CNN 함수 정의

def url_to_image(url, gray=False): # url 이미지 불러오기
  resp = urllib.request.urlopen(url)
  image = np.asarray(bytearray(resp.read()), dtype="uint8")

  if gray == True:
    image = cv2.imdecode(image, cv2.IMREAD_GRAYSCALE)
  else:
    image = cv2.imdecode(image, cv2.IMREAD_COLOR)
    image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)

  return image
# 이미지 필터
def filtered_image(image, filter, output_size):
  filtered_img = np.zeros((output_size, output_size))
  filter_size = filter.shape[0]

  for i in range(output_size):
    for j in range(output_size):
      multiply_values = image[i:(i + filter_size), j:(j+filter_size)] * filter
      sum_value = np.sum(multiply_values)

      if (sum_value > 255):
        sum_value = 255

      filtered_img[i, j] = sum_value

  return filtered_img
# 이미지를 2차원 배열로 변환해주는 함수
def im2col(input_data, filter_h, filter_w, stride=1, pad=0): 

  N, C, H, W = input_data.shape
  out_h = (H + 2*pad - filter_h) // stride + 1
  out_w = (W + 2*pad - filter_w) // stride + 1

  img = np.pad(input_data, [(0,0), (0,0), (pad, pad), (pad, pad)], 'constant')
  col = np.zeros((N, C, filter_h, filter_w, out_h, out_w))

  for y in range(filter_h):
    y_max = y + stride * out_h
    for x in range(filter_w):
      x_max = x + stride * out_w
      col[:, :, y, x, :, :] = img[:, :, y:y_max:stride, x:x_max:stride]

  col = col.transpose(0, 4, 5, 1, 2, 3).reshape(N * out_h * out_w, -1)
  return col 

# 2차원 배열을 이미지로 변환해주는 함수
def col2im(col, input_shape, filter_h, filter_w, stride=1, pad=0):

  N, C, H, W = input_shape
  out_h = (H + 2*pad - filter_h) // stride + 1
  out_w = (W + 2*pad - filter_w) // stride + 1
  col = col.reshape(N, out_h, out_w, C, filter_h, filter_w).transpose(0, 3, 4, 5, 1, 2)

  img = np.zeros((N, C, H + 2*pad + stride - 1, W + 2*pad + stride -1))
  for y in range(filter_h):
    y_max = y + stride * out_h
    for x in range(filter_w):
      x_man = x + stride * out_w
      img[:, :, y:y_max:stride, x:x_man:stride] = img[:, :, y, x, :, :]

  return img[:, :, pad:H + pad, pad:W + pad]
class conv2D: #
  def __init__(self, W, b, stride=1, pad=0):
    self.W = W
    self.b = b
    self.stride = stride
    self.pad = pad

    self.input_data = None
    self.col = None
    self.col_w = None

    self.dW = None
    self.db = None

  def forward(self, input_data): # 순전파
    FN, C, FH, FW = self.W.shape
    N, C, H, W = input_data.shape
    out_h = (H + 2*self.pad - FH) // self.stride + 1
    out_w = (W + 2*self.pad - FW) // self.stride + 1

    col = im2col(input_data, FH, FW, self.stride, self.pad )
    col_W = self.W.reshape(FN, -1).T

    out = np.dot(col, col_W) + self.b
    output = out.reshape(N, out_h, out_w, -1).transpose(0, 3, 1, 2)

    self.input_data = input_data
    self.col = col 
    self.col_W = col_W
    
    return output

  def backward(self, dout): # 역전파
    FN, C, FH, FW = self.W.shape
    dout = dout.transpose(0, 2, 3, 1).reshape()

    self.db = np.sum(dout, axis=0)
    self.dW - np.dot(self.col.T, dout)
    self.dW = self.dw.transpose(1, 0).reshape(FN, C, FH, FW)

    dcol = np.dot(dout, self.col_W.T)
    dx = col2im(dcol, self.input_data.shape, FH, FW, self.stride, self.pad)

    return dx
def init_weight(num_filters, data_dim, kernel_size, stride=1, pad=0, weight_std=0.01):

  weights = weight_std * np.random.randn(num_filters, data_dim, kernel_size, kernel_size)
  biases = np.zeros(num_filters)

  return weights, biases

2. 동일한 이미지 여러장 테스트 (배치처리, gray)

img_url = "https://upload.wikimedia.org/wikipedia/ko/thumb/2/24/Lenna.png/440px-Lenna.png"
image_gray = url_to_image(img_url, gray=True)
image_gray = image_gray.reshape(image_gray.shape[0], -1, 1)
print("image.shape:", image_gray.shape)

image_gray = image_gray.transpose(2, 0, 1)
print("image_gray.shape: ", image_gray.shape)
image.shape: (440, 440, 1)
image_gray.shape:  (1, 440, 440)
batch_image_gray = np.repeat(image_gray[np.newaxis, :, :, :], 15, axis=0)
print(batch_image_gray.shape)
(15, 1, 440, 440)
W, b = init_weight(10, 1, 3, stride=2)
conv = conv2D(W, b)
output = conv.forward(batch_image_gray)

print("Conv Layer size: ", output.shape)
Conv Layer size:  (15, 10, 438, 438)
plt.figure(figsize=(10,10))

plt.subplot(1,3,1)
plt.title("Filter 3")
plt.imshow(output[3, 2, :, :], cmap='gray')

plt.subplot(1, 3, 2)
plt.title("Filter 6")
plt.imshow(output[3, 5, :, :], cmap='gray')

plt.subplot(1, 3, 3)
plt.title("Filter 9")
plt.imshow(output[3, 9, :, :], cmap='gray')

plt.show()

3. 동일한 이미지 배치처리(color)

img_url = "https://upload.wikimedia.org/wikipedia/ko/thumb/2/24/Lenna.png/440px-Lenna.png"
image_color = url_to_image(img_url)
print("image.shape:", image_color.shape)

image_color = image_color.transpose(2, 0, 1)
print("image_color.shape: ", image_color.shape)
image.shape: (440, 440, 3)
image_color.shape:  (3, 440, 440)
batch_image_color = np.repeat(image_color[np.newaxis, :, :, :], 15, axis=0)
print(batch_image_color.shape)
(15, 3, 440, 440)
W2, b2 = init_weight(64, 3, 5)
conv2 = conv2D(W2, b2)
output2 = conv2.forward(batch_image_color)

print("Conv Layer size:", output2.shape)
Conv Layer size: (15, 64, 436, 436)
plt.figure(figsize=(10,10))

plt.subplot(1,3,1)
plt.title("Filter 50")
plt.imshow(output2[10, 49, :, :])

plt.subplot(1, 3, 2)
plt.title("Filter 31")
plt.imshow(output2[10, 30, :, :])

plt.subplot(1, 3, 3)
plt.title("Filter 1")
plt.imshow(output2[10, 0, :, :])

plt.show()

profile
Data & PM

0개의 댓글