A股上市公司传智教育(股票代码 003032)旗下技术交流社区北京昌平校区

 找回密码
 加入黑马

QQ登录

只需一步,快速开始

简介和CelebA类似,使用ACGAN生成二次元头像,并且用多个条件进行控制
数据图片爬取自getchu.com/,是一个日本二次元游戏网站,包含大量游戏人物立绘,共爬取31,970张


头像截取之前介绍的dlib可用于提取人脸,但不适用于二次元头像
使用OpenCV从每张图片中截取头像部分,用到以下项目,github.com/nagadomi/lb…
对于检测结果适当放大范围,以包含更多人物细节
# -*- coding: utf-8 -*-import cv2cascade = cv2.CascadeClassifier('lbpcascade_animeface.xml')image = cv2.imread('imgs/二次元头像示例.jpg')gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)gray = cv2.equalizeHist(gray)faces = cascade.detectMultiScale(gray, scaleFactor=1.1, minNeighbors=5, minSize=(64, 64))for i, (x, y, w, h) in enumerate(faces):    cx = x + w // 2    cy = y + h // 2    x0 = cx - int(0.75 * w)    x1 = cx + int(0.75 * w)    y0 = cy - int(0.75 * h)    y1 = cy + int(0.75 * h)    if x0 < 0:        x0 = 0    if y0 < 0:        y0 = 0    if x1 >= image.shape[1]:        x1 = image.shape[1] - 1    if y1 >= image.shape[0]:        y1 = image.shape[0] - 1    w = x1 - x0    h = y1 - y0    if w > h:        x0 = x0 + w // 2 - h // 2        x1 = x1 - w // 2 + h // 2        w = h    else:        y0 = y0 + h // 2 - w // 2        y1 = y1 - h // 2 + w // 2        h = w    face = image[y0: y0 + h, x0: x0 + w, :]    face = cv2.resize(face, (128, 128))    cv2.imwrite('face_%d.jpg' % i, face)复制代码

标签提取使用Illustration2Vec从二次元图片中抽取丰富的标签,github.com/rezoo/illus…
Illustration2Vec用到chainer这个深度学习框架,以及一些其他库,如果没有则安装
pip install chainer Pillow scikit-image复制代码Illustration2Vec可以完成以下三项功能:
  • 将每张图片表示为一个4096维的向量
  • 指定阈值,并提取概率高于阈值的标签
  • 指定一些标签,并返回对应的概率
举个例子,提取全部可能的标签,以0.5为阈值
# -*- coding: utf-8 -*-import i2vfrom imageio import imreadillust2vec = i2v.make_i2v_with_chainer('illust2vec_tag_ver200.caffemodel', 'tag_list.json')img = imread('imgs/二次元头像示例.jpg')tags = illust2vec.estimate_plausible_tags([img], threshold=0.5)print(tags)tags = illust2vec.estimate_specific_tags([img], ['blue eyes', 'red hair'])print(tags)复制代码也可以指定标签并获取对应的概率
[{'blue eyes': 0.9488178491592407, 'red hair': 0.0025324225425720215}]复制代码预处理在服务器上处理全部图片,即截取头像、提取标签
加载库
# -*- coding: utf-8 -*-import i2vimport cv2import globimport osfrom imageio import imreadfrom tqdm import tqdmimport matplotlib.pyplot as plt%matplotlib inlineimport numpy as npimport pickle复制代码读取图片路径
images = glob.glob('characters/*.jpg')print(len(images))复制代码加载两个模型
illust2vec = i2v.make_i2v_with_chainer('illust2vec_tag_ver200.caffemodel', 'tag_list.json')cascade = cv2.CascadeClassifier('lbpcascade_animeface.xml')OUTPUT_DIR = 'faces/'if not os.path.exists(OUTPUT_DIR):    os.mkdir(OUTPUT_DIR)复制代码提取全部头像,共检测到27772张
num = 0for x in tqdm(range(len(images))):    img_path = images[x]    image = cv2.imread(img_path)    gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)    gray = cv2.equalizeHist(gray)    faces = cascade.detectMultiScale(gray, scaleFactor=1.1, minNeighbors=5, minSize=(64, 64))    for (x, y, w, h) in faces:        cx = x + w // 2        cy = y + h // 2        x0 = cx - int(0.75 * w)        x1 = cx + int(0.75 * w)        y0 = cy - int(0.75 * h)        y1 = cy + int(0.75 * h)        if x0 < 0:            x0 = 0        if y0 < 0:            y0 = 0        if x1 >= image.shape[1]:            x1 = image.shape[1] - 1        if y1 >= image.shape[0]:            y1 = image.shape[0] - 1        w = x1 - x0        h = y1 - y0        if w > h:            x0 = x0 + w // 2 - h // 2            x1 = x1 - w // 2 + h // 2            w = h        else:            y0 = y0 + h // 2 - w // 2            y1 = y1 - h // 2 + w // 2            h = w                face = image[y0: y0 + h, x0: x0 + w, :]        face = cv2.resize(face, (128, 128))        cv2.imwrite(os.path.join(OUTPUT_DIR, '%d.jpg' % num), face)        num += 1print(num)复制代码感兴趣的标签包括以下34个:
  • 13种头发颜色:blonde hair, brown hair, black hair, blue hair, pink hair, purple hair, green hair, red hair, silver hair, white hair, orange hair, aqua hair, grey hair
  • 5种发型:long hair, short hair, twintails, drill hair, ponytail
  • 10种眼睛颜色:blue eyes, red eyes, brown eyes, green eyes, purple eyes, yellow eyes, pink eyes, aqua eyes, black eyes, orange eyes
  • 6种其他属性:blush, smile, open mouth, hat, ribbon, glasses
头发颜色、发型和眼睛颜色取概率最高的一种,其他属性概率高于0.25则以存在处理
fw = open('face_tags.txt', 'w')tags = ['blonde hair', 'brown hair', 'black hair', 'blue hair', 'pink hair', 'purple hair', 'green hair',         'red hair', 'silver hair', 'white hair', 'orange hair', 'aqua hair', 'grey hair',        'long hair', 'short hair', 'twintails', 'drill hair', 'ponytail',        'blue eyes', 'red eyes', 'brown eyes', 'green eyes', 'purple eyes', 'yellow eyes', 'pink eyes',         'aqua eyes', 'black eyes', 'orange eyes',        'blush', 'smile', 'open mouth', 'hat', 'ribbon', 'glasses']fw.write('id,' + ','.join(tags) + '\n')images = glob.glob(os.path.join(OUTPUT_DIR, '*.jpg'))for x in tqdm(range(len(images))):    img_path = images[x]    image = imread(img_path)    result = illust2vec.estimate_specific_tags([image], tags)[0]        hair_colors = [[h, result[h]] for h in tags[0:13]]    hair_colors.sort(key=lambda x:x[1], reverse=True)    for h in tags[0:13]:        if h == hair_colors[0][0]:            result[h] = 1        else:            result[h] = 0        hair_styles = [[h, result[h]] for h in tags[13:18]]    hair_styles.sort(key=lambda x:x[1], reverse=True)    for h in tags[13:18]:        if h == hair_styles[0][0]:            result[h] = 1        else:            result[h] = 0        eye_colors = [[h, result[h]] for h in tags[18:28]]    eye_colors.sort(key=lambda x:x[1], reverse=True)    for h in tags[18:28]:        if h == eye_colors[0][0]:            result[h] = 1        else:            result[h] = 0        for h in tags[28:]:        if result[h] > 0.25:            result[h] = 1        else:            result[h] = 0            fw.write(img_path + ',' + ','.join([str(result[t]) for t in tags]) + '\n')    fw.close()复制代码这样一来,便得到了27772张二次元头像,以及每张头像对应的34个标签值
获取每张头像的4096维向量表示
illust2vec = i2v.make_i2v_with_chainer("illust2vec_ver200.caffemodel")img_all = []vec_all = []for x in tqdm(range(len(images))):    img_path = images[x]    image = imread(img_path)    vector = illust2vec.extract_feature([image])[0]    img_all.append(image / 255.)    vec_all.append(vector)img_all = np.array(img_all)vec_all = np.array(vec_all)复制代码随机选择2000张头像,进行tSNE降维可视化
from sklearn.manifold import TSNEfrom imageio import imsavedata_index = np.arange(img_all.shape[0])np.random.shuffle(data_index)data_index = data_index[:2000]tsne = TSNE(perplexity=30, n_components=2, init='pca', n_iter=5000)two_d_vectors = tsne.fit_transform(vec_all[data_index, :])puzzles = np.ones((6400, 6400, 3))xmin = np.min(two_d_vectors[:, 0])xmax = np.max(two_d_vectors[:, 0])ymin = np.min(two_d_vectors[:, 1])ymax = np.max(two_d_vectors[:, 1])for i, vector in enumerate(two_d_vectors):    x, y = two_d_vectors[i, :]    x = int((x - xmin) / (xmax - xmin) * (6400 - 128) + 64)    y = int((y - ymin) / (ymax - ymin) * (6400 - 128) + 64)    puzzles[y - 64: y + 64, x - 64: x + 64, :] = img_all[data_index]imsave('二次元头像降维可视化.png', puzzles)复制代码可视化结果如下,相似的头像确实被聚到了一起


模型使用ACGAN结构,但和CelebA中用的DCGAN不同,这次使用更深更复杂的网络来实现G和D,参考自SRGAN,arxiv.org/abs/1609.04…
生成器结构如下:


Sub-pixel CNN原理如下,把多个层拼接成一个层,从而达到增加高度和宽度、减少深度的目的


判别器结构如下,使用10个残差块,输出端包括两支,分别完成判别和分类任务


实现加载库
# -*- coding: utf-8 -*-import tensorflow as tfimport numpy as npimport pandas as pdimport matplotlib.pyplot as plt%matplotlib inlineimport osfrom imageio import imread, imsave, mimsaveimport globfrom tqdm import tqdm复制代码加载图片
images = glob.glob('faces/*.jpg')print(len(images))复制代码加载标签
tags = pd.read_csv('face_tags.txt')tags.index = tags['id']tags.head()复制代码定义一些常量、网络tensor、辅助函数,批大小设为2的幂比较合适,这里设为64,考虑学习率衰减
batch_size = 64z_dim = 128WIDTH = 128HEIGHT = 128LABEL = 34LAMBDA = 0.05BETA = 3OUTPUT_DIR = 'samples'if not os.path.exists(OUTPUT_DIR):    os.mkdir(OUTPUT_DIR)X = tf.placeholder(dtype=tf.float32, shape=[batch_size, HEIGHT, WIDTH, 3], name='X')X_perturb = tf.placeholder(dtype=tf.float32, shape=[batch_size, HEIGHT, WIDTH, 3], name='X_perturb')Y = tf.placeholder(dtype=tf.float32, shape=[batch_size, LABEL], name='Y')noise = tf.placeholder(dtype=tf.float32, shape=[batch_size, z_dim], name='noise')noise_y = tf.placeholder(dtype=tf.float32, shape=[batch_size, LABEL], name='noise_y')is_training = tf.placeholder(dtype=tf.bool, name='is_training')global_step = tf.Variable(0, trainable=False)add_global = global_step.assign_add(1)initial_learning_rate = 0.0002learning_rate = tf.train.exponential_decay(initial_learning_rate, global_step=global_step, decay_steps=20000, decay_rate=0.5)def lrelu(x, leak=0.2):    return tf.maximum(x, leak * x)def sigmoid_cross_entropy_with_logits(x, y):    return tf.nn.sigmoid_cross_entropy_with_logits(logits=x, labels=y)def conv2d(inputs, kernel_size, filters, strides, padding='same', use_bias=True):    return tf.layers.conv2d(inputs=inputs, kernel_size=kernel_size, filters=filters, strides=strides, padding=padding, use_bias=use_bias)    def batch_norm(inputs, is_training=is_training, decay=0.9):    return tf.contrib.layers.batch_norm(inputs, is_training=is_training, decay=decay)复制代码判别器部分
def d_block(inputs, filters):    h0 = lrelu(conv2d(inputs, 3, filters, 1))    h0 = conv2d(h0, 3, filters, 1)    h0 = lrelu(tf.add(h0, inputs))    return h0def discriminator(image, reuse=None):    with tf.variable_scope('discriminator', reuse=reuse):        h0 = image                f = 32        for i in range(5):            if i < 3:                h0 = lrelu(conv2d(h0, 4, f, 2))            else:                h0 = lrelu(conv2d(h0, 3, f, 2))            h0 = d_block(h0, f)            h0 = d_block(h0, f)            f = f * 2                h0 = lrelu(conv2d(h0, 3, f, 2))        h0 = tf.contrib.layers.flatten(h0)        Y_ = tf.layers.dense(h0, units=LABEL)        h0 = tf.layers.dense(h0, units=1)        return h0, Y_复制代码生成器部分
def g_block(inputs):    h0 = tf.nn.relu(batch_norm(conv2d(inputs, 3, 64, 1, use_bias=False)))    h0 = batch_norm(conv2d(h0, 3, 64, 1, use_bias=False))    h0 = tf.add(h0, inputs)    return h0def generator(z, label):    with tf.variable_scope('generator', reuse=None):        d = 16        z = tf.concat([z, label], axis=1)        h0 = tf.layers.dense(z, units=d * d * 64)        h0 = tf.reshape(h0, shape=[-1, d, d, 64])        h0 = tf.nn.relu(batch_norm(h0))        shortcut = h0                for i in range(16):            h0 = g_block(h0)                    h0 = tf.nn.relu(batch_norm(h0))        h0 = tf.add(h0, shortcut)                for i in range(3):            h0 = conv2d(h0, 3, 256, 1, use_bias=False)            h0 = tf.depth_to_space(h0, 2)            h0 = tf.nn.relu(batch_norm(h0))                h0 = tf.layers.conv2d(h0, kernel_size=9, filters=3, strides=1, padding='same', activation=tf.nn.tanh, name='g', use_bias=True)        return h0复制代码损失函数,这里的gp项来自DRAGAN,arxiv.org/abs/1705.07…,WGAN使用真实样本和合成样本的插值,而DRAGAN使用真实样本和干扰样本的插值
g = generator(noise, noise_y)d_real, y_real = discriminator(X)d_fake, y_fake = discriminator(g, reuse=True)loss_d_real = tf.reduce_mean(sigmoid_cross_entropy_with_logits(d_real, tf.ones_like(d_real)))loss_d_fake = tf.reduce_mean(sigmoid_cross_entropy_with_logits(d_fake, tf.zeros_like(d_fake)))loss_g_fake = tf.reduce_mean(sigmoid_cross_entropy_with_logits(d_fake, tf.ones_like(d_fake)))loss_c_real = tf.reduce_mean(sigmoid_cross_entropy_with_logits(y_real, Y))loss_c_fake = tf.reduce_mean(sigmoid_cross_entropy_with_logits(y_fake, noise_y))loss_d = loss_d_real + loss_d_fake + BETA * loss_c_realloss_g = loss_g_fake + BETA * loss_c_fakealpha = tf.random_uniform(shape=[batch_size, 1, 1, 1], minval=0., maxval=1.)interpolates = alpha * X + (1 - alpha) * X_perturbgrad = tf.gradients(discriminator(interpolates, reuse=True)[0], [interpolates])[0]slop = tf.sqrt(tf.reduce_sum(tf.square(grad), axis=[1]))gp = tf.reduce_mean((slop - 1.) ** 2)loss_d += LAMBDA * gpvars_g = [var for var in tf.trainable_variables() if var.name.startswith('generator')]vars_d = [var for var in tf.trainable_variables() if var.name.startswith('discriminator')]复制代码定义优化器
update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)with tf.control_dependencies(update_ops):    optimizer_d = tf.train.AdamOptimizer(learning_rate=learning_rate, beta1=0.5).minimize(loss_d, var_list=vars_d)    optimizer_g = tf.train.AdamOptimizer(learning_rate=learning_rate, beta1=0.5).minimize(loss_g, var_list=vars_g)复制代码合成图片的函数
def montage(images):        if isinstance(images, list):        images = np.array(images)    img_h = images.shape[1]    img_w = images.shape[2]    n_plots = int(np.ceil(np.sqrt(images.shape[0])))    if len(images.shape) == 4 and images.shape[3] == 3:        m = np.ones(            (images.shape[1] * n_plots + n_plots + 1,             images.shape[2] * n_plots + n_plots + 1, 3)) * 0.5    elif len(images.shape) == 4 and images.shape[3] == 1:        m = np.ones(            (images.shape[1] * n_plots + n_plots + 1,             images.shape[2] * n_plots + n_plots + 1, 1)) * 0.5    elif len(images.shape) == 3:        m = np.ones(            (images.shape[1] * n_plots + n_plots + 1,             images.shape[2] * n_plots + n_plots + 1)) * 0.5    else:        raise ValueError('Could not parse image shape of {}'.format(images.shape))    for i in range(n_plots):        for j in range(n_plots):            this_filter = i * n_plots + j            if this_filter < images.shape[0]:                this_img = images[this_filter]                m[1 + i + i * img_h:1 + i + (i + 1) * img_h,                  1 + j + j * img_w:1 + j + (j + 1) * img_w] = this_img    return m复制代码整理数据
X_all = []Y_all = []for i in tqdm(range(len(images))):    image = imread(images)    image = (image / 255. - 0.5) * 2    X_all.append(image)        y = list(tags.loc[images])    Y_all.append(y[1:])X_all = np.array(X_all)Y_all = np.array(Y_all)print(X_all.shape, Y_all.shape)复制代码定义随机产生标签的函数,原始数据中标签分布不均匀,但我们希望G能学到各种标签,所以均匀地生成各类标签
def get_random_tags():    y = np.random.uniform(0.0, 1.0, [batch_size, LABEL]).astype(np.float32)    y[y > 0.75] = 1    y[y <= 0.75] = 0    for i in range(batch_size):        hc = np.random.randint(0, 13)        hs = np.random.randint(13, 18)        ec = np.random.randint(18, 28)        y[i, :28] = 0        y[i, hc] = 1 # hair color        y[i, hs] = 1 # hair style        y[i, ec] = 1 # eye color    return y复制代码训练模型,CelebA中男女比例均衡,因此每次迭代随机取一批数据训练即可。但现在由于原始数据中各类标签分布不均匀,所以需要完整地迭代数据
sess = tf.Session()sess.run(tf.global_variables_initializer())z_samples = np.random.uniform(-1.0, 1.0, [batch_size, z_dim]).astype(np.float32)y_samples = get_random_tags()for i in range(batch_size):    y_samples[i, :28] = 0    y_samples[i, i // 8 % 13] = 1 # hair color    y_samples[i, i // 8 % 5 + 13] = 1 # hair style    y_samples[i, i // 8 % 10 + 18] = 1 # eye colorsamples = []loss = {'d': [], 'g': []}offset = 0for i in tqdm(range(60000)):    if offset + batch_size > X_all.shape[0]:        offset = 0    if offset == 0:        data_index = np.arange(X_all.shape[0])        np.random.shuffle(data_index)        X_all = X_all[data_index, :, :, :]        Y_all = Y_all[data_index, :]    X_batch = X_all[offset: offset + batch_size, :, :, :]    Y_batch = Y_all[offset: offset + batch_size, :]    X_batch_perturb = X_batch + 0.5 * X_batch.std() * np.random.random(X_batch.shape)    offset += batch_size        n = np.random.uniform(-1.0, 1.0, [batch_size, z_dim]).astype(np.float32)    ny = get_random_tags()    _, d_ls = sess.run([optimizer_d, loss_d], feed_dict={X: X_batch, X_perturb: X_batch_perturb, Y: Y_batch, noise: n, noise_y: ny, is_training: True})            n = np.random.uniform(-1.0, 1.0, [batch_size, z_dim]).astype(np.float32)    ny = get_random_tags()    _, g_ls = sess.run([optimizer_g, loss_g], feed_dict={noise: n, noise_y: ny, is_training: True})        loss['d'].append(d_ls)    loss['g'].append(g_ls)        _, lr = sess.run([add_global, learning_rate])        if i % 500 == 0:        print(i, d_ls, g_ls, lr)        gen_imgs = sess.run(g, feed_dict={noise: z_samples, noise_y: y_samples, is_training: False})        gen_imgs = (gen_imgs + 1) / 2        imgs = [img[:, :, :] for img in gen_imgs]        gen_imgs = montage(imgs)        plt.axis('off')        plt.imshow(gen_imgs)        imsave(os.path.join(OUTPUT_DIR, 'sample_%d.jpg' % i), gen_imgs)        plt.show()        samples.append(gen_imgs)plt.plot(loss['d'], label='Discriminator')plt.plot(loss['g'], label='Generator')plt.legend(loc='upper right')plt.savefig('Loss.png')plt.show()mimsave(os.path.join(OUTPUT_DIR, 'samples.gif'), samples, fps=10)复制代码生成的二次元头像如下,每一行对应的头发颜色、发型、眼睛颜色相同,其他属性随机。少部分结果不太好,可能是某些噪音或条件的问题


保存模型
saver = tf.train.Saver()saver.save(sess, './anime_acgan', global_step=60000)复制代码在单机上加载模型,进行以下三项尝试:
  • 按原始标签分布随机生成样本
  • 生成指定标签的样本
  • 固定噪音,按原始标签分布生成样本
# -*- coding: utf-8 -*-import tensorflow as tfimport numpy as npimport matplotlib.pyplot as pltfrom imageio import imsavebatch_size = 64z_dim = 128LABEL = 34def montage(images):        if isinstance(images, list):        images = np.array(images)    img_h = images.shape[1]    img_w = images.shape[2]    n_plots = int(np.ceil(np.sqrt(images.shape[0])))    if len(images.shape) == 4 and images.shape[3] == 3:        m = np.ones(            (images.shape[1] * n_plots + n_plots + 1,             images.shape[2] * n_plots + n_plots + 1, 3)) * 0.5    elif len(images.shape) == 4 and images.shape[3] == 1:        m = np.ones(            (images.shape[1] * n_plots + n_plots + 1,             images.shape[2] * n_plots + n_plots + 1, 1)) * 0.5    elif len(images.shape) == 3:        m = np.ones(            (images.shape[1] * n_plots + n_plots + 1,             images.shape[2] * n_plots + n_plots + 1)) * 0.5    else:        raise ValueError('Could not parse image shape of {}'.format(images.shape))    for i in range(n_plots):        for j in range(n_plots):            this_filter = i * n_plots + j            if this_filter < images.shape[0]:                this_img = images[this_filter]                m[1 + i + i * img_h:1 + i + (i + 1) * img_h,                  1 + j + j * img_w:1 + j + (j + 1) * img_w] = this_img    return mdef get_random_tags():    y = np.random.uniform(0.0, 1.0, [batch_size, LABEL]).astype(np.float32)    p_other = [0.6, 0.6, 0.25, 0.04488882, 0.3, 0.05384738]    for i in range(batch_size):        for j in range(len(p_other)):            if y[i, j + 28] < p_other[j]:                y[i, j + 28] = 1            else:                y[i, j + 28] = 0    phc = [0.15968645, 0.21305391, 0.15491921, 0.10523116, 0.07953927, 0.09508879, 0.03567429, 0.07733163, 0.03157895, 0.01833307, 0.02236442, 0.00537514, 0.00182371]    phs = [0.52989922,  0.37101264,  0.12567589,  0.00291153,  0.00847864]    pec = [0.28350664, 0.15760678, 0.17862742, 0.13412254, 0.14212126, 0.0543913, 0.01020637, 0.00617501, 0.03167493, 0.00156775]    for i in range(batch_size):        y[i, :28] = 0        hc = np.random.random()        for j in range(len(phc)):            if np.sum(phc[:j]) < hc < np.sum(phc[:j + 1]):                y[i, j] = 1                break        hs = np.random.random()        for j in range(len(phs)):            if np.sum(phs[:j]) < hs < np.sum(phs[:j + 1]):                y[i, j + 13] = 1                break        ec = np.random.random()        for j in range(len(pec)):            if np.sum(pec[:j]) < ec < np.sum(pec[:j + 1]):                y[i, j + 18] = 1                break    return ysess = tf.Session()sess.run(tf.global_variables_initializer())saver = tf.train.import_meta_graph('./anime_acgan-60000.meta')saver.restore(sess, tf.train.latest_checkpoint('./'))graph = tf.get_default_graph()g = graph.get_tensor_by_name('generator/g/Tanh:0')noise = graph.get_tensor_by_name('noise:0')noise_y = graph.get_tensor_by_name('noise_y:0')is_training = graph.get_tensor_by_name('is_training:0')# 随机生成样本z_samples = np.random.uniform(-1.0, 1.0, [batch_size, z_dim]).astype(np.float32)y_samples = get_random_tags()gen_imgs = sess.run(g, feed_dict={noise: z_samples, noise_y: y_samples, is_training: False})gen_imgs = (gen_imgs + 1) / 2imgs = [img[:, :, :] for img in gen_imgs]gen_imgs = montage(imgs)gen_imgs = np.clip(gen_imgs, 0, 1)imsave('1_二次元头像随机生成.jpg', gen_imgs)# 生成指定标签的样本all_tags = ['blonde hair', 'brown hair', 'black hair', 'blue hair', 'pink hair', 'purple hair', 'green hair', 'red hair', 'silver hair', 'white hair', 'orange hair', 'aqua hair', 'grey hair', 'long hair', 'short hair', 'twintails', 'drill hair', 'ponytail', 'blue eyes', 'red eyes', 'brown eyes', 'green eyes', 'purple eyes', 'yellow eyes', 'pink eyes', 'aqua eyes', 'black eyes', 'orange eyes', 'blush', 'smile', 'open mouth', 'hat', 'ribbon', 'glasses']for i, tags in enumerate([['blonde hair', 'twintails', 'blush', 'smile', 'ribbon', 'red eyes'], ['silver hair', 'long hair', 'blush', 'smile', 'open mouth', 'blue eyes']]):    z_samples = np.random.uniform(-1.0, 1.0, [batch_size, z_dim]).astype(np.float32)    y_samples = np.zeros([1, LABEL])    for tag in tags:        y_samples[0, all_tags.index(tag)] = 1    y_samples = np.repeat(y_samples, batch_size, 0)    gen_imgs = sess.run(g, feed_dict={noise: z_samples, noise_y: y_samples, is_training: False})    gen_imgs = (gen_imgs + 1) / 2    imgs = [img[:, :, :] for img in gen_imgs]    gen_imgs = montage(imgs)    gen_imgs = np.clip(gen_imgs, 0, 1)    imsave('%d_二次元头像指定标签.jpg' % (i + 2), gen_imgs)# 固定噪音随机标签z_samples = np.random.uniform(-1.0, 1.0, [1, z_dim]).astype(np.float32)z_samples = np.repeat(z_samples, batch_size, 0)y_samples = get_random_tags()gen_imgs = sess.run(g, feed_dict={noise: z_samples, noise_y: y_samples, is_training: False})gen_imgs = (gen_imgs + 1) / 2imgs = [img[:, :, :] for img in gen_imgs]gen_imgs = montage(imgs)gen_imgs = np.clip(gen_imgs, 0, 1)imsave('4_二次元头像固定噪音.jpg', gen_imgs)复制代码按原始标签分布随机生成样本


生成金发、双马尾、脸红、微笑、系丝带、红眼睛的头像


生成银发、长发、脸红、微笑、张嘴、蓝眼睛的头像


固定噪音随机标签,使得头像主体大致相同但各种细节不一样


掌握以上内容后,也可以在CelebA上训练受40个01属性控制的ACGAN模型,而且比二次元头像更简单一些



链接:https://juejin.im/post/5ba302386fb9a05d265957de



2 个回复

正序浏览
回复 使用道具 举报
奈斯
回复 使用道具 举报
您需要登录后才可以回帖 登录 | 加入黑马