为什么谷歌colab TPU很慢?

问题描述 投票:0回答:1

我在使用 塔洛斯 运行超参数调整 Keras 模型。在Google colab上运行这个短代码 TPU 是很慢的。我想这和数据的类型有关。我是否应该将其转换为张力,使之成为 TPU 更快?

%tensorflow_version 2.x
import os
import tensorflow as tf
import talos as ta
from tensorflow.keras.models import Sequential
from tensorflow.keras.layers import Dense
from tensorflow.keras.optimizers import Adam
from sklearn.model_selection import train_test_split

def iris_model(x_train, y_train, x_val, y_val, params):

    # Specify a distributed strategy to use TPU
    resolver = tf.distribute.cluster_resolver.TPUClusterResolver(tpu='grpc://' + os.environ['COLAB_TPU_ADDR'])
    tf.config.experimental_connect_to_host(resolver.master())
    tf.tpu.experimental.initialize_tpu_system(resolver)
    strategy = tf.distribute.experimental.TPUStrategy(resolver)

    # Use the strategy to create and compile a Keras model
    with strategy.scope():
      model = Sequential()
      model.add(Dense(32, input_shape=(4,), activation=tf.nn.relu, name="relu"))
      model.add(Dense(3, activation=tf.nn.softmax, name="softmax"))
      model.compile(optimizer=Adam(learning_rate=0.1), loss=params['losses'])

    # Convert data type to use TPU
    x_train = x_train.astype('float32')
    x_val = x_val.astype('float32')

    # Fit the Keras model on the dataset
    out = model.fit(x_train, y_train, batch_size=params['batch_size'], epochs=params['epochs'], validation_data=[x_val, y_val], verbose=0, steps_per_epoch=0)

    return out, model

# Load dataset
X, y = ta.templates.datasets.iris()

# Train and test set
x_train, x_val, y_train, y_val = train_test_split(X, y, test_size=0.30, shuffle=False)

# Create a hyperparameter distributions 
p = {'losses': ['logcosh'], 'batch_size': [128, 256, 384, 512, 1024], 'epochs': [10, 20]}

# Use Talos to scan the best hyperparameters of the Keras model
scan_object = ta.Scan(x_train, y_train, params=p, model=iris_model, experiment_name='test', x_val=x_val, y_val=y_val, fraction_limit=0.5)
tensorflow keras google-colaboratory google-cloud-tpu talos
1个回答
2
投票

谢谢你的问题。

不幸的是,我没能让你的代码样本在TensorFlow 2.2上运行,所以我不知道你原来看到的性能如何。我可以通过以下修改来修复它,并让它在TPU上运行。

  • 替换 tf.config.experimental_connect_to_host(resolver.master())tf.config.experimental_connect_to_cluster(resolver)
  • 将TPU初始化移至外部 iris_model().
  • 使用 tf.data.Dataset 为TPU输入。

这是修改后的Colab代码。

# Run this to install Talos before running the rest of the code.
!pip install git+https://github.com/autonomio/[email protected]
%tensorflow_version 2.x
import os
import tensorflow as tf
import talos as ta
from tensorflow.keras.models import Sequential
from tensorflow.keras.layers import Dense
from tensorflow.keras.optimizers import Adam
from sklearn.model_selection import train_test_split

print(tf.__version__) # TF 2.2.0 in my case

resolver = tf.distribute.cluster_resolver.TPUClusterResolver(tpu='grpc://' + os.environ['COLAB_TPU_ADDR'])
tf.config.experimental_connect_to_cluster(resolver)
tf.tpu.experimental.initialize_tpu_system(resolver)

def iris_model(x_train, y_train, x_val, y_val, params):
    # Use the strategy to create and compile a Keras model
    strategy = tf.distribute.experimental.TPUStrategy(resolver)
    with strategy.scope():
      model = Sequential()
      model.add(Dense(32, input_shape=(4,), activation=tf.nn.relu, name="relu"))
      model.add(Dense(3, activation=tf.nn.softmax, name="softmax"))
      model.compile(optimizer=Adam(learning_rate=0.1), loss=params['losses'])

    train_dataset = tf.data.Dataset.from_tensor_slices((x_train, y_train)).batch(params['batch_size'])
    val_dataset = tf.data.Dataset.from_tensor_slices((x_val, y_val)).batch(params['batch_size'])

    # Fit the Keras model on the dataset
    out = model.fit(train_dataset, epochs=params['epochs'], validation_data=val_dataset)

    return out, model

# Load dataset
X, y = ta.templates.datasets.iris()

# Train and test set
x_train, x_val, y_train, y_val = train_test_split(X, y, test_size=0.30, shuffle=False)

# Create a hyperparameter distributions 
p = {'losses': ['logcosh'], 'batch_size': [128, 256, 384, 512, 1024], 'epochs': [10, 20]}

# Use Talos to scan the best hyperparameters of the Keras model
scan_object = ta.Scan(x_train, y_train, params=p, model=iris_model, experiment_name='test', x_val=x_val, y_val=y_val, fraction_limit=0.5)

对我来说,最后一次调用花了不到两分钟的时间。

对于知名的数据集,你可以跳过创建自己的 tf.data.Dataset 借助 TensorFlow数据集库. TFDS确实有: 虹膜数据集 的库中。有关使用TFDS与TPU的端到端示例,请参见TensorFlow的 官方TPU指南.

© www.soinside.com 2019 - 2024. All rights reserved.