import tensorflow as tf import numpy as np from transformers import ( AutoModelForSequenceClassification, BertTokenizer, TFBertModel, ) visible_default = tf.config.list_physical_devices("GPU") try: tf.config.set_visible_devices([], "GPU") visible_devices = tf.config.get_visible_devices() for device in visible_devices: assert device.device_type != "GPU" except: # Invalid device or cannot modify virtual devices once initialized. pass ##################### Tensorflow Hugging Face LM Models ################################### MAX_SEQUENCE_LENGTH = 512 BATCH_SIZE = 1 # Create a set of 2-dimensional inputs tf_bert_input = [ tf.TensorSpec(shape=[BATCH_SIZE, MAX_SEQUENCE_LENGTH], dtype=tf.int32), tf.TensorSpec(shape=[BATCH_SIZE, MAX_SEQUENCE_LENGTH], dtype=tf.int32), tf.TensorSpec(shape=[BATCH_SIZE, MAX_SEQUENCE_LENGTH], dtype=tf.int32), ] class TFHuggingFaceLanguage(tf.Module): def __init__(self, hf_model_name): super(TFHuggingFaceLanguage, self).__init__() # Create a BERT trainer with the created network. self.m = TFBertModel.from_pretrained(hf_model_name, from_pt=True) # Invoke the trainer model on the inputs. This causes the layer to be built. self.m.predict = lambda x, y, z: self.m.call( input_ids=x, attention_mask=y, token_type_ids=z, training=False ) @tf.function(input_signature=tf_bert_input) def forward(self, input_ids, attention_mask, token_type_ids): return self.m.predict(input_ids, attention_mask, token_type_ids) def get_TFhf_model(name): # gpus = tf.config.experimental.list_physical_devices("GPU") # for gpu in gpus: # tf.config.experimental.set_memory_growth(gpu, True) model = TFHuggingFaceLanguage(name) tokenizer = BertTokenizer.from_pretrained( "microsoft/MiniLM-L12-H384-uncased" ) text = "Replace me by any text you'd like." encoded_input = tokenizer( text, padding="max_length", truncation=True, max_length=MAX_SEQUENCE_LENGTH, ) for key in encoded_input: encoded_input[key] = tf.expand_dims( tf.convert_to_tensor(encoded_input[key]), 0 ) test_input = ( encoded_input["input_ids"], encoded_input["attention_mask"], encoded_input["token_type_ids"], ) actual_out = model.forward(*test_input) return model, test_input, actual_out # Utility function for comparing two tensors (tensorflow). def compare_tensors_tf(tf_tensor, numpy_tensor): # setting the absolute and relative tolerance rtol = 1e-02 atol = 1e-03 tf_to_numpy = tf_tensor.numpy() return np.allclose(tf_to_numpy, numpy_tensor, rtol, atol) ##################### Tensorflow Hugging Face Masked LM Models ################################### from transformers import TFAutoModelForMaskedLM, AutoTokenizer import tensorflow as tf visible_default = tf.config.list_physical_devices("GPU") try: tf.config.set_visible_devices([], "GPU") visible_devices = tf.config.get_visible_devices() for device in visible_devices: assert device.device_type != "GPU" except: # Invalid device or cannot modify virtual devices once initialized. pass # The max_sequence_length is set small for testing purpose. BATCH_SIZE = 1 MAX_SEQUENCE_LENGTH = 16 # Create a set of input signature. inputs_signature = [ tf.TensorSpec(shape=[BATCH_SIZE, MAX_SEQUENCE_LENGTH], dtype=tf.int32), tf.TensorSpec(shape=[BATCH_SIZE, MAX_SEQUENCE_LENGTH], dtype=tf.int32), ] # For supported models please see here: # https://huggingface.co/docs/transformers/model_doc/auto#transformers.TFAutoModelForCasualLM def preprocess_input( model_name, text="This is just used to compile the model" ): tokenizer = AutoTokenizer.from_pretrained(model_name) inputs = tokenizer( text, padding="max_length", return_tensors="tf", truncation=True, max_length=MAX_SEQUENCE_LENGTH, ) return inputs class MaskedLM(tf.Module): def __init__(self, model_name): super(MaskedLM, self).__init__() self.m = TFAutoModelForMaskedLM.from_pretrained( model_name, output_attentions=False, num_labels=2 ) self.m.predict = lambda x, y: self.m(input_ids=x, attention_mask=y)[0] @tf.function(input_signature=inputs_signature) def forward(self, input_ids, attention_mask): return self.m.predict(input_ids, attention_mask) def get_causal_lm_model(hf_name, text="Hello, this is the default text."): # gpus = tf.config.experimental.list_physical_devices("GPU") # for gpu in gpus: # tf.config.experimental.set_memory_growth(gpu, True) model = MaskedLM(hf_name) encoded_input = preprocess_input(hf_name, text) test_input = (encoded_input["input_ids"], encoded_input["attention_mask"]) actual_out = model.forward(*test_input) return model, test_input, actual_out ##################### Tensorflow Hugging Face Image Classification Models ################################### from transformers import TFAutoModelForImageClassification from transformers import ConvNextFeatureExtractor, ViTFeatureExtractor from transformers import BeitFeatureExtractor, AutoFeatureExtractor import tensorflow as tf from PIL import Image import requests # Create a set of input signature. inputs_signature = [ tf.TensorSpec(shape=[1, 3, 224, 224], dtype=tf.float32), ] class AutoModelImageClassfication(tf.Module): def __init__(self, model_name): super(AutoModelImageClassfication, self).__init__() self.m = TFAutoModelForImageClassification.from_pretrained( model_name, output_attentions=False ) self.m.predict = lambda x: self.m(x) @tf.function(input_signature=inputs_signature) def forward(self, inputs): return self.m.predict(inputs) fail_models = [ "facebook/data2vec-vision-base-ft1k", "microsoft/swin-tiny-patch4-window7-224", ] supported_models = [ "facebook/convnext-tiny-224", "google/vit-base-patch16-224", ] img_models_fe_dict = { "facebook/convnext-tiny-224": ConvNextFeatureExtractor, "facebook/data2vec-vision-base-ft1k": BeitFeatureExtractor, "microsoft/swin-tiny-patch4-window7-224": AutoFeatureExtractor, "google/vit-base-patch16-224": ViTFeatureExtractor, } def preprocess_input_image(model_name): # from datasets import load_dataset # dataset = load_dataset("huggingface/cats-image") # image1 = dataset["test"]["image"][0] # # print("image1: ", image1) # url = "http://images.cocodataset.org/val2017/000000039769.jpg" # image = Image.open(requests.get(url, stream=True).raw) feature_extractor = img_models_fe_dict[model_name].from_pretrained( model_name ) # inputs: {'pixel_values': } inputs = feature_extractor(images=image, return_tensors="tf") return [inputs[str(*inputs)]] def get_causal_image_model(hf_name): model = AutoModelImageClassfication(hf_name) test_input = preprocess_input_image(hf_name) # TFSequenceClassifierOutput(loss=None, logits=, hidden_states=None, attentions=None) actual_out = model.forward(*test_input) return model, test_input, actual_out