This article provides a research viewpoint on the training procedure of a Bidirectional Encoder Representations from Transformers (BERT) model starting from the beginning, employing the Tensor Processing Unit (TPU), with a particular emphasis on the Spanish language. This article focuses on the techniques and strategies, such as the whole word masking (WWM) technique, that are essential for achieving successful BERT training in multiple languages, including English. The goal is to provide efficient development of natural language processing applications for understanding the Spanish language. The BEThiz model exhibits exceptional proficiency in comprehending contextual cues and producing accurate responses for masked inputs, surpassing other models within the same category and language. This characteristic sets our model apart and solidifies its role as a particularly potent tool in the field of natural language processing.