Knowledge representation learning represents entities and relations of knowledge graph in a continuous low-dimensional semantic space. Recently, various representation learning models have successfully been developed to infer novel relations in general-purpose knowledge bases such as FreeBase and WordNet. However, few studies have used such models for biomedical data for inferring useful relations among biomedical entities such as genes, chemicals, diseases, and symptoms. This study aimed to compare the potential of representation learning models in extracting biomedical relations by using four different types of representation learning models, viz., TransE, PTransE, TransR, and TransH. For training and evaluating the models, we collected and utilized manually curated data from public databases, including relations among chemicals, genes, diseases, and symptoms. Overall, TransE, the most efficient translation-based monolingual knowledge graph embedding model, displayed the best performance with a higher learning speed for large-scale biomedical data. Using TransE, we inferred new relations among chemicals, genes, diseases, and symptoms, and evaluated the reliability of these inferred relations. Furthermore, TransE outperformed an existing statistical method used in the Comparative Toxicogenomics Database for inferring new chemical-disease relations. Together, the present results show that the representation learning model is useful for inferring new biological data from numerous existing biomedical data.