In the recent years, deep learning has become one of the most important topics in computer science. Deep learning is a growing trend in the edge of technology and its applications are now seen in many aspects of our life such as object detection, speech recognition, natural language processing, etc. Currently, almost all major sciences and technologies are benefiting from the advantages of deep learning such as high accuracy, speed and flexibility. Therefore, any efforts for improving performance of related techniques is valuable. Deep learning accelerators are considered as hardware architecture, which are designed and optimized for increasing the speed, efficiency and accuracy of computers that are running deep learning algorithms. In this paper, after reviewing some backgrounds about deep learning, a well-known accelerator architecture named MAERI is investigated. By using an open source tool called MAESTRO, the performance of a deep learning task is measured and compared on two different data flow strategies: NLR and NVDLA. Measured performance indicators of novel optimized architecture, NVDLA shows higher L1 and L2 computation reuse and lower total runtime (cycles) in comparison to the other one.