This paper presents the results of a study of the effects of integer (fixed-point) arithmetic implementation on classification accuracy of a popular open-source people detection system based on Histogram of Oriented Gradients. It is investigated how the system performance deviates from the reference algorithm performance as integer arithmetic is introduced with different bit-width in several critical parts of the system. In performed experiments, the effects of different bit-width integer arithmetic implementation for four key operations were separately considered: HoG descriptor magnitude calculation, HoG descriptor angle calculation, normalization and SVM classification. It is found that a 13-bit representation of variables is more than sufficient to accurately implement this system in integer arithmetic. The experiments in the paper are conducted for pedestrian detection and the methodology and the lessons learned from this study allow generalization of conclusions to a broader class of applications.