The automatic recognition of human gestures is a complex multidisciplinary problem that has not yet been completely solved. Since the advent of digital video capture technologies, there have been attempts to recognize dynamic gestures for different purposes. In the recent years, new technologies such as depth sensors or highresolution cameras were incorporated as well as the high processing capacity of the current devices emerged, allowing the new technologies development capable of detecting different movements and acting in real time. Unlike the recognition of the spoken voice, which has been researched for more than forty years, the topic of this thesis is relatively new in the scientific area and it evolves rapidly as new devices appear as well as new computer vision algorithms.It is necessary to tackle many different tasks to be able to use an automatic sign language recognition system to translate the interpreter gestures. First, there are different approaches depending on the sensing device to use. Once the gesture is captured, several pre-processing stages are required to identify regions of interest such as the hands and face of the interpreter, and then identify the different trajectories of the performed gesture.The sign language presents a huge variability in the different postures or configurations that a hand can have, which makes this discipline a particularly complex problem. To deal with this, a correct generation of the static and dynamic descriptors is necessary. In addition, because each region has specific language grammars, it is required the provision of an Argentine Sign Language (LSA) database, which has not been available yet. Based on the reasons mentioned above, this thesis aims to develop a complete process of interpretation and translation of the Argentinian Sign Language through videos obtained with an RGB camera.First, a state of the art study about the gesture recognition was carried out. Intelligent techniques for image and video processing as well as the different descriptors types were researched. As a preliminary work, a strategy capable of processing human actions captured with an MS Kinect device [4] was developed. This strategy implements a probabilistic SOM neural network (ProbSOM) with a descriptor specifically designed to retain temporal information. This work allowed to overcome the existing results so far for two recognized databases.As a result of this thesis, two main contributions in the sign language field were made. In the first place, a specific database for the recognition of the Argentinian Sign Language was developed. This included an image database with the 16 configurations most used in the language [3], along with a database of high-resolution videos with 64 different signs, with a total of 3200 videos [2]. These databases were recorded with 10 different interpreters and several repetitions, allowing their use with classic techniques of machine learning. In addition, in these databases, the interpreters have worn colored gloves in the form of a marker. Thi...