Gesture posture prediction method and system based on multi-view images
1. A gesture posture prediction method based on multi-view images is characterized by comprising the following steps: the method comprises the following steps:
acquiring gesture image data of at least two visual angles;
obtaining each gesture image prediction angle according to each acquired gesture image at each view angle and a preset gesture prediction model corresponding to each view angle;
taking the average value of the predicted angles of all the gesture images as a final gesture angle;
in the training of the preset attitude prediction model, the maximum error of the kth convolutional neural network on a training set, the relative error of each sample, the regression error rate of the kth convolutional neural network and the weight coefficient of the weak learner are calculated, and the sample weight is updated according to the obtained maximum error, relative error, regression error rate and the weight coefficient of the weak learner.
2. The multi-view image-based gesture posture prediction method of claim 1, characterized in that:
constructing a posture prediction model by adopting an Adaboost algorithm and combining a convolutional neural network;
the convolutional neural network adopts a Resnet or Alexnet model as a base learner, iteration times and learning rate are set, a mean square error function is selected as a loss function, a gradient descent method is used for updating and optimizing model parameters, a posture prediction model is fitted, and the base learner of the kth convolutional neural network is obtained through training.
3. The multi-view image-based gesture posture prediction method of claim 2, characterized in that:
after iteration is carried out for K times, K convolutional neural network regression prediction models are obtained, weight coefficients of K base learners are obtained, the median of the weight of each weak learner is selected, the weak learner corresponding to the median of the weight is used as a strong learner of the prediction model, and a final posture prediction model is obtained according to the obtained strong learner.
4. The multi-view image-based gesture posture prediction method of claim 1, characterized in that:
the updated sample weights are:
wherein, WkiIs an original template weight, αkIs the weight coefficient of the weak learner,ekiis the relative error of the sample.
5. The multi-view image-based gesture posture prediction method of claim 1, characterized in that:
constructing a posture prediction model by adopting a Bagging algorithm;
the convolutional neural network adopts a Resnet or Alexnet model as a base learner, iteration times and learning rate are set, a mean square error function is selected as a loss function, a gradient descent method is used for updating and optimizing model parameters, a posture prediction model is fitted, and the base learner of the kth convolutional neural network is obtained through training.
6. The multi-view image-based gesture posture prediction method of claim 5, characterized in that:
and after iterating for K times, obtaining K convolutional neural network regression prediction models, taking the K convolutional neural network prediction models as a base learner, synthesizing the base learner by using a Bagging algorithm, and taking the prediction average value of the K convolutional neural network base learners as the output of the attitude prediction model.
7. The multi-view image-based gesture posture prediction method of claim 1, characterized in that:
in the training of the preset posture prediction model, the types of the gesture images comprise color images under illumination and infrared gray images without illumination, and the color images shot in the daytime are processed into gray images;
the gesture image content includes at least gesture images of different situations of a bare hand, wearing a half-finger glove, and wearing a full-finger glove.
8. A gesture posture prediction system based on multi-view images is characterized in that: the method comprises the following steps:
a data acquisition module configured to: acquiring gesture image data of at least two visual angles;
a pose prediction module configured to: obtaining each gesture image prediction angle according to each acquired gesture image at each view angle and a preset gesture prediction model corresponding to each view angle;
a pose determination module configured to: taking the average value of the predicted angles of all the gesture images as a final gesture angle;
in the training of the preset attitude prediction model, the maximum error of the kth convolutional neural network on a training set, the relative error of each sample, the regression error rate of the kth convolutional neural network and the weight coefficient of the weak learner are calculated, and the sample weight is updated according to the obtained maximum error, relative error, regression error rate and the weight coefficient of the weak learner.
9. A computer-readable storage medium, on which a program is stored, which, when being executed by a processor, carries out the steps of the multi-view image based gesture pose prediction method according to any one of claims 1 to 7.
10. An electronic device comprising a memory, a processor and a program stored on the memory and executable on the processor, wherein the processor implements the steps of the multi-view image based gesture pose prediction method according to any one of claims 1-7 when executing the program.
Background
The statements in this section merely provide background information related to the present disclosure and may not necessarily constitute prior art.
A gesture is a silent language. The gestures have extremely strong information expression and transmission functions, and people borrow the gestures in life to express various emotions. With the continuous development of artificial intelligence, people can also perform gesture interaction with computers. And performing man-machine interaction with the computer by means of gestures, so that the computer can execute the set commands. The gesture is used for man-machine interaction, so that the gesture control system has the advantages of convenience in operation, separation from physical contact, capability of performing remote control, richer and more natural interactive operation and the like, and plays a great role in various fields such as smart home, intelligent transportation, biomedical treatment, entertainment, military apparatus and the like. Therefore, research on the recognition and estimation of gesture poses is essential.
Because human gestures have a great number of degrees of freedom, the gestures can make a great number of gestures in a three-dimensional space, so that the problem of gesture self-occlusion is caused, and in addition, the computer is difficult to acquire three-dimensional information of an object only through a two-dimensional image, so that the computer is difficult to directly predict a gesture angle only by using the two-dimensional image of the gestures.
The inventor finds that, in the related art of gesture posture prediction, the estimation of the gesture posture is mostly realized by means of a depth picture acquired by a depth camera, and the method has the advantages that the acquired image contains depth information of an object, but most depth cameras are sensitive to light and have the defect of being easily influenced by sunlight, and the method is often applied indoors and has poor outdoor experience.
Disclosure of Invention
In order to solve the defects of the prior art, the method and the system for predicting the gesture posture based on the multi-view images are provided, the multi-view gesture images are adopted for training, prediction errors caused by gesture shielding are greatly reduced, prediction models are respectively established for the multi-view gesture images, the average value of the prediction results is taken as the final prediction result, and the accuracy of gesture posture prediction is improved.
In order to achieve the purpose, the following technical scheme is adopted in the disclosure:
the disclosure provides a gesture posture prediction method based on multi-view images in a first aspect.
A gesture posture prediction method based on multi-view images comprises the following processes:
acquiring gesture image data of at least two visual angles;
obtaining each gesture image prediction angle according to each acquired gesture image at each view angle and a preset gesture prediction model corresponding to each view angle;
taking the average value of the predicted angles of all the gesture images as a final gesture angle;
in the training of the preset attitude prediction model, the maximum error of the kth convolutional neural network on a training set, the relative error of each sample, the regression error rate of the kth convolutional neural network and the weight coefficient of the weak learner are calculated, and the sample weight is updated according to the obtained maximum error, relative error, regression error rate and the weight coefficient of the weak learner.
Further, an Adaboost algorithm is combined with a convolutional neural network to construct a posture prediction model;
the convolutional neural network adopts a Resnet or Alexnet model as a base learner, iteration times and learning rate are set, a mean square error function is selected as a loss function, a gradient descent method is used for updating and optimizing model parameters, a posture prediction model is fitted, and the base learner of the kth convolutional neural network is obtained through training.
Further, after K iterations, K convolutional neural network regression prediction models are obtained, weight coefficients of K base learners are obtained, a median of the weight of each weak learner is selected, the weak learner corresponding to the median of the weight is used as a strong learner of the prediction model, and a final posture prediction model is obtained according to the obtained strong learner.
Further, the updated sample weight is:
wherein, WkiIs an original template weight, αkIs the weight coefficient of the weak learner,ekiis the relative error of the sample. Further, a posture prediction model is constructed by adopting a Bagging algorithm;
the convolutional neural network adopts a Resnet or Alexnet model as a base learner, iteration times and learning rate are set, a mean square error function is selected as a loss function, a gradient descent method is used for updating and optimizing model parameters, a posture prediction model is fitted, and the base learner of the kth convolutional neural network is obtained through training.
Furthermore, after K times of iteration, K convolutional neural network regression prediction models are obtained, the K convolutional neural network prediction models are used as basis learners, a Bagging algorithm is used for synthesizing the basis learners, and the prediction average value of the K convolutional neural network basis learners is used as the output of the attitude prediction model.
Further, in the training of the preset posture prediction model, the gesture image types comprise a color image under illumination and an infrared gray image without illumination, and the color image shot in the daytime is processed into a gray image;
the gesture image content includes at least gesture images of different situations of a bare hand, wearing a half-finger glove, and wearing a full-finger glove.
A second aspect of the present disclosure provides a multi-view image-based gesture pose prediction system.
A multi-view image based gesture pose prediction system, comprising:
a data acquisition module configured to: acquiring gesture image data of at least two visual angles;
a pose prediction module configured to: obtaining each gesture image prediction angle according to each acquired gesture image at each view angle and a preset gesture prediction model corresponding to each view angle;
a pose determination module configured to: taking the average value of the predicted angles of all the gesture images as a final gesture angle;
in the training of the preset attitude prediction model, the maximum error of the kth convolutional neural network on a training set, the relative error of each sample, the regression error rate of the kth convolutional neural network and the weight coefficient of the weak learner are calculated, and the sample weight is updated according to the obtained maximum error, relative error, regression error rate and the weight coefficient of the weak learner.
A third aspect of the present disclosure provides a computer-readable storage medium, on which a program is stored, which when executed by a processor, implements the steps in the multi-view image-based gesture posture prediction method according to the first aspect of the present disclosure.
A fourth aspect of the present disclosure provides an electronic device, including a memory, a processor, and a program stored on the memory and executable on the processor, wherein the processor implements the steps of the multi-view image-based gesture posture prediction method according to the first aspect of the present disclosure when executing the program.
Compared with the prior art, the beneficial effect of this disclosure is:
1. according to the method, the system, the medium or the electronic equipment, the gesture pictures collected by the multi-view camera are used for training, prediction errors caused by shielding of gestures can be greatly reduced, prediction models are respectively established for the gesture images at all views, the average value of the prediction results is taken as the final prediction result, and the accuracy of gesture posture prediction is improved.
2. Compared with the depth camera used in the prior art, the method, the system, the medium or the electronic equipment preferably uses the double-view-angle camera consisting of two most common day and night dual-purpose cameras, has a simple structure, saves the cost, and is more economical and convenient.
3. The method, the system, the medium or the electronic equipment preferably use a day and night dual-purpose camera, can shoot color images when the day is lighted, and shoot infrared gray-scale images when the night is not lighted, so that the gesture posture prediction all day can be realized without time limitation, and the gesture posture prediction all day long-term use can be realized and the gesture posture prediction method can be used outdoors.
4. The method, the system, the medium or the electronic equipment disclosed by the disclosure are not limited to predicting completely naked fingers, and can also establish a gesture data set comprising a gray scale image and an infrared gray scale image of a naked hand, a half-finger glove and a full-finger glove, so that the gesture posture prediction range is expanded, and the method, the system, the medium or the electronic equipment can be used for a special environment for gesture prediction.
5. Compared with the method based on image depth information, which needs to segment the gesture contour region and match the gesture key points, the method, the system, the medium or the electronic equipment disclosed by the disclosure reduces the complexity of a training model and improves the training and predicting speed.
6. The method, the system, the medium or the electronic equipment disclosed by the disclosure are based on the convolutional neural network and combined with the idea of integrated learning, so that the gesture posture prediction algorithm is improved, and the prediction precision is improved.
Advantages of additional aspects of the disclosure will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the disclosure.
Drawings
The accompanying drawings, which are included to provide a further understanding of the disclosure, illustrate embodiments of the disclosure and together with the description serve to explain the disclosure and are not to limit the disclosure.
Fig. 1 is a schematic flowchart of a gesture posture prediction method provided in embodiment 1 of the present disclosure.
Fig. 2 is a structural diagram of a deep-integration gesture regression prediction model constructed by using a convolutional neural network and an Adaboost integration algorithm according to embodiment 1 of the present disclosure.
Fig. 3 is a structural diagram of a deep-integration gesture regression prediction model constructed by using a convolutional neural network and a Bagging integration algorithm according to embodiment 1 of the present disclosure.
Fig. 4 is a first schematic diagram of a training phase of the wearable gesture prediction system and the prediction method provided in embodiment 1 of the present disclosure;
wherein, M1 is a double-view camera, M2 is a head-mounted virtual vision device, M3 is a wireless transmission module, M4 is a controller of the system, and M5 is an attitude sensor.
Fig. 5 is a schematic diagram ii of a prediction stage of the wearable gesture prediction system and the prediction method provided in embodiment 1 of the present disclosure;
wherein, M1 is a dual-view camera, M2 is a head-mounted virtual vision device, M3 is a wireless transmission module, and M4 is a system controller.
Fig. 6 is a schematic diagram of right-handed cartesian coordinate system hand type prediction provided in embodiment 1 of the present disclosure.
Fig. 7 is a block diagram of a wearable gesture prediction system provided in embodiment 2 of the present disclosure.
Detailed Description
The present disclosure is further described with reference to the following drawings and examples.
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the disclosure. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present disclosure. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
The embodiments and features of the embodiments in the present disclosure may be combined with each other without conflict.
Example 1:
the embodiment 1 of the present disclosure provides a gesture posture prediction method based on a multi-view camera, which is different from a previous gesture posture prediction method based on a depth image including depth information.
In the embodiment, preferably, a binocular camera is used for gesture prediction, as shown in fig. 1, a convolutional neural network is used as a base learner for regression prediction of a preset gesture, and an integrated learning algorithm is combined to establish a deep integrated gesture prediction model.
The embodiment specifically describes the proposed gesture posture prediction system and method for a right-handed cartesian coordinate system gesture, and mainly includes the following technical solutions:
step 1: and (3) establishing connection of the system, namely establishing connection of the system controller M4 with the wearable double-view-angle camera M1, the head-mounted virtual vision equipment M2 and the wireless data transmission device M3.
Step 2: the gesture of the posture to be predicted is determined, the gesture of the posture to be predicted is a right-hand cartesian coordinate system hand type, and a gesture image data set containing angle data is established for the hand type. The method comprises the following specific steps:
step 2.1: fig. 4 is a schematic diagram of a training phase of the gesture posture prediction system and method for a wearable dual-view camera according to this embodiment, and as shown in fig. 4, a cartesian coordinate system gesture of a right hand to be predicted is maintained, and a posture sensor is attached to the inner side of a palm, so that the positive directions of the X, Y, Z axes of the posture sensor are respectively kept consistent with the directions of the index finger, the middle finger and the thumb of the right hand, and a corresponding relationship between the coordinate axes of the posture sensor and the gesture is established.
Step 2.2: the system controller M4 controls the double-view camera M1 to collect right-hand Cartesian coordinate system gesture images of left and right views, and meanwhile, the controller collects angle data output by the attitude sensor M5 in real time and uses the angle data output by the attitude sensorThe images are named, wherein alpha, beta and gamma are angle data of the three fingers of the right-handed Cartesian coordinate system gesture in the space respectively.
Step 2.3: and transforming different angles of the preset gesture in the space, repeating the steps 2.1 and 2.2, and collecting 20000 gesture images and angle data. The acquired gesture image types comprise color images under illumination and infrared gray images without illumination, the acquired gesture image contents comprise gesture images under different conditions of wearing a bare hand, wearing a half-finger glove and wearing a full-finger glove, and the gesture prediction universality and the gesture prediction scene adaptability can be expanded.
Step 2.4: and (3) arranging the gesture images with the attitude angle data acquired in the steps, processing all the shot color images into gray images, enabling the data set to only contain single-channel images with the same data types of the gray images and the infrared gray images, and finishing the establishment of the data set after the processing.
And step 3: and (3) training a deep integrated gesture posture prediction model by using the Cartesian coordinate system gesture image data set established in the step (2), wherein the prediction model takes a convolutional neural network as a base learning device to perform regression prediction, combines the thought of integrated learning, trains to obtain a plurality of convolutional neural network base learning devices, and synthesizes the output results of all the base learning devices by adopting the integrated learning combination strategy to obtain the deep integrated gesture posture prediction model. In the method described in this embodiment, a dual-view camera composed of a day-night dual-purpose camera is used, and gesture images including left and right views are acquired, so that a depth-integrated gesture posture prediction model can be respectively established for the gesture images at the left and right views, and the average value of the prediction output results of the two prediction models is used as final prediction data.
In some embodiments, the ensemble learning may use Adaboost algorithm, as shown in fig. 2, in combination with a convolutional neural network to construct a deep integrated gesture pose prediction model, which includes the following specific steps:
step 3.1.1: and (3) dividing the Cartesian coordinate system gesture image data set established in the step (2) into a training set S and a test set T according to the 8:2 ratio by using a leaving method, wherein the training set S comprises 16000 gestures and angle data, and the test set comprises 4000 gesture images and corresponding angle data.
Training set samplexiA two-dimensional image of the gesture is represented,representing the current angle data of the gesture, whereinRespectively the angle of the finger corresponding to the three axes of xyz.
Step 3.1.2: assigning an initial sample weight W (1) — (W) to all the left-view gesture image samples in the training set S11,W12,..,W1m) Wherein, m is 16000,
step 3.1.3: training a convolutional neural network regression prediction model on a sample set with W (1) weight, and using a Resnet or Alexnet model of a classical model structure of a convolutional neural network as a base learner for realizing regression prediction on gestures. Inputting the image into a convolutional neural network structure, setting the iteration times to be 100 times, setting the learning rate to be 0.001, selecting a mean square error function as a loss function, updating and optimizing model parameters by using a gradient descent method, fitting a gesture posture prediction model, and training to obtain a first convolutional neural network base learner.
Step 3.1.4: in this embodiment model prediction estimationCalculating the maximum error of the first convolutional neural network on a training set for the angles of three axes of the xyz axis, and taking the maximum value of the mean value of the three angle errors as the maximum error:
wherein f is1(xαi),f1(xβi),f1(xγi) Respectively representing the angles of the first prediction model to the x, y and z axes predicted by the ith sample;
calculate the relative error for each sample:
calculating the regression error rate of the first convolutional neural network:
calculating the coefficients of weak learning:
step 3.1.5: updating the sample weights according to the data calculated in step 3.1.4:
wherein the content of the first and second substances,a second convolutional neural network regression prediction model is obtained using training with new sample weights.
Step 3.1.6: repeating the steps, iterating for k times, training to obtain k convolutional neural network regression prediction models, according to the combination strategy of ensemble learning, firstly taking the median of the weights of k weak learners, then adopting the weak learners corresponding to the median of the weights as the strong learners of the prediction model to generate a final strong regressor F (x), and obtaining a deep ensemble gesture regression prediction model F based on the left visual angle gesture imageL(x)。
Step 3.1.7: according to the steps, the depth integration gesture posture prediction model F based on the right visual angle gesture image can be obtained by using the same methodR(x)。
In other embodiments, step 3 may also use a Bagging algorithm in ensemble learning to construct a deep integration gesture posture prediction model, and fig. 3 is a structural diagram of the deep integration gesture posture regression prediction model constructed by using a convolutional neural network and the Bagging integration algorithm in the present invention, and its specific steps are as follows:
step 3.2.1: and (3) dividing the Cartesian coordinate system gesture image data set in the step (2) into a training set S and a test set T according to a certain proportion by using a leave-out method, wherein the training set S comprises 16000 gestures and angle data, and the test set comprises 4000 gesture images and corresponding angle data.
Training set samplexiA two-dimensional image of the gesture is represented,representing the current angle data of the gesture, whereinRespectively the angle of the finger corresponding to the three axes of xyz.
Step 3.2.2: and (3) extracting samples from the left visual angle gesture images in the training set S by using a self-sampling method, namely carrying out uniform and replaced random extraction 16000 times.
Step 3.2.3: and loading the extracted gesture image and the angle data of the gesture, and performing preprocessing operations such as saturation improvement, contrast improvement, clipping and the like on the gesture image.
Step 3.2.4: training a regression prediction model of the convolutional neural network on the extracted sample data, wherein a model of a classical model structure of the convolutional neural network, such as Resnet or Alexnet, can be used as a base learner for realizing regression prediction on the gesture. Inputting an image into a convolutional neural network structure, setting the iteration times to be 100 times, setting the learning rate to be 0.001, selecting a mean square error function as a loss function, updating and optimizing model parameters by using a gradient descent method, fitting a gesture posture prediction model, and training to obtain a convolutional neural network base learner f1(x)。
Step 3.2.5: repeating the operations of the steps 3.2.1, 3.2.2, 3.2.3 and 3.2.4, and training to obtain 20 different convolutional neural network prediction models fk(x)。
Step 3.2.6: taking the 20 convolutional neural network prediction models obtained by training in the step 3.2.5 as base learners, synthesizing the base learners by using a Bagging algorithm, and taking the prediction average value of the 20 convolutional neural network base learners as the output of the deep integration gesture posture prediction model, namelyObtaining a depth integration gesture posture prediction model F based on left visual angle gesture imagesL(x)。
Step 3.2.7: repeating the step 3.2, and obtaining a depth integration gesture posture prediction model F based on the right visual angle gesture imageR(x)。
And 4, step 4: completing the training stage of the model after the steps 1, 2 and 3 are operated, and obtaining a deep integration gesture posture model F based on the left visual angle gesture image trainingL(x) And a depth integration gesture posture prediction model F based on right visual angle gesture image trainingR(x) In that respect After the trained prediction model is obtained, the prediction order is shown in the schematic diagram of the prediction stage in FIG. 5The gesture sensor is not needed to be used for collecting angle data, the model can directly predict the gesture angle data of the gesture through a two-dimensional image of the gesture, and the specific steps in the prediction stage are as follows:
step 4.1: and controlling the double-view camera to shoot images of the left view angle and the right view angle for the predicted Cartesian coordinate system gestures by the system controller. If the color image is shot under the illumination, the color image needs to be processed into a single-channel gray image and other preprocessing operations, and the image which meets the input requirement of the depth integration gesture posture prediction model is obtained.
Step 4.2: inputting the processed gesture image of the left visual angle into a trained deep integration gesture posture prediction model FL(x) The prediction is carried out, and the prediction angle data of the left visual angle image is output Inputting the processed gesture image of the right visual angle into a trained deep integration gesture posture prediction model FR(x) Predicting, and outputting the predicted angle data of the right view angle imageTaking the average value of the prediction results of the deep integration gesture posture prediction models of the left view angle and the right view angle as the final gesture posture prediction resultThe prediction result can be more accurate, and the prediction data result can be described by the following formula:
and 5: virtually displaying the photographed gesture image on the virtual vision device M2, and displaying the angle data output by the depth-integrated gesture posture prediction modelThe prediction result can be conveniently observed by the staff, and the method is more visual and intuitive. Meanwhile, the angle data output by the model prediction is transmitted to the slave end robot through the wireless transmission module M3, so that the corresponding control operation of the slave end robot is realized.
The embodiment provides a gesture posture prediction method based on a dual-view camera, which is different from the previous gesture posture prediction method based on a depth image containing depth information, and the gesture prediction can be realized only by using common two-dimensional images of left and right visual angles of a gesture shot by the dual-view camera.
The prediction method is characterized in that a deep integrated gesture posture prediction model is established for a preset hand type, and the prediction method mainly comprises two stages of model training and prediction. Before training, a two-dimensional image dataset containing angle data needs to be established for a preset gesture.
The data set image comprises a gesture color image shot by a camera with double visual angles under the condition of illumination in the daytime and an infrared gray image shot under the condition of no illumination at night, and the color image shot in the daytime is processed into a gray image, so that the images in the data set are single-channel images with consistent data formats; the acquired image content comprises gesture images under different conditions of bare hands, wearing half-finger gloves, wearing full-finger gloves and the like.
A convolutional neural network is used as a base learner to perform regression prediction, and a deep integrated gesture posture prediction model is established by combining an integrated learning algorithm. After the prediction model is obtained through training, the two-dimensional image of the gesture is input into the depth integration gesture posture prediction model in the prediction stage, and the output is the current angle posture data of the gesture.
It is to be noted that the hand type of this embodiment is only an example hand type, and the prediction method can be used for predicting the gesture of any hand type, and the specific hand type can be set according to specific needs, a two-dimensional image data set with angle data is established for the specific hand type, and the gesture prediction of the gesture can be realized by training to obtain a corresponding deep-integration gesture prediction model.
Example 2:
the embodiment 2 of the disclosure provides a wearable gesture posture prediction system based on a dual-view camera, solves the problem that a computer only predicts the spatial posture of the computer through a two-dimensional gesture image, and can predict different angular postures of any gesture in the space.
In this embodiment, a typical gesture named as a right-handed cartesian coordinate system gesture is taken as an example, as shown in fig. 6, that is, a gesture in which a thumb, an index finger and a middle finger of a right hand are perpendicular to each other is obtained by predicting angle data of the thumb, the index finger and the middle finger of the right hand in spaceThe gesture posture prediction system and the gesture posture prediction method are further explained for the hand model.
Fig. 7 is a block diagram of the wearable gesture posture prediction system, and fig. 4 and 5 are schematic diagrams of two stages of training and prediction of the wearable dual-view camera gesture posture prediction system and method, respectively.
As shown in fig. 4 and 5, the system mainly includes a system controller M4, a wearable dual-view camera M1, an attitude sensor M5, a wireless transmission module M3, and a head-mounted virtual vision device M2.
As shown in fig. 7, in the system, a controller of the system may adopt some kind of embedded controller, a depth-integrated gesture prediction model trained in advance for a preset gesture is loaded in the controller, the controller is used to collect gesture image data of left and right two viewing angles photographed by the wearable dual-view camera in real time, prediction of a gesture posture can be realized by inputting the collected gesture image into the depth-integrated gesture prediction model trained in advance, and a prediction result is transmitted to the head-mounted virtual visual device, so that an operator can conveniently view the gesture prediction result.
The wearable double-view camera is a double-view camera consisting of day and night dual-purpose cameras and is used for collecting gesture images of a left view angle and a right view angle of a gesture posture. The day and night dual-purpose camera can shoot color images under the condition of illumination in the day and infrared gray images under the condition of no illumination at night, can be used all day long, and can be applied outdoors. Wear virtual vision equipment, can show virtual gesture image and gesture angle gesture data, make things convenient for the staff to observe the prediction result and more visual and more.
For a specific prediction method, see embodiment 1, which is not described herein again.
Example 3:
the embodiment 3 of the present disclosure provides a gesture posture prediction system based on a multi-view image, including:
a data acquisition module configured to: acquiring gesture image data of at least two visual angles;
a pose prediction module configured to: obtaining each gesture image prediction angle according to each acquired gesture image at each view angle and a preset gesture prediction model corresponding to each view angle;
a pose determination module configured to: taking the average value of the predicted angles of all the gesture images as a final gesture angle;
in the training of the preset attitude prediction model, the maximum error of the kth convolutional neural network on a training set, the relative error of each sample, the regression error rate of the kth convolutional neural network and the weight coefficient of the weak learner are calculated, and the sample weight is updated according to the obtained maximum error, relative error, regression error rate and the weight coefficient of the weak learner.
The working method of the system is the same as that provided in embodiment 1, and is not described herein again.
Example 4:
the embodiment 4 of the present disclosure provides a computer-readable storage medium, on which a program is stored, which when executed by a processor, implements the steps in the multi-view image-based gesture posture prediction method according to the embodiment 1 of the present disclosure.
Example 5:
the embodiment 5 of the present disclosure provides an electronic device, which includes a memory, a processor, and a program stored in the memory and executable on the processor, and when the processor executes the program, the processor implements the steps in the multi-view image-based gesture posture prediction method according to embodiment 1 of the present disclosure.
As will be appreciated by one skilled in the art, embodiments of the present disclosure may be provided as a method, system, or computer program product. Accordingly, the present disclosure may take the form of a hardware embodiment, a software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present disclosure may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, optical storage, and the like) having computer-usable program code embodied therein.
The present disclosure is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the disclosure. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by a computer program, which can be stored in a computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. The storage medium may be a magnetic disk, an optical disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), or the like.
The above description is only a preferred embodiment of the present disclosure and is not intended to limit the present disclosure, and various modifications and changes may be made to the present disclosure by those skilled in the art. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present disclosure should be included in the protection scope of the present disclosure.