CN101377813B - Method for real time tracking individual human face in complicated scene - Google Patents

Method for real time tracking individual human face in complicated scene Download PDF

Info

Publication number
CN101377813B
CN101377813B CN2008102003102A CN200810200310A CN101377813B CN 101377813 B CN101377813 B CN 101377813B CN 2008102003102 A CN2008102003102 A CN 2008102003102A CN 200810200310 A CN200810200310 A CN 200810200310A CN 101377813 B CN101377813 B CN 101377813B
Authority
CN
China
Prior art keywords
face
area
people
human face
pixel
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN2008102003102A
Other languages
Chinese (zh)
Other versions
CN101377813A (en
Inventor
寇超
白琮
陈泉林
王华红
王少波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Haian Soochow University Advanced Robot Research Institute
Original Assignee
University of Shanghai for Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Shanghai for Science and Technology filed Critical University of Shanghai for Science and Technology
Priority to CN2008102003102A priority Critical patent/CN101377813B/en
Publication of CN101377813A publication Critical patent/CN101377813A/en
Application granted granted Critical
Publication of CN101377813B publication Critical patent/CN101377813B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a method for tracing the face of a single person in real time in complex scenes. The method comprises the following steps: (1) pretreating frames: linearly transforming the frames captured by a video and filtrating waves with Gaussian templates; (2) detecting the face by adopting complexion features: detecting the face by adopting the convergence of complexion in YCrCb colored space and morphologic operation of the face; (3) elimination of areas similar to the complexion: elimination of non-face areas similar to complexion in combination with the face area and the face geometrical features to get a perfect face area before the barycenter of the face area is computed with the perfect face area and the initial search area is set around the barycenter; (4) tracing the face: tracing the face by adopting the continuous adaptive mean deflection method with the face detection results as novel back projection drawings and features. In the method, the face is abstracted and tracked in combination with multiple features; the effects are stable and the complexity of computation is low. The method can be widely applied in the fields of video monitoring, man computer interaction interfaces and so on.

Description

The method for real time tracking of individual human face in the complex scene
Technical field
The present invention relates to the mutual and field of video monitoring of man-machine interface, particularly to the real-time tracing of individual human face in the complex scene.Specifically be a kind ofly to take all factors into consideration many features, utilize the continuous average drifting method of novel trans to perspective view.
Background technology
In man-machine interaction and video monitoring etc. are used,, will be directly connected to the overall performance of system to the detection and tracking effect of people's face as one of notable feature of discerning personal identification.Recent decades, existing countless scholars did extensive studies and discussion to the face tracking problem, wherein mainly were divided into: based on the faceform, based on characteristic matching, based on movable information and based on the method for the colour of skin.Yet owing to exist such as head rotation, all multifactor influences such as illumination, shade and shelter cause the face tracking in the actual complex environment always all to be difficult to effective realization.
To consulting as can be known of prior art document, in various trackings, the average drifting algorithm is more satisfactory continuously.It is proposed and is successfully applied in the computer game human-computer interaction interface machine vision to the seizure of people's face motion in the article of a piece that delivered on the Intel technology quarterly in 1998 " Computer Vision Face Trackingfor Use in a Perceptual User Interface " by name by G.R.Bradski..Article is based on the average drifting algorithm, utilizes statistics with histogram in the inherent H color channel in region of search to calculate the similarity degree of follow-up each frame pixel of video sequence and skin color probability, and as feature people's face followed the tracks of.Yet, at the H color channel interference of background pixel in the region of search is carried out being subject in the statistics with histogram process in the region of search, influence the accuracy of calculating pixel class skin color probability, and then can not reach good tracking effect.Simultaneously, the initial search area in the literary composition must manually be set.
Summary of the invention
The objective of the invention is to overcome the deficiency of above-mentioned prior art, the method for real time tracking of individual human face in a kind of complex scene is provided, the binaryzation that detects with people's face is the probability of the pixel class colour of skin in the token image as a result, and as feature people's face is followed the tracks of.Simultaneously, utilize people's face testing result of first two field picture to determine initial search area automatically, algorithm is moved automatically.
For achieving the above object, the present invention adopts following technical proposals:
A kind of method for real time tracking at individual human face in the complex scene, its step comprises:
(1) picture pre-service: for reducing the influence of ambient lighting and picture noise, the present invention at first implements segmentation grey level fractal transform and gaussian filtering to image, and purpose is to reduce the interference of ambient lighting and picture noise;
(2) utilizing features of skin colors to carry out people's face detects: utilize the polymerism of the colour of skin in the YCrCb color space to carry out people's face and detect, and remove detection noise by morphological operation;
(3) elimination of class area of skin color: at the class area of skin color, bonded area thresholding and people's face height, widely further eliminate than restriction condition, determine the central point of facial zone by the space weight of calculating human face region, and be center setting initial search area with it;
(4) face tracking: people's face is detected the binaryzation result make tracked feature, utilize continuous adaptive mean shift method to realize following the tracks of.
Experimental result shows that this method does not need manual intervention to carry out real-time follow-up to the people's face in the picture automatically, and accuracy rate can reach more than 90%.
Below each step is described further:
(1) picture pre-service: comprise grey level fractal transform and the filtering of Gauss's template.Because the influence of complex illumination can make face some high light or dim zone occur, and these zones all can exert an influence to the judgement of the colour of skin in the environment.Therefore the brightness of picture need be carried out linear adjustment, so that it reaches the overall uniform effect.And because the hardware reason of video capture device can noise occur unavoidably in image, this can exert an influence to the judgement of pixel class area of skin color equally.Utilize Gauss's template that image is carried out convolution, can under the prerequisite that keeps the edge, remove noise preferably.
(2) utilizing features of skin colors to carry out people's face detects: in various color spaces, the colour of skin in the Cr among the YCrCb, the Cb passage has good polymerism, and observation data can obtain more satisfactory segmentation threshold by experiment, and then produces good skin color segmentation result.Simultaneously the initial detecting result is carried out morphological operation, eliminate some tiny cavity and detection noise.
(3) elimination of class area of skin color: utilize colour of skin information that face is detected and no doubt have many advantages, yet only rely on this single information always to be subjected to the interference of class area of skin color, and these zones can't rely on shape filtering to eliminate merely.Because some class area of skin color is bigger, needs repeatedly corrosion could eliminate, and repeatedly can cause the facial zone atrophy processing time.This method is further eliminated this class zone with geometric properties in conjunction with people's face is long-pending.At first be when zone during, think that this zone does not belong to people's face less than certain area; Next is that the long and short axial ratio of human face region is limited, and the zone in this scope all will not be eliminated, to reduce the influence of class area of skin color as far as possible.Simultaneously, in detected human face region, calculate centre of gravity place, and be that initial search area is set at the center, algorithm is moved automatically with it.
(4) face tracking: with people's face testing result as novel trans to perspective view, spatial moment in the search window is carried out iteration, make feature in the current search frame trend towards the distribution pattern of target signature, thereby instruct continuous average drifting method to adjust the position of search window and major axis angle automatically to dope the information of target in next frame, make track algorithm continue to carry out, when guaranteeing good tracking effect, also have simple calculated amount concurrently.
The pretreated specific implementation method of above-mentioned picture is as follows:
(1) linear transformation
If the brightness of piece image by f (x, y) expression, (x, the y) locus of representative image pixel, Min[f (x, y)], Max[f (x, y)] be the maximum of linear transformation, minimum thresholding, then pass through brightness of image G after the linear transformation (x y) is:
G ( x , y ) = 0 f ( x , y ) &le; Min [ f ( x , y ) ] f ( x , y ) - Min [ f ( x , y ) ] Max [ f ( x , y ) ] - Min [ f ( x , y ) ] &times; 256 Min [ f ( x , y ) ] < f ( x , y ) < Max [ f ( x , y ) ] 255 f ( x , y ) &GreaterEqual; Max [ f ( x , y ) ] ,
(2) Gauss's masterplate filtering: filtering adopts discretize Gauss template that image is carried out convolution, and discretize Gauss's template is as follows:
Figure DEST_PATH_GSB00000057788900012
Above-mentioned to utilize features of skin colors to carry out the specific implementation method that people's face detects as follows: digital picture is as follows to the conversion formula of YCrCb color space by RGB, wherein R, G, B, Y, C r, C bThe pixel value of difference represent pixel in corresponding color channel:
Y = 0.299 &times; R + 0.587 &times; G + 0.114 &times; B Cr = R - Y Cb = B - Y ,
Utilize following formula afterwards, the pixel value that satisfies condition is set to 255, otherwise pixel value is set to 0, obtains people's face Preliminary detection result with this.C wherein rRepresent the pixel value of digital picture in respective channel, C RMax, C RMinC when defining pixel and belonging to human face region rThe maximal value that is satisfied, minimum value scope; C bRepresent the pixel value of digital picture in respective channel, C BMax, C BMinC when defining pixel and belonging to human face region bThe maximal value that is satisfied, minimum value scope.
(C rMin<C r<C rMax)∩(C bMin<C b<C bMax),
The elimination specific implementation method of above-mentioned class area of skin color is as follows:
(1) takes all factors into consideration the long-pending and geometric properties elimination class area of skin color of people's face
Each isolated area is asked its area, in digital picture, be the number of pixels in the zone, when region area is just eliminated it less than 100 the time from the Preliminary detection result; Simultaneously, remaining isolated area is calculated long and short axial ratio respectively,, the zone that long and short axial ratio does not change in 1.0~2.3 scopes is further eliminated according to facial characteristics and observation experience.
(2) initial search area determines
Initial search area is determined by following formula:
x o = &Sigma; x &Sigma; y xI ( x , y ) &Sigma; x &Sigma; y I ( x , y ) y o = &Sigma; x &Sigma; y yI ( x , y ) &Sigma; x &Sigma; y I ( x , y ) ,
Wherein, (x o, y o) be the center of initial search area, (and x, y) representative's face detects the locus of binary image pixel, and (x is that it is at (x, the pixel value of y) locating y) to I; With (x o, y o) be the center, the rectangular window of setting one 200 * 100 is as initial search area.
The present invention has following conspicuous outstanding substantive distinguishing features and remarkable advantage compared with prior art:
Face tracking method of the present invention can be when having low calculated amount to complex scene in the situations such as rotation, inclination of people's face successfully realize following the tracks of, after following the tracks of failure, also can carry out again automatically, and to not obvious such as the interference of hand class area of skin color.Simultaneously, the present invention has also realized the automatic setting of initial search area, and whole process can be carried out under unmanned situation of interfering automatically.The present invention extracts people's face in conjunction with multiple characteristics and realizes following the tracks of, and effect stability and computation complexity are low, can be widely used in fields such as video monitoring, human-computer interaction interface.
Description of drawings
Fig. 1 is a FB(flow block) of the present invention.
Fig. 2 is for carrying out the comparison of morphological operation front and back to the initial detecting result among the present invention.
The Cr that Fig. 3 utilizes for the present invention, the colour of skin polymerization property in the Cb color channel are showed.
Fig. 4 is that final tracking effect of the present invention is showed.
Embodiment
A preferred embodiment of the present invention accompanying drawings is as follows: referring to Fig. 1, this is at the method for real time tracking of individual human face in the complex scene, and its step comprises:
(1) picture pre-service: the picture to Video Capture carries out image linear transformation and the filtering of Gauss's template;
(2) utilizing features of skin colors to carry out people's face detects: utilize the colour of skin people's face to be detected in the CrCb passage at the polymerism of YCrCb color space, and utilize the noise after morphological operation is tentatively removed detection;
(3) elimination of class area of skin color: at the class area of skin color in the scene, further eliminate, calculate the human face region barycenter, and be center setting initial search area with it in conjunction with human face region area and geometric properties;
(4) face tracking: with people's face testing result as novel trans to perspective view, utilize continuous adaptive mean shift method to realize face tracking.
The pretreated implementation method of above-mentioned picture is as follows:
Before people's face is detected, need carry out pre-service to picture.Specifically comprise gray scale linear transformation and Gauss's masterplate convolution.Wherein gray scale linear transformation purpose is the overall brightness adjustment of picture even.Adjustment process is calculated according to formula (1):
G ( x , y ) = 0 f ( x , y ) &le; Min [ f ( x , y ) ] f ( x , y ) - Min [ f ( x , y ) ] Max [ f ( x , y ) ] - Min [ f ( x , y ) ] &times; 256 Min [ f ( x , y ) ] < f ( x , y ) < Max [ f ( x , y ) ] 255 f ( x , y ) &GreaterEqual; Max [ f ( x , y ) ] - - - ( 1 )
Wherein, (x y) represents gray-scale value in the original image to f, and ((x y) is gray-scale value after the linear transformations to G for x, the y) locus of representative image pixel.Min[f (x, y)], Max[f (x, y)] be the maximum, the minimum thresholding that set.
Afterwards, adopt following Gauss's template and original image to carry out convolution, to eliminate the noise that brings by video capture device.
Figure G2008102003102D00052
The above-mentioned implementation method of utilizing features of skin colors to carry out the detection of people's face is as follows: in various color spaces, the colour of skin in Cr among the YCrCb, the Cb passage has good polymerism (seeing that Fig. 3 is that colour of skin piece polymerism in Cr, Cb passage is showed), therefore image need be transformed into the YCrCb color space from RGB, wherein R, G, B, Y, Cr, Cb distinguish the pixel value of represent pixel in respective channel, conversion formula such as formula (2):
Y = 0.299 &times; R + 0.587 &times; G + 0.114 &times; B Cr = R - Y Cb = B - Y - - - ( 2 )
As table 1, show according to experimental observation, under different periods, illumination condition, by setting different thresholdings, can utilize formula (3) in Cr, Cb passage, people's face to be detected, and the pixel value that satisfies condition is changed to 255, otherwise pixel value is changed to 0, obtains people's face Preliminary detection result with this
(C rMin<C r<C rMax)∩(C bMin<C b<C bMax) (3)
Table 1
Figure G2008102003102D00061
Carry out morphological operation according to table 1 couple initial detecting result afterwards, eliminate some tiny cavities and detection noise (Fig. 2 is seen in morphological operation result displaying).
The implementation method of the elimination of above-mentioned class area of skin color is as follows: after utilizing features of skin colors to finish people's face Preliminary detection, comprise that some pixel values of human face region meeting generation are 255 zone, these class area of skin color to be eliminated: at first each isolated area is asked its area (being the number of pixels in the zone in digital picture), when region area is just eliminated it less than 100 the time from the Preliminary detection result.Secondly, more remaining isolated area is calculated long and short axial ratio respectively.According to facial geometric properties and observation experience, the zone that long and short axial ratio does not change in 1.0~2.3 scopes is further eliminated.Through above-mentioned processing, most class area of skin color can both be eliminated smoothly.
Set initial search area:
For tracing process is moved automatically, utilize formula (4) to calculate the facial zone barycenter and be center automatic setting initial search area with it:
x o = &Sigma; x &Sigma; y xI ( x , y ) &Sigma; x &Sigma; y I ( x , y ) y o = &Sigma; x &Sigma; y yI ( x , y ) &Sigma; x &Sigma; y I ( x , y ) - - - ( 4 )
Wherein, (x o, y o) be the center of human face region, (and x, y) representative's face detects the locus of binary image pixel, and (x is that it is at (x, the brightness value of y) locating y) to I.With (x o, y o) be the center, the rectangular window of setting one 200 * 100 is as initial search area.
The implementation method of above-mentioned face tracking is as follows:
By above-mentioned processing, face tracking just can move automatically.This is the computation process of an iteration, and concrete computation process is as follows:
(a) in first frame of sequence of frames of video, carry out people's face and detect,, then determine initial search area automatically if detect successfully; Otherwise then continue next frame is detected, till detecting the success of people's face
(b) in the binary picture that people's face detects, utilize 0 ~ 2 rank square of average drifting algorithm, in order to determine to follow the tracks of position of window, major axis deflection angle by one or many iterative computation zone
(c) keep following the tracks of the window size constancy, change the center and the deflection angle of search box, it is moved to definite position by (b)
(d) follow-up every two field picture is repeated (b), (c) step,, realize face tracking up to the window's position convergence.Fig. 4 is that final tracking effect of the present invention is showed, the result shows that the present invention has not only successfully realized aligning the tracking of face, under offside face and the head deflection situation good tracking effect is arranged also, in addition, the interference such as class area of skin color such as hands is also had stronger robustness.

Claims (4)

1. method for real time tracking at individual human face in the complex scene, its step comprises:
A. picture pre-service: the picture to Video Capture carries out image linear transformation and the filtering of Gauss's template;
B. utilizing features of skin colors to carry out people's face detects: utilize the colour of skin people's face to be detected in the CrCb passage at the polymerism of YCrCb color space, and utilize the noise after morphological operation is tentatively removed detection;
C. the elimination of class area of skin color: at the class area of skin color in the scene, further eliminate, calculate the human face region barycenter, and be center setting initial search area with it in conjunction with human face region area and geometric properties;
D. face tracking: with people's face testing result as novel trans to perspective view, utilize continuous adaptive mean shift method to realize face tracking.
2. the method for real time tracking at individual human face in the complex scene according to claim 1 is characterized in that the pretreated specific implementation method of described picture is as follows:
(1) linear transformation
If the brightness of piece image by f (x, y) expression, (x, the y) locus of representative image pixel, Min[f (x, y)], Max[f (x, y)] be the maximum of linear transformation, minimum thresholding, then pass through brightness of image G after the linear transformation (x y) is:
Figure F2008102003102C00011
(2) Gauss's masterplate filtering
Filtering adopts discretize Gauss template that image is carried out convolution, and discretize Gauss's template is as follows:
Figure F2008102003102C00012
3. the method for real time tracking at individual human face in the complex scene according to claim 1 is characterized in that the described specific implementation method of utilizing features of skin colors to carry out the detection of people's face is as follows:
Digital picture is as follows to the conversion formula of YCrCb color space by RGB, wherein R, G, B, Y, C r, C bThe pixel value of difference represent pixel in corresponding color channel:
Figure DEST_PATH_FSB00000057788800011
Utilize following formula afterwards, the pixel value that satisfies condition is set to 255, otherwise pixel value is set to 0, obtains people's face Preliminary detection result, wherein C with this rRepresent the pixel value of digital picture in respective channel, C RMax, C RMinC when defining pixel and belonging to human face region rThe maximal value that is satisfied, minimum value scope; C bRepresent the pixel value of digital picture in respective channel, C BMax, C BMinC when defining pixel and belonging to human face region bThe maximal value that is satisfied, minimum value scope,
(C rMin<C r<C rMax)∩(C bMin<C b<C bMax)。
4. the method for real time tracking at individual human face in the complex scene according to claim 1 is characterized in that the specific implementation method of elimination of described class area of skin color is as follows:
(1) takes all factors into consideration the long-pending and geometric properties elimination class area of skin color of people's face
Each isolated area is asked its area, and this area is the number of pixels in the zone in digital picture, when region area is just eliminated it less than 100 the time from the Preliminary detection result; Simultaneously, remaining isolated area is calculated long and short axial ratio respectively,, the zone that long and short axial ratio does not change in 1.0~2.3 scopes is further eliminated according to facial characteristics and observation experience;
(2) initial search area determines
Initial search area is determined by following formula:
Figure FSB00000057788800012
Wherein, (x O, y O) be the center of initial search area, (and x, y) representative's face detects the locus of binary image pixel, and (x is that it is at (x, the pixel value of y) locating y) to I; With (x O, y O) be the center, the rectangular window of setting one 200 * 100 is as initial search area.
CN2008102003102A 2008-09-24 2008-09-24 Method for real time tracking individual human face in complicated scene Expired - Fee Related CN101377813B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2008102003102A CN101377813B (en) 2008-09-24 2008-09-24 Method for real time tracking individual human face in complicated scene

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2008102003102A CN101377813B (en) 2008-09-24 2008-09-24 Method for real time tracking individual human face in complicated scene

Publications (2)

Publication Number Publication Date
CN101377813A CN101377813A (en) 2009-03-04
CN101377813B true CN101377813B (en) 2010-08-11

Family

ID=40421350

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2008102003102A Expired - Fee Related CN101377813B (en) 2008-09-24 2008-09-24 Method for real time tracking individual human face in complicated scene

Country Status (1)

Country Link
CN (1) CN101377813B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI401963B (en) * 2009-06-25 2013-07-11 Pixart Imaging Inc Dynamic image compression method for face detection
CN102054159B (en) * 2009-10-28 2014-05-28 腾讯科技(深圳)有限公司 Method and device for tracking human faces
CN102214291B (en) * 2010-04-12 2013-01-16 云南清眸科技有限公司 Method for quickly and accurately detecting and tracking human face based on video sequence
CN102096823A (en) * 2011-02-12 2011-06-15 厦门大学 Face detection method based on Gaussian model and minimum mean-square deviation
CN102431034B (en) * 2011-09-05 2013-11-20 天津理工大学 Color recognition-based robot tracking method
CN102750672A (en) * 2011-11-30 2012-10-24 新奥特(北京)视频技术有限公司 Method for realizing special effect through brightness adjustment
CN105069431B (en) * 2015-08-07 2018-09-14 成都明图通科技有限公司 The localization method and device of face
CN106127108B (en) * 2016-06-14 2019-07-16 中国科学院软件研究所 A kind of manpower image region detection method based on convolutional neural networks

Also Published As

Publication number Publication date
CN101377813A (en) 2009-03-04

Similar Documents

Publication Publication Date Title
CN101377813B (en) Method for real time tracking individual human face in complicated scene
Qian et al. A robust real-time face tracking algorithm
CN103761514B (en) The system and method for recognition of face is realized based on wide-angle gunlock and many ball machines
CN101673338B (en) Fuzzy license plate identification method based on multi-angle projection
Hongo et al. Focus of attention for face and hand gesture recognition using multiple cameras
JP2002501234A (en) Human face tracking system
CN103955949A (en) Moving target detection method based on Mean-shift algorithm
Amarnag et al. Real time eye tracking for human computer interfaces
Hidayatullah et al. CAMSHIFT improvement on multi-hue and multi-object tracking
Mu Ear detection based on skin-color and contour information
CN113592911A (en) Apparent enhanced depth target tracking method
Chen et al. Fast eye detection using different color spaces
CN105844235A (en) Visual saliency-based complex environment face detection method
CN111161219B (en) Robust monocular vision SLAM method suitable for shadow environment
KR20040042501A (en) Face detection based on template matching
Talea et al. Automatic combined lip segmentation in color images
Byun et al. Robust face detection and tracking for real-life applications
Huang et al. Development of the joint attention with a new face tracking method for multiple people
Qian et al. Face tracking using robust statistical estimation
CN111627047A (en) Underwater fish dynamic visual sequence moving target detection method
Yuen et al. Mouth state estimation in mobile computing environment
Huang et al. Multi-camshift for multi-view faces tracking and recognition
Sun et al. Real-time non-rigid object tracking using CAMShift with weighted back projection
Zhou et al. Adaptive object detection and recognition based on a feedback strategy
Kashima et al. An iris detection method using the hough transform and its evaluation for facial and eye movement

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
ASS Succession or assignment of patent right

Owner name: HAIAN SUSHI TECHNOLOGY TRANSFORMATION CENTER CO.,

Free format text: FORMER OWNER: SHANGHAI UNIVERSITY

Effective date: 20140512

C41 Transfer of patent application or patent right or utility model
COR Change of bibliographic data

Free format text: CORRECT: ADDRESS; FROM: 200444 BAOSHAN, SHANGHAI TO: 226600 NANTONG, JIANGSU PROVINCE

TR01 Transfer of patent right

Effective date of registration: 20140512

Address after: 226600 Yingbin Road 8, Haian Town Development Zone, Jiangsu, Nantong

Patentee after: Haian Su Fu Technology Transfer Center Co.,Ltd.

Address before: 200444 Baoshan District Road, Shanghai, No. 99

Patentee before: Shanghai University

ASS Succession or assignment of patent right

Owner name: HAIAN SUZHOU UNIVERSITY ADVANCED ROBOT RESEARCH IN

Free format text: FORMER OWNER: HAIAN SUSHI TECHNOLOGY TRANSFORMATION CENTER CO., LTD.

Effective date: 20141125

C41 Transfer of patent application or patent right or utility model
TR01 Transfer of patent right

Effective date of registration: 20141125

Address after: No. 8 Haian County in Jiangsu province was 226600 Avenue in Nantong City

Patentee after: HAIAN SOOCHOW UNIVERSITY ADVANCED ROBOT Research Institute

Address before: 226600 Yingbin Road 8, Haian Town Development Zone, Jiangsu, Nantong

Patentee before: Haian Su Fu Technology Transfer Center Co.,Ltd.

CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20100811

CF01 Termination of patent right due to non-payment of annual fee