Hostname: page-component-7bb8b95d7b-pwrkn Total loading time: 0 Render date: 2024-09-25T11:18:55.411Z Has data issue: false hasContentIssue false

Image-based anti-interference robotic Chinese character writing system

Published online by Cambridge University Press:  22 January 2024

Xian Li
Affiliation:
College of Automation Science and Engineering, South China University of Technology, Guangzhou, China
Chenguang Yang*
Affiliation:
College of Automation Science and Engineering, South China University of Technology, Guangzhou, China
Sheng Xu
Affiliation:
Guangdong Provincial Key Lab of Robotics and Intelligent System, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
Yongsheng Ou
Affiliation:
Faculty of Electronic Information and Electrical Engineering, Dalian University of Technology, Dalian, China
*
Corresponding author: Chenguang Yang; Email: cyang@ieee.org
Rights & Permissions [Opens in a new window]

Abstract

This article designs a robotic Chinese character writing system that can resist random human interference. Firstly, an innovative stroke extraction method of Chinese characters was devised. A basic Chinese character stroke extraction method based on cumulative direction vectors is used to extract the components that make up the strokes of Chinese characters. The components are then stitched together into strokes based on the sequential base stroke joining method. To enable the robot to imitate handwriting Chinese character skills, we utilised stroke information as the demonstration and modelled the skills using dynamic movement primitives (DMPs). To suppress random human interference, this article combines improved DMPs and conductance control to adjust robot trajectories based on real-time visual measurements. The experimental results show that the proposed method can accurately extract the strokes of most Chinese characters. The designed trajectory adjustment method offers better smoothness and robustness than direct rotating and translating curves. The robot is able to adjust its posture and trajectory in real time to eliminate the negative impacts of human interference.

Type
Research Article
Copyright
© The Author(s), 2024. Published by Cambridge University Press

1. Introduction

Today, robots are used in a wide range of applications. Robots show great advantages in industrial areas where repetitive work is performed because of their low cost, high efficiency and low scrap rate [Reference Li, Li and Luo1]. But in areas where motion trajectories do not repeat, robots have difficulty completing tasks. One of the classic tasks is handwriting. Handwriting actions are unavoidable in human life and play an important role in scenarios such as teaching and businesses both requiring signatures [Reference Yin, Alves-Oliveira, Melo, Billard and Paiva2, Reference Zeng, Huang, Chao and Zhou3]. English words are made up of 26 letters arranged in a row, but Chinese characters are two-dimensional pictures made up of Chinese strokes written in specific positions superimposed on each other [Reference Chao, Huang, Zhang, Shang, Yang, Zhou, Hu and Lin4]. How to encode Chinese characters is a challenge. As a result, the robotic writing of Chinese characters has also been studied by numerous researchers. At the same time, the structure of the strokes of Chinese characters has evolved over millennia to conform to the human mind and is the scientifically best way to segment Chinese characters. In other words, the segmentation of Chinese character strokes can also be applied to tasks that require the segmentation of superimposed trajectory images into executable trajectories, such as the grinding and cleaning of complex recesses in workpieces. At the same time, writing is a delicate manipulation task that would be difficult to accomplish if interference were encountered. Therefore, there is an urgent need to design an interference-resistant robotic Chinese character writing system for diverse applications.

The first step in imitating human writing is to disassemble Chinese characters into Chinese strokes. Currently, three methods are widely applied for Chinese character extraction, i.e., (a) computer font reproduction, (b) imitation of human writing trajectories and (c) Chinese character image decomposition. Computer font reproduction involves using the stroke information that comes with a standard font library [Reference Wang, Chen, Deng, Hutchinson and Dellaert5Reference Chuanjian, Chunmei and Jun8] and handling of word posters [Reference Yang, Chen, Zhou, Zheng, Xiao, Huang and Sun9Reference Lam and Yam13], but these methods are database dependent and cannot imitate the individual writing of a particular person. Imitation of human writing trajectories includes recording the trajectory of the pen tip when a person is writing [Reference Li, Sun, Zhou and Dai14, Reference Lin and Tang15]. The motion sensing input devices are utilised to capture the gestures and trajectory of the human hand [Reference Chao, Huang, Zhang, Shang, Yang, Zhou, Hu and Lin4], and the physical contact demonstrations [Reference Sun, Qian and Xu16]. However, these methods rely on human teaching and cannot break down Chinese strokes through pictures. Chinese character image decomposition means that it is possible to analyse the strokes of Chinese characters directly through pictures without relying on word library and human teaching. Chinese character image decomposition, the most difficult but widely used method, has been studied by many researchers. Consequently, many methods were developed, such as corner detection algorithm [Reference Chao, Huang, Lin, Yang, Hu and Zhou17, Reference Cao, Wu, Ao and Zhou18], point to boundary orientation distance of one triangular mesh [Reference Wang, Liang, Sun and Liu19], character library template matching segmentation [Reference Gan, Fang, Chao, Zhou, Yang, Lin and Shang20, Reference Wang, Jiang and Liu21], B-spline curve matching [Reference Ju, Ji, Li and Liu22], extraction of strokes using ambiguous zone information [Reference Cao, Wu, Ao and Zhou18], extraction of strokes using optimum paths [Reference Yao, Shao and Yi11], stroke speed feature and stroke vector feature segmentation [Reference Chang23], etc. However, these methods require a lot of computation. There is still a need for a simple picture-based method to extract strokes from Chinese characters.

The robot can write the corresponding Chinese characters according to the extracted strokes. However, in order to eliminate interference during the writing process, an algorithm that can adaptive the writing trajectory according to the position of the writing board is required. Dynamic movement primitive (DMP) is an effective method for modelling robot movement behaviour and biological phenomena [Reference Ijspeert, Nakanishi and Schaal24]. It has the advantage of being stable, simple and easy to generalise. It is possible to generalise a trajectory simply by modifying the position of the start and target points, and the shape of the generalised trajectory is same to the original demonstration. Because of its significant advantages, DMPs have been widely used and many studies have improved the learning and generalisation capabilities of DMPs [Reference Koutras and Doulgeri25]. In [Reference Yang, Chen, He, Cui and Li26], Gaussian mixture model and Gaussian mixture regression were integrated to enhance DMPs’ learning skills from multiple demonstrations. In [Reference Yang, Zeng, Cong, Wang and Wang27], the learning skills are split into a series of sub-skills, thus the generalisation ability of the learned skills is improved. In [Reference Anand, Østvik, Grøtli, Vagia and Gravdahl28], the framework of the DMPs is extended to accommodate real-time changes during task execution. In [Reference Liao, Jiang, Zhao, Wu, Yue and Mei29], a Riemannian-based DMP framework is proposed to learn and generalise multi-space data. In [Reference Zeng, Chen, Wang and Yang30], a biomimetic controller is integrated with DMPs to facilitate the learning and adaptation of compliant skills. The authors of [Reference Koutras and Doulgeri31] proposed a new DMPs to solve the problems arising from the spatial scaling of DMPs in Cartesian space. The authors of [Reference Abu-Dakka and Kyrki32] used Riemannian metrics to reformulate DMPs such that the resulting DMPs can be directly employed with quantities expressed as SymmSymmetric Positive Definite (SPD) matrices. The study in [Reference Wang, Yan, Wang, Gao, Du and Chen33] introduced a deep neural network in DMPs to address the invalidation of DMPs forcing terms. The authors of [Reference Yang, Zeng, Fang, He and Li34] developed a framework for the robot to learn both movement and muscle stiffness features. However, all these methods are difficult to adapt to the case where the start and target points are rotated simultaneously. The writing trajectory of the robot may be distorted when the writing board is rotated and flipped.

At the same time, the contact force between the pen tip and the writing board is an important piece of information during writing, especially when disturbed. Therefore, it is extremely necessary to design a writing system that combines stroke segmentation, trajectory generalisation, visual information and force information.

To address the above issues, the following work has been done. Firstly, inspired by the human mindset of splitting strokes, a direction-based algorithm for extending and stitching stroke components was designed to extract the strokes of a Chinese character picture in a simple and fast way. Subsequently, some improvements were made to the DMPs. The rotation and translation matrices were fused into the framework of the DMPs so that the DMPs could accommodate the simultaneous rotation of the start and target points, i.e., it could accommodate the rotation and flipping of the writing board. Finally, an interference-resistant robotic Chinese character writing system was developed. The system uses segmented strokes as demonstrations and then combines the rotation and translation matrices calculated from visual information with improved DMPs to generalise the trajectories in real time. At the same time, the position is regulated online using admittance control based on force sensor information. Thus, the robot can write complete characters on the writing board despite a human’s random movement, rotation and flipping.

The rest of this article is presented below. Section 2 describes the proposed method in this article, including stroke division, force-position hybrid control and the designed system. The stroke division algorithm is divided into stroke component extraction and stroke component connection. Modified DMPs and admittance control are introduced in force-position hybrid control. Finally, a jam-resistant robotic writing system was designed based on the above algorithm. Section 3 provides sufficient experimental examples to demonstrate the effectiveness of the stroke division algorithm, modified DMPs, admittance control, and the designed system. Section 4 is the conclusion.

2. Proposed method

2.1. Stroke division

Stroke division is one of the difficulties in the study of Chinese characters because it is tough to learn the connections and intersections in Chinese characters. When the endpoints of two strokes are at the same point, as shown in Fig. 1, the point is defined as a connection. Improper recognising this point will tend to misidentify multiple strokes as the same stroke, as shown in Fig. 2. The intersection is the point formed when two strokes intersect, and improper handling of this point will result in incorrect recognition of the strokes, as shown in Fig. 3.

Figure 1. The connection and intersection of Chinese character “zhong”.

Figure 2. (a) The correct stroke division for the Chinese character “zhong”. (b) The incorrect stroke division after the improper handling of the connections.

Figure 3. (a) The correct stroke division for the Chinese character “shi”. (b) and (c) The incorrect stroke division after the improper handling of the intersection.

The Chinese character strokes are shown in Fig. 4. Analysing the strokes of Chinese characters, we can see that each stroke can be divided into one or more approximately straight curves. These curves can be summarised as dot, horizontal, vertical, left-falling and right-falling as shown in Fig. 5. These curves are named stroke components. Thus, each Chinese character stroke can be formed by combining one or more Chinese character stroke components. For example, as shown in Fig. 6, a Chinese character stroke “dot” can be seen as a combination of a Chinese character stroke component “dot”. A Chinese character stroke “vertical-turning” can be decomposed as a combination of a Chinese character stroke component “vertical” and a Chinese character stroke component “horizontal”. For another example, a Chinese character “horizontal-break-hook” can be seen as a combination of a “horizontal”, a “vertical” and a “dot”.

Figure 4. Chinese character strokes.

Figure 5. Chinese character stroke components.

Figure 6. Chinese character stroke can be formed by combining one or more Chinese character stroke components.

As the Chinese character stroke components all extend in the same direction with no turning points, the Chinese character stroke components can be extracted when the direction of extension is known. At the same time, the determination of the direction of extension also deals with the issue of connections and intersections. A turning place connects two stroke components, and thus the turning identification is important in the basic component extraction. The intersection occurs when the strokes have been extended to different directions. When the different stroke components’ writing directions are determined, the stroke component extensions can cross the intersection in the specified direction to prevent misidentification at the intersection. Therefore, the stroke extraction algorithm has two parts: stroke components extraction and stroke components connection.

2.1.1. Stroke components extraction

Firstly, the handwritten Chinese character images were binarised and refined. The algorithm for stroke components extraction is built on the basis of a refined image of a white Chinese character on a black background. The image is then traversed to find the white pixel, which is recorded as the starting point $P_{sta}$ and treated as $P_{0}$ . As shown in Fig. 7, for a set of eight neighbourhoods with $P_{0}$ as the centroid, we set $P_{0}$ as the origin and the length of the pixel as the unit length. The pixel points in this eight-neighbourhood are therefore named by the coordinates in which they are located. Starting from the pixel point above $P_{0}$ , the pixels in a clockwise circle are each named $P_{n}=(x_n,y_n)(n=1,2,3,\ldots,8)$ . The process of extracting the strokes component is shown in Fig. 8. Define a direction vector $\textbf{D} =(\alpha,\beta )$ with an initial value of $(0,0)$ . $\textbf{D}$ describes the direction from the start pixel to the end pixel of the searched stroke component. The values of $\textbf{D}$ are normalised to obtain the coordinate points $P_{D}$ in the eight-neighborhood coordinate system.

(1) \begin{equation} {P_{D} =(x_{D},y_{D})} \end{equation}
(2) \begin{equation} {x_{D} = \frac{\alpha }{\sqrt{\alpha ^{2}+\beta ^{2} } } } \end{equation}
(3) \begin{equation} {y_{D} = \frac{\beta }{\sqrt{\alpha ^{2}+\beta ^{2} } } } \end{equation}

Figure 7. A set of eight neighbourhoods with $P_{0}$ as the centroid.

Figure 8. The process of extracting the strokes component.

The distance between $P_{D}$ and the coordinates of each pixel in the eight neighbourhoods is

(4) \begin{equation} {d_{n} =|P_{n}-P_{D}|} \end{equation}

Arrange $d_1, d_2\ldots, d_n$ according to their values (from small to large values). To ensure that the stroke component extends in the direction indicated by $\textbf{D}$ , perform a sequential search for the $P_{n}$ corresponding to the first 4 $d_{n}$ . If all the $P_{n}$ picked out are black, the search stops. If $P_{n}$ is a white pixel, extend the base stroke to $P_{n}$ and update direction vector

(5) \begin{equation} {\textbf{D}_{\textbf{new}} =(\alpha _{new},\beta _{new})} \end{equation}
(6) \begin{equation} {\alpha _{new} = \alpha +x_n} \end{equation}
(7) \begin{equation} {\beta _{new} = \beta +y_n} \end{equation}

Subsequently, $P_{n}$ is used as the next $P_{0}$ and $\textbf{D}_{\textbf{new}}$ as the next $\textbf{D}$ in order to continue extending the stroke component until the search stops.

However, the starting point $P_{sta}$ is not always the endpoint of the stroke component. So after the extension along direction $\textbf{D}$ is complete, the direction vector $\textbf{D}$ needs to be reversed

(8) \begin{equation} {\textbf{D}_{\textbf{R}} = -\textbf{D}} \end{equation}

It is then extended to the opposite direction from the starting point $P_{sta}$ , and finally, the two curves are joined to form a complete stroke component.

In order to prevent the same stroke component from being searched repeatedly, each pixel point is blacked out after it has been recorded as a point in the stroke component. However, this creates the interruption problem. As shown in Fig. 9, an intersection point is shared by multiple stroke components and when the intersection is blacked out, another stroke component passing through this intersection will be interrupted at the intersection. Therefore, we use two identical refined images $I_{1}$ and $I_{2}$ of white characters on a black background. Image $I_{1}$ is blacked out at the recorded pixels and is used to find the starting point $P_{sta}$ . This prevents repeated searching of the same basic stroke. Image $I_{2}$ is not blacked out at the recorded pixels and is used to extend the stroke component so that the stroke components are not interrupted at the intersection point.

2.1.2. Stroke components connection

When the end point of stroke component $SC_{1}$ and the start point of stroke component $SC_{2}$ are connected, $SC_{1}$ and $SC_{2}$ will make a stroke. Therefore, the starting and ending points of the stroke components need to be identified, i.e. the points between the connected stroke components. According to the conventional writing rules, the strokes are written from left to right or from top to bottom. The specific sorting algorithm is shown in Algorithm 1.

Algorithm 1: Sorting of points

Figure 9. (a) The darkening at the intersection divides the vertical into two parts. (b) Using two identical images $I_{1}$ and $I_{2}$ can split the stroke components correctly.

However, there is a special Chinese stroke called “hook” as shown in Fig. 10, which is written from bottom to top and from right to left. Similar to the stroke component “dot”, it is much shorter than the other basic strokes. It is also connected to the other stroke components. Thus, a stroke component is considered to be a ”hook” if it satisfies the following conditions:

Figure 10. The “hook” in Chinese character “horizontal-break-hook”.

  1. 1. the end point is connected to the end points of the other stroke components;

  2. 2. the start point is not connected to the end points of any of the other stroke components;

  3. 3. the length is less than one-half the length of the average stroke component.

When combining strokes, the points of the “hook” need to be reverse ordered and then connected to other stroke components.

2.2. Force-position hybrid control

For a robotic writing system, the position control mode provides a limited performance, e.g. nonsmooth and nonsimilar to human’s writing characters. To write smoothly on the writing board, the contact force between the pen tip and the writing board needs to be controlled. Therefore, we use the force-position hybrid control. A modified DMP is used to control the position and posture of the robot’s end-effector. At the same time, the contact force between the pen tip and the writing board is controlled using admittance control. As a result, the robot can perform writing tasks.

2.2.1. Modified DMPs

Based on a spring damping model, the DMPs use a non-linear adjusting term to achieve the desired point attractor behaviour. It has the properties of a second-order dynamical system such as convergence at the target point, robustness to disturbances and generalisation in time and space. At the same time, the non-linear terms can be made to produce a continuous smooth trajectory. DMPs are classified as discrete DMPs and periodic DMPs. We use the discrete DMPs to model writing skills. The discrete DMPs model can be expressed as:

(9) \begin{equation} {\tau \dot{v}=\alpha \left (y_{g}-y\right )-\beta v+f(x)} \end{equation}
(10) \begin{equation} {\tau \dot{y}=v} \end{equation}

where $\tau$ is the temporal scaling constant, and changing it allows the trajectory to generalise in the time dimension. $\alpha$ and $\beta$ represent the damping factor and the spring constant, respectively, and $\alpha$ can generally be set to $\alpha =\beta ^{2}/4$ . $y_{g}$ is the target position. $y$ is the position of the demonstration. $v$ is the velocity of the demonstration. The forcing term $f(x)$ can be expressed as:

(11) \begin{equation} {f\left (x,y_{g}\right )=\frac{\sum _{i=1}^{N}{\psi _i\cdot \omega _i}}{\sum _{i=1}^{N}\psi _i}\cdot x\cdot (y_{g}-y_0)} \end{equation}
(12) \begin{equation} {\psi _i=\exp ({-}h_i(x-c_i)^2)} \end{equation}
(13) \begin{equation} {\tau \dot{x}=-\alpha _x\cdot x} \end{equation}

where $\omega _i$ is the weight of the Gaussian kernel function, $y_0$ is the starting point of the trajectory, $c_i$ and $h_i$ are the centre and width of the Gaussian kernel function, respectively, $N$ denotes the number of Gaussian kernel functions, and $\alpha _x$ is a positive gain coefficient.

However, as shown in Fig. 11, when the start and target points are rotated, the original DMP cannot rotate the generalised trajectory in its original shape and the trajectory is distorted. In order to apply the DMPs to this case where both the start and target points are rotated, we have improved the original DMPs inspired by [Reference Koutras and Doulgeri31]. The generalised space is obtained by rotating the original space. Once the rotation matrix between the two spaces is known, the corresponding rotation can be applied to the non-linear term. Thus, equation (9) becomes

(14) \begin{equation} {\tau \ddot{y}=\alpha _y(\beta _y\left (y_{g,1}-y\right )-\dot{y})+s\cdot \textbf{R} \cdot f } \end{equation}
(15) \begin{equation} {y_{g,1}=\textbf{R}\cdot y_{g,0}+\textbf{t}} \end{equation}
(16) \begin{equation} {y_{0,1}=\textbf{R}\cdot y_{0,0}+\textbf{t}} \end{equation}
(17) \begin{equation} {s=\frac{\left \| y_{g,1}-y_{0,1} \right \| }{\left \| y_{g,0}-y_{0,0} \right \| } } \end{equation}

where $y_{g,1}$ and $y_{g,0}$ are the target points in the generalised space and the original space, respectively. $y_{0,1}$ and $y_{0,0}$ are the starting points in the generalised space and the original space, respectively. $s$ is the scaling constant. $\textbf{R}$ is the rotation matrix and $\textbf{t}$ is the translation matrix. Throughout this article, $\textbf{R}$ and $\textbf{t}$ are calculated from visual information.

Figure 11. Comparison of original DMPs and modified DMPs. For linear and curvilinear demonstrations, generalisations are carried out using the original DMPs and modified DMPs, respectively. The trajectories generalised from the original DMPs have distortions. The trajectories generalised by the improved DMPs conform to the expected trajectories.

2.2.2. Admittance control

The admittance model is defined as follows:

(18) \begin{equation} {M_{A}\ddot{x}_{A}+D_{A}\dot{x}_A+K_{A}(x_A-x_0)=f_h} \end{equation}

where $M_{A}$ is the quality matrix, $D_{A}$ is the damping matrix and $K_{A}$ is the spring factor matrix. $f_h$ is the traction of the demonstrator. $x_A$ is the pose of the robot end-effector, $\dot{x}_{A}$ is the speed and $\ddot{x}_{A}$ is the acceleration. $x_0$ is the original point of the pose. The robot end-effector needs to move on the writing board during the writing process, so $x_A=x_0$ . The admittance model can be simplified as:

(19) \begin{equation} {M_{A}\ddot{x}_{A}+D_{A}\dot{x}_A=f_h}. \end{equation}

Therefore, the acceleration can be calculated using the traction force $f_h$ and then the velocity and position can be calculated by integration to obtain the trajectory. In this article, $f_h$ is the contact force between the pen tip and the writing board, with an expected value of $2N$ .

2.3. Designed system

Based on the above approach, we have designed a multi-sensor-based robotic writing system. As shown in Fig. 12, at the demonstration stage, the strokes are extracted from the demonstrations. During the trajectory generalisation phase, the camera detects changes in the pose of the writing board and calculates the rotation matrix $\textbf{R}$ and the translation matrix $\textbf{t}$ . The DMPs model calculates the generalised trajectory points based on the $\textbf{R}$ and $\textbf{t}$ and feeds the trajectory points into the admittance control model. The admittance control algorithm calculates the execution position based on the given trajectory points and real-time force information and outputs the execution position to the robot. At the same time, the execution position will be different from the desired position and in the next step this differential will be used to correct the trajectory points in advance to improve accuracy.

Figure 12. Robotic writing system.

3. Experiment

3.1. Stroke extraction

In order to verify the accuracy of Chinese character stroke extraction, we randomly selected 100 handwriting Chinese characters with different structures and complexities for experiments. As shown in Fig. 13, the experiments show that the algorithm works well. Statistically, 93% of the strokes of handwritten Chinese characters can be extracted accurately, with an average speed of 0.4 s per image. The characters boxed in Fig. 13 are characters with inaccurate stroke recognition. After studying them, we have divided them into three categories. Category 1 is for those boxed in red, whose strokes are broken. This means that a single stroke has been split into multiple strokes or is incompletely recognised. Those boxed in green are in category 2: misconnected strokes. This means that a stroke has been incorrectly connected to another stroke. Boxed in blue is category 3, there are redundant strokes. Excess strokes take the form of burrs consisting of just one or two pixels. Analysis of these three categories shows that almost all recognition errors happen in the refinement process. When inaccurate refinement leads to a change in the direction of the strokes, deformations are observed at the intersection and junction points of the Chinese character’s skeleton, errors in category 1 are likely to occur. During the writing process, variations in pen pressure or ink leakage can result in changes in the thickness of strokes, leading to the occurrence of jaggies in the refinement process, errors in category 3 are likely to occur. Errors in category 2 are caused by inaccurate refinement and the poor writing where multiple interconnected or intersecting stroke components share a consistent direction.

Figure 13. Stroke extraction results. Different colours indicate different strokes.

3.2. Modified DMPs

There are two methods to generalise a trajectory based on disturbances in real time. Method 1 firstly generalises the new trajectory using the original DMPs and then adjusts the new trajectory using rotation and translation matrices, $\textbf{R}$ and $\textbf{t}$ . Method 2 uses the modified DMPs to generalise the trajectory directly based on $\textbf{R}$ and $\textbf{t}$ . We set up a writing track and use visual information to obtain the pose of the writing board to calculate $\textbf{R}$ and $\textbf{t}$ . These two methods are then used to make the trajectory follow the desired pose on the writing board. The effects of the two methods are shown in Fig. 14. It can be seen that the trajectory output of method 1 is more oscillating, not smooth enough and has large errors. Method 2 outputs a smoother trajectory with less oscillation, which is more effective. In comparison to Method 1, the trajectories produced by Method 2 exhibit a difference of up to 7.1977 mm in the z-direction, which poses a potential safety hazard in practical applications, as it could result in the inadvertent puncturing of the writing surface. We analysed the reasons for the large differences in the results between the two methods. When using method 1, outputs of the original DMP have errors in practice, such as distortion of the trajectory and robot movement errors. Also, $\textbf{R}$ and $\textbf{t}$ are calculated from visual information and have errors. Because $\textbf{R}$ and $\textbf{t}$ do not fall within the framework of DMPs, the multiplication with the trajectory points is not constrained by DMPs, leaving the trajectory with a large offset. Whereas method 2 incorporates $\textbf{R}$ and $\textbf{t}$ into the framework of DMPs, the generalisation of DMPs will change with $\textbf{R}$ and $\textbf{t}$ . At the same time, the trajectory does not produce large offsets due to the constraints of the DMPs framework.

Figure 14. The effect of the two methods.

3.3. Admittance control

We have measured the contact force between the pen tip and the writing board when a person writes. We find that the best writing performance is achieved when the force sensor has $2N$ contact force on the z-axis. If the contact force is not $2N$ when the robot arm moves to the specified position, the admittance control algorithm can adjust the end position of the robot arm in time to guarantee the contact force always be $2N$ . As shown in Fig. 15, when the initial contact force is 0, i.e. when the end-effector is not touching the writing board, the admittance control moves the end-effector closer to the writing board until the contact force converges to $2N$ . When the initial contact force is $6N$ , the admittance control policy steers the end-effector moving away from the writing board until the contact force is stabilised at $2N$ . The admittance control also adjusts the contact force in time when there is any interference during the adjustment process.

Figure 15. The effect of the admittance control.

3.4. Application

In this experiment, we verify the effectiveness of the designed interference-resistant robotic Chinese character writing system. A human randomly writes a Chinese character. Then the robot imitates the human handwriting on the writing board. While the robot writes, the human randomly moves, turns and flips the board, and the robot can still follow the posture of the board to adjust its trajectory. As shown in Fig. 16, the experimental platform is composed of an Elite Collaborative Robot EC66, an Intel RealSense Depth Camera D435 and a LCD Writing Board. The Elite Collaborative Robot EC66 is a lightweight and flexible collaborative robot. Its weight is 17.5 kg and the largest payload is 6 kg. It runs efficiently and can reach a maximum speed of 2.8 m/s, with a soft and consistent trajectory. As a collaborative robot, it enables safe human–robot interaction with a high degree of reliability and safety. It can therefore complete different tasks while ensuring the safety of humans. With a range of 10 m and a wide field of view, the Intel RealSense Depth Camera D435 has a wide range of applications in robotics development. By affixing four calibration papers to the four corners of the writing board, it is able to estimate the posture of the writing board very well in this experiment, i.e. to recognise human interference. During the robot’s writing process, the human randomly and continuously moves, rotates and flips the writing board. The robot can write complete Chinese characters on the writing board.

Figure 16. Experimental platform.

The results of the experiment are shown in Fig. 17. The robot is able to adjust its end trajectory in time to write complete and accurate Chinese characters on the writing pad under human interferences.

Figure 17. The results of the experiment.

4. Conclusion

In this article, we designed a robotic writing system that allows a robot to imitate human handwriting even with unexpected interruptions. We first designed a new stroke extraction algorithm that extracts stroke components based on the direction of extension, and then connected the stroke components according to the writing direction of the stroke. To cope with random interference, the original DMPs model was improved such that the generalisation function could be adapted to the situations when the writing board was rotated and flipped. The writing system was then designed by combining visual information, admittance control and an improved DMP to enable the robot to accurately imitate human handwriting in the face of interference.

In this article, we have presented an interference-resistant robotic writing system that can also be used in applications such as grinding and cleaning of complex recesses in workpieces. In the future, we will improve the current visual measurement module to make this system more accurate and better applicable.

Author contributions

Xian Li, Chenguang Yang, Sheng Xu and Yongsheng Ou proposed the research project; Xian Li conducted the review and wrote the first draft; Xian Li, Chenguang Yang, Sheng Xu and Yongsheng Ou revised the manuscript; and Chenguang Yang, Sheng Xu and Yongsheng Ou provided advice and supervision.

Financial support

This work was supported in part by National Nature Science Foundation of China (NSFC) under Grant U20A20200 and Major Research Grant No. 92148204, in part by Guangdong Basic and Applied Basic Research Foundation under Grants 2019B1515120076 and 2020B1515120054, in part by Industrial Key Technologies R & D Program of Foshan under Grant 2020001006308 and Grant 2020001006496.

Competing interests

The authors declare no competing interests exist.

Ethical approval

Not applicable.

References

Li, Z., Li, S. and Luo, X., “An overview of calibration technology of industrial robots,” IEEE/CAA J. Autom. Sin. 8(1), 2336 (2021).Google Scholar
Yin, H., Alves-Oliveira, P., Melo, F. S., Billard, A. and Paiva, A., “Synthesizing Robotic Handwriting Motion by Learning from Human Demonstrations.” Proceedings of the 25th International Joint Conference on Artificial Intelligence, No. CONF (2016).Google Scholar
Zeng, H., Huang, Y., Chao, F. and Zhou, C., “Survey of robotic calligraphy research,” CAAI Trans. Intell. Syst. 11(1), 1526 (2016).Google Scholar
Chao, F., Huang, Y., Zhang, X., Shang, C., Yang, L., Zhou, C., Hu, H. and Lin, C.-M., “A robot calligraphy system: From simple to complex writing by human gestures,” Eng. Appl. Artif. Intel. 59, 114 (2017).Google Scholar
Wang, S., Chen, J., Deng, X., Hutchinson, S. and Dellaert, F., “Robot Calligraphy Using Pseudospectral Optimal Control in Conjunction with a Novel Dynamic Brush Model.” 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE (2020) pp. 66966703.Google Scholar
Wang, G., Liao, Z. and Chen, X., “Robot manipulator lettering technology and motion,” J Chongq. Univ. 26(12), 69 (2003).Google Scholar
Gong, X. and Wang, J., “Drawing Chinese character based on continuous path control of IRB 140 industrial robot,” Modern Manuf. Eng. (2), 135–137(2010).Google Scholar
Chuanjian, Z., Chunmei, L. and Jun, H., “Functional design and realization of writing for motoman-up6 robots,” Mach. Buildi. Autom. 40(2), 119121 (2011).Google Scholar
Yang, Y., Chen, W., Zhou, L., Zheng, B., Xiao, W., Huang, Y. and Sun, Z., “A Hybrid Control Framework Teaching Robot to Write Chinese Characters: From Image to Handwriting,” 2021 IEEE 17th International Conference on Automation Science and Engineering (CASE), IEEE (2021) pp. 11611166.Google Scholar
Yao, F. and Shao, G., “Modeling of Ancient-style Chinese Character and Its Application to CCC Robot,” 2006 IEEE International Conference on Networking, Sensing and Control., IEEE (2006) pp. 7277.Google Scholar
Yao, F., Shao, G. and Yi, J., “Extracting the trajectory of writing brush in Chinese character calligraphy,” Eng. Appl. Artif. Intel. 17(6), 631644 (2004).Google Scholar
Lam, J. H. and Yam, Y., “Application of Brush Footprint Geometric Model for Realization of Robotic Chinese Calligraphy, ” 2011 2nd International Conference on Cognitive Infocommunications (CogInfoCom), IEEE (2011) pp. 15.Google Scholar
Lam, J. H. and Yam, Y., “Stroke Trajectory Generation Experiment for a Robotic Chinese Calligrapher Using a Geometric Brush Footprint Model,” 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, IEEE (2009) pp. 23152320.Google Scholar
Li, J., Sun, W., Zhou, M. and Dai, X., “Teaching a Calligraphy Robot Via a Touch Screen,” 2014 IEEE International Conference on Automation Science and Engineering (CASE), IEEE (2014) pp. 221226.Google Scholar
Lin, F. and Tang, X., “Dynamic Stroke Information Analysis for Video-based Handwritten Chinese Character Recognition,” Proceedings Ninth IEEE International Conference on Computer Vision, IEEE (2003) pp. 695700.Google Scholar
Sun, Y., Qian, H. and Xu, Y., “Robot Learns Chinese Calligraphy from Demonstrations,” 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, IEEE (2014) pp. 44084413.Google Scholar
Chao, F., Huang, Y., Lin, C.-M., Yang, L., Hu, H. and Zhou, C., “Use of automatic Chinese character decomposition and human gestures for Chinese calligraphy robots,” IEEE Trans. Hum. Mach. Syst. 49(1), 4758 (2018).Google Scholar
Cao, F., Wu, Z., Ao, X. and Zhou, M., “Vectorization of Qi Gong calligraphy,” J. Chin. Inf. Process. 24(6), 97102 (2010).Google Scholar
Wang, X., Liang, X., Sun, L. and Liu, M., “Triangular Mesh Based Stroke Segmentation for Chinese Calligraphy,” 2013 12th International Conference on Document Analysis and Recognition, IEEE (2013) pp. 11551159,Google Scholar
Gan, L., Fang, W., Chao, F., Zhou, C., Yang, L., Lin, C.-M. and Shang, C., “Towards a Robotic Chinese Calligraphy Writing Framework,” 2018 IEEE International Conference on Robotics and Biomimetics (ROBIO), IEEE (2018) pp. 493498,Google Scholar
Wang, T.-Q., Jiang, X. and Liu, C.-L., “Query pixel guided stroke extraction with model-based matching for offline handwritten Chinese characters,” Pattern Recogn. 123, 108416 (2022).Google Scholar
Ju, Z., Ji, X., Li, J. and Liu, H., “An integrative framework of human hand gesture segmentation for human–robot interaction,” IEEE Syst. J. 11(3), 13261336 (2015).Google Scholar
Chang, C.-C., “A library for support vector machines. 2001. http://www.csie.ntu.edu.tw/~cjlin/libsvm.Google Scholar
Ijspeert, A., Nakanishi, J. and Schaal, S., “Learning Attractor Landscapes for Learning Motor Primitives,” In: Advances in Neural Information Processing Systems. vol. 15 (2002).Google Scholar
Koutras, L. and Doulgeri, Z., “Dynamic Movement Primitives for Moving Goals with Temporal Scaling Adaptation,” 2020 IEEE International Conference on Robotics and Automation (ICRA), IEEE (2020) pp. 144150,Google Scholar
Yang, C., Chen, C., He, W., Cui, R. and Li, Z., “Robot learning system based on adaptive neural control and dynamic movement primitives,” IEEE Trans. Neur. Netw. Learn. Syst. 30(3), 777787 (2018).Google Scholar
Yang, C., Zeng, C., Cong, Y., Wang, N. and Wang, M., “A learning framework of adaptive manipulative skills from human to robot,” IEEE Trans. Ind. Inform. 15(2), 11531161 (2018).Google Scholar
Anand, A. S., Østvik, A., Grøtli, E. I., Vagia, M. and Gravdahl, J. T., “Real-time temporal adaptation of dynamic movement primitives for moving targets,” 2021 20th International Conference on Advanced Robotics (ICAR), IEEE (2021) pp. 261268,Google Scholar
Liao, Z., Jiang, G., Zhao, F., Wu, Y., Yue, Y. and Mei, X., “Dynamic skill learning from human demonstration based on the human arm stiffness estimation model and Riemannian DMP,” IEEE/ASME Trans. Mechatron. 28(2), 11491160 (2022).Google Scholar
Zeng, C., Chen, X., Wang, N. and Yang, C., “Learning compliant robotic movements based on biomimetic motor adaptation,” Robot. Auton. Syst. 135, 103668 (2021).Google Scholar
Koutras, L. and Doulgeri, Z., “A Novel DMP Formulation for Global and Frame Independent Spatial Scaling in the Task Space,” 2020 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), (2020) pp. 727732,Google Scholar
Abu-Dakka, F. J. and Kyrki, V., “Geometry-aware Dynamic Movement Primitives,” 2020 IEEE International Conference on Robotics and Automation (ICRA), IEEE (2020) pp. 44214426,Google Scholar
Wang, T., Yan, L., Wang, G., Gao, X., Du, N. and Chen, I.-M., “Learning from Demonstration Using Improved Dynamic Movement Primitives,” 2021 IEEE 16th Conference on Industrial Electronics and Applications (ICIEA), (2021) pp. 21302135,Google Scholar
Yang, C., Zeng, C., Fang, C., He, W. and Li, Z., “A DMPs-based framework for robot learning and generalization of humanlike variable impedance skills,” IEEE/ASME Trans. Mechatron. 23(3), 11931203 (2018).Google Scholar
Figure 0

Figure 1. The connection and intersection of Chinese character “zhong”.

Figure 1

Figure 2. (a) The correct stroke division for the Chinese character “zhong”. (b) The incorrect stroke division after the improper handling of the connections.

Figure 2

Figure 3. (a) The correct stroke division for the Chinese character “shi”. (b) and (c) The incorrect stroke division after the improper handling of the intersection.

Figure 3

Figure 4. Chinese character strokes.

Figure 4

Figure 5. Chinese character stroke components.

Figure 5

Figure 6. Chinese character stroke can be formed by combining one or more Chinese character stroke components.

Figure 6

Figure 7. A set of eight neighbourhoods with $P_{0}$ as the centroid.

Figure 7

Figure 8. The process of extracting the strokes component.

Figure 8

Algorithm 1: Sorting of points

Figure 9

Figure 9. (a) The darkening at the intersection divides the vertical into two parts. (b) Using two identical images $I_{1}$ and $I_{2}$ can split the stroke components correctly.

Figure 10

Figure 10. The “hook” in Chinese character “horizontal-break-hook”.

Figure 11

Figure 11. Comparison of original DMPs and modified DMPs. For linear and curvilinear demonstrations, generalisations are carried out using the original DMPs and modified DMPs, respectively. The trajectories generalised from the original DMPs have distortions. The trajectories generalised by the improved DMPs conform to the expected trajectories.

Figure 12

Figure 12. Robotic writing system.

Figure 13

Figure 13. Stroke extraction results. Different colours indicate different strokes.

Figure 14

Figure 14. The effect of the two methods.

Figure 15

Figure 15. The effect of the admittance control.

Figure 16

Figure 16. Experimental platform.

Figure 17

Figure 17. The results of the experiment.