This application claims priority to Japanese Patent Application No. 2010-064437, filed Mar. 19, 2010, the content of which is hereby incorporated herein by reference.
The present disclosure relates to a sewing machine that includes a projection portion and an image capture portion and to a non-transitory computer-readable medium that stores a sewing machine control program.
A sewing machine is known that is provided with a function that detects the thickness of a work cloth that is an object of sewing. In this sort of sewing machine, the thickness of the work cloth is detected by an angle sensor that is provided on a member that presses the work cloth, for example. Then, a point mark at a position that corresponds to the cloth thickness is illuminated by a marking light. A cloth stage detector detects the thickness of the work cloth based on the position of a beam of light that is projected onto the work cloth by a light-emitting portion and reflected by the work cloth.
In a case where the thickness of the work cloth is detected using the angle sensor, the thickness may not be detected in a state where the work cloth is not being pressed. For example, in a case where the work cloth tends to contract and in a case where the work cloth is a quilted material that is filled with cotton batting, the thickness may not be properly detected by the known sewing machine in a state where the work cloth is not being pressed. In a case where the thickness is detected based on the position of a beam of light that is reflected by the work cloth, an area within which the thickness can be detected may be extremely narrow. Therefore, in order to detect the thickness at the desired position, a user may need to perform a complicated operation of positioning the portion of the work cloth where the thickness is to be detected in the small area onto which the light will be shone.
Various exemplary embodiments of the broad principles derived herein provide a sewing machine and a non-transitory computer-readable medium storing a sewing machine control program that enables detecting, by a simple operation, the thickness of a sewing object that is not being pressed.
Exemplary embodiments provide the sewing machine that includes a creating portion that creates a projection image being an image that includes a characteristic point and that is to be projected onto a sewing object, and a projecting portion that projects onto the sewing object the projection image created by the creating portion. The sewing machine also includes an image capture portion that is mounted in a position being different from a position of the projecting portion and that creates a captured image by image capture of the characteristic point projected by the projecting portion, and a computing portion that computes a thickness of the sewing object based on the projection image created by the creating portion and the captured image created by the image capture portion.
Exemplary embodiments also provide a non-transitory computer-readable medium storing a control program executable on a sewing machine. The program includes instructions that cause a computer of the sewing machine to perform the steps of creating a projection image being an image that includes a characteristic point and that is to be projected onto a sewing object, acquiring a captured image created by image capture of the characteristic point projected on the sewing object, and computing a thickness of the sewing object based on the projection image and the captured image.
Exemplary embodiments will be described below in detail with reference to the accompanying drawings in which:
Hereinafter, a sewing machine 1 according to first and second embodiments of the present disclosure will be explained in order with reference to the drawings. The drawings are used for explaining technical features that can be used in the present disclosure, and the device configuration, the flowcharts of various types of processing, and the like that are described are simply explanatory examples that does not limit the present disclosure to only the configuration, the flowcharts, and the like.
A physical configuration and an electrical configuration of the sewing machine 1 according to the first and second embodiments will be explained with reference to
A feed dog front-and-rear moving mechanism (not shown in the drawings), a feed dog up-and-down moving mechanism (not shown in the drawings), a pulse motor 78 (refer to
A sewing machine motor 79 (refer to
A top cover 21 is provided in the longitudinal direction of the arm 4. The top cover 21 is axially supported at the rear upper edge of the arm 4 such that the top cover 21 may be opened and closed around the left-right directional shaft. A thread spool housing 23 is provided close to the middle of the top of the arm 4 under the top cover 21. The thread spool housing 23 is a recessed portion for accommodating a thread spool 20 that supplies a thread to the sewing machine 1. A spool pin 22, which projects toward the head 5, is provided on an inner face of the thread spool housing 23 on the pillar 3 side. The thread spool 20 may be attached to the spool pin 22 when the spool pin 22 is inserted through the insertion hole (not shown in the drawings) that is formed in the thread spool 20. Although not shown in the drawings, the thread of the thread spool 20 may be supplied as an upper thread to the needle 7 (refer to
A pulley (not shown in the drawings) is provided on a right side surface of the sewing machine 1. The pulley is used to manually rotate the drive shaft (not shown in the drawings). The pulley causes the needle bar 6 to be moved up and down. A front cover 19 is provided on a front surface of the head 5 and the arm 4. A group of switches 40 is provided on the front cover 19. The group of switches 40 includes a sewing start/stop switch 41 and a speed controller 43, for example. The sewing start/stop switch 41 is used to issue a command to start or stop sewing. If the sewing start/stop switch 41 is pressed when the sewing machine 1 is stopped, the operation of the sewing machine 1 is started. If the sewing start/stop switch 41 is pressed when the sewing machine 1 is operating, the operation of the sewing machine 1 is stopped. The speed controller 43 is used for controlling the revolution speed of the drive shaft. An image sensor 50 (refer to
The image sensor 50 will be explained with reference to
As shown in
As shown in
The embroidery unit 30 will be explained with reference to
Based on an amount of movement that is expressed by coordinates in an embroidery coordinate system, drive commands for the Y axis motor 82 and the X axis motor 81 are output by a CPU 61 (refer to
A main electrical configuration of the sewing machine 1 will be explained with reference to
The CPU 61 conducts main control over the sewing machine 1, and performs various types of computation and processing in accordance with programs stored in the ROM 62 and the like. The ROM 62 includes a plurality of storage areas including a program storage area. Programs that are executed by the CPU 61 are stored in the program storage area. The RAM 63 is a storage element that can be read from and written to as desired. The RAM 63 stores, for example, data that is required when the CPU 61 executes a program and computation results that is obtained when the CPU 61 performs computation. The EEPROM 64 is a storage element that can be read from and written to. The EEPROM 64 stores various parameters that are used when various types of programs stored in the program storage area are executed. Storage areas of the EEPROM 64 will be described in detail below. A card slot 17 is connected to the external access RAM 65. The card slot 17 can be connected to a memory card 18. The sewing machine 1 can read and write information from and to the memory card 18 by connecting the card slot 17 and the memory card 18.
The sewing start/stop switch 41, the speed controller 43, the touch panel 16, the image sensor 50, drive circuits 70 to 76, and the light source 56 are electrically connected to the input/output interface 66. The drive circuit 70 drives the pulse motor 77. The pulse motor 77 is a drive source of the needle bar swinging mechanism (not shown in the drawings). The drive circuit 71 drives the pulse motor 78 for adjusting a feed amount. The drive circuit 72 drives the sewing machine motor 79. The sewing machine motor 79 is a drive source of the drive shaft (not shown in the drawings). The drive circuit 73 drives the X axis motor 81. The drive circuit 74 drives the Y axis motor 82. The drive circuit 75 drives the LCD 10. The drive circuit 76 drives the liquid crystal panel 57 of the projector 53. Another element (not shown in the drawings) may be connected to the input/output interface 66 as appropriate.
The storage areas of the EEPROM 64 will be explained. The EEPROM 64 includes a settings storage area, an internal variables storage area, and an external variables storage area, which are not shown in the drawings. Setting values that are used when the sewing machine 1 performs various types of processing are stored in the settings storage area. The setting values that are stored may include, for example, correspondences between the types of embroidery frames and the sewing areas.
Internal variables for the image sensor 50 and the projector 53 are stored in the internal variables storage area. The internal variables for the image sensor 50 are parameters to correct a shift in focal length, a shift in principal point coordinates, and distortion of a captured image due to properties of the image sensor 50. An X-axial focal length, a Y-axial focal length, an X-axial principal point coordinate, a Y-axial principal point coordinate, a first coefficient of distortion, and a second coefficient of distortion are stored as internal variables in the internal variables storage area. The X-axial focal length represents an X-axis directional shift of the focal length of the image sensor 50. The Y-axial focal length represents a Y-axis directional shift of the focal length of the image sensor 50. The X-axial principal point coordinate represents an X-axis directional shift of the principal point of the image sensor 50. The Y-axial principal point coordinate represents a Y-axis directional shift of the principal point of the image sensor 50. The first coefficient of distortion and the second coefficient of distortion represent distortion due to the inclination of a lens of the image sensor 50. The internal variables may be used, for example, in processing that converts the image that the sewing machine 1 has captured into a normalized image and in processing in which the sewing machine 1 computes information on a position on the sewing object 34. The normalized image is an image that would presumably be captured by a normalized camera. The normalized camera is a camera for which the distance from the optical center to a screen surface is a unit distance.
The optical models for the image sensor 50 and the projector 53 are the same. Therefore, the projector 53 can be considered to have the same external variables and internal variables as the image sensor 50. The internal variables for the projector 53 are stored in the internal variables storage area in the same manner as the internal variables for the image sensor 50.
External variables for the image sensor 50 and the projector 53 are stored in the external variables storage area. The external variables for the image sensor 50 are parameters that indicate the installed state (the position and the orientation) of the image sensor 50 with respect to a world coordinate system 100. Accordingly, the external variables indicate a shift of a camera coordinate system 200 with respect to the world coordinate system 100. The camera coordinate system is a three-dimensional coordinate system for the image sensor 50. The camera coordinate system 200 is schematically shown in
An X-axial rotation vector, a Y-axial rotation vector, a Z-axial rotation vector, an X-axial translation vector, a Y-axial translation vector, and a Z-axial translation vector are stored as the external variables for the image sensor 50 in the external variables storage area. The X-axial rotation vector represents a rotation of the camera coordinate system 200 around the X-axis with respect to the world coordinate system 100. The Y-axial rotation vector represents a rotation of the camera coordinate system 200 around the Y-axis with respect to the world coordinate system 100. The Z-axial rotation vector represents a rotation of the camera coordinate system 200 around the Z-axis with respect to the world coordinate system 100. The X-axial rotation vector, the Y-axial rotation vector, and the Z-axial rotation vector are used for determining a conversion matrix that is used for converting three-dimensional coordinates in the world coordinate system 100 into three-dimensional coordinates in the camera coordinate system 200, and vice versa. The X-axial translation vector represents an X-axial shift of the camera coordinate system 200 with respect to the world coordinate system 100. The Y-axial translation vector represents a Y-axial shift of the camera coordinate system 200 with respect to the world coordinate system 100. The Z-axial translation vector represents a Z-axial shift of the camera coordinate system 200 with respect to the world coordinate system 100. The X-axial translation vector, the Y-axial translation vector, and the Z-axial translation vector are used for determining a translation vector that is used for converting three-dimensional coordinates in the world coordinate system 100 into three-dimensional coordinates in the camera coordinate system 200, and vice versa. A 3-by-3 rotation matrix that is determined based on the X-axial rotation vector, the Y-axial rotation vector, and the Z-axial rotation vector and that is used for converting the three-dimensional coordinates of the world coordinate system 100 into the three-dimensional coordinates of the camera coordinate system 200 is defined as a rotation matrix Rc for the image sensor 50. A 3-by-1 translation vector that is determined based on the X-axial translation vector, the Y-axial translation vector, and the Z-axial translation vector and that is used for converting the three-dimensional coordinates of the world coordinate system 100 into the three-dimensional coordinates of the camera coordinate system 200 is defined as a translation vector tc for the image sensor 50.
The external variables for the projector 53 are parameters that indicate the installed state (the position and the orientation) of the projector 53 with respect to the world coordinate system 100. That is, the external variables for the projector 53 are parameters that indicate a shift of a projector coordinate system 300 with respect to the world coordinate system 100. The projector coordinate system 300 is a three-dimensional coordinate system for the projector 53. The projector coordinate system 300 is schematically shown in
Thickness detection processing that is performed by the sewing machine 1 according to the first embodiment will be explained with reference to
As shown in
An image of the sewing object 34 is captured before the projection image is projected onto the sewing object 34. The image that is created by the image capture is stored in the RAM 63 as an initial image (Step S20). The initial image that is created in the processing at Step S20 is used in processing that identifies a characteristic point in the image that is captured of the image that is being projected. Hereinafter, the image that is captured of the image that is being projected is referred to as the “captured image”. Next, image coordinates of the characteristic point are computed in order for the projector 53 to project the characteristic point onto the sewing object 34, and the computed image coordinates of the characteristic point are stored in the RAM 63 (Step S30). The image coordinates that are computed in the processing at Step S30 are image coordinates for the projection image. The image coordinates are coordinates that are determined according to a position within the image. In the present embodiment, in a case where the projector 53 projects a characteristic point 501 at the center of the projection area Q, the coordinates of the characteristic point 501 are computed. In the processing at Step S30, the coordinates are computed on the assumption that the thickness of the sewing object 34 is the value that was set in the processing at Step S10.
In a case where the three-dimensional coordinates of the characteristic point in the world coordinate system 100 are defined as Mw (Xw, Yw, Zw), Xw and Yw are predetermined values. Zw is the initial value that was set in the processing at Step S10. The image coordinates in the projection image, m′=(u′, v′)T, are computed by the procedure described below. (u′, v′)T is a transposed matrix for (u′, v′). First, the three-dimensional coordinates Mw (Xw, Yw, Zw) of the characteristic point in the world coordinate system 100 are converted into the three-dimensional coordinates Mp (Xp, Yp, Zp) of the point in the projector coordinate system 300, based on Equation (1).
Mp=RpMw+tp Equation (1)
In Equation (1), Rp is the rotation matrix that is used for converting the three-dimensional coordinates of the world coordinate system 100, which is stored in the EEPROM 64, into the three-dimensional coordinates of the projector coordinate system 300. tp is the translation vector that is used for converting the three-dimensional coordinates of the world coordinate system 100, which is stored in the EEPROM 64, into the three-dimensional coordinates of the projector coordinate system 300.
Next, the three-dimensional coordinates of the characteristic point in the projector coordinate system 300 are converted into coordinates (x′, y′) in the normalized image in the projector coordinate system 300, based on Equations (2) and (3).
x′=Xp/Zp Equation (2)
y′=Yp/Zp Equation (3)
In addition, coordinates (x″, y″) are computed for a normalized projector, based on Equations (4) and (5), by taking into account the distortion of a projector lens of the projector 53. The normalized projector is a projector for which the distance from the optical center to a screen surface is a unit distance.
x″=x′×(1+k1×r2+k2×r4) Equation (4)
y″=y′×(1+k1×r2+k2×r4) Equation (5):
In Equations (4) and (5), k1 and k2 are respectively the first coefficient of distortion and the second coefficient of distortion for the projector 53. The equation r2=x′2+y′2 holds true.
Next, the coordinates (x″, y″) are converted into the image coordinates (u′, v′) of the projection image, based on Equations (6) and (7).
u′=fx×x″+cx Equation (6)
v′=fy×y″+cy Equation (7)
In Equations (6) and (7), fx, cx, fy, and cy are internal variables for the projector 53. Specifically, fx is the X-axial focal length. cx is the X-axial principal point coordinate. fy is the Y-axial focal length. cy is the Y-axial principal point coordinate.
Next, the projection image is created based on the image coordinates of the characteristic point that were computed in the processing at Step S30, and the created projection image is stored in the RAM 63 (Step S40). Specifically, an image is created in which the characteristic point is placed at the position described by the image coordinates that were computed in the processing at Step S30. Next, the projecting onto the sewing object 34 of the projection image that was created in the processing at Step S40 is started (Step S50). Specifically, the light source 56 of the projector 53 is turned ON, the liquid crystal panel 57 is operated based on the projection image that was created in the processing at Step S40, and the projecting of a projected image 500 onto the sewing object 34 in the projection area Q (refer to
Next, an image of the image capture area is captured by the image sensor 50. The image that is acquired by the image capture is stored in the RAM 63 as the captured image (Step S60). In the present embodiment, the image capture area for the image sensor 50 and the projection area Q for the projector 53 are congruent. However, due to the thickness of the sewing object 34, the projection area Q and the image capture area may be partially non-congruent. An image that shows the characteristic point 501 that is projected by the projector 53 is included in the captured image.
Next, the thickness of the sewing object 34 is computed, and the computed thickness is stored in the RAM 63 (Step S80). Specifically, the thickness of the sewing object 34 is computed based on the coordinates of the characteristic point 501 in the projection image that were computed in the processing at Step S30, the coordinates of the characteristic point 501 in the captured image that was acquired in the processing at Step S60, the parameters for the image sensor 50, and the parameters for the projector 53.
In the processing at Step S80, the three-dimensional coordinates of the characteristic point in the world coordinate system 100 are computed. The three-dimensional coordinates of the characteristic point in the world coordinate system 100 are computed by a method that applies a method that computes three-dimensional coordinates for a corresponding point (the characteristic point) of which images have been captured by cameras that are disposed at two different positions, by utilizing the parallax between the two camera positions. In the computation method that utilizes parallax, the three-dimensional coordinates for the corresponding point in the world coordinate system 100 are computed as hereinafter described. Under conditions in which the position of the sewing object 34 is not changed, if the image coordinates m=(u, v)T and m′=(u′, v′)T are known for the corresponding point of which the images have been captured by the two cameras that are disposed at the different positions, then Equations (8) and (9) can be derived.
smav=PMwav Equation (8)
s′mav′=P′Mwav Equation (9)
In Equation (8), P is a camera projection matrix that yields the image coordinates m=(u, v)T. In Equation (9), P′ is a camera projection matrix that yields the image coordinates m′=(u′, v′)T. The projection matrices are matrices that include the internal variables and the external variables for the cameras. mav, mav′, and Mwav are augmented vectors of m, m′, and Mw, respectively. Mw represents the three-dimensional coordinates in the world coordinate system 100. The augmented vectors are derived by adding an element 1 to given vectors. For example, the augmented vector of m=(u, v)T is mav=(u, v, I)T. s and s′ are scalars.
Equation (10) is derived from Equations (8) and (9).
BMw=b Equation (10)
In Equation (10), B is a matrix with four rows and three columns. An element Bij at row i and column j of the matrix B is expressed by Equation (11). b is expressed by Equation (12).
(B11,B21,B31,B41,B12,B22,B32,B42,B13,B23,B33,B43)=(up31−p11,vp31−p21,u′p31′−p11′,v′p31′−p21′,up32−p12,vp32−p22,u′p32′−p12′,v′p32′−p22′,up33−p13,vp33−p23,u′p33′−p13′,v′p33′−p23′) Equation (11)
b=[p14−up34,p24−vp34,p14′−u′p34′,p24′−v′p34′]T Equation (12):
In Equations (11) and (12), pij is the element at row i and column j of the matrix P. pij′ is the element at row i and column j of the matrix P′. [p14−up34, p24−vp34, p14′−u′p34′, p24′−v′p34′]T is a transposed matrix for [p14−up34, p24−vp34, p14′−u′p34′, p24′−v′p34′].
Accordingly, Mw is expressed by Equation (13).
Mw=B+b Equation (13)
In Equation (13), B+ expresses a pseudoinverse matrix for the matrix B.
The optical models for the image sensor 50 and the projector 53 are the same, so the case where there are two cameras is applicable. The characteristic point is defined as the corresponding point. The image coordinates of the characteristic point in the captured image are defined as m=(u, v)T. The characteristic point in the captured image is specified by taking the difference between the captured image and the initial image. The image coordinates of the characteristic point in the projection image are defined as m′=(u′, v′)T. In Equation (8), the projection matrix for the image sensor 50 is set for P. The projection matrix for the image sensor 50 is expressed by Equation (14). In the same manner, in Equation (9), the projection matrix for the projector 53 is set for P′. The projection matrix for the projector 53 is expressed by Equation (15).
P=Ac[Rc,tc] Equation (14)
P′=Ap[Rp,tp] Equation (15)
In Equation (14), Ac is an internal variable for the image sensor 50. Rc is a rotation matrix for converting the three-dimensional coordinates of the world coordinate system 100 into the three-dimensional coordinates of the camera coordinate system 200. tc is a translation vector for converting the three-dimensional coordinates of the world coordinate system 100 into the three-dimensional coordinates of the camera coordinate system 200. In Equation (15), Ap is an internal variable for the projector 53. Rp is a rotation matrix for converting the three-dimensional coordinates of the world coordinate system 100 into the three-dimensional coordinates of the projector coordinate system 300. tp is a translation vector for converting the three-dimensional coordinates of the world coordinate system 100 into the three-dimensional coordinates of the projector coordinate system 300. Ac, Rc, tc, Ap, Rp, and tp are stored in the EEPROM 64. The three-dimensional coordinates Mw in the world coordinate system 100 are computed based on Equation (13), using m, m′, P, and P′, which are derived as described above. Of the three-dimensional coordinates Mw (Xw, Yw, Zw) of the characteristic point in the world coordinate system 100, Zw denotes the thickness of the sewing object 34. The thickness detection processing is then terminated.
According to the sewing machine 1 according to the first embodiment, the thickness of the sewing object 34 can be computed in a state in which the sewing object 34 is not being pressed. The thickness of the sewing object 34 at the desired position can be computed by the simple operation of placing the sewing object 34 within the area where the image sensor 50 can capture an image of the pattern that the projector 53 projects within the projection area Q.
Projection processing that is performed by the sewing machine 1 according to the second embodiment will be explained with reference to
As shown in
Next, the processing at Steps S50 to S80 is performed in the same manner as in the thickness detection processing shown in
In order for the characteristic point to be projected accurately in the position that is indicated by the three-dimensional coordinates of the world coordinate system 100, it is necessary for the thickness of the sewing object 34 to be set accurately. Therefore, in the known sewing machine, the three-dimensional coordinates of the characteristic point are computed on the assumption that the thickness value is a specified value. Alternatively, in the known sewing machine, the three-dimensional coordinates of the characteristic point are computed using a device that detects the thickness of the sewing object. In the known sewing machine, if the height coordinate for the characteristic point is not set accurately, the characteristic point may not be projected accurately in the position that is indicated by the three-dimensional coordinates of the world coordinate system 100. The sewing machine 1 according to the second embodiment creates the projection image based on the thickness of the sewing object 34 that is computed based on the projection image and the captured image. The sewing machine 1 is therefore able to accurately project a pattern of a specified size in a specified position on the sewing object 34. In a case where a projection image is projected that includes a pattern that indicates the needle drop position, the user is able to know the needle drop position accurately based on the projected image. It is therefore possible to prevent a stitch from being formed in a position where the user does not intend to form the stitch. In a case where a projection image is projected that includes an embroidery pattern that is to be sewn, the user is able to accurately know the position where the embroidery pattern is to be sewn, based on the projected image, before the sewing is performed. It is therefore possible to prevent the embroidery pattern to be sewn in a position where the user does not intend to sew the embroidery pattern.
The sewing machine 1 of the present disclosure is not limited to the embodiments that have been described above, an various types of modifications can be made within the scope of the claims of the present disclosure. For example, the modifications described in (A) to (D) below may be made as desired.
(A) The configuration of the sewing machine 1 may be modified as desired. For example, the sewing machine 1 may be one of a multi-needle sewing machine and an industrial sewing machine. For example, the sewing machine 1 may be modified as described in (A-1) to (A-3) below.
(A-1) The image sensor 50 that the sewing machine 1 includes may be one of a CCD camera and another image capture element. The mounting position of the image sensor 50 can be modified as desired, as long as the image sensor 50 is able to acquire an image of an area on the bed 2.
(A-2) The projector 53 which the sewing machine 1 includes may be any device that is capable of projecting an image onto the bed 2. The position in which the projector 53 is mounted and the projection area of the projector 53 can be modified as desired. In the present embodiment, the projection area Q of the projector 53 is congruent with the image capture area of the image sensor 50. However, the projection area Q of the projector 53 and the image capture area of the image sensor 50 may be partially non-congruent areas. In that case, the characteristic point may be projected in an area where the projection area Q of the projector 53 and the image capture area of the image sensor 50 overlap.
(A-3) The embroidery unit 30 may be attached to the sewing machine 1. However, it is acceptable for the embroidery unit 30 not to be attachable to the sewing machine 1. Different initial values are set for the thickness value in a case where the embroidery unit 30 is attached to the sewing machine 1 and in a case where the side table 49 is attached to the sewing machine 1. However, it is acceptable for the initial values that are set not to be different. The same value may be set for the thickness value in a case where the embroidery unit 30 is attached to the sewing machine 1 and in a case where the side table 49 is attached to the sewing machine 1, as long as the position of the surface of the sewing object is the same.
(B) The camera coordinate system, the projector coordinate system, and the world coordinate system may be associated with one another by parameters that are stored in the sewing machine 1. The methods for defining the camera coordinate system, the projector coordinate system, and the world coordinate system may be modified as desired. For example, the world coordinate system may be defined such that the upper portion of the up-down direction of the sewing machine 1 is defined as positive on the Z axis.
(C) Any given pattern may be projected in the thickness detection processing and the projection processing. For example, one of an embroidery pattern and a stitch that the sewing machine 1 is to sew may be projected in the position where the one of the embroidery pattern and the stitch is to be sewn. In accordance with the image that is projected onto the sewing object, the user is easily able to know the position where the one of the embroidery pattern and the stitch will be formed. For example, any pattern that indicates a specified position, such as a cross-shaped mark that indicates the needle drop position, may be projected.
(D) The processing that is performed in the thickness detection processing and the projection processing may be modified as desired. For example, the method for computing the three-dimensional coordinates of the characteristic point in the world coordinate system 100 may be modified as desired. The three-dimensional coordinates of the characteristic point in the world coordinate system 100 may be computed based on the assumption that the three-dimensional coordinates of the characteristic point in the world coordinate system 100 that are specified based on the projection image are equal to the three-dimensional coordinates of the characteristic point in the world coordinate system 100 that are specified based on the captured image, with the thickness of the sewing object defined as an unknown value. In a case where a plurality of the characteristic points are included in the projection image, the thickness of the sewing object may be computed for one of the characteristic points and may also be computed for the plurality of the characteristic points. In a case where the thickness of the sewing object can be assumed to be uniform, a representative value for the thickness may be computed based on a plurality of thicknesses that are computed for the plurality of the characteristic points. The representative value may be one of a mean value and a mode value, for example. In a case where the thickness of the sewing object can be assumed not to be uniform, the projection image may be created based on each of the plurality of the thicknesses that are computed for the plurality of the characteristic points.
The apparatus and methods described above with reference to the various embodiments are merely examples. It goes without saying that they are not confined to the depicted embodiments. While various features have been described in conjunction with the examples outlined above, various alternatives, modifications, variations, and/or improvements of those features and/or examples may be possible. Accordingly, the examples, as set forth above, are intended to be illustrative. Various changes may be made without departing from the broad spirit and scope of the underlying principles.
Number | Date | Country | Kind |
---|---|---|---|
2010-064437 | Mar 2010 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
4998489 | Hisatake et al. | Mar 1991 | A |
5042410 | Nakashima | Aug 1991 | A |
5195451 | Nakashima | Mar 1993 | A |
5323722 | Goto et al. | Jun 1994 | A |
5537946 | Sadeh et al. | Jul 1996 | A |
5553559 | Inoue et al. | Sep 1996 | A |
5855176 | Takenoya et al. | Jan 1999 | A |
5865133 | Herbach et al. | Feb 1999 | A |
5911182 | Uyama et al. | Jun 1999 | A |
6000350 | Koike et al. | Dec 1999 | A |
6158366 | Codos | Dec 2000 | A |
6167822 | Miyasako et al. | Jan 2001 | B1 |
6407745 | Yamada et al. | Jun 2002 | B1 |
6715435 | Wieczorek et al. | Apr 2004 | B1 |
7155302 | Muto et al. | Dec 2006 | B2 |
7702415 | Roos | Apr 2010 | B2 |
7848842 | Suzuki et al. | Dec 2010 | B2 |
7854209 | Yamasaki | Dec 2010 | B2 |
20060096510 | Kuki et al. | May 2006 | A1 |
20070221107 | Kato et al. | Sep 2007 | A1 |
20090188413 | Hirata et al. | Jul 2009 | A1 |
20090188415 | Tokura | Jul 2009 | A1 |
20090217850 | Tokura | Sep 2009 | A1 |
20110146553 | Wilhelmsson et al. | Jun 2011 | A1 |
Number | Date | Country |
---|---|---|
1 777 331 | Apr 2007 | EP |
A-3-254782 | Nov 1991 | JP |
A-3-254783 | Nov 1991 | JP |
A-5-269285 | Oct 1993 | JP |
A-6-241748 | Sep 1994 | JP |
A-9-81779 | Mar 1997 | JP |
A-2000-55636 | Feb 2000 | JP |
A-2001-229388 | Aug 2001 | JP |
A-2001-330413 | Nov 2001 | JP |
A-2004-28728 | Jan 2004 | JP |
A-2004-144644 | May 2004 | JP |
A-2004-151883 | May 2004 | JP |
A-2005-115420 | Apr 2005 | JP |
A-2008-188148 | Aug 2008 | JP |
A-2009-172123 | Aug 2009 | JP |
A-2009-174981 | Aug 2009 | JP |
A-2009-201704 | Sep 2009 | JP |
Entry |
---|
European Search Report issued in Application No. 11158244.1; Dated Aug. 2, 2011. |
U.S. Appl. No. 13/042,008, filed Mar. 7, 2011, by Tokura. |
Feb. 12, 2013 Office Action issued in U.S. Appl. No. 13/042,008. |
Number | Date | Country | |
---|---|---|---|
20110226170 A1 | Sep 2011 | US |