UNMANNED PARKING SPACE DETECTION METHOD BASED ON PANORAMIC SURROUND VIEW

Information

  • Patent Application
  • 20250086983
  • Publication Number
    20250086983
  • Date Filed
    October 29, 2024
    11 months ago
  • Date Published
    March 13, 2025
    6 months ago
  • CPC
    • G06V20/586
    • G06V10/757
    • G06V10/806
    • G06V10/82
  • International Classifications
    • G06V20/58
    • G06V10/75
    • G06V10/80
    • G06V10/82
Abstract
An unmanned parking space detection method based on a panoramic surround view is provided. The unmanned parking space detection method adopts a Laplace fusion algorithm to splice images to obtain the panoramic surround view, which effectively realize fusion of the images. By using an image quality evaluation function, complexity of the unmanned parking space detection method is reduced while ensuring a quality of the splicing. The panoramic surround view generated by the present disclosure has a more natural transition in the areas being spliced and fused, and the panoramic surround view display is more complete and beautiful. Compared with conventional machine vision detection technologies, the unmanned parking space detection method adopts a target detection technology for detecting parking space entrance lines and adopts a more lightweight network design to ensure real-time performance of a detection algorithm thereof, which has higher detection accuracy and stronger robustness.
Description
TECHNICAL FIELD

The present disclosure relates to a technical field of intelligent vehicles, and in particular to an unmanned parking space detection method based on a panoramic surround view.


BACKGROUND

With development of automotive technology, more and more intelligent vehicles are equipped with advanced driver assistance systems. As one of core technologies of assisted driving, perception technology has become an important way for transition of the intelligent vehicles from assisted driving to unmanned driving.


Currently, gradual increase in the number of urban vehicles has brought challenges to urban transportation, making an urban traffic situation more and more harsh and complex. Since the number of urban parking spaces is difficult to keep up with a growth in the number of automobiles, tension in parking space resources is exacerbated. Parking space detection technology is a key technology used in an unmanned parking environment sensing system. An effect of parking space detection greatly affects safety and stability of a parking process. Through the parking space detection technology, a driverless vehicle is able to obtain information of surrounding parking spaces in real time according to parking space marking lines on a parking lot and accurately judge whether a parking space thereof is valid or not.


In prior art, the driverless vehicle mainly obtains information about a surrounding environment through a certain number of cameras and obtains parking space information based on machine vision detection technology. However, a method of fusing images captured by the cameras increases cost, the images are unable to be effectively fused in the prior art, and detection accuracy of an algorithm thereof is poor.


SUMMARY

An object of the present disclosure is to provide an unmanned parking space detection method based on a panoramic surround view to solve technical problems of high cost, ineffective fusion of images, and poor detection accuracy of the prior art.


The unmanned parking space detection method based on the panoramic surround view comprises:

    • a step 1: respectively acquiring a front original image, a rear original image, a left original image, and a right original image of a surrounding environment of an automobile through a front-view camera, a rear-view camera, a left-view camera and a right-view camera arranged on the automobile;
    • a step 2: respectively performing distortion removal processing and perspective transformation processing on the front original image, the rear original image, the left original image, and the right original image to obtain a front view, a rear view, a left view and a right view;
    • a step 3: extracting feature points from the front view, the rear view, the left view and the right view by a random sample consensus (RANSAC) algorithm, and determining areas to be spliced and fused;
    • a step 4: fusing the areas to be spliced and fused by a Laplace fusion algorithm to obtain the panoramic surround view, and in the fusion process, adjusting a quantity of layers of a Laplace pyramid according to an image quality evaluation function; and a step 5: inputting the panoramic surround view into a trained parking space detection model based on deep learning to obtain a parking space attribute result and a parking space identification point result.


The trained parking space detection model based on deep learning uses a repeated visual geometry group network (REP-VGG) network as a network trunk, and after the network trunk is output, the network trunk passes through a Sim-CSPSPPF module and residual convolution modules, branches of the network trunk pass through the residual convolution modules, and different parking space results are output on feature maps with different resolutions. The parking space attribute result is predicted by a first feature map with a lower resolution, and the parking space identification point result is predicted by a second feature map with a higher resolution.


The unmanned parking space detection method based on the panoramic surround view of the present disclosure has following characteristics.


The unmanned parking space detection method provided by the present disclosure only requires four cameras to complete tasks of image acquisition and splicing, and cost thereof is inexpensive.


The present disclosure adopts the Laplace fusion algorithm to splice the front view, the rear view, the left view, and the right view to obtain the panoramic surround view, which effectively realize fusion of the front view, the rear view, the left view, and the right view. By using the image quality evaluation function, complexity of the unmanned parking space detection method is reduced while ensuring a quality of the splicing. The panoramic surround view generated by the present disclosure has a more natural transition in the areas being spliced and fused, and the panoramic surround view display is more complete and beautiful.


Compared with conventional machine vision detection technologies, the unmanned parking space detection method based on the panoramic surround view of the present disclosure adopts a target detection technology for detecting parking space entrance lines, which has higher detection accuracy and stronger robustness. The present disclosure adopts a more lightweight network design to ensure real-time performance of a detection algorithm thereof.


The unmanned parking space detection method based on the panoramic surround view of the present disclosure does not rely on relative spatial locations of vehicles around a target parking space, but only requires a presence of the parking entrance lines on the ground in a parking area that meets parking standards to accomplish a parking space detection task.


The unmanned parking space detection method based on the panoramic surround view of the present disclosure is applicable to a wide range of parking environments, and is able to detect whether the target parking space is a vertical parking space, a parallel parking space, or an inclined parking space, and is able to obtain information about whether the target parking space is occupied.





BRIEF DESCRIPTION OF DRAWINGS


FIG. 1 is a flow chart of an unmanned parking space detection method based on a panoramic surround view of the present disclosure.



FIG. 2 is a schematic diagram showing mounting positions of cameras of the present disclosure.





DETAILED DESCRIPTION

In order to make objectives, technical solutions, and advantages of the embodiments of the present disclosure clear, technical solutions in the embodiments of the present disclosure will be described clearly and completely in conjunction with the accompanying drawings in the embodiments of the present disclosure. Obviously, the described embodiments are only a part of the embodiments of the present disclosure, rather than all of the embodiments. Based on the embodiments of the present disclosure, all other embodiments obtained by those of ordinary skill in the art without creative work shall fall within the protection scope of the present disclosure.


As shown in FIG. 1, an unmanned parking space detection method based on a panoramic surround view of the present disclosure comprises steps 1-5.


The step 1 comprises respectively acquiring a front original image, a rear original image, a left original image, and a right original image of a surrounding environment of an automobile through a front-view camera, a rear-view camera, a left-view camera and a right-view camera arranged on the automobile.


In the embodiment, mounting positions of the front-view camera, the rear-view camera, the left-view camera, and the right-view camera are shown in FIG. 2, and the four cameras are fisheye cameras with a viewing angle of 190°, which enable a sensing range thereof to completely cover the surrounding environment of the automobile.


The step 2 comprises respectively performing distortion removal processing and perspective transformation processing on the front original image, the rear original image, the left original image, and the right original image to obtain a front view, a rear view, a left view and a right view.


In the embodiment, only radial distortions and tangential distortions of the fisheye cameras are considered.


The step 2 comprises steps 201-205.


The step 201 comprises converting physical coordinates in a world coordinate system to pixel coordinates in a pixel coordinate system by following formula:







Z



(



u




v




1



)


=


D



(




f
x



0



C
x





0



f
y




C
y





0


0


1



)




(



R


T




0


1



)




(



U




V




W




1



)


=

DAE




(



U




V




W




1



)

.







Z is a scale factor, D is a distortion matrix, R is a rotation matrix, T is a translation vector, A is an intrinsic parameter matrix,







A
=

(




f
x



0



C
x





0



f
y




C
y





0


0


1



)



,




E is an extrinsic parameter matrix,







E
=

(



R


T




0


1



)


,

(

U
,
V
,
W

)





is physical coordinates of a target point in the world coordinate system, (u, v) is pixel coordinates of the target point in the pixel coordinate system, the pixel coordinate system is ideal and undistorted, fx is a normalized focal length on an x-axis, fy is a normalized focal length on a y-axis, fx=f/dx, fy=f/dy, f represents a focal length of a corresponding camera, dx and dy are pixel units, and (Cx, Cy) is an origin of a corresponding image.


The step 202 comprises fixing the world coordinate system on a checkerboard, so physical coordinates of any point W on the checkerboard is 0, and obtaining a plurality of (u, v), (U, V) by calibrating to obtain parameters of the intrinsic parameter matrix A.


The step 203 comprises obtaining an ideal matrix form of a conversion relationship between image coordinates and the pixel coordinates based on radial distortion formulas and tangential distortion formulas:









D

1



[




k
1






k
2






k
3




]





D

2



[




p
1






p
2




]


=



D

[




k
1






k
2






k
3




]


[




p
1






p
2




]

=


[





u
^

-
u







v
^

-
v




]

.







The radial distortion formulas are:









x
ˆ

=

x



(

1
+


k
1



r
2


+


k
2



r
4


+


k
3



r
6



)



,

and





y
ˆ

=

y




(

1
+


k
1



r
2


+


k
2



r
4


+


k
3



r
6



)

.







The tangential distortion formulas are:









x
ˆ

=

x
+

(


2


p
1


y

+


p
2

(


r
2

+

2


x
2



)


)



,
and





y
ˆ

=

y
+


(



p
1

(


r
2

+

2


y
2



)

+

2


p
2


x


)

.







D1 and D2 are intermediate values of the distortion matrix, (x, y) are ideal undistorted normalized image coordinates, ({circumflex over (x)},ŷ) are distorted normalized image coordinates, r is a distance from an image pixel to an image center, r=√{square root over (x2+y2)}, k1, k2, k3, p1, p2 are distortion parameters, (û, {circumflex over (v)}) are distorted pixel coordinates, and the conversion relationship between the image coordinates and the pixel coordinates is







u
^

=




x
ˆ


d

x


+


C
x



and



v
^



=



y
ˆ

dy

+


C
y

.







The step 204 comprises determining D1 and D2 by obtaining the distortion parameters through offline calibration images, so as to determine the distortion matrix D, and respectively performing the distortion removal processing on the front original image, the rear original image, the left original image, and the right original image according to the distortion matrix to obtain a dedistorted front original image, a dedistorted rear original image, a dedistorted left original image, and a dedistorted right original image.


The step 205 comprises respectively performing the perspective transformation processing on the dedistorted front original image, the dedistorted rear original image, the dedistorted left original image, and the dedistorted right original image according to the extrinsic parameter matrix E obtained by extrinsic parameter calibration to obtain the front view, the rear view, the left view, and the right view.


As shown in FIG. 2, the step 2 is performed to obtain the front view A, the right view B, the rear view C, and the left view D.


The step 3 comprises extracting feature points from the front view, the rear view, the left view, and the right view by a random sample consensus (RANSAC) algorithm, and determining areas to be spliced and fused.


As shown in FIG. 2, a feature-based image splicing method is configured for the areas that need image registration, and the RANSAC algorithm is configured to extract the feature points to determine the areas A-B, B-A, B-C, C-B, C-D, D-C, D-A, and A-D that need to be spliced and fused.


The step 4 comprises fusing the areas to be spliced and fused by a Laplace fusion algorithm to obtain the panoramic surround view, and in the fusion process, adjusting a quantity of layers of a Laplace pyramid according to an image quality evaluation function.


In the step 4, fusing the areas to be spliced and fused by the Laplace fusion algorithm to obtain the panoramic surround view comprises steps 401a-406a.


The step 401a comprises performing Gaussian smoothing and down-sampling on a first image P1 in two images to be spliced and fused and performing Gaussian smoothing and up-sampling on the first image P1, and respectively repeating for N times to obtain an (N+1)-layer Gaussian pyramid of the first image P1, obtaining a first intermediate image P_downiof an i-th layer of the (N+1)-layer Gaussian pyramid during an i-th Gaussian smoothing and down-sampling, obtaining a second intermediate image P_upi of the i-th layer of the (N+1)-layer Gaussian pyramid during an i-th Gaussian smoothing and up-sampling.


i≤N and the (N+1)-layer Gaussian pyramid of the first image P1 comprises N second intermediate images P_upi and the first image PT.


The step 402a comprises constructing the Laplace pyramid of the first image P1 by following formula:







P1_L
i

=


P_down
i

-

UP




(

P_down

i
+
1


)

.







P1_Li represents a Laplacian residual image of an i-th layer of the Laplace pyramid of the first image P1, UP represent an up-sampling operation, N Laplacian residual images P1_Li constitute the Laplacian pyramid having N layers of the first image P1, and a third intermediate image P1_Ltop of the first image P1 is obtained by performing the Gaussian smoothing and down-sampling on a Laplacian residual image P1_LN of an N-th layer of the Laplace pyramid of the first image PT.


The step 403a comprises processing a second image P2 in the two images to be spliced and fused according to the steps 401a-402a to obtain a Laplace residual image P2_Li of an i-th layer of the second image P2 and a Laplacian pyramid of the second image P2.


The Laplacian pyramid of the second image P2 has N+1 layers and comprises a third intermediate image P2_Ltop of the second image P2.


The step 404a comprises setting binary mask values of a mask matrix, where the mask matrix represents a weight ratio of the first image P1 that needs to be fused and a weight ratio of the second image P2 that needs to be fused, and performing down-sampling on the mask matrix for N times to obtain a down-sampled matrix mask_downi, so as to construct an (N+1)-layer Gaussian pyramid of the mask matrix, and performing image mask fusing on P1_Li and P2_Li to obtain P12fusei according to the (N+1)-layer Gaussian pyramid of the mask matrix.


P12fsei represents a mask fusion image of an i-th layer of the (N+1)-layer Gaussian pyramid of the mask matrix, and an expression thereof is:







P

1


2

fuse
i



=



mask_down
i

*

P1_L
i


+


mask_down
i

*


P2_L
i

.







* is an image layer fusion operator.


The step 405a comprises reconstructing P12fusei by following formula to obtain a reconstructed fusion effect image P12constructi+1 of the i-th layer of the (N+1)-layer Gaussian pyramid of the mask matrix:







P

1


2

c

o

n

s

t

r

u

c


t

i
+
1





=

{







UP



(

P


12

fuse
i



)

*
P

1


2

fuse

i
+
1








i
=
0






UP



(

P


12

c

o

n

s

t

r

u

c


t
i




)

*
P

1


2

fuse

i
+
1









i
>
0





.






UP represents the up-sampling operation, P12constructi+1 represents the reconstructed fusion effect image of the i-th layer of the (N+1)-layer Gaussian pyramid of the mask matrix, P12constructi represents a reconstructed fusion effect image of the (i−1)th layer of the (N+1)-layer Gaussian pyramid of the mask matrix, and * is an image layer fusion operator.


The step 406a comprises starting from i=0 to repeatedly execute the step 405a until i=N, constructing a reconstructed fusion effect image P12constructN+1 of an N-th layer of the (N+1)-layer Gaussian pyramid of the mask matrix, and taking P12constructN+1 as a final result of image fusion.


In the step 4, in the fusion process, adjusting the quantity of the layers of the Laplace fusion algorithm according to the image quality evaluation function; comprises steps 401b-402b.


The step 401b comprises performing an image splicing quality evaluation on the areas to be spliced and fused, and using a structural similarityinde (SSIM) evaluation method to evaluate an image splicing quality from three aspects of brightness, contrast, and structure similarity.


Expressions of an SSIM evaluation function are:









S


(


I

o

r

i


,

I

f

u

s

e



)


=

(



l
α

(


I

o

r

i


,

I
fuse


)

,


c
β

(


I

o

r

i


,

I
fuse


)

,


s
γ

(


I

o

r

i


,

I
fuse


)


)


;






l


(


I

o

r

i


,

I
fuse


)


=



2


μ
0



μ
f


+

c
1




μ
0
2

+

μ
f
2

+

c
1




;






c


(


I

o

r

i


,

I
fuse


)


=



2


σ
o



σ
f


+

c
2




σ
o
2

+

σ
f
2

+

c
2




;

and








s


(


I

o

r

i


,

I
fuse


)


=




σ

o
-
f


+

c
3





σ
o



σ
f


+

c
3



.





S(Iori, Ifuse) is an evaluation coefficient, S(Iori, Ifuse)∈(0, 1), la(Iori,Ifuse) represents brightness similarity, CR(Iori,Ifuse) represents contrast similarity, sy(Iori, Ifuse) represents structural similarity, Iori represents an original image, Ifuse represents an image to be evaluated, α, β, γ respectively represent a weighted contribution value on a score of the brightness similarity, a weighted contribution value on a score of the contrast similarity, and a weighted contribution value on a score of the structural similarity. μo represents an average intensity of the original image, μf represents an average intensity of the image to be evaluated, σo represents a pixel standard deviation of the original image, of represents a pixel standard deviation of the image to be evaluated, σo-f represents a mutual correlation coefficient between the original image and the image to be evaluated, and C1, C2, and C3 are initial constants.


The step 402b comprises adjusting the quantity of the layers of the Laplace fusion algorithm according to an adjustment function A(N, S) based on the evaluation coefficient.


An expression of the adjustment function is:







A

(

N
,
S

)

=


20


e




(

S
-
1

)

4


+

10




(

S
-
1

)

2


+

1
.






e is a constant, S is an adjustable parameter, and S∈(0, 1).


The areas A-B, B-A, B-C, C-B, C-D, D-C, D-A, and A-D that need to be spliced and fused are respectively input as the first image P1 and the second image P2 of the Laplace fusion algorithm. Then above operations are performed to obtain four smooth fused images of an upper left smooth fused image, an upper right smooth fused image, a lower right smooth fused image, and a lower left smooth fused image, and the four smooth fused images are spliced with other areas that do not need to be fused to obtain the panoramic surround view that is complete, smooth, and clear.


The step 5 comprises inputting the panoramic surround view into a trained parking space detection model based on deep learning to obtain a parking space attribute result and a parking space identification point result.


The trained parking space detection model uses a repeated visual geometry group network (REP-VGG) network as a network trunk. After the network trunk is output, the network trunk passes through a Sim-CSPSPPF module and residual convolution modules, branches of the network trunk pass through the residual convolution modules, and different parking space results are output on feature maps with different resolutions. The parking space attribute result is predicted by a first feature map with a lower resolution, and the parking space identification point result is predicted by a second feature map with a higher resolution.


The step 5 comprises steps 501-503.


The step 501 comprises providing a total loss function for the trained parking space detection model based on deep learning.


An expression of the total loss function Loss is:






Loss
=



λ
PKL

·

L
PKL


+


λ
MP

·


L
MP

.







λPKL is a weight hyperparameter of a parking space loss function LPKL, λMP is a weight hyperparameter of a parking space identification point loss function LMP, the parking space loss function LPKL is designed on the first feature map with the lower resolution, and the parking space identification point loss function LMP is designed on the second feature map with the higher resolution.


The step 502 comprises setting an initial learning rate to 0.001, selecting AdamW as a gradient descent optimizer, setting BATCH to be 32, setting a quantity of iterations to be 200, and obtaining the trained parking space detection model based on deep learning when a value of the total loss function Loss converges.


The step 503 comprises inputting the panoramic surround view into the trained parking space detection model based on deep learning to obtain information of two parking space identification points of a parking space entrance line, combining the parking space attribute result and the parking space identification point result to obtain coordinate information M1 and M2 of the two parking space identification points of the parking space entrance line, and obtaining coordinate information M3 and M4 of another two parking space identification points of a target parking space that do not appear in the panoramic surround view by parking space identification point derivation expressions.


The parking space identification point derivation expressions are:








M

3

=



[




cos


α
n





sin


α
n








-
sin



α
n





cos


α
n





]





M

1

M

2







M

1

M

2








L
n


+

M

2



,





and






M

4

=



[




cos


α
n





sin


α
n








-
sin



α
n





cos


α
n





]





M

1

M

2







M

1

M

2








L
n


+

M

1.






αn represents an angle of an n-th parking space in the panoramic surround view, and Ln represents a depth of the n-th parking space in the panoramic surround view.


An expression of the parking space loss function LPKL is:







L
PKL

=



W
T

·

L
T


+


W
O

·

L
O


+


W
C

·


L
C

.







WT represents a weight hyperparameter of a parking space category loss function LT, Wo represents a weight hyperparameter corresponding to a parking space occupancy loss function Lo, and Wc represents a weight hyperparameter corresponding to a parking location confidence loss function LC.


Expressions of the parking space category loss function LT are:








L
T

=




g
=
1


w


1
·
h


1





P
T
i

·

-

1
3










t
=
1




3



[



y
t

·

log

(

σ

(

x
t

)

)


+


(

1
-

y
t


)

·

log

(

1
-

σ

(

x
t

)


)



]





,





and






σ

(

x
t

)

=


1

1
+

exp

(

-

x
t


)



.





w1 represents a width of the first feature map, h1 represents a height of the first feature map, PT determines whether a center of a g-th grid belongs to a parking space area, PTi=1 represents that the center of the g-th grid belongs to the parking space area, and PTi=0 represents that the center of the g-th grid does not belong to the parking space area. xt represents a probability that a predicted parking space is a t-th type parking space. yt represents a true value that the predicted parking space belongs to the t-th type parking space, yt=(0, 0, 1) represents that the predicted parking space is a vertical parking space, yt=(0, 1, 0) represents that the predicted parking space is a parallel parking space, yt=(1, 0, 0) represents that the predicted parking space is an inclined parking space, and σ(xt) is an intermediate function.


An expression of the parking space occupancy loss function LO is:







L
O

=




g
=
1


w


1
·
h


1




[


α
·



P
A
g

(


S

o
x

g

-

S

o
y

g


)

2


+



P
IN
g

(


S

o
x

g

-

S

o
y

g


)

2


]

.






α is a weight value for balancing valid parking spaces and invalid parking spaces, PAg determines whether the center of the g-th grid is within a valid parking space area, PAg=1 represents that the center of the g-th grid is within the valid parking space area, and PAg=0 represents that the center of the g-th grid is located in a non-valid parking space area. PINg determines whether the g-th grid is located in an invalid parking space area, PINg=1=1 represents that the center of the g-th grid is located in the invalid parking space area, and PINg=0 indicates that the center of the g-th grid is located in a non-invalid parking space area. Soxg represents a probability of whether the predicted parking space on the g-th grid is valid, Soyg represents a true value of whether the predicted parking space on the g-th grid is valid, Soyg=0 represents that the predicted parking space is a valid parking space, and Soyg=1 represents that the predicted parking space is an invalid parking space.


An expression of the parking location confidence loss function LC is:







L
C

=




g
=
1


w


1
·
h


1




[



P
C
g

·


(


C
x
g

-

C
y
g


)

2


+



α
P

(

1
-

P
C
g


)




(


C
x
g

-

C
y
g


)

2



]

.






PCg determines whether the center of the g-th grid belongs to the parking space area, PCg=1 represents that the center of the g-th grid belongs to the parking space area, and PCg=0 represents that the center of the g-th grid does not belong to the parking space area. Cxg represents a predicted probability value that the g-th grid belongs to the parking space. Cyg represents a true value that the g-th grid belongs to the parking space, and αp is a weight value for balancing the parking space area and a non-parking space area.


An expression of the parking space identification point loss function LMP is:







L
MP

=



W

M
xy


·

L

M
xy



+


W
θ

·


L
θ

.







WMxy represents a weight hyperparameter corresponding to a parking space identification point coordinate loss function LMxy, and Wθ represents a weight hyperparameter corresponding to a parking space identification point angle orientation loss function Lθ.


An expression of the parking space identification point coordinate loss function LMxy is:







L

M
xy


=




g
=
1


w


2
·
h


2




P

M
xy

g

·


[



(


X
x
g

-


X
y
g


w

2



)

2

+


(


Y
x
g

-


Y
y
g


h

2



)

2


]

.







w2 represents a width of the second feature map, h2 represents a height of the second feature map, PMxy determines whether a g-th grid contains one of the parking space identification points, PMxyg=1 represents that the g-th grid contains the one of the parking space identification points, and PMxyg=0 represents that the g-th grid does not contain the one of the parking space identification points. Xxg represents a predicted value of the one of the parking space identification points in an x-coordinate direction of the g-th grid, Xyg represents a true value of the one of the parking space identification points in the x-coordinate direction of the g-th grid. Yxg represents a predicted value of the one of the parking space identification points in a y-coordinate direction of the g-th grid, and Yyg represents a true value of the one of the parking space identification points in the y-coordinate direction of the g-th grid.


An expression of the parking space identification point angle orientation loss function Lθ is:







L
θ

=




g
=
1


w


2
·
h


2




P

M
xy

g

·


[



(


cos


θ
x
g


-

cos


θ
y
g



)

2

+


(


sin


θ
x
g


-

sin


θ
y
g



)

2


]

.







cosθxg represents a predicted cosine value of an angle orientation of the one of the parking space identification points on the g-th grid, cosθyg represents a true cosine value of the angle orientation of the one of the parking space identification points on the g-th grid. sinθxg represents a predicted sine value of the angle orientation of the one of the parking space identification points on the g-th grid, and sinθyg represents a true sine value of the angle orientation of the one of the parking space identification points on the g-th grid.


Finally, by combining information of the four parking space identification points of each of the parking spaces in the panoramic surround view and the parking space attribute information, detailed information of all of the parking spaces around the automobile is obtained in real time, and the detailed information of all of the parking spaces is transmitted to a next level of autonomous driving path planning module and a decision-making control module to realize an automatic parking function of the automobile.


In summary, the unmanned parking space detection method based on the panoramic surround view of the present disclosure has following characteristics.


The unmanned parking space detection method provided by the present disclosure only requires four cameras to complete tasks of image acquisition and splicing, and cost thereof is inexpensive.


The present disclosure adopts the Laplace fusion algorithm to splice the front view, the rear view, the left view, and the right view to obtain the panoramic surround view, which effectively realize fusion of the front view, the rear view, the left view, and the right view. By using the image quality evaluation function, complexity of the unmanned parking space detection method is reduced while ensuring a quality of the splicing. The panoramic surround view generated by the present disclosure has a more natural transition in the areas being spliced and fused, and the panoramic surround view display is more complete and beautiful.


Compared with conventional machine vision detection technologies, the unmanned parking space detection method based on the panoramic surround view of the present disclosure adopts a target detection technology for detecting parking space entrance lines, which has higher detection accuracy and stronger robustness. The present disclosure adopts a more lightweight network design to ensure real-time performance of a detection algorithm thereof.


The unmanned parking space detection method based on the panoramic surround view of the present disclosure does not rely on relative spatial locations of vehicles around a target parking space, but only requires a presence of the parking entrance lines on the ground in a parking area that meets parking standards to accomplish a parking space detection task.


The unmanned parking space detection method based on the panoramic surround view of the present disclosure is applicable to a wide range of parking environments, and is able to detect whether the target parking space is a vertical parking space, a parallel parking space, or an inclined parking space, and is able to obtain information about whether the target parking space is occupied.


In the description of the present disclosure, the description of reference terms “one embodiment”, “some embodiments”, “examples”, “particular examples”, “some examples”, etc. mean that particular features, structures, materials, or characteristics described in connection with the embodiments or examples are included in at least one embodiment or example of the present disclosure. In the specification, schematic representations of the above terms do not necessarily refer to the same embodiment or example. Furthermore, the specific features, structures, materials or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.


Although the embodiments of the present disclosure have been shown and described, those of ordinary skill in the art can understand that various changes, modifications, substitutions, and variations can be made to these embodiments without departing from the principle and spirit of the present disclosure. The scope of the present disclosure is defined by the appended claims and their equivalents.

Claims
  • 1. An unmanned parking space detection method based on a panoramic surround view, comprising: a step 1: respectively acquiring a front original image, a rear original image, a left original image, and a right original image of a surrounding environment of an automobile through a front-view camera, a rear-view camera, a left-view camera and a right-view camera arranged on the automobile;a step 2: respectively performing distortion removal processing and perspective transformation processing on the front original image, the rear original image, the left original image, and the right original image to obtain a front view, a rear view, a left view, and a right view;a step 3: extracting feature points from the front view, the rear view, the left view and the right view by a random sample consensus (RANSAC) algorithm, and determining areas to be spliced and fused;a step 4: fusing the areas to be spliced and fused by a Laplace fusion algorithm to obtain the panoramic surround view, and adjusting a quantity of layers of a Laplace pyramid according to an image quality evaluation function in a fusion process; anda step 5: inputting the panoramic surround view into a trained parking space detection model based on deep learning to obtain a parking space attribute result and a parking space identification point result;wherein the trained parking space detection model based on deep learning uses a repeated visual geometry group network (REP-VGG) network as a network trunk, and after the network trunk is output, the network trunk passes through a Sim-CSPSPPF module and residual convolution modules, branches of the network trunk pass through the residual convolution modules, and different parking space results are output on feature maps with different resolutions; the parking space attribute result is predicted by a first feature map with a lower resolution, and the parking space identification point result is predicted by a second feature map with a higher resolution.
  • 2. The unmanned parking space detection method based on the panoramic surround view according to claim 1, wherein the step 2 comprises: a step 201: converting physical coordinates in a world coordinate system to pixel coordinates in a pixel coordinate system by following formula:
  • 3. The unmanned parking space detection method based on the panoramic surround view according to claim 1, wherein in the step 4, fusing the areas to be spliced and fused by the Laplace fusion algorithm to obtain the panoramic surround view comprises: a step 401a: performing Gaussian smoothing and down-sampling on a first image P1 in two images to be spliced and fused and performing Gaussian smoothing and up-sampling on the first image P1, and respectively repeating for N times to obtain an (N+1)-layer Gaussian pyramid of the first image P1, obtaining a first intermediate image P_downiof an i-th layer of the (N+1)-layer Gaussian pyramid during an i-th Gaussian smoothing and down-sampling, obtaining a second intermediate image P_upi of the i-th layer of the (N+1)-layer Gaussian pyramid during an i-th Gaussian smoothing and up-sampling, i≤N, and the (N+1)-layer Gaussian pyramid of the first image P1 comprises N second intermediate images P_upiand the first image P1;a step 402a: constructing the Laplace pyramid of the first image P1 by following formula:
  • 4. The unmanned parking space detection method based on the panoramic surround view according to claim 3, wherein in the step 4, in the fusion process, adjusting the quantity of the layers of the Laplace fusion algorithm according to the image quality evaluation function; comprises: a step 401b: performing an image splicing quality evaluation on the areas to be spliced and fused, and using a structural similarityinde (SSIM) evaluation method to evaluate an image splicing quality from three aspects of brightness, contrast, and structure similarity, and expressions of an SSIM evaluation function are:
  • 5. The unmanned parking space detection method based on the panoramic surround view according to claim 1, wherein the step 5 comprises: a step 501: providing a total loss function for the trained parking space detection model based on deep learning, where an expression of the total loss function Loss is:
  • 6. The unmanned parking space detection method based on the panoramic surround view according to claim 5, wherein an expression of the parking space loss function LPKL is:
  • 7. The unmanned parking space detection method based on the panoramic surround view according to claim 5, wherein an expression of the parking space identification point loss function LMP is:
Priority Claims (1)
Number Date Country Kind
202310511825.9 May 2023 CN national
Continuations (1)
Number Date Country
Parent PCT/CN2023/111468 Aug 2023 WO
Child 18931002 US