CN112349227A - Projection display system and method for aircraft takeoff and landing identification - Google Patents
Projection display system and method for aircraft takeoff and landing identification Download PDFInfo
- Publication number
- CN112349227A CN112349227A CN202011367767.XA CN202011367767A CN112349227A CN 112349227 A CN112349227 A CN 112349227A CN 202011367767 A CN202011367767 A CN 202011367767A CN 112349227 A CN112349227 A CN 112349227A
- Authority
- CN
- China
- Prior art keywords
- real
- landing
- take
- time
- aircraft
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 45
- 230000003287 optical effect Effects 0.000 claims abstract description 22
- 238000012545 processing Methods 0.000 claims description 18
- 230000003044 adaptive effect Effects 0.000 claims description 10
- 230000000630 rising effect Effects 0.000 claims description 7
- 230000000694 effects Effects 0.000 abstract description 20
- 230000000875 corresponding effect Effects 0.000 description 76
- 230000007613 environmental effect Effects 0.000 description 27
- 238000005516 engineering process Methods 0.000 description 19
- 230000000007 visual effect Effects 0.000 description 19
- 238000003860 storage Methods 0.000 description 13
- 238000013459 approach Methods 0.000 description 10
- 238000003384 imaging method Methods 0.000 description 10
- 238000012423 maintenance Methods 0.000 description 10
- 239000003086 colorant Substances 0.000 description 9
- 238000009434 installation Methods 0.000 description 9
- 230000008859 change Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 8
- 230000008569 process Effects 0.000 description 8
- 238000004364 calculation method Methods 0.000 description 7
- 230000005540 biological transmission Effects 0.000 description 6
- 230000002567 autonomic effect Effects 0.000 description 5
- 238000004891 communication Methods 0.000 description 5
- 230000007423 decrease Effects 0.000 description 5
- 230000002596 correlated effect Effects 0.000 description 4
- 238000009826 distribution Methods 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 230000004297 night vision Effects 0.000 description 3
- 230000035515 penetration Effects 0.000 description 3
- 230000008447 perception Effects 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 239000011248 coating agent Substances 0.000 description 2
- 238000000576 coating method Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000004313 glare Effects 0.000 description 2
- 238000003331 infrared imaging Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000002834 transmittance Methods 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 230000033228 biological regulation Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000002265 prevention Effects 0.000 description 1
- 238000005096 rolling process Methods 0.000 description 1
- 150000003839 salts Chemical class 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 230000000087 stabilizing effect Effects 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
- 238000003466 welding Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09F—DISPLAYING; ADVERTISING; SIGNS; LABELS OR NAME-PLATES; SEALS
- G09F19/00—Advertising or display means not otherwise provided for
- G09F19/12—Advertising or display means not otherwise provided for using special optical effects
- G09F19/18—Advertising or display means not otherwise provided for using special optical effects involving the use of optical projection means, e.g. projection of images on clouds
Landscapes
- Business, Economics & Management (AREA)
- Accounting & Taxation (AREA)
- Marketing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Traffic Control Systems (AREA)
Abstract
The application discloses aircraft take-off and landing identification projection display system and method, and the system comprises: the aircraft sensor is configured to sense a target aircraft in a take-off and landing area entering an airspace and acquire real-time position information of the target aircraft; a controller configured to generate a real-time presentation scheme of the takeoff and landing identification adapted to real-time position information of the target aircraft; a projection device configured to project an optical image of the take-off and landing indicator to the take-off and landing area according to a real-time presentation scheme to display the take-off and landing indicator indicating the take-off and landing of the target aircraft. The application solves the technical problems that in the prior art, the landing mark in the landing area of the aircraft is poor in display effect and high in layout cost.
Description
Technical Field
The application relates to the technical field of aircrafts, in particular to a projection display system and method for takeoff and landing identification of an aircraft.
Background
In order to ensure the safety of taking off and landing and enable the aircraft driver to accurately position the taking off and landing point, the aircraft, such as a helicopter, puts certain requirements on the identification of the taking off and landing area, and the identification in the taking off and landing area is clearly specified in domestic and foreign standards. Under the condition of good visibility in daytime, the surface of the take-off and landing area is required to clearly show identification lines, identifiers, characters and the like; especially, under the condition of low visibility at night or in daytime, the helicopter taking-off and landing area is required to be provided with a visual prompting lighting system, and the display of the marks of the taking-off and landing area is enhanced, so that a helicopter pilot can accurately identify the taking-off and landing area and assist in safe taking-off and landing.
For example, to achieve both distance visual guidance and near glare prevention, domestic and foreign standards make explicit requirements on the distribution of light intensity along the spatial pitch angle of indicator lights, as shown in table 1. When the helicopter is sailed back at a long distance, the pitch slope angle of the gliding and landing route is usually 2-10 degrees, and the indication type lamp is required to have the maximum light intensity in the range so as to provide the acting distance as far as possible; in the process from approaching a lifting deck to landing, the spatial pitch angle of the helicopter relative to the center of the deck is generally gradually increased from a downward sliding slope angle to approach 90 degrees, and in order to avoid glare possibly generated by over-strong light rays to interfere a pilot, the standard requires that the light intensity of the lamp in the range is gradually reduced along with the increase of the pitch angle. The difference between the maximum and minimum light intensity values distributed according to the pitch angle of the same lamp is 6-20 times.
In order to meet the requirement of light intensity distribution, the solution idea of the existing product is to set different light transmittances for the emergent light path of the same light source according to different pitch angles, so that low-angle light intensity is maintained, and high-angle light intensity is weakened. The specific product realization form is usually to carry out customized design on a light-transmitting cover of the lamp, and different processes such as light transmission, frosting, texture and the like are adopted at different pitch angle positions to realize the change of light transmittance along the pitch angle
However, in the existing solutions, a plurality of lamps are arranged continuously along the identification line, the identifier, the text, and the like for display, and meanwhile, in order to ensure that the identification is clear and recognizable, a certain arrangement number of the lamps is required to be maintained, and in the existing solutions, the total arrangement number of the identification lamps in each type of the take-off and landing area usually exceeds hundreds. That is to say, the aircraft takes off and land needs the pilot lamp in the region very much, is difficult to guarantee that numerous pilot lamps maintain unified display effect, and single pilot lamp still need satisfy requirements such as high surface strength, light transmission scope are big, luminousness along pitch angle change simultaneously, leads to the technical difficulty and the manufacturing cost of single lamp higher, problem such as daily maintenance work load is big. Aiming at the technical problems of poor display effect and high layout cost of the take-off and landing marks in the take-off and landing area of the aircraft in the prior art, an effective solution is not provided at present.
TABLE 1 requirements of standards of take-off and landing areas of domestic and foreign aircrafts for distribution of light intensity of lamps along spatial pitch angle
Disclosure of Invention
The embodiment of the application provides an aircraft take-off and landing identification projection display system and method, which can effectively reduce the number of lamp equipment and improve the display effect so as to at least solve the technical problems of poor take-off and landing identification display effect and high layout cost in the take-off and landing area of an aircraft in the prior art.
According to an aspect of an embodiment of the present application, there is provided an aircraft take-off and landing identification display projection system, including: the aircraft sensor is configured to sense a target aircraft in a take-off and landing area entering an airspace and acquire real-time position information of the target aircraft; a controller configured to generate a real-time presentation scheme of the takeoff and landing identification adapted to real-time position information of the target aircraft; a projection device configured to project an optical image of the take-off and landing indicator to the take-off and landing area according to a real-time presentation scheme to display the take-off and landing indicator indicating the take-off and landing of the target aircraft.
On the basis of the above embodiment, the real-time presentation scheme of the take-off and landing identification includes at least one of the following: the display pattern of the take-off and landing mark, the display color of the take-off and landing mark and the display light intensity of the take-off and landing mark.
On the basis of the above embodiment, the controller is configured to: receiving first real-time location information of a target aircraft from an aircraft sensor; acquiring the corresponding relation between the preset real-time position information of the target aircraft and the real-time presentation scheme of the take-off and landing identification; and determining a first real-time presentation scheme corresponding to the first real-time position information according to the first real-time position information and the corresponding relation of the target aircraft, wherein the first real-time presentation scheme is used for indicating that the take-off and landing marks in the take-off and landing area are presented according to at least one of a first presentation pattern, a first presentation color and a first presentation light intensity.
On the basis of the above-described embodiment, the real-time position information of the target aircraft includes a real-time relative distance of the target aircraft from the take-off and landing area, wherein the controller is configured to: and setting the corresponding relation between the real-time relative distance and the real-time presentation scheme of the take-off and landing identifier, so that the real-time relative distance is related to the presentation light intensity of the take-off and landing identifier.
On the basis of the above embodiment, the controller is configured to: and setting the corresponding relation between the real-time relative distance and the real-time presentation scheme of the take-off and landing identifier, so that the real-time relative distance is related to the size of the presentation pattern of the take-off and landing identifier.
On the basis of the above embodiment, the controller is configured to receive second real-time position information of the target aircraft from the aircraft sensor; determining the real-time relative distance, the real-time relative altitude and the real-time relative azimuth of the target aircraft and the take-off and landing area according to the second real-time position information, and obtaining a second real-time pitch angle numerical value of the target aircraft and the take-off and landing area according to the real-time relative distance, the real-time relative altitude and the real-time relative azimuth; acquiring a corresponding relation between a preset real-time pitch angle value and a real-time presentation scheme of a take-off and landing identifier; and determining a second real-time presentation scheme corresponding to the second real-time pitch angle value according to the second real-time pitch angle value and the corresponding relation, wherein the second real-time presentation scheme is used for indicating the take-off and landing identifier in the take-off and landing area to be presented according to at least one of a second presentation pattern, a second presentation color and a second presentation light intensity.
On the basis of the above embodiments, the projection apparatus is a laser projection apparatus, and has the advantages of high brightness, low power consumption, long service life, and the like. The real-time presentation scheme of the take-off and landing identifier further comprises a projection angle of the take-off and landing identifier, wherein the projection device comprises: the laser projector is used for projecting laser beams according to at least one of a first presentation pattern, a first presentation color and a first presentation light intensity; the base is used for driving the laser projector to move so as to adjust the projection angle of the laser beam to the surface of the take-off and landing area, so that the projection scheme is consistent with the mark of the take-off and landing area, and the standard requirement of the take-off and landing area and/or the actual requirement of a user are met.
On the basis of the embodiment, the real-time position information of the target aircraft comprises the real-time relative distance, the real-time relative altitude and the real-time relative azimuth of the target aircraft and the take-off and landing area, and accordingly, the real-time pitch angle numerical value of the target aircraft and the take-off and landing area is obtained; wherein the projection device is configured to: and adjusting the projection angle and the presentation scheme according to the real-time pitch angle value. Specifically, when the real-time pitch angle values of the target aircraft and the take-off and landing area are within the preset interval range as shown in table 1, the corresponding presentation scheme is displayed according to the projection angle corresponding to the preset interval range.
On the basis of the above-described embodiment, the aircraft sensor includes: the image collector is used for collecting real-time image information of a rising and landing area entering an airspace; the sensor is used for determining the view field position and/or view field pitch angle of the image collector; and the image processor is used for processing the real-time image information, identifying the target aircraft in the real-time image information, and determining the real-time relative position of the target aircraft and the image collector according to the view field position and/or the view field pitch angle so as to obtain the real-time position information of the target aircraft.
On the basis of the above embodiment, the real-time position information of the target aircraft includes at least one of the following: the real-time relative distance between the target aircraft and the image collector, the real-time relative height between the target aircraft and the image collector, and the real-time relative orientation between the target aircraft and the image collector; wherein the image processor is configured to: calculating the real-time relative distance between the target aircraft and the image collector in the real-time image information; acquiring a view field position and/or a view field pitch angle of an image collector; calculating the real-time relative height and the real-time relative orientation between the target aircraft and the image collector according to the view field orientation and/or the view field pitch angle of the image collector and the real-time relative distance between the target aircraft and the image collector to obtain the real-time position information of the target aircraft
On the basis of the embodiment, the aircraft sensor further comprises an aircraft characteristic database for recording the corresponding relation between the external dimension parameter and the image characteristic parameter of the aircraft; wherein the image processor is further configured to: extracting image characteristic parameters of the target aircraft from the real-time image information; reading an aircraft characteristic database, and determining the apparent size parameter of the target aircraft corresponding to the image characteristic parameter; and determining the real-time relative distance between the target aircraft and the image collector according to the apparent size parameter of the target aircraft and the size of the target aircraft in the real-time image information.
On the basis of the above embodiment, the system further includes an environment sensor for acquiring ambient brightness and/or ambient visibility, and the controller is configured to: and setting the corresponding relation between the environment brightness and/or the environment visibility and the real-time presentation scheme of the take-off and landing identifier, so that the environment brightness and/or the environment visibility are at least related to the presentation light intensity of the take-off and landing identifier.
According to another aspect of the embodiments of the present application, there is also provided a projection display method for aircraft takeoff and landing identification, including: sensing a target aircraft in a take-off and landing area entering an airspace, and acquiring real-time position information of the target aircraft; generating a real-time presentation scheme of the take-off and landing identification adaptive to the real-time position information of the target aircraft; an optical image of the takeoff and landing indicator is projected onto the takeoff and landing area according to a real-time presentation scheme to display the takeoff and landing indicator indicating the takeoff and landing of the target aircraft.
According to another aspect of the embodiments of the present application, there is also provided an aircraft take-off and landing identification projection display device, including: the acquisition module is used for sensing a target aircraft in a take-off and landing area entering an airspace and acquiring real-time position information of the target aircraft; the generating module is used for generating a real-time presentation scheme of the take-off and landing identification adaptive to the real-time position information of the target aircraft; and the projection module is used for projecting the optical image of the take-off and landing identifier to the take-off and landing area according to the real-time presentation scheme so as to display the take-off and landing identifier for indicating the take-off and landing of the target aircraft.
According to another aspect of the embodiments of the present application, there is also provided a storage medium, wherein the storage medium includes a stored program, and when the program runs, the apparatus on which the storage medium is located is controlled to execute the above-mentioned projection display method for aircraft takeoff and landing identification.
According to another aspect of the embodiments of the present application, there is also provided a computing device including a processor, wherein the processor is configured to execute a program, and the program executes the method for projecting and displaying the takeoff and landing identifier of the aircraft.
In the embodiment of the application, a target aircraft entering an approaching airspace in a take-off and landing area is sensed through an aircraft sensor, and real-time position information of the target aircraft is obtained; the controller generates a real-time presentation scheme of the take-off and landing identification which is adaptive to the real-time position information of the target aircraft; the projection device projects the optical image of the take-off and landing identification to the take-off and landing area according to the real-time presentation scheme to display the take-off and landing identification used for indicating the take-off and landing of the target aircraft. And then solved the technical problem that the take-off and landing sign in the aircraft take-off and landing area among the prior art shows the effect poor, lay with high costs.
In other words, the display effect of the take-off and landing identification of the take-off and landing area can be dynamically adjusted according to the real-time position information of the aircraft in the whole process of approaching and landing of the target aircraft, so that the display of the take-off and landing identification meets the requirements of various standards, the aircraft can be ensured to safely land in the designated area, meanwhile, the laying cost of the take-off and landing identification is remarkably reduced, the number of devices is reduced, the installation complexity is reduced, the workload of use and maintenance is reduced, and the cost is effectively saved for users.
Drawings
The accompanying drawings, which are included to provide a further understanding of the application and are incorporated in and constitute a part of this application, illustrate embodiment(s) of the application and together with the description serve to explain the application and not to limit the application. In the drawings:
FIG. 1 is a schematic illustration of a take-off and landing zone indication for a helicopter in accordance with an embodiment of the present application;
FIG. 2 is a schematic view of a helicopter take-off and landing area lighting arrangement according to an embodiment of the present application;
FIG. 3 is a block diagram of an aircraft takeoff and landing identification projection display system according to an embodiment of the present application;
FIG. 4 is a schematic diagram of an aircraft takeoff and landing identification projection display system according to an embodiment of the present application;
FIG. 5 is a schematic structural diagram of a particular helicopter take-off and landing indicator projection display system according to an embodiment of the present application;
FIG. 6 is a flow chart of a method for projection display of aircraft takeoff and landing identification in accordance with an embodiment of the present application; and
fig. 7 is a schematic structural diagram of an aircraft takeoff and landing identification projection display device according to an embodiment of the application.
Detailed Description
In order to make the technical solutions better understood by those skilled in the art, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only partial embodiments of the present application, but not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
It should be noted that the terms "first," "second," and the like in the description and claims of this application and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the application described herein are capable of operation in sequences other than those illustrated or described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
Before the embodiments of the present application are explained and explained in detail, some terms, nouns and technical contents related to the embodiments of the present application are explained.
Helicopter Take-Off and landing zone, FATO, Final Approach and Take-Off area.
Helicopter ground-off area, TLOF, Touchdown and lift-off area.
MH5013-2014, technical Standard for civil helicopters flying ground.
CAP437, Standards for offset helicopter landing area, Standard for landing areas on the sea.
Boundary lamp: the system is arranged on the boundary lines of FATO and TLOF and used for remarkably identifying the outlines of the FATO and TLOF areas in a lighting mode under the condition of low visibility at night or in the daytime.
Lifting ring lamp: and the marking lines are arranged on the aircraft landing circle marking lines in FATO and TLOF and are used for remarkably marking the outline of the landing circle in a lighting mode in the low visibility condition at night or in the daytime.
H-shaped lamp: the H-shaped mark is arranged on the H-shaped mark in the FATO and TLOF, and the outline of the H-shaped mark is obviously marked in a lighting mode under the condition of low visibility at night or in the daytime.
Aircraft (aircraft) are aircraft capable of controlled flight in the atmosphere, such as balloons, airships, airplanes, gliders, gyroplanes, helicopters, ornithopters, tiltrotors, etc. In order to accurately control the take-off and landing of an aircraft and guarantee the take-off and landing safety, a corresponding take-off and landing identifier is generally required to be equipped in a take-off and landing area of the aircraft. For the sake of brevity, in the embodiment of the present application, a helicopter is taken as an example for specific description, but those skilled in the art should understand that the aircraft to which the technical solution of the present application is applied is not limited to helicopters.
The helicopter has the characteristics of vertical take-off and landing, hovering and the like, is suitable for field, mountainous area, urban area and other scene operations with narrow space and no fixed runway, and is particularly suitable for performing various operations such as search and rescue, patrol, large facility maintenance, personnel material delivery and the like on ships and ocean engineering platforms. In order to ensure the operation safety of the helicopter on ships, ocean engineering platforms, ground surface airports and temporary take-off and landing points, the national and foreign standards make clear provisions for the marks of the take-off and landing areas of the helicopter. Generally, marks of different types and colors, such as straight lines, arc lines, characters and the like, are coated on the surface of the take-off and landing area so as to facilitate visual observation of a helicopter driver in the take-off and landing process; especially, under the condition of low visibility at night or in daytime, a visual prompting lighting system is required to be arranged in the take-off and landing area of the helicopter, so that a helicopter pilot can accurately identify the take-off and landing area and assist in safe take-off and landing. Fig. 1 is a schematic diagram of a take-off and landing area identifier of a helicopter according to an embodiment of the present application, and as shown in fig. 1, the take-off and landing area of the helicopter should be equipped with an H-shaped identifier, a take-off and landing platform name identifier, a take-off and landing circle identifier, a TLOF area boundary line identifier, a maximum bearing weight identifier, and the like.
FIG. 2 is a schematic view of a helicopter take-off and landing area lighting arrangement according to an embodiment of the present application; as shown in fig. 2, the existing lighting devices arranged in the helicopter take-off and landing area mainly have the following requirements according to the standard: boundary light, lifting ring light, H-shaped light and floodlight. Wherein, the first to the third are mark lamps and the fourth are lighting lamps. The lamps are generally arranged as shown in fig. 2 according to standard requirements, wherein (a) the lamps are arranged at the edge of a rising and landing area, and (b) the lamps are arranged near the center of the rising and landing area. In order to ensure the safety of the helicopter in taking off and landing, the height of the surface of the lamp protruding out of the taking off and landing area arranged on the edge is not higher than 25cm, and the height of the lamp arranged in the taking off and landing area is not more than 3 cm. Meanwhile, in order to bear the impact load when the helicopter lands, the strength of the circle lifting lamp and the H-shaped lamp which are arranged in the lifting area is required to be not less than 1.655 Mpa.
The existing sign lamps are mainly in the form of single-point light sources, point light source arrays, light-emitting plates and the like according to the light-emitting mode classification, and most floodlights are in the form of point light sources. The sign lamp forms a dotted line or a sign symbol visually by a plurality of separated bright spots in a mode that a single lamp or a group of array lamps are arranged on a sign line at certain intervals; the lighting lamps are arranged at intervals on the periphery of the take-off and landing area according to the light intensity distribution range of the lighting lamps, and provide background lighting for a deck of the take-off and landing area.
As shown in fig. 1 and 2, the configuration of the helicopter take-off and landing area lighting equipment has the following problems:
1. the difference between the visual effect at night and the daytime is large
The visual identification of the helicopter take-off and landing area consists of two parts, namely surface coating and an indicating lamp, wherein the surface coating of the take-off and landing area comprises a ground color, an identification line and an identification symbol, and the ground surface is usually a large-area monochromatic color block, the identification line and the symbol are continuous colored solid lines with certain width, so that the visual indication of the whole perfection and clear details is provided for pilots in daytime.
The indicating lamps are arranged at intervals along each marking line and cannot present continuous marking lines and accurate widths. No matter single-point light source, pointolite array, luminescent plate of current lighting apparatus can not provide the bulk lighting, and receive the high restriction of district's barrier of taking off and land, and floodlight can only arrange in the district of taking off and land periphery, and every floodlight can only provide near local blast, finally causes the whole luminance of district of taking off and land to distribute inhomogeneous, and night visual effect is big with daytime difference, increases helicopter operation risk at night.
2. Large quantity of equipment, high purchasing and installation cost and inconvenient use and maintenance
In order to ensure the visual effect at night, the domestic and foreign standards make clear regulations on the lower limit of the arrangement quantity of the lamps, and the configuration quantity of the lamps is required to be increased along with the increase of the take-off and landing area. Wherein, the boundary lamps are not less than 24, the lifting ring lamps are not less than 14, the H-shaped lamps are not less than 24 groups of 144, the floodlight lamps are not less than 20, and the total number of the separation equipment is more than one hundred.
As mentioned above, the lamps in the take-off and landing area of the helicopter need to meet the requirements of high surface strength, large light transmission range, variation of light transmission along the pitch angle and the like, so that the technical difficulty and the manufacturing cost of a single lamp are high. A plurality of devices are separately arranged and centrally controlled, when the helicopter is installed, besides the technical methods of embedding, welding, ground foundation construction or deck hole opening and the like of single devices, cables and connectors among all the single devices need to be laid, and the installation accessories of the type also need to be protected by surface covering, so that a plurality of protrusions exist on the surface of a take-off and landing area, personnel passing and helicopter take-off and landing safety are affected, and the integral installation cost is directly high.
The large number of separate devices also increases the amount of routine maintenance. In order to guarantee the headroom requirement, the equipment in the take-off and landing area needs to endure the open-air operation environment, especially the helicopter take-off and landing area on ships and ocean engineering platforms, the equipment still faces severe environments such as sea damp and hot, salt fog, and the daily maintenance is indispensable, and the equipment is large in quantity and directly leads to the increase of maintenance workload. The taking-off and landing areas of urban building elevated helicopters, ships and ocean engineering platform helicopters are often arranged at the open positions around the taking-off and landing areas, the difference between the taking-off and landing platforms and the ground/water surface is large, personnel falling accidents can be caused by carelessness, and frequent maintenance work is implemented at the edges of the taking-off and landing areas, so that the safety risk of operators is increased seriously.
3. The imaging effect of the lamp is fixed, and the lamp is difficult to be compatible with different standard requirements and different use requirements
Because the lamps are fixedly arranged on the surface of the helicopter take-off and landing area and cannot move, the pattern formed by a plurality of lamps is determined and cannot be changed once the lamps are arranged, so that the difference between domestic and foreign standards cannot be compatible. On the other hand, a helicopter platform as a guarantee facility is often required to serve helicopters of various models, the sizes, weights and configurations of different models are greatly different, the models have different requirements on the marks of the take-off and landing areas, and the fixed indicating lamp cannot be compatible with the requirements of different models.
Example 1
According to the embodiment of the application, the embodiment of the projection display system for the takeoff and landing identification of the aircraft is further provided. FIG. 3 is a block diagram of an aircraft takeoff and landing identification projection display system according to an embodiment of the present application; FIG. 4 is a schematic diagram of an aircraft takeoff and landing identification projection display system according to an embodiment of the present application; as shown in fig. 3 and 4, the aircraft take-off and landing identification projection display system 30 includes:
the aircraft sensor 31 is configured to sense a target aircraft in a take-off and landing area entering an approaching airspace and acquire real-time position information of the target aircraft;
in particular, the aircraft sensor may be implemented in a variety of ways. In an alternative, the aircraft sensor may include a communication device for communicating with the target aircraft to obtain the reported information of the target aircraft, so as to capture the aircraft entering the approaching airspace and its specific location. In another alternative, the aircraft sensors may include a radar system that rapidly captures information about aircraft entering the approach airspace by continuously scanning the take-off and landing area into the approach airspace. In another alternative, the aircraft sensor may include a single image capturing device, such as a large-field camera, a large wide-angle camera, a panoramic camera, or the like, and the background image in the approaching airspace may be acquired through the single image capturing device, so as to capture the target aircraft entering the approaching airspace based on an image recognition technology; the aircraft sensor may also include a combination of multiple image capture devices, such as a combination of multiple conventional cameras oriented to different viewing angles, and image processing techniques to stitch and correct the images or videos of the multiple cameras to capture a target aircraft entering the near airspace. In other optional schemes, the aircraft sensor may further include a night vision imaging camera such as infrared and low-light, and by cooperation with other image acquisition devices, night imaging capability and visible light interference resistance of the target sensing system are further improved. For example, images captured by a large field of view camera may be corrected by images captured by an infrared imaging camera to improve the nighttime capture capability of the large field of view camera.
Specifically, the real-time position information of the target aircraft may include absolute position information of the target aircraft, for example, self position information obtained and reported by the target aircraft according to positioning equipment installed on the target aircraft, and may also include relative position information of the target aircraft, for example, relative position information between the target aircraft and a radar obtained by a radar technology, relative position information between the target aircraft and an image acquisition device obtained by an image processing technology, and meanwhile, based on known position information of the radar or the image acquisition device and a take-off and landing area, relative position information between the target aircraft and the take-off and landing area may be further calculated.
A controller 32 configured to generate a real-time presentation scheme of the takeoff and landing identification adapted to the real-time location information of the target aircraft;
specifically, the visual requirements of the target aircraft on the take-off and landing identifier at different positions are different, so that the presentation effect of the take-off and landing identifier can be dynamically adjusted in real time according to the real-time position information of the target aircraft. The controller 32 may receive the real-time location information from the aircraft sensor 31 in real time, or may receive the real-time location information from the aircraft sensor 31 at intervals (e.g., every second) that are set to meet real-time requirements and to conserve processor resources. In another alternative, the aircraft sensors 31 may send real-time location information to the controller 32 in real-time, or may send real-time location information to the controller 32 at intervals (e.g., every second) that are set to balance the requirements of real-time performance and conservation of aircraft sensor resources.
Specifically, the real-time presentation scheme of the take-off and landing indicator is used for indicating the presentation effect of the take-off and landing indicator in the take-off and landing area, such as the ground color, the boundary line, the take-off and landing circle, the H character and various indicator characters of the take-off and landing area of the helicopter. The specific scheme may include, for example, which landing indicators are presented, the presentation pattern style and pattern size of the landing indicators, the presentation color of the landing indicators, the presentation light intensity, i.e., the presentation brightness, of the landing indicators, and the like.
A projection device 33 configured to project an optical image of a take-off and landing indicator to the take-off and landing area according to the real-time presentation scheme to display a take-off and landing indicator indicating the take-off and landing of the target aircraft.
In particular, the projection device 33 may be implemented based on laser projection technology. The laser projection imaging technology has gradually replaced the traditional light source projection products by the characteristics of high brightness, low power consumption, high reliability and the like, and is widely applied to scenes of high-end visual display such as digital cinema, live performance and the like. According to the specific conditions of buildings around the helicopter taking-off and landing area, the laser projection equipment can be erected at high points around the helicopter taking-off and landing area and projected obliquely downwards, and can also be arranged at the edge of the taking-off and landing area to perform low-angle projection. The ground color, the boundary line, the lifting circle, the H-shaped mark and various mark characters of the helicopter lifting area are directly projected to the helicopter lifting area, a light path pointing to the area near the air where the helicopter navigates is established through surface reflection, and the helicopter can be accurately identified and safely navigated when arriving at the approaching air path.
In conclusion, sensing a target aircraft entering an approaching airspace in a take-off and landing area through an aircraft sensor, and acquiring real-time position information of the target aircraft; the controller generates a real-time presentation scheme of the take-off and landing identification which is adaptive to the real-time position information of the target aircraft; the projection device projects the optical image of the take-off and landing identification to the take-off and landing area according to the real-time presentation scheme to display the take-off and landing identification used for indicating the take-off and landing of the target aircraft. And then solved the technical problem that the take-off and landing sign in the aircraft take-off and landing area among the prior art shows the effect poor, lay with high costs.
Further, the real-time presentation scheme of the take-off and landing identification comprises at least one of the following: the display pattern of the take-off and landing mark, the display color of the take-off and landing mark and the display light intensity of the take-off and landing mark.
Specifically, the presentation pattern of the landing marks is used for determining which landing marks are displayed, and the landing marks are displayed in which pattern, for example, a boundary line of a landing area of the helicopter, a landing circle, an H-word, various mark characters and the like are specified. The patterns can be presented corresponding to different take-off and landing marks according to different types of aircrafts, and meanwhile, simplified patterns and complete and complex patterns can be presented so as to meet the use requirements of different occasions. Therefore, in a specific embodiment, an image of the take-off and landing area may be obtained by the image acquisition device, the environmental ground color of the take-off and landing area may be obtained by the image processing technology, a projection color with a large contrast with the environmental ground color of the take-off and landing area may be obtained according to the color contrast relationship, and the display color of the take-off and landing indicator may be determined according to the projection color to enhance the display effect of the take-off and landing indicator.
Further, the take-off and landing indicator comprises at least one of the following: take-off and landing platform boundaries, FAT0 boundaries, TL0F boundaries, take-off and landing circles, H-words, text logos, etc. In a specific scheme, different rising and falling identifiers can independently set different presentation schemes. For example, the light intensity of the ring and the H may be set higher than the light intensity of other marks for accurate recognition.
Further, the aircraft takeoff and landing identification projection display system 30 may further include an ambient sensor for acquiring ambient brightness information, the controller being configured to: the corresponding relation of the real-time presentation scheme of the environment brightness and the take-off and landing identification is set, so that the environment brightness is related to the presentation light intensity of the take-off and landing identification, for example, the self-adaptive performance of the take-off and landing identification display projection system to the natural environment is enhanced under the condition that the external environment light intensity is weak, the presentation light intensity of the take-off and landing identification is enhanced, so that the identification of a driver or image equipment is facilitated, the self-adaptive performance of the take-off and landing identification display projection system to the natural environment is enhanced by the means, the light intensity can be automatically adjusted.
Further, the aircraft takeoff and landing identification projection display system 30 may further include an environmental sensor for acquiring environmental visibility information, the controller being configured to: setting the corresponding relation between the environmental visibility and the real-time display scheme of the take-off and landing identifier to enable the environmental visibility to be related to the display light intensity of the take-off and landing identifier, and/or setting the corresponding relation between the environmental visibility and the real-time display scheme of the take-off and landing identifier to enable the environmental visibility to be related to the visible light wavelength of the display color of the take-off and landing identifier. For example, in visible light, the wavelengths of visible light of various colors are sequentially reduced in the order of red, orange, yellow, green, blue, ingot and purple, and in the case of low environmental visibility, visible light with longer wavelength can be selected and higher light intensity can be used for facilitating the penetration and the propagation distance of the light wave. So, can strengthen this take-off and land sign projection display system to natural environment's autonomic adaptability, can adjust light intensity or color according to environmental change automatically, the pilot of being convenient for more observes to realize taking off and land smoothly.
Further, the controller is configured to: the method comprises the steps of obtaining the ground color of a take-off and landing area, and setting the corresponding relation between the ground color of the take-off and landing area and a real-time presentation scheme, so that the ground color of the take-off and landing area and the presentation color of the take-off and landing identification are opposite in color, and the display effect of the take-off and landing identification is enhanced.
Further, the controller is configured to: receiving first real-time location information of a target aircraft from an aircraft sensor; acquiring the corresponding relation between the preset real-time position information of the target aircraft and the real-time presentation scheme of the take-off and landing identification; and determining a first real-time presentation scheme corresponding to the first real-time position information according to the first real-time position information and the corresponding relation of the target aircraft, wherein the first real-time presentation scheme is used for indicating that the take-off and landing marks in the take-off and landing area are presented according to at least one of a first presentation pattern, a first presentation color and a first presentation light intensity.
Specifically, the corresponding relationship between the real-time position information of the target aircraft and the real-time presentation scheme of the take-off and landing identifier may be recorded in the configuration file in the form of a corresponding relationship table, and when the real-time position information is acquired, the real-time presentation scheme corresponding to the recording may be acquired in the form of a table lookup. The corresponding relation between the real-time position information of the target aircraft and the real-time presentation scheme of the take-off and landing identification can also be a functional relation, and the corresponding real-time presentation scheme can be obtained through calculation according to the function by inputting the real-time position information. In addition, the corresponding relationship can be realized in a combination manner of a corresponding relationship table and a functional relationship, for example, the display pattern and the display color corresponding to the real-time position information can be obtained according to the corresponding relationship table, and the display light intensity corresponding to the real-time position information is calculated according to the functional relationship, so that the display scheme can be flexibly configured, and the display effect can be enhanced.
Specifically, it can be seen from the above-mentioned scheme that the presentation scheme of the take-off and landing identification is closely related to the real-time position information of the target aircraft, and the take-off and landing identification can be dynamically projected according to the real-time position information of the target aircraft. The product technology is upgraded and updated, the number of equipment is reduced, the installation complexity is reduced, the workload of use and maintenance is reduced, and the cost is effectively saved for users.
Further, the real-time location information of the target aircraft includes a real-time relative distance of the target aircraft from the takeoff and landing area, wherein the controller is configured to: and setting the corresponding relation between the real-time relative distance and the real-time presentation scheme of the take-off and landing identifier, so that the real-time relative distance is related to the presentation light intensity of the take-off and landing identifier, and/or the real-time relative distance is related to the size of the presentation pattern of the take-off and landing identifier.
Specifically, the real-time relative distance includes a horizontal distance and a vertical distance.
Specifically, when the target aircraft is far away from the real-time relative distance of the take-off and landing area, the take-off and landing mark can be projected by large light intensity, so that the pilot can observe remotely, and in the process that the target aircraft is gradually close to the take-off and landing area, the light intensity of the take-off and landing mark can be gradually reduced, so that the visual interference of the light to the pilot is reduced under the condition of meeting the observation requirement of the pilot. It should be further noted that the correlation in the present application may be a positive correlation or a negative correlation, and the correlation may be a linear relationship or a nonlinear relationship. For a linear relationship, for example, as the real-time relative distance of the target aircraft from the takeoff and landing area decreases, an equal proportional decrease in light intensity is exhibited; for a non-linear relationship, for example, as the real-time relative distance of the target aircraft from the takeoff and landing area decreases, but the real-time relative distance is still greater than the preset threshold value, the apparent light intensity decreases slowly by a first factor, and when the real-time relative distance is less than the preset threshold value, the apparent light intensity decreases rapidly by a second factor. According to the scheme, the transmission characteristics of the laser beams in the laser projection mode are considered, and the influence of the reflection of the close-range laser beams on a pilot can be better avoided.
Further, the controller is configured to: receiving second real-time location information of the target aircraft from the aircraft sensor; determining the real-time relative distance, the real-time relative altitude and the real-time relative azimuth of the target aircraft and the take-off and landing area according to the second real-time position information, and obtaining a second real-time pitch angle numerical value of the target aircraft and the take-off and landing area according to the real-time relative distance, the real-time relative altitude and the real-time relative azimuth; acquiring a corresponding relation between a preset real-time pitch angle value and a real-time presentation scheme of a take-off and landing identifier; and determining a second real-time presentation scheme corresponding to the second real-time pitch angle value according to the second real-time pitch angle value and the corresponding relation, wherein the second real-time presentation scheme is used for indicating the take-off and landing identifier in the take-off and landing area to be presented according to at least one of a second presentation pattern, a second presentation color and a second presentation light intensity. Specifically, table 1 lists the requirements of different projection areas in different real-time pitch values on light intensity in the projection area, and the corresponding light intensity can be determined according to the real-time pitch and presented by the projection device.
Further, the controller is configured to: receiving second real-time location information of the target aircraft from the aircraft sensor
Further, projection arrangement is laser projection arrangement, and the real-time scheme that presents of take-off and landing sign still includes the projection angle of take-off and landing sign, and wherein, projection arrangement includes: the laser projector is used for projecting laser beams according to at least one of a first presentation pattern, a first presentation color and a first presentation light intensity; and the base is used for driving the laser projector to move so as to adjust the projection angle of the laser beam to the surface of the take-off and landing area.
Specifically, the laser projection device comprises a laser projection indicator and an adjustable base, the laser projection indicator projects mark lines, symbols, characters and ground background colors required by a helicopter taking-off and landing area to the corresponding position of the helicopter taking-off and landing area, and the adjustable base is used for fixing the laser projection indicator and adjusting the projection light emergent angle within a certain range to adapt to different installation positions.
Specifically, the laser projection device can adapt to a plurality of arrangement positions, and the projection lamp can be installed at a high position near a take-off and landing area and also can be installed at the edge of the take-off and landing area so as to adapt to the requirements of different scenes. When the periphery of the take-off and landing area is provided with a projecting building/structure, such as an upper-layer hull structure near the take-off and landing area of a ship helicopter, an elevated structure near the take-off and landing area of a marine engineering platform and the like, the projection lamp can be erected at a high position to project the required patterns of the take-off and landing area at an elevation angle of-15 degrees to-65 degrees; when no building/structure protrudes from the periphery of the take-off and landing area, the projection device can be arranged at the edge of the take-off and landing area, the required pattern of the take-off and landing area is projected at a low elevation angle of-0.5 degrees to-5 degrees, and the surface height of the projection device protruding out of the take-off and landing area integrally is not more than 25cm at the moment, so that the standard requirement is met.
Specifically, the adjustable base can be adapted to different installation positions, and can also precisely adjust the projection direction of the laser beam, for example, in the case that the take-off and landing area can realize specular reflection, the laser beam after specular reflection in the take-off and landing area can be continuously transmitted along the direction indicated by the real-time relative position by adjusting the projection angle of the laser beam, and can be accurately observed by a pilot in the aircraft.
Further, the aircraft sensor includes: the image collector is used for collecting real-time image information of a rising and landing area entering an airspace; the sensor is used for determining the view field position and/or view field pitch angle of the image collector; and the image processor is used for processing the real-time image information, identifying the target aircraft in the real-time image information, and determining the real-time relative position of the target aircraft and the image collector according to the view field position and/or the view field pitch angle so as to obtain the real-time position information of the target aircraft.
Specifically, the technology of video monitoring by using a camera is deep in the aspects of daily production and life, and the large-view-field and panoramic video acquisition technology is gradually matured and pushed to industrial and commercial application. According to the scheme, a large-view-field camera is selected to collect aerial video information, and a helicopter target is autonomously identified and continuously tracked from a background environment through an artificial intelligence image processing algorithm (hereinafter referred to as an algorithm); in an optional mode, the aircraft sensor is calibrated through the sensor during installation, the azimuth and pitch angle reference parameters of the image collector field of view can be determined according to calibration data under the condition that the aircraft sensor is not moved subsequently, and in another optional mode, the azimuth and pitch angle reference parameters of the image collector field of view can be acquired in real time through the sensor installed on the image collector. The relative angle of the helicopter target in the air can be determined by the azimuth and the pitch angle reference parameters of the view field of the image collector and combining with an image processing algorithm, so that the position information of the helicopter target can be acquired.
Further, the real-time position information of the target aircraft includes at least one of: the real-time relative distance between the target aircraft and the image collector, the real-time relative height between the target aircraft and the image collector, and the real-time relative orientation between the target aircraft and the image collector; wherein the image processor is configured to: calculating the real-time relative distance between the target aircraft and the image collector in the real-time image information; acquiring a view field position and/or a view field pitch angle of an image collector; and calculating the real-time relative height or the real-time relative orientation between the target aircraft and the image collector according to the view field orientation and/or the view field pitch angle of the image collector and the real-time relative distance between the target aircraft and the image collector so as to obtain the real-time position information of the target aircraft.
Specifically, under the condition that the target azimuth and the pitching angle of the helicopter are known, the flight height and the horizontal distance information of the helicopter can be mastered by acquiring the distance between the helicopter and the camera, the flight speed and the flight direction of the helicopter can be mastered by the superposed time sequence information, and the flight track approaching the landing stage can be further generated.
Further, the aircraft sensor also comprises an aircraft characteristic database which is used for recording the corresponding relation between the external dimension parameter and the image characteristic parameter of the aircraft; wherein the image processor is further configured to: extracting image characteristic parameters of the target aircraft from the real-time image information; reading an aircraft characteristic database, and determining the apparent size parameter of the target aircraft corresponding to the image characteristic parameter; and determining the real-time relative distance between the target aircraft and the image collector according to the apparent size parameter of the target aircraft and the size of the target aircraft in the real-time image information.
Specifically, the distance between the helicopter and the camera is determined through an artificial intelligence image processing algorithm. At present, about 80 types of helicopter models are used in the civil market, each type of helicopter has individual differences in overall dimension and configuration layout, the three-dimensional modeling is carried out on the overall shape of each type of helicopter, characteristic parameters are extracted, and a helicopter target characteristic database is generated. When the helicopter flies from far to near when navigating to near, a few pixels appear in video imaging of the camera, the algorithm identifies the moving target and lists the moving target as a suspected target for attention, and as the helicopter approaches to present more and more abundant appearance characteristics in the video imaging, the algorithm compares parameters of the suspected target with a helicopter target characteristic database, so that the target model information can be quickly acquired. After the model is determined, the algorithm can calculate the distance of the target by combining the target angle parameter according to the proportional relation between the external dimension of the model and the video imaging dimension.
It should be noted here that, based on the acquired relative position information between the target aircraft and the image capturing device, and based on the known position information between the image capturing device and the take-off and landing area, the relative position information between the target aircraft and the take-off and landing area may be further calculated.
The aircraft take-off and landing identification projection display system further comprises a communication device for receiving a presentation scheme setting instruction sent by the aircraft, and the controller is configured to generate a real-time presentation scheme of the corresponding take-off and landing identification according to the presentation scheme setting instruction.
Specifically, the system overall control can be matched with a wireless control function besides a wired control mode, and image-text data and video signals are transmitted through technologies such as a satellite, a digital mobile public network, 5G, an industrial local area network and the like, so that remote control is realized. Besides local control in a helicopter take-off and landing area, personnel on the helicopter can implement control and remote online control, unattended operation in the helicopter take-off and landing area is realized, use and maintenance cost is further reduced, emergency deployment capability is improved, and basic support is provided for realizing cooperative guarantee of multiple helicopters in the future.
Furthermore, the aircraft take-off and landing identification projection display system also comprises an attitude sensor used for sensing the motion attitude of a platform to which the take-off and landing area belongs, such as a helicopter take-off and landing platform on a ship and an ocean platform; the aircraft sensor is configured to compensate the acquired real-time image information according to the motion posture of the platform to which the take-off and landing area belongs, so as to determine the real-time relative position of the target aircraft and the image acquisition unit. By the method, the influence of the motion of the platform belonging to the take-off and landing area on the image data can be counteracted, the image processing precision is improved, and more accurate real-time position information of the target aircraft is obtained.
Furthermore, the aircraft take-off and landing identification projection display system also comprises an attitude sensor and a server, wherein the attitude sensor is used for sensing the motion attitude of the platform to which the take-off and landing area belongs, and the server is used for stabilizing the image collector according to the motion attitude and ensuring the fixed position of the image collector. By the method, the use requirements of dynamic scenes such as ships, ocean engineering platforms, field maneuvering guarantee and the like can be met, and the attitude sensor is configured to sense the motion attitudes such as pitching, rolling, heaving and the like of the platform; and configuring a servo platform for stably acquiring the position information of the aerial target in the platform motion state.
Furthermore, the aircraft taking-off and landing identification projection display system is matched with a ceilometer, an visibility meter, meteorological sensors for wind speed, wind direction, temperature and humidity and the like, can realize the autonomous adaptation of a light indicating system to the natural environment, autonomously adjust light intensity and change content according to environmental changes, directly project information influencing the safe taking-off and landing of the helicopter in a taking-off and landing area, and is more convenient for a pilot to observe.
FIG. 5 is a schematic structural diagram of a particular helicopter take-off and landing indicator projection display system according to an embodiment of the present application; as shown in fig. 5, the system is composed of three parts, namely a target sensing system, a laser projection indicating system and a control device. Wherein:
the target sensing system mainly comprises a large-view-field camera with a night vision function and an image processing unit, and is used for finding and locking an aerial target, calculating the spatial position (height, azimuth angle and pitch angle) of the target and sending the information of the target position to the control device.
The laser projection indicating system consists of a laser projection indicator and an adjustable base, the laser projection indicator projects mark lines, symbols, characters and ground background colors required by the take-off and landing area of the helicopter to the corresponding position of the take-off and landing area of the helicopter, and the adjustable base is used for fixing the laser projection indicator and adjusting the projection light emergent angle within a certain range to adapt to different installation positions.
The control device is used for system control, receiving target information from the target perception system, and adjusting the light intensity, the projection pattern and the color of the laser projection indicator so as to meet the visual indication requirements of different helicopter models in various take-off and landing areas.
The workflow of the system comprises: a large-view-field camera of the target perception system monitors the approach airspace of a helicopter taking-off and landing area in a day and night environment, a laser projection indication system adjusts the light intensity according to standard requirements or the maximum visual distance customized by a user, and complete images such as ground colors, identification lines, identifiers and the like are projected on the surface of the helicopter taking-off and landing area; after the target appears in the field of view of the large-field-of-view camera, the image processing unit can identify the target helicopter from the background image through an image identification algorithm, determine information such as the type and relative spatial position of the target helicopter and send the information to the control device; the target sensing system continuously tracks the target helicopter and continuously sends the related information of the target helicopter to the control device according to a certain frequency; in the whole process of approaching and landing of the target helicopter, the control device adjusts the emergent light intensity of the laser projection indicating system according to the position information (distance, height and direction) of the target helicopter relative to the space position according to the position information of the target helicopter continuously sent by the target sensing system so as to meet the requirements of relevant standards until the target helicopter safely lands in a designated area.
Through the embodiment, patterns, colors and brightness of visual indication of the helicopter take-off and landing area can be dynamically adjusted by acquiring target information in real time, compatibility with domestic and foreign standards is achieved, and richer information presentation means are provided for users. Meanwhile, the laser projection technology is used for realizing the night integral visual indication of the helicopter take-off and landing area, the light emitting surface is used for replacing the existing point light source indication mode, the better visual effect compared with the existing scheme is realized, and the night operation safety of the helicopter is improved. Moreover, the method and the device sense the state of the aerial target by using the image recognition technology and acquire the distance, height, direction and speed information of the aerial helicopter in real time. Compared with the existing civil airport solution, the aerial target perception is realized by using an image recognition technology instead of a radar system, so that a large amount of cost can be saved, and the basic requirements of civil general aviation airports, helicopter take-off and landing points and field temporary take-off and landing points are met.
Example 2
There is also provided, in accordance with an embodiment of the present application, an aircraft takeoff and landing identification projection display method embodiment, it should be noted that the steps illustrated in the flowchart of the accompanying drawings may be implemented in a computer system, such as a set of computer-executable instructions, and that, although a logical order is illustrated in the flowchart, in some cases the steps illustrated or described may be performed in an order different than here.
Fig. 6 is a flowchart of a method for displaying a takeoff and landing indicator in a projection manner according to an embodiment of the present application, and as can be seen from fig. 6, the method for displaying a takeoff and landing indicator in a projection manner may include:
s602: sensing a target aircraft in a take-off and landing area entering an airspace, and acquiring real-time position information of the target aircraft;
in step S602, the aircraft sensor may be implemented in various ways. In an alternative, the aircraft sensor may include a communication device for communicating with the target aircraft to obtain the reported information of the target aircraft, so as to capture the aircraft entering the approaching airspace and its specific location. In another alternative, the aircraft sensors may include a radar system that rapidly captures information about aircraft entering the approach airspace by continuously scanning the take-off and landing area into the approach airspace. In another alternative, the aircraft sensor may include a single image capturing device, such as a large-field camera, a large wide-angle camera, a panoramic camera, or the like, and the background image in the approaching airspace may be acquired through the single image capturing device, so as to capture the target aircraft entering the approaching airspace based on an image recognition technology; the aircraft sensor may also include a combination of multiple image capture devices, such as a combination of multiple conventional cameras oriented to different viewing angles, and image processing techniques to stitch and correct the images or videos of the multiple cameras to capture a target aircraft entering the near airspace. In other optional schemes, the aircraft sensor may further include a night vision imaging camera such as infrared and low-light, and by cooperation with other image acquisition devices, night imaging capability and visible light interference resistance of the target sensing system are further improved. For example, images captured by a large field of view camera may be corrected by images captured by an infrared imaging camera to improve the nighttime capture capability of the large field of view camera.
In step S602, the real-time position information of the target aircraft may include absolute position information of the target aircraft, for example, self-position information that is obtained and reported by the target aircraft according to positioning equipment installed on the target aircraft, and may also include relative position information of the target aircraft, for example, relative position information between the target aircraft and a radar that is obtained by a radar technology, relative position information between the target aircraft and an image capturing device that is obtained by an image processing technology, and meanwhile, based on known position information between the radar or the image capturing device and a take-off and landing area, relative position information between the target aircraft and the take-off and landing area may be further calculated.
S604: generating a real-time presentation scheme of a take-off and landing identifier adaptive to the real-time position information of the target aircraft;
in step S604, since the visual requirements of the target aircraft on the takeoff and landing indicator at different positions are different, the presentation effect of the takeoff and landing indicator can be dynamically adjusted in real time according to the real-time position information of the target aircraft. The controller 32 may receive the real-time location information from the aircraft sensor 31 in real time, or may receive the real-time location information from the aircraft sensor 31 at intervals (e.g., every second) that are set to meet real-time requirements and to conserve processor resources. In another alternative, the aircraft sensors 31 may send real-time location information to the controller 32 in real-time, or may send real-time location information to the controller 32 at intervals (e.g., every second) that are set to balance the requirements of real-time performance and conservation of aircraft sensor resources.
In step S604, the real-time presentation scheme of the landing indicator is used to indicate the presentation effect of the landing indicator in the landing area, such as the ground color, the boundary line, the landing circle, the H word, and various identifier characters of the helicopter landing area. The specific scheme may include, for example, which landing indicators are presented, the presentation pattern style and pattern size of the landing indicators, the presentation color of the landing indicators, the presentation light intensity, i.e., the presentation brightness, of the landing indicators, and the like.
S606: and projecting an optical image of a take-off and landing identifier to the take-off and landing area according to the real-time presentation scheme so as to display the take-off and landing identifier for indicating the take-off and landing of the target aircraft.
In step S606, the method may be implemented based on a laser projection technology. The laser projection imaging technology has gradually replaced the traditional light source projection products by the characteristics of high brightness, low power consumption, high reliability and the like, and is widely applied to scenes of high-end visual display such as digital cinema, live performance and the like. According to the specific conditions of buildings around the helicopter taking-off and landing area, the laser projection equipment can be erected at high points around the helicopter taking-off and landing area and projected obliquely downwards, and can also be arranged at the edge of the taking-off and landing area to perform low-angle projection. The ground color, the boundary line, the lifting circle, the H-shaped mark and various mark characters of the helicopter lifting area are directly projected to the helicopter lifting area, a light path pointing to the area near the air where the helicopter navigates is established through surface reflection, and the helicopter can be accurately identified and safely navigated when arriving at the approaching air path.
In conclusion, sensing a target aircraft entering an approaching airspace in a take-off and landing area through an aircraft sensor, and acquiring real-time position information of the target aircraft; the controller generates a real-time presentation scheme of the take-off and landing identification which is adaptive to the real-time position information of the target aircraft; the projection device projects the optical image of the take-off and landing identification to the take-off and landing area according to the real-time presentation scheme to display the take-off and landing identification used for indicating the take-off and landing of the target aircraft. And then solved the technical problem that the take-off and landing sign in the aircraft take-off and landing area among the prior art shows the effect poor, lay with high costs.
Further, the real-time presentation scheme of the take-off and landing identification comprises at least one of the following: the display pattern of the take-off and landing mark, the display color of the take-off and landing mark and the display light intensity of the take-off and landing mark.
Further, the step S604 of generating a real-time presentation scheme of the takeoff and landing identifier adapted to the real-time position information of the target aircraft includes:
step S604 a: acquiring first real-time position information of a target aircraft; acquiring the corresponding relation between the preset real-time position information of the target aircraft and the real-time presentation scheme of the take-off and landing identification; and determining a first real-time presentation scheme corresponding to the first real-time position information according to the first real-time position information and the corresponding relation of the target aircraft, wherein the first real-time presentation scheme is used for indicating that the take-off and landing marks in the take-off and landing area are presented according to at least one of a first presentation pattern, a first presentation color and a first presentation light intensity.
The real-time location information of the target aircraft includes the real-time relative distance of the target aircraft from the takeoff and landing area, and before step S604a, the method further includes the steps of: and setting the corresponding relation between the real-time relative distance and the real-time presentation scheme of the take-off and landing identifier, so that the real-time relative distance is related to the presentation light intensity of the take-off and landing identifier, and/or the real-time relative distance is related to the size of the presentation pattern of the take-off and landing identifier.
Step S604 b: receiving second real-time location information of the target aircraft from the aircraft sensor; determining the real-time relative distance, the real-time relative altitude and the real-time relative azimuth of the target aircraft and the take-off and landing area according to the second real-time position information, and obtaining a second real-time pitch angle numerical value of the target aircraft and the take-off and landing area according to the real-time relative distance, the real-time relative altitude and the real-time relative azimuth; acquiring a corresponding relation between a preset real-time pitch angle value and a real-time presentation scheme of a take-off and landing identifier; and determining a second real-time presentation scheme corresponding to the second real-time pitch angle value according to the second real-time pitch angle value and the corresponding relation, wherein the second real-time presentation scheme is used for indicating the take-off and landing identifier in the take-off and landing area to be presented according to at least one of a second presentation pattern, a second presentation color and a second presentation light intensity. Specifically, table 1 lists the requirements of different projection areas in different real-time pitch values on light intensity in the projection area, and the corresponding light intensity can be determined according to the real-time pitch and presented by the projection device.
Step S604 c: acquiring ambient brightness information; and setting the corresponding relation between the ambient brightness and the real-time presentation scheme of the take-off and landing mark, so that the ambient brightness is inversely related to the presentation light intensity of the take-off and landing mark. So, can strengthen this take-off and land sign projection display system to natural environment's autonomic adaptability, can adjust the light intensity according to environmental change automatically, the pilot of being convenient for more observes to realize taking off and land smoothly.
Step S604 d: obtaining environmental visibility information; setting the corresponding relation between the environmental visibility and the real-time display scheme of the take-off and landing identifier to make the environmental visibility and the display light intensity of the take-off and landing identifier negatively correlated, and/or setting the corresponding relation between the environmental visibility and the real-time display scheme of the take-off and landing identifier to make the environmental visibility and the visible light wavelength of the display color of the take-off and landing identifier negatively correlated. For example, in visible light, the wavelengths of visible light of various colors are sequentially reduced in the order of red, orange, yellow, green, blue, ingot and purple, and in the case of low environmental visibility, visible light with longer wavelength can be selected and higher light intensity can be used for facilitating the penetration and the propagation distance of the light wave. So, can strengthen this take-off and land sign projection display system to natural environment's autonomic adaptability, can adjust light intensity or color according to environmental change automatically, the pilot of being convenient for more observes to realize taking off and land smoothly.
Step S604 e: the method comprises the steps of obtaining the ground color of a take-off and landing area, and setting the corresponding relation between the ground color of the take-off and landing area and a real-time presentation scheme, so that the ground color of the take-off and landing area and the presentation color of the take-off and landing identification are opposite in color, and the display effect of the take-off and landing identification is enhanced.
Further, the step S606 of projecting the optical image of the landing logo to the landing area according to the real-time presentation scheme includes: determining a projection angle of the light beam projected to the surface of the take-off and landing area; and performing laser beam projection according to at least one of a presentation pattern, a presentation color and a presentation light intensity to display an optical image of the take-off and landing mark.
Further, the step S602 of sensing a target aircraft in a take-off and landing area approaching to the airspace and acquiring the real-time position information of the target aircraft includes:
step S6022: acquiring real-time image information of a rising and landing area entering an airspace;
step S6024: acquiring a view field position and/or a view field pitch angle of an image collector;
step S6026: and processing the real-time image information, identifying the target aircraft in the real-time image information, and determining the real-time relative position of the target aircraft and the image collector according to the view field position and/or the view field pitch angle so as to obtain the real-time position information of the target aircraft.
Further, the real-time position information of the target aircraft includes at least one of: the real-time relative distance between the target aircraft and the image collector, the real-time relative height between the target aircraft and the image collector, and the real-time relative orientation between the target aircraft and the image collector.
Further, step S6026 further includes:
step S60262: calculating the real-time relative distance between the target aircraft and the image collector in the real-time image information;
step S60264: and calculating the real-time relative height or the real-time relative orientation between the target aircraft and the image collector according to the view field orientation and/or the view field pitch angle of the image collector and the real-time relative distance between the target aircraft and the image collector so as to obtain the real-time position information of the target aircraft.
Further, step S60262 further includes:
step S602622: extracting image characteristic parameters of the target aircraft from the real-time image information;
step S602624: reading an aircraft characteristic database, and determining the apparent size parameter of the target aircraft corresponding to the image characteristic parameter, wherein the aircraft characteristic database is used for recording the corresponding relation between the apparent size parameter of the aircraft and the image characteristic parameter;
step S602626: and determining the real-time relative distance between the target aircraft and the image collector according to the apparent size parameter of the target aircraft and the size of the target aircraft in the real-time image information.
Further, before sensing a target aircraft in a takeoff and landing area approaching to an airspace and acquiring real-time position information of the target aircraft in step S602, the method further includes:
acquiring a real-time presentation scheme of a default take-off and landing identifier;
and when the target aircraft is not sensed in the approach airspace of the take-off and landing area, projecting an optical image of the take-off and landing identification to the take-off and landing area according to a default take-off and landing identification presentation scheme.
It should be noted that, for simplicity of description, the above-mentioned method embodiments are described as a series of acts or combination of acts, but those skilled in the art will recognize that the present application is not limited by the order of acts described, as some steps may occur in other orders or concurrently depending on the application. Further, those skilled in the art should also appreciate that the embodiments described in the specification are preferred embodiments and that the acts and modules referred to are not necessarily required in this application.
Through the above description of the embodiments, those skilled in the art can clearly understand that the method for displaying the takeoff and landing identification of an aircraft according to the above embodiments can be implemented by software plus a necessary general hardware platform, and of course, may also be implemented by hardware, but in many cases, the former is a better implementation. Based on such understanding, the technical solutions of the present application may be embodied in the form of a software product, which is stored in a storage medium (e.g., ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal device (e.g., a mobile phone, a computer, a server, or a network device) to execute the method of the embodiments of the present application.
Example 3
According to an embodiment of the present application, there is also provided an aircraft takeoff and landing indication projection display device for implementing the aircraft takeoff and landing indication projection display method, as shown in fig. 7, where the device 70 includes: an acquisition module 702, a generation module 704, and a projection module 706. Wherein:
an obtaining module 702, configured to sense a target aircraft in a take-off and landing area approaching an airspace, and obtain real-time position information of the target aircraft;
a generating module 704, configured to generate a real-time presentation scheme of the takeoff and landing identifier adapted to the real-time position information of the target aircraft;
a projecting module 706 configured to project an optical image of the take-off and landing indicator to the take-off and landing area according to the real-time presentation scheme, so as to display the take-off and landing indicator indicating the take-off and landing of the target aircraft.
In conclusion, sensing a target aircraft entering an approaching airspace in a take-off and landing area through an aircraft sensor, and acquiring real-time position information of the target aircraft; the controller generates a real-time presentation scheme of the take-off and landing identification which is adaptive to the real-time position information of the target aircraft; the projection device projects the optical image of the take-off and landing identification to the take-off and landing area according to the real-time presentation scheme to display the take-off and landing identification used for indicating the take-off and landing of the target aircraft. And then solved the technical problem that the take-off and landing sign in the aircraft take-off and landing area among the prior art shows the effect poor, lay with high costs.
Here, it should be noted that the acquiring module 702, the generating module 704 and the projecting module 706 correspond to steps S602 to S606 in embodiment 2, and the three modules are the same as the corresponding steps in the implementation example and the application scenario, but are not limited to the disclosure in embodiment 2.
Further, the real-time presentation scheme of the take-off and landing identification comprises at least one of the following: the display pattern of the take-off and landing mark, the display color of the take-off and landing mark and the display light intensity of the take-off and landing mark.
Further, the generating module 704 includes:
a first generating unit configured to acquire first real-time position information of a target aircraft; acquiring the corresponding relation between the preset real-time position information of the target aircraft and the real-time presentation scheme of the take-off and landing identification; and determining a first real-time presentation scheme corresponding to the first real-time position information according to the first real-time position information and the corresponding relation of the target aircraft, wherein the first real-time presentation scheme is used for indicating that the take-off and landing marks in the take-off and landing area are presented according to at least one of a first presentation pattern, a first presentation color and a first presentation light intensity.
Further, the device further comprises a setting unit, which is used for setting the corresponding relation between the real-time relative distance and the real-time presentation scheme of the take-off and landing mark, so that the real-time relative distance is related to the presentation light intensity of the take-off and landing mark, and/or the real-time relative distance is related to the size of the presentation pattern of the take-off and landing mark.
A second generating unit configured to acquire second real-time position information of the target aircraft; determining the real-time relative distance, the real-time relative altitude and the real-time relative azimuth of the target aircraft and the take-off and landing area according to the second real-time position information, and obtaining a second real-time pitch angle numerical value of the target aircraft and the take-off and landing area according to the real-time relative distance, the real-time relative altitude and the real-time relative azimuth; acquiring a corresponding relation between a preset real-time pitch angle value and a real-time presentation scheme of a take-off and landing identifier; and determining a second real-time presentation scheme corresponding to the second real-time pitch angle value according to the second real-time pitch angle value and the corresponding relation, wherein the second real-time presentation scheme is used for indicating the take-off and landing identifier in the take-off and landing area to be presented according to at least one of a second presentation pattern, a second presentation color and a second presentation light intensity. Specifically, table 1 lists the requirements of different projection areas in different real-time pitch values on light intensity in the projection area, and the corresponding light intensity can be determined according to the real-time pitch and presented by the projection device.
A third generating unit configured to acquire ambient brightness information; and setting the corresponding relation between the ambient brightness and the real-time presentation scheme of the take-off and landing mark, so that the ambient brightness is inversely related to the presentation light intensity of the take-off and landing mark. So, can strengthen this take-off and land sign projection display system to natural environment's autonomic adaptability, can adjust the light intensity according to environmental change automatically, the pilot of being convenient for more observes to realize taking off and land smoothly.
A fourth generating unit configured to acquire environmental visibility information; setting the corresponding relation between the environmental visibility and the real-time display scheme of the take-off and landing identifier to make the environmental visibility and the display light intensity of the take-off and landing identifier negatively correlated, and/or setting the corresponding relation between the environmental visibility and the real-time display scheme of the take-off and landing identifier to make the environmental visibility and the visible light wavelength of the display color of the take-off and landing identifier negatively correlated. For example, in visible light, the wavelengths of visible light of various colors are sequentially reduced in the order of red, orange, yellow, green, blue, ingot and purple, and in the case of low environmental visibility, visible light with longer wavelength can be selected and higher light intensity can be used for facilitating the penetration and the propagation distance of the light wave. So, can strengthen this take-off and land sign projection display system to natural environment's autonomic adaptability, can adjust light intensity or color according to environmental change automatically, the pilot of being convenient for more observes to realize taking off and land smoothly.
And the fifth generating unit is configured to acquire the ground color of the take-off and landing area, and set a corresponding relation between the ground color of the take-off and landing area and the real-time presentation scheme, so that the ground color of the take-off and landing area and the presentation color of the take-off and landing identifier are opposite colors, and the display effect of the take-off and landing identifier is enhanced.
Here, it should be noted that the first to fifth generating units correspond to steps S604a to S604e in embodiment 2, and the five units are the same as the examples and application scenarios realized by the corresponding steps, but are not limited to the disclosure of embodiment 2.
Further, the projection module 706 is configured to determine a projection angle of the light beam to the surface of the take-off and landing area; and performing laser beam projection according to at least one of a presentation pattern, a presentation color and a presentation light intensity to display an optical image of the take-off and landing mark.
Further, the obtaining module 702 includes;
a collecting unit: the system is used for acquiring real-time image information of a take-off and landing area in an approaching airspace;
an acquisition unit: the system comprises a field of view azimuth and/or a field of view elevation angle used for acquiring an image collector;
a calculation unit: the real-time image acquisition device is used for processing the real-time image information, identifying the target aircraft in the real-time image information, and determining the real-time relative position of the target aircraft and the image acquisition device according to the view field position and/or the view field pitch angle so as to obtain the real-time position information of the target aircraft.
Here, it should be noted that the above-mentioned acquisition unit, and calculation unit correspond to steps S6022 to S6026 in embodiment 2, and the above-mentioned three units are the same as the examples and application scenarios realized by the corresponding steps, but are not limited to the disclosure of embodiment 2.
Further, the real-time position information of the target aircraft includes at least one of: the real-time relative distance between the target aircraft and the image collector, the real-time relative height between the target aircraft and the image collector, and the real-time relative orientation between the target aircraft and the image collector.
Further, the calculation unit includes;
a first calculation subunit: calculating the real-time relative distance between the target aircraft and the image collector in the real-time image information;
a second calculation subunit: and calculating the real-time relative height or the real-time relative orientation between the target aircraft and the image collector according to the view field orientation and/or the view field pitch angle of the image collector and the real-time relative distance between the target aircraft and the image collector so as to obtain the real-time position information of the target aircraft.
Here, it should be noted that the first and second calculating sub-units correspond to steps S60262 to S60264 in embodiment 2, and the two units are the same as the example and application scenarios realized by the corresponding steps, but are not limited to the disclosure of embodiment 2.
Further, the first calculation subunit includes;
an extraction subunit: extracting image characteristic parameters of the target aircraft from the real-time image information;
reading the subunit: reading an aircraft characteristic database, and determining the apparent size parameter of the target aircraft corresponding to the image characteristic parameter, wherein the aircraft characteristic database is used for recording the corresponding relation between the apparent size parameter of the aircraft and the image characteristic parameter;
determining a subunit: and determining the real-time relative distance between the target aircraft and the image collector according to the apparent size parameter of the target aircraft and the size of the target aircraft in the real-time image information.
Here, it should be noted that the above-mentioned extraction subunit, reading subunit and determination subunit correspond to steps S602622 to S602626 in embodiment 2, and the above-mentioned two units are the same as the example and application scenarios realized by the corresponding steps, but are not limited to what is disclosed in embodiment 2.
Further, the apparatus further comprises: the default display module is used for acquiring a default real-time presentation scheme of the take-off and landing identifier; and when the target aircraft is not sensed in the approach airspace of the take-off and landing area, projecting an optical image of the take-off and landing identification to the take-off and landing area according to a default take-off and landing identification presentation scheme.
Example 4
Embodiments of the present application may provide a computing device, which may be any one of computer terminal devices in a computer terminal group. Optionally, in this embodiment, the computing device may also be replaced with a terminal device such as a mobile terminal.
Optionally, in this embodiment, the computing device may be located in at least one network device of a plurality of network devices of a computer network.
Optionally, in this embodiment, the above-mentioned computing device includes one or more processors, a memory, and a transmission device. The memory may be used to store software programs and modules, such as program instructions/modules corresponding to the aircraft takeoff and landing identification display method and apparatus in the embodiments of the present application. The processor executes various functional applications and data processing by running software programs and modules stored in the memory, namely, the aircraft take-off and landing identification projection display method is realized.
Alternatively, the memory may include high speed random access memory, and may also include non-volatile memory, such as one or more magnetic storage devices, flash memory, or other non-volatile solid-state memory. In some instances, the memory may further include memory located remotely from the processor, which may be connected to the computing device over a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
In this embodiment, when the processor in the above-mentioned computing device runs the stored program code, the following method steps may be executed: sensing a target aircraft in a take-off and landing area entering an airspace, and acquiring real-time position information of the target aircraft; generating a real-time presentation scheme of the take-off and landing identification adaptive to the real-time position information of the target aircraft; an optical image of the takeoff and landing indicator is projected onto the takeoff and landing area according to a real-time presentation scheme to display the takeoff and landing indicator indicating the takeoff and landing of the target aircraft.
Further, in this embodiment, when the processor in the computing device runs the stored program code, any method step listed in embodiment 2 may be executed, which is not described in detail herein for reasons of brevity.
Example 5
Embodiments of the present application also provide a storage medium. Optionally, in this embodiment, the storage medium may be configured to store program codes executed by the aircraft takeoff and landing identification projection display method.
Optionally, in this embodiment, the storage medium may be located in any one of computer terminals in a computer terminal group in a computer network, or in any one of mobile terminals in a mobile terminal group.
Optionally, in this embodiment, the storage medium is configured to store program code for performing the following steps: sensing a target aircraft in a take-off and landing area entering an airspace, and acquiring real-time position information of the target aircraft; generating a real-time presentation scheme of the take-off and landing identification adaptive to the real-time position information of the target aircraft; an optical image of the takeoff and landing indicator is projected onto the takeoff and landing area according to a real-time presentation scheme to display the takeoff and landing indicator indicating the takeoff and landing of the target aircraft.
Further, in this embodiment, the storage medium is configured to store the program code for executing any one of the method steps listed in embodiment 2, which is not described in detail herein for brevity.
The above-mentioned serial numbers of the embodiments of the present application are merely for description and do not represent the merits of the embodiments.
In the above embodiments of the present application, the descriptions of the respective embodiments have respective emphasis, and for parts that are not described in detail in a certain embodiment, reference may be made to related descriptions of other embodiments.
In the embodiments provided in the present application, it should be understood that the disclosed technology can be implemented in other ways. The above-described embodiments of the apparatus are merely illustrative, and for example, a division of a unit is merely a division of a logic function, and an actual implementation may have another division, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interface units, and may be an electrical or other form.
Units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present application may be substantially implemented or contributed to by the prior art, or all or part of the technical solution may be embodied in a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present application. And the aforementioned storage medium includes: a U-disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a removable hard disk, a magnetic or optical disk, and other various media capable of storing program codes.
The foregoing is only a preferred embodiment of the present application and it should be noted that those skilled in the art can make several improvements and modifications without departing from the principle of the present application, and these improvements and modifications should also be considered as the protection scope of the present application.
Claims (10)
1. An aircraft takeoff and landing identification projection display system, comprising:
the aircraft sensor is configured to sense a target aircraft in a take-off and landing area entering an airspace and acquire real-time position information of the target aircraft;
a controller configured to generate a real-time presentation scheme of a takeoff and landing identification adapted to real-time location information of the target aircraft;
a projection device configured to project an optical image of a take-off and landing indicator to the take-off and landing area according to the real-time presentation scheme to display a take-off and landing indicator indicating the take-off and landing of the target aircraft.
2. The system of claim 1, wherein the real-time presentation of the landing identifiers comprises at least one of: the display pattern of the take-off and landing mark, the display color of the take-off and landing mark and the display light intensity of the take-off and landing mark.
3. The system of claim 2, wherein the controller is configured to:
receiving first real-time location information of the target aircraft from the aircraft sensor;
acquiring the corresponding relation between preset real-time position information of the target aircraft and a real-time presentation scheme of the take-off and landing identification;
and determining a first real-time presentation scheme corresponding to the first real-time position information according to the first real-time position information of the target aircraft and the corresponding relation, wherein the first real-time presentation scheme is used for indicating that the take-off and landing marks in the take-off and landing area are presented according to at least one of a first presentation pattern, a first presentation color and a first presentation light intensity.
4. The system of claim 3, wherein the real-time location information of the target aircraft includes a real-time relative distance of the target aircraft from the takeoff and landing area, wherein the controller is configured to: and setting the corresponding relation between the real-time relative distance and the real-time presentation scheme of the take-off and landing identifier, so that the real-time relative distance is at least related to the presentation light intensity of the take-off and landing identifier.
5. The system of claim 2, wherein the controller is configured to:
receiving second real-time location information of the target aircraft from the aircraft sensor;
determining a real-time relative distance, a real-time relative altitude and a real-time relative azimuth of the target aircraft and the take-off and landing area according to the second real-time position information, and obtaining a second real-time pitch angle value of the target aircraft and the take-off and landing area according to the real-time relative distance, the real-time relative altitude and the real-time relative azimuth;
acquiring a corresponding relation between a preset real-time pitch angle value and a real-time presentation scheme of the take-off and landing identifier;
and determining a second real-time presentation scheme corresponding to the second real-time pitch angle value according to the second real-time pitch angle value and the corresponding relation, wherein the second real-time presentation scheme is used for indicating that the take-off and landing identifier in the take-off and landing area is presented according to at least one of a second presentation pattern, a second presentation color and a second presentation light intensity.
6. The system of claim 2, wherein the projection device is a laser projection device, the real-time presentation scheme of the landing indicator further comprises a projection angle of the landing indicator, wherein the projection device comprises:
the laser projector is used for projecting laser beams according to at least one of the presentation pattern, the presentation color and the presentation light intensity;
and the base is used for driving the laser projector to move so as to adjust the projection angle of the laser beam to the surface of the take-off and landing area.
7. The system of any of claims 1-6, wherein the aircraft sensor comprises:
the image collector is used for collecting real-time image information of the rising and landing area in the approaching airspace;
the sensor is used for determining the view field position and/or view field pitch angle of the image collector;
and the image processor is used for processing the real-time image information, identifying a target aircraft in the real-time image information, and determining the real-time relative position of the target aircraft and the image collector according to the view field position and/or the view field pitch angle so as to obtain the real-time position information of the target aircraft.
8. The system of claim 7, wherein the real-time location information of the target aircraft includes at least one of: the real-time relative distance between the target aircraft and the image collector, the real-time relative height between the target aircraft and the image collector, and the real-time relative orientation between the target aircraft and the image collector; wherein the image processor is configured to:
calculating the real-time relative distance between the target aircraft and the image collector in the real-time image information;
acquiring a view field position and/or a view field pitch angle of the image collector;
and calculating the real-time relative altitude or the real-time relative orientation between the target aircraft and the image collector according to the view field orientation and/or the view field pitch angle of the image collector and the real-time relative distance between the target aircraft and the image collector to obtain the real-time position information of the target aircraft.
9. The system of claim 8, wherein the aircraft sensor further comprises an aircraft feature database for recording the correspondence between the external dimension parameter and the image feature parameter of the aircraft; wherein the image processor is further configured to:
extracting image characteristic parameters of the target aircraft from the real-time image information;
reading the aircraft characteristic database, and determining the apparent size parameter of the target aircraft corresponding to the image characteristic parameter;
and determining the real-time relative distance between the target aircraft and the image collector according to the apparent size parameter of the target aircraft and the size of the target aircraft in the real-time image information.
10. An aircraft take-off and landing identification projection display method is characterized by comprising the following steps:
sensing a target aircraft in a take-off and landing area entering an airspace, and acquiring real-time position information of the target aircraft;
generating a real-time presentation scheme of a take-off and landing identifier adaptive to the real-time position information of the target aircraft;
and projecting an optical image of a take-off and landing identifier to the take-off and landing area according to the real-time presentation scheme so as to display the take-off and landing identifier for indicating the take-off and landing of the target aircraft.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011367767.XA CN112349227A (en) | 2020-11-27 | 2020-11-27 | Projection display system and method for aircraft takeoff and landing identification |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011367767.XA CN112349227A (en) | 2020-11-27 | 2020-11-27 | Projection display system and method for aircraft takeoff and landing identification |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112349227A true CN112349227A (en) | 2021-02-09 |
Family
ID=74365136
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011367767.XA Pending CN112349227A (en) | 2020-11-27 | 2020-11-27 | Projection display system and method for aircraft takeoff and landing identification |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112349227A (en) |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102756808A (en) * | 2011-04-28 | 2012-10-31 | 株式会社拓普康 | Taking-off and landing target instrument and automatic taking-off and landing system |
CN105059533A (en) * | 2015-08-14 | 2015-11-18 | 深圳市多翼创新科技有限公司 | Aircraft and landing method thereof |
CN105501457A (en) * | 2015-12-16 | 2016-04-20 | 南京航空航天大学 | Infrared vision based automatic landing guidance method and system applied to fixed-wing UAV (unmanned aerial vehicle) |
CN107409174A (en) * | 2015-03-31 | 2017-11-28 | 深圳市大疆创新科技有限公司 | System and method for the operation of control unmanned vehicle |
CN111332485A (en) * | 2018-12-19 | 2020-06-26 | 福特全球技术公司 | Accurate unmanned aerial vehicle landing system and method based on thermal image |
CN214705387U (en) * | 2020-11-27 | 2021-11-12 | 海丰通航科技有限公司 | Aircraft take-off and landing identification projection display system |
-
2020
- 2020-11-27 CN CN202011367767.XA patent/CN112349227A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102756808A (en) * | 2011-04-28 | 2012-10-31 | 株式会社拓普康 | Taking-off and landing target instrument and automatic taking-off and landing system |
CN107409174A (en) * | 2015-03-31 | 2017-11-28 | 深圳市大疆创新科技有限公司 | System and method for the operation of control unmanned vehicle |
CN105059533A (en) * | 2015-08-14 | 2015-11-18 | 深圳市多翼创新科技有限公司 | Aircraft and landing method thereof |
CN105501457A (en) * | 2015-12-16 | 2016-04-20 | 南京航空航天大学 | Infrared vision based automatic landing guidance method and system applied to fixed-wing UAV (unmanned aerial vehicle) |
CN111332485A (en) * | 2018-12-19 | 2020-06-26 | 福特全球技术公司 | Accurate unmanned aerial vehicle landing system and method based on thermal image |
CN214705387U (en) * | 2020-11-27 | 2021-11-12 | 海丰通航科技有限公司 | Aircraft take-off and landing identification projection display system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Alam et al. | A survey of safe landing zone detection techniques for autonomous unmanned aerial vehicles (UAVs) | |
CN106203265B (en) | A kind of Construction Fugitive Dust Pollution source monitors automatically and coverage forecasting system and method | |
US20190112048A1 (en) | Systems and methods for unmanned aerial vehicles | |
CN202976376U (en) | Forest fire monitoring and emergency command system based unmanned aerial vehicle | |
US10068506B2 (en) | System, device for creating an aerial image | |
CN106127201B (en) | A kind of unmanned plane landing method of view-based access control model positioning landing end | |
US10384801B2 (en) | Device for assisting the piloting of a rotorcraft, associated display, and a corresponding method of assisting piloting | |
CN110362109A (en) | A kind of cross-domain shutdown library landing method of unmanned plane and landing platform | |
US20120026190A1 (en) | Methods and systems for attitude differentiation in enhanced vision images | |
CN214705387U (en) | Aircraft take-off and landing identification projection display system | |
CN106813900B (en) | A kind of civil airport navigational lighting aid flight check method based on unmanned air vehicle technique | |
CN108399797A (en) | The system and method landed for assisting in flying device and corresponding aircraft | |
EP3702869B1 (en) | Autonomous aircraft sensor-based positioning and navigation system using markers | |
CN106210627A (en) | A kind of unmanned plane fire dispatch system | |
CN112068539A (en) | Unmanned aerial vehicle automatic driving inspection method for blades of wind turbine generator | |
CN111709994B (en) | Autonomous unmanned aerial vehicle visual detection and guidance system and method | |
Claro et al. | ArTuga: A novel multimodal fiducial marker for aerial robotics | |
US20190276146A1 (en) | Method and system for capturing images of asset using unmanned aerial vehicles | |
KR20140030610A (en) | Surveillance method for using unmanned aerial vehicles and ground observation equipments | |
CN115580708A (en) | Unmanned aerial vehicle inspection method for optical cable line | |
CN209487090U (en) | A kind of unmanned plane lands marker automatically | |
CN112349227A (en) | Projection display system and method for aircraft takeoff and landing identification | |
US20200211403A1 (en) | Apparatus for presenting information to an airborne aircraft and method of using the same | |
Huang et al. | Vision-based autonomous landing of uav on moving platform using a new marker | |
CN104157105B (en) | Runway is boarded a plane the detection warning system of state |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |