{"id":3543,"date":"2022-03-23T10:00:34","date_gmt":"2022-03-23T10:00:34","guid":{"rendered":"https:\/\/autowise.ai\/?p=3543"},"modified":"2022-06-17T10:57:34","modified_gmt":"2022-06-17T10:57:34","slug":"autowise-cvpr-paper","status":"publish","type":"post","link":"https:\/\/autowise.ai\/en\/2022\/03\/23\/autowise-cvpr-paper\/","title":{"rendered":"CVPR 2022 | Latest research results published by Autowise.ai and Zhang Tong"},"content":{"rendered":"[vc_row type=&#8221;full_width_content&#8221; full_screen_row_position=&#8221;middle&#8221; column_margin=&#8221;default&#8221; column_direction=&#8221;default&#8221; column_direction_tablet=&#8221;default&#8221; column_direction_phone=&#8221;default&#8221; scene_position=&#8221;center&#8221; top_padding=&#8221;80px&#8221; left_padding_desktop=&#8221;15%&#8221; constrain_group_2=&#8221;yes&#8221; right_padding_desktop=&#8221;15%&#8221; top_padding_tablet=&#8221;60px&#8221; left_padding_tablet=&#8221;5%&#8221; constrain_group_4=&#8221;yes&#8221; right_padding_tablet=&#8221;5%&#8221; top_padding_phone=&#8221;30px&#8221; left_padding_phone=&#8221;20px&#8221; constrain_group_6=&#8221;yes&#8221; right_padding_phone=&#8221;20px&#8221; text_color=&#8221;dark&#8221; text_align=&#8221;left&#8221; row_border_radius=&#8221;none&#8221; row_border_radius_applies=&#8221;bg&#8221; overflow=&#8221;visible&#8221; advanced_gradient_angle=&#8221;0&#8243; overlay_strength=&#8221;0.3&#8243; gradient_direction=&#8221;left_to_right&#8221; shape_divider_position=&#8221;bottom&#8221; bg_image_animation=&#8221;none&#8221; gradient_type=&#8221;default&#8221; shape_type=&#8221;&#8221;][vc_column column_padding=&#8221;no-extra-padding&#8221; column_padding_tablet=&#8221;inherit&#8221; column_padding_phone=&#8221;inherit&#8221; column_padding_position=&#8221;all&#8221; left_margin=&#8221;auto&#8221; constrain_group_2=&#8221;yes&#8221; right_margin=&#8221;auto&#8221; column_element_spacing=&#8221;default&#8221; background_color_opacity=&#8221;1&#8243; background_hover_color_opacity=&#8221;1&#8243; column_shadow=&#8221;none&#8221; column_border_radius=&#8221;none&#8221; column_link_target=&#8221;_self&#8221; max_width_desktop=&#8221;1000px&#8221; column_position=&#8221;default&#8221; advanced_gradient_angle=&#8221;0&#8243; gradient_direction=&#8221;left_to_right&#8221; overlay_strength=&#8221;0.3&#8243; width=&#8221;1\/1&#8243; tablet_width_inherit=&#8221;default&#8221; tablet_text_alignment=&#8221;default&#8221; phone_text_alignment=&#8221;default&#8221; animation_type=&#8221;default&#8221; bg_image_animation=&#8221;none&#8221; border_type=&#8221;simple&#8221; column_border_width=&#8221;none&#8221; column_border_style=&#8221;solid&#8221; gradient_type=&#8221;default&#8221;][vc_column_text css_animation=&#8221;fadeInUp&#8221; max_width=&#8221;1000px&#8221;]On March 1, the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2022, announced the list of accepted papers. The research paper <em><i>Exploring <\/i><\/em><em><i>Geometry Consistency <\/i><\/em><em><i>for <\/i><\/em><em><i>Monocular <\/i><\/em><em><i>3D <\/i><\/em><em><i>Object Detection<\/i><\/em> by the Autowise.ai-Hong Kong University of Science and Technology Joint Laboratory was accepted at CVPR2022.[\/vc_column_text][image_with_animation image_url=&#8221;3553&#8243; image_size=&#8221;full&#8221; animation_type=&#8221;entrance&#8221; animation=&#8221;Fade In&#8221; hover_animation=&#8221;none&#8221; alignment=&#8221;center&#8221; border_radius=&#8221;none&#8221; box_shadow=&#8221;none&#8221; image_loading=&#8221;default&#8221; max_width=&#8221;100%&#8221; max_width_mobile=&#8221;default&#8221; margin_bottom=&#8221;20px&#8221;][vc_column_text css_animation=&#8221;fadeInUp&#8221; max_width=&#8221;1000px&#8221;]The work focuses on monocular 3D object detection which aims to detect 3D obstacles through only 2D monocular images. With the difficulty in inferring accurate depth from images, monocular 3D object detection task is an ill-posed and challenging task. To start with, the research analyzes how existing methods use visual cues to locate obstacles, and then proposes some data augmentation methods to enhance robustness.[\/vc_column_text][image_with_animation image_url=&#8221;3569&#8243; image_size=&#8221;full&#8221; animation_type=&#8221;entrance&#8221; animation=&#8221;Fade In&#8221; hover_animation=&#8221;none&#8221; alignment=&#8221;center&#8221; border_radius=&#8221;none&#8221; box_shadow=&#8221;none&#8221; image_loading=&#8221;default&#8221; max_width=&#8221;100%&#8221; max_width_mobile=&#8221;default&#8221; margin_bottom=&#8221;20px&#8221;][vc_column_text css_animation=&#8221;fadeInUp&#8221; max_width=&#8221;1000px&#8221;]As seen in the illustration above, visual cues that can be used by neural networks include the apparent size of the object in the image and its vertical position in the image (the closer the object is, the larger it is in the image, and the lower it appears \u00a0in the image). By apllying various perturbation to images, researchers found that the neural network tends to use the object size information to predict depth, but the network is not robust enough to use this information (as shown in the illustration below):[\/vc_column_text][image_with_animation image_url=&#8221;3562&#8243; image_size=&#8221;full&#8221; animation_type=&#8221;entrance&#8221; animation=&#8221;Fade In&#8221; hover_animation=&#8221;none&#8221; alignment=&#8221;center&#8221; border_radius=&#8221;none&#8221; box_shadow=&#8221;none&#8221; image_loading=&#8221;default&#8221; max_width=&#8221;100%&#8221; max_width_mobile=&#8221;default&#8221; margin_bottom=&#8221;20px&#8221;][vc_column_text css_animation=&#8221;fadeInUp&#8221; max_width=&#8221;1000px&#8221;]Motivated by the analysis, the researchers designed four different levels of the data augmentation methods to generate additional training data. The training data can effectively enhance neural network robustness to relevant visual cues, by ensuring the geometry consistency before and after image perturbation.To the best of our knowledge, no related data augmentation methods has been proposed before in the monocular 3D object detection field. Extensive experiments conducted on Kitti and nuScene datasets show the effectiveness of the proposed data augmentation methods.[\/vc_column_text][image_with_animation image_size=&#8221;full&#8221; animation_type=&#8221;entrance&#8221; animation=&#8221;Fade In&#8221; hover_animation=&#8221;none&#8221; alignment=&#8221;&#8221; border_radius=&#8221;none&#8221; box_shadow=&#8221;none&#8221; image_loading=&#8221;default&#8221; max_width=&#8221;100%&#8221; max_width_mobile=&#8221;default&#8221;][\/vc_column][\/vc_row][vc_row type=&#8221;full_width_content&#8221; full_screen_row_position=&#8221;middle&#8221; column_margin=&#8221;default&#8221; column_direction=&#8221;default&#8221; column_direction_tablet=&#8221;default&#8221; column_direction_phone=&#8221;default&#8221; scene_position=&#8221;center&#8221; top_padding=&#8221;80px&#8221; left_padding_desktop=&#8221;15%&#8221; constrain_group_2=&#8221;yes&#8221; right_padding_desktop=&#8221;15%&#8221; top_padding_tablet=&#8221;60px&#8221; left_padding_tablet=&#8221;5%&#8221; constrain_group_4=&#8221;yes&#8221; right_padding_tablet=&#8221;5%&#8221; top_padding_phone=&#8221;30px&#8221; left_padding_phone=&#8221;20px&#8221; constrain_group_6=&#8221;yes&#8221; right_padding_phone=&#8221;20px&#8221; text_color=&#8221;dark&#8221; text_align=&#8221;left&#8221; row_border_radius=&#8221;none&#8221; row_border_radius_applies=&#8221;bg&#8221; overflow=&#8221;visible&#8221; advanced_gradient_angle=&#8221;0&#8243; overlay_strength=&#8221;0.3&#8243; gradient_direction=&#8221;left_to_right&#8221; shape_divider_position=&#8221;bottom&#8221; bg_image_animation=&#8221;none&#8221; gradient_type=&#8221;default&#8221; shape_type=&#8221;&#8221;][vc_column column_padding=&#8221;no-extra-padding&#8221; column_padding_tablet=&#8221;inherit&#8221; column_padding_phone=&#8221;inherit&#8221; column_padding_position=&#8221;all&#8221; left_margin=&#8221;auto&#8221; constrain_group_2=&#8221;yes&#8221; right_margin=&#8221;auto&#8221; column_element_spacing=&#8221;default&#8221; background_color_opacity=&#8221;1&#8243; background_hover_color_opacity=&#8221;1&#8243; column_shadow=&#8221;none&#8221; column_border_radius=&#8221;none&#8221; column_link_target=&#8221;_self&#8221; max_width_desktop=&#8221;1000px&#8221; column_position=&#8221;default&#8221; advanced_gradient_angle=&#8221;0&#8243; gradient_direction=&#8221;left_to_right&#8221; overlay_strength=&#8221;0.3&#8243; width=&#8221;1\/1&#8243; tablet_width_inherit=&#8221;default&#8221; tablet_text_alignment=&#8221;default&#8221; phone_text_alignment=&#8221;default&#8221; animation_type=&#8221;default&#8221; bg_image_animation=&#8221;none&#8221; border_type=&#8221;simple&#8221; column_border_width=&#8221;none&#8221; column_border_style=&#8221;solid&#8221; gradient_type=&#8221;default&#8221;][vc_column_text css_animation=&#8221;fadeInUp&#8221; max_width=&#8221;1000px&#8221;]The Autowise.ai-Hong Kong University of Science and Technology Joint Laboratory, which was co-founded by Autowise.ai and the Hong Kong University of Science and Technology, aims to give full play to both sides\u2019 advantages, jointly promote machine learning technology industrialization, and give rise to innovative applications to improve environmental health in an autonomous way. Professor Zhang Tong, an internationally renowned scholar in the field of machine-learning, serves as the laboratory head. Professor Zhang Tong is currently a chair professor in the Department of Mathematics and the Department of Computer Science at Hong Kong University of Science and Technology. Previously, he held the positions of chief scientist at the Yahoo Research Institute, vice president and big data laboratory head at the Baidu Research Institute, and lab head at Tencent AI. He is also an ASA fellow and an IEEE fellow, and has served as the chari or area-chair in major machine learning conferences such as NIPS, ICML, and COLT, and has been in the editorial boards of leading machine learning journals such as PAMI, JMLR, and Machine Learning Journal.[\/vc_column_text][image_with_animation image_url=&#8221;3550&#8243; image_size=&#8221;large&#8221; animation_type=&#8221;entrance&#8221; animation=&#8221;Fade In&#8221; hover_animation=&#8221;none&#8221; alignment=&#8221;center&#8221; border_radius=&#8221;none&#8221; box_shadow=&#8221;none&#8221; image_loading=&#8221;default&#8221; max_width=&#8221;100%&#8221; max_width_mobile=&#8221;default&#8221; margin_bottom=&#8221;20px&#8221; margin_bottom_tablet=&#8221;20px&#8221; el_class=&#8221;full_image&#8221;][vc_column_text css_animation=&#8221;fadeInUp&#8221; max_width=&#8221;1000px&#8221;]\n<p style=\"text-align: center;\"><em>Online meeting between Mr. Huang Chao, CEO of Autowise.ai, and Professor Zhang Tong, head of Joint Laboratory.<\/em><\/p>\n[\/vc_column_text][image_with_animation image_size=&#8221;full&#8221; animation_type=&#8221;entrance&#8221; animation=&#8221;Fade In&#8221; hover_animation=&#8221;none&#8221; alignment=&#8221;&#8221; border_radius=&#8221;none&#8221; box_shadow=&#8221;none&#8221; image_loading=&#8221;default&#8221; max_width=&#8221;100%&#8221; max_width_mobile=&#8221;default&#8221;][\/vc_column][\/vc_row][vc_row type=&#8221;full_width_content&#8221; full_screen_row_position=&#8221;middle&#8221; column_margin=&#8221;default&#8221; column_direction=&#8221;default&#8221; column_direction_tablet=&#8221;default&#8221; column_direction_phone=&#8221;default&#8221; scene_position=&#8221;center&#8221; top_padding=&#8221;80px&#8221; left_padding_desktop=&#8221;15%&#8221; constrain_group_2=&#8221;yes&#8221; right_padding_desktop=&#8221;15%&#8221; top_padding_tablet=&#8221;60px&#8221; left_padding_tablet=&#8221;5%&#8221; constrain_group_4=&#8221;yes&#8221; right_padding_tablet=&#8221;5%&#8221; top_padding_phone=&#8221;30px&#8221; left_padding_phone=&#8221;20px&#8221; constrain_group_6=&#8221;yes&#8221; right_padding_phone=&#8221;20px&#8221; text_color=&#8221;dark&#8221; text_align=&#8221;left&#8221; row_border_radius=&#8221;none&#8221; row_border_radius_applies=&#8221;bg&#8221; overflow=&#8221;visible&#8221; advanced_gradient_angle=&#8221;0&#8243; overlay_strength=&#8221;0.3&#8243; gradient_direction=&#8221;left_to_right&#8221; shape_divider_position=&#8221;bottom&#8221; bg_image_animation=&#8221;none&#8221; gradient_type=&#8221;default&#8221; shape_type=&#8221;&#8221;][vc_column column_padding=&#8221;no-extra-padding&#8221; column_padding_tablet=&#8221;inherit&#8221; column_padding_phone=&#8221;inherit&#8221; column_padding_position=&#8221;all&#8221; left_margin=&#8221;auto&#8221; constrain_group_2=&#8221;yes&#8221; right_margin=&#8221;auto&#8221; column_element_spacing=&#8221;default&#8221; background_color_opacity=&#8221;1&#8243; background_hover_color_opacity=&#8221;1&#8243; column_shadow=&#8221;none&#8221; column_border_radius=&#8221;none&#8221; column_link_target=&#8221;_self&#8221; max_width_desktop=&#8221;1000px&#8221; column_position=&#8221;default&#8221; advanced_gradient_angle=&#8221;0&#8243; gradient_direction=&#8221;left_to_right&#8221; overlay_strength=&#8221;0.3&#8243; width=&#8221;1\/1&#8243; tablet_width_inherit=&#8221;default&#8221; tablet_text_alignment=&#8221;default&#8221; phone_text_alignment=&#8221;default&#8221; animation_type=&#8221;default&#8221; bg_image_animation=&#8221;none&#8221; border_type=&#8221;simple&#8221; column_border_width=&#8221;none&#8221; column_border_style=&#8221;solid&#8221; gradient_type=&#8221;default&#8221;][vc_column_text css_animation=&#8221;fadeInUp&#8221; max_width=&#8221;1000px&#8221;]According to Professor Zhang Tong, \u201cmonocular camera plays an increasingly important role as\u00a0a perception component in autonomous driving, and in recent years, relevant research has gradually drawn extensive attention to it, from both academic communities and industries.\u201d Compared to using LIDAR, using monocular images to gain accurate depth information for 3D perception-based tasks lacks results, which makes these tasks extremely difficult. Based on the hypothesis that &#8220;the same obstacle\u2019s 3D geometric characteristics should be consistent between mutltiple-view images,&#8221; the Joint Laboratory took the lead in designing data augmentation methods for monocular 3D object detection, achieving significant improvement. Furthermore, the exploration into geometry consistency will be an important research direction to improve the perception performance of monocular images. To do this, both sides will expand relevant technologies to more application scenarios. For example, they will jointly explore innovative applications for geometry consistency in multi-view data and promote this academic research for industrial usage.[\/vc_column_text][image_with_animation image_size=&#8221;full&#8221; animation_type=&#8221;entrance&#8221; animation=&#8221;Fade In&#8221; hover_animation=&#8221;none&#8221; alignment=&#8221;&#8221; border_radius=&#8221;none&#8221; box_shadow=&#8221;none&#8221; image_loading=&#8221;default&#8221; max_width=&#8221;100%&#8221; max_width_mobile=&#8221;default&#8221;][\/vc_column][\/vc_row]\n","protected":false},"excerpt":{"rendered":"<p>[vc_row type=&#8221;full_width_content&#8221; full_screen_row_position=&#8221;middle&#8221; column_margin=&#8221;default&#8221; column_direction=&#8221;default&#8221; column_direction_tablet=&#8221;default&#8221; column_direction_phone=&#8221;default&#8221; scene_position=&#8221;center&#8221; top_padding=&#8221;80px&#8221; left_padding_desktop=&#8221;15%&#8221; constrain_group_2=&#8221;yes&#8221; right_padding_desktop=&#8221;15%&#8221; top_padding_tablet=&#8221;60px&#8221; left_padding_tablet=&#8221;5%&#8221; constrain_group_4=&#8221;yes&#8221; right_padding_tablet=&#8221;5%&#8221; top_padding_phone=&#8221;30px&#8221; left_padding_phone=&#8221;20px&#8221; constrain_group_6=&#8221;yes&#8221; right_padding_phone=&#8221;20px&#8221; text_color=&#8221;dark&#8221; text_align=&#8221;left&#8221; row_border_radius=&#8221;none&#8221; row_border_radius_applies=&#8221;bg&#8221; overflow=&#8221;visible&#8221; advanced_gradient_angle=&#8221;0&#8243; overlay_strength=&#8221;0.3&#8243; gradient_direction=&#8221;left_to_right&#8221; shape_divider_position=&#8221;bottom&#8221; bg_image_animation=&#8221;none&#8221;&#8230;<\/p>\n","protected":false},"author":2,"featured_media":3569,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[],"class_list":{"0":"post-3543","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-news"},"_links":{"self":[{"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/posts\/3543","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/comments?post=3543"}],"version-history":[{"count":8,"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/posts\/3543\/revisions"}],"predecessor-version":[{"id":4013,"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/posts\/3543\/revisions\/4013"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/media\/3569"}],"wp:attachment":[{"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/media?parent=3543"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/categories?post=3543"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/autowise.ai\/en\/wp-json\/wp\/v2\/tags?post=3543"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}