Browse Topic: Active safety systems
Artificial Intelligence has gained lot of traction and importance in the 21st century with use cases ranging from speech recognition, learning, planning, problem solving to search engines etc. Artificial Intelligence also has played a key role in the development of autonomous vehicles and robots ranging from perception, localization, decision to controls. Within the big AI umbrella there is machine learning which is all about using your computer to "learn" how to deal with problems without “programming". Deep learning is a branch of machine learning based on a set of algorithms that learn to represent the data directly from the input such as an image, text, Sound, etc. Within deep learning there are Convolutional Neural Networks and Recurrent Neural Networks (CNN/RNN). The study here used convolutional neural network approach to perform image/object recognition. Given that the objective of the autonomous or semi-autonomous vehicle is to promote safety and reduce number of accidents, it
With the surge in adoption of artificial intelligence (AI) in automotive systems, especially Advanced Driver Assistance Systems (ADAS) and autonomous vehicles (AV), comes an increase of AI-related incidents–several of which have ended in injuries and fatalities. These incidents all share a common deficiency: insufficient coverage towards safety, ethical, and/or legal requirements. Responsible AI (RAI) is an approach to developing AI-enabled systems that systematically take such requirements into account. Existing published international standards like ISO 21448:2022 (Safety of the Intended Functionality) and ISO 26262:2018 (Road Vehicles – Functional Safety) do offer some guidance in this regard but are far from being sufficient. Therefore, several technical standards are emerging concurrently to address various RAI-related challenges, including but not limited to ISO 8800 for the integration of AI in automotive systems, ISO/IEC TR 5469:2024 for the integration of AI in functional
A total of 148 tests were conducted to evaluate the Forward Collision Warning (FCW) and Automatic Emergency Braking (AEB) systems in five different Tesla Model 3 vehicles between model years 2018 and 2020 across four calendar years. These tests involved stationary vehicle targets, including a foam Stationary Vehicle Target (SVT), a Deformable Stationary Vehicle Target (DSVT), a live vehicle with brake lights, and a SoftCar360 designed for high-speed impact tests. The evaluations were conducted at speeds of 35, 50, 60, 65, 70, 75, and 80 miles per hour (mph) during both daytime and nighttime conditions and utilized early and medium FCW settings. These settings, part of Tesla's Collision Avoidance AssistTM, modify object detection alerts and the timing of visual and auditory warnings issued to drivers. The 2018 to 2020 vehicles initially utilized cameras, radar and ultrasonic sensors (USS) for object detection. Tesla updated their Autoilot software and detection algorithms to a vision
The integrated bracket is a plastic part that packages functional components such as the ADAS (Advanced Driver Assistance System) camera, rain light sensor, and the mounting provisions of the auto-dimming IRVM (Inner Rear View Mirror). This part is fixed on the windshield of an automobile using double-sided adhesive tapes and glue. ADAS, rain light sensors, and auto-dimming IRVM play an important part in the safety of the driver and everyone present in the automobile. This makes proper functioning of the integrated bracket very integral to occupant safety. Prior to this work, the following literature; Integrated Bracket for Rain Light Sensor/ADAS/Auto-Dimming IRVM with provision of mounting for Aesthetic Cover [1] outlines the design considerations and advantages of mounting several components on the same bracket. It follows the theme where the authors first define the components packaged on the integrated bracket and then the advantages of packaging multiple components on a single
Items per page:
50
1 – 50 of 1854