Browse Topic: Cameras

Items (570)
The U-Shift IV represents the latest evolution in modular urban mobility solutions, offering significant advancements over its predecessors. This innovative vehicle concept introduces a distinct separation between the drive module, known as the driveboard, and the transport capsules. The driveboard contains all the necessary components for autonomous driving, allowing it to operate independently. This separation not only enables versatile applications - such as easily swapping capsules for passenger or goods transportation - but also significantly improves the utilization of the driveboard. By allowing a single driveboard to be paired with different capsules, operational efficiency is maximized, enabling continuous deployment of driveboards while the individual capsules are in use. The primary focus of U-Shift IV was to obtain a permit for operating at the Federal Garden Show 2023. To achieve this goal, we built the vehicle around the specific requirements for semi-public road
Pohl, EricScheibe, SebastianMünster, MarcoOsebek, ManuelKopp, GerhardSiefkes, Tjark
In order to comply with increasingly stringent emission regulations and ensure clean air, wall-flow particulate filters are predominantly used in exhaust gas aftertreatment systems of combustion engines to remove reactive soot and inert ash particles from exhaust gases. These filters consist of parallel porous channels with alternately closed ends, effectively separating particles by forming a layer on the filter surface. However, the accumulated particulate layer increases the pressure drop across the filter, requiring periodic filter regeneration. During regeneration, soot oxidation breaks up the particulate layer, while resuspension and transport of individual agglomerates can occur. These phenomena are influenced by gas temperature and velocity, as well as by the dispersity and reactivity of the soot particles. Renewable and biomass based fuels can produce different types of soot with different reactivities and dispersities. Therefore, this study focuses on the influences of soot
Desens, OleHagen, Fabian P.Meyer, JörgDittler, Achim
The video systems include a camera, display, and lights. Video is the recording, reproducing, or broadcasting of moving visual images as illustrated in Figure 1. A camera video imaging system is a system composed of a camera and a monitor, as well as other components, in which the monitor provides a real-time or near real-time visual image of the scene captured by the camera. Such systems are capable of providing remote views to the pilot and can therefore be used to provide improved visibility (for example, coverage of blind spots). In general, camera video systems may be used in the pilot’s work position for purposes of improving airplane and corresponding environmental visibility. Examples of aircraft video system applications include: Ground maneuver or taxi camera system Flight deck entry video surveillance system Cargo loading and unloading Cargo compartment livestock monitoring Monitoring systems that are used to track the external, internal, and security functions of an
A-20B Exterior Lighting Committee
With 2D cameras and space robotics algorithms, astronautics engineers at Stanford have created a navigation system able to manage multiple satellites using visual data only. They recently tested it in space for the first time. Stanford University, Stanford, CA Someday, instead of large, expensive individual space satellites, teams of smaller satellites - known by scientists as a “swarm” - will work in collaboration, enabling greater accuracy, agility, and autonomy. Among the scientists working to make these teams a reality are researchers at Stanford University's Space Rendezvous Lab, who recently completed the first-ever in-orbit test of a prototype system able to navigate a swarm of satellites using only visual information shared through a wireless network. “It's a milestone paper and the culmination of 11 years of effort by my lab, which was founded with this goal of surpassing the current state of the art and practice in distributed autonomy in space,” said Simone D'Amico
In October 2024, Kongsberg NanoAvionics discovered damage to their MP42 satellite, and used the discovery as an opportunity to raise awareness on the need to reduce space debris generated by satellites. Kongsberg NanoAvionics, Vilnius, Lithuania Our MP42 satellite, which launched into low Earth orbit (LEO) two and a half years ago aboard the SpaceX Transporter-4 mission, recently took an unexpected hit from a small piece of space debris or micrometeoroid. The impact created a 6 mm hole, roughly the size of a chickpea, in one of its solar panels. Despite this damage, the satellite continued performing its mission without interruption, and we only discovered the impact thanks to an image taken by its onboard selfie camera in October of 2024. It is challenging to pinpoint exactly when the impact occurred because MP42's last selfie was taken a year and a half ago, in April of 2023.
Design verification and quality control of automotive components require the analysis of the source location of ultra-short sound events, for instance the engaging event of an electromechanical clutch or the clicking noise of the aluminium frame of a passenger car seat under vibration. State-of-the-art acoustic cameras allow for a frame rate of about 100 acoustic images per second. Considering that most of the sound events introduced above can be far less than 10ms, an acoustic image generated at this rate resembles an hard-to-interpret overlay of multiple sources on the structure under test along with reflections from the surrounding test environment. This contribution introduces a novel method for visualizing impulse-like sound emissions from automotive components at 10x the frame rate of traditional acoustic cameras. A time resolution of less than 1ms eventually allows for the true localization of the initial and subsequent sound events as well as a clear separation of direct from
Rittenschober, Thomas
In active noise control, the control region size (same meaning as zone of control) decreases as the frequency increases, so that even a small moving of the passenger's head causes the ear position to go out of the control region. To increase the size of the control region, many speakers and microphones are generally required, but it is difficult to apply it in a vehicle cabin due to space and cost constraints. In this study, we propose moving zone of quiet active noise control technique. A 2D image-based head tracking system captured by a camera to generate the passenger's 0head coordinates in real time with deep learning algorithm. In the controller, the control position is moved to the ear position using a multi-point virtual microphone algorithm according to the generated ear position. After that, the multi-point adaptive filter training system applies the optimal control filter to the current position and maintains the control performance. Through this study, it is possible to
Oh, ChiSungKang, JonggyuKim, Joong-Kwan
This study presents a novel methodology for optimizing the acoustic performance of rotating machinery by combining scattered 3D sound intensity data with numerical simulations. The method is demonstrated on the rear axle of a truck. Using Scan&Paint 3D, sound intensity data is rapidly acquired over a large spatial area with the assistance of a 3D sound intensity probe and infrared stereo camera. The experimental data is then integrated into far-field radiation simulations, enabling detailed analysis of the acoustic behavior and accurate predictions of far-field sound radiation. This hybrid approach offers a significant advantage for assessing complex acoustic sources, allowing for quick and reliable evaluation of noise mitigation solutions.
Fernandez Comesana, DanielVael, GeorgesRobin, XavierOrselli, JosephSchmal, Jared
The segment manipulator machine, a large custom-built apparatus, is used for assembling and disassembling heavy tooling, specifically carbon fiber forms. This complex yet slow-moving machine had been in service for nineteen years, with many control components becoming obsolete and difficult to replace. The customer engaged Electroimpact to upgrade the machine using the latest state-of-the-art controls, aiming to extend the system's operational life by at least another two decades. The program from the previous control system could not be reused, necessitating a complete overhaul.
Luker, ZacharyDonahue, Michael
Industries that require high-accuracy automation in the creation of high-mix/low-volume parts, such as aerospace, often face cost constraints with traditional robotics and machine tools due to the need for many pre-programmed tool paths, dedicated part fixtures, and rigid production flow. This paper presents a new machine learning (ML) based vision mapping and planning technique, created to enhance flexibility and efficiency in robotic operations, while reducing overall costs. The system is capable of mapping discrete process targets in the robot work envelope that the ML algorithms have been trained to identify, without requiring knowledge of the overall assembly. Using a 2D camera, images are taken from multiple robot positions across the work area and are used in the ML algorithm to detect, identify, and predict the 6D pose of each target. The algorithm uses the poses and target identifications to automatically develop a part program with efficient tool paths, including
Langan, DanielHall, MichaelGoldberg, EmilySchrandt, Sasha
This paper explores the integration of two deep learning models that are currently being used for object detection, specifically Mask R-CNN and YOLOX, for two distinct driving environments: urban cityscapes and highway settings. The hypothesis underlying this work is that different methods of object detection will work best in different driving environments, due to the differences in their unique strengths as well as the key differences in those driving environments. Some of these differences in the driving environment include varying traffic densities, diverse object classes, and differing scene complexities, including specific differences such as the types of signs present, the presence or absence of stoplights, and the limited-access nature of highways as compared to city streets. As part of this work, a scene classifier has also been developed to categorize the driving context into the two categories of highway and urban driving, in order to allow the overall object detection
Patel, KrunalPeters, Diane
This study experimentally investigates the liquid jet breakup process in a vaporizer of a microturbine combustion chamber under equivalent operating conditions, including temperature and air mass flow rate. A high-speed camera experimental system, coupled with an image processing code, was developed to analyze the jet breakup length. The fuel jet is centrally positioned in a vaporizer with an inner diameter of 8mm. Airflow enters the vaporizer at controlled pressures, while thermal conditions are maintained between 298 K and 373 K using a PID-controlled heating system. The liquid is supplied through a jet with a 0.4 mm inner diameter, with a range of Reynolds numbers (Reliq = 2300÷3400), and aerodynamic Weber numbers (Weg = 4÷10), corresponding to the membrane and/or fiber breakup modes of the liquid jet. Based on the results of jet breakup length, a new model has been developed to complement flow regimes by low Weber and Reynolds numbers. The analysis of droplet size distribution
Ha, NguyenQuan, NguyenManh, VuPham, Phuong Xuan
Off-road vehicles are required to traverse a variety of pavement environments, including asphalt roads, dirt roads, sandy terrains, snowy landscapes, rocky paths, brick roads, and gravel roads, over extended periods while maintaining stable motion. Consequently, the precise identification of pavement types, road unevenness, and other environmental information is crucial for intelligent decision-making and planning, as well as for assessing traversability risks in the autonomous driving functions of off-road vehicles. Compared to traditional perception solutions such as LiDAR and monocular cameras, stereo vision offers advantages like a simple structure, wide field of view, and robust spatial perception. However, its accuracy and computational cost in estimating complex off-road terrain environments still require further optimization. To address this challenge, this paper proposes a terrain environment estimating method for off-road vehicle anticipated driving area based on stereo
Zhao, JianZhang, XutongHou, JieChen, ZhigangZheng, WenboGao, ShangZhu, BingChen, Zhicheng
Vehicle ADAS Systems majorly comprises of two functions: Driving and Parking. The most common form of damage to the vehicle which goes unnoticed with unidentified cause are parking damages. A vehicle once parked at a certain location may get damaged without knowledge of the user. In this work developed a solution that not only pre-warns the driver but also prepares the vehicle beforehand if it suspects a damage may occur. This eliminates the latency between damage and information capture, detects small damages such as scratches, classifies the type of damage and informs the user beforehand. This is solution is different from our competitors as the existing solutions informs the user about the scratches/damages, but these solutions are expensive, have high response time, and the damage information is captured after the damage has occurred. The solution consists of the following check blocks: Precondition, Sensor Control and Action Module. The Precondition Module observes the vehicle
Debnath, SarnabPatil, PrasadBelur Subramanya, SheshagiriGovinda, Shiva Prasad
Accurate reconstruction of vehicle collisions is essential for understanding incident dynamics and informing safety improvements. Traditionally, vehicle speed from dashcam footage has been approximated by estimating the time duration and distance traveled as the vehicle passes between reference objects. This method limits the resolution of the speed profile to an average speed over given intervals and reduces the ability to determine moments of acceleration or deceleration. A more detailed speed profile can be calculated by solving for the vehicle’s position in each video frame; however, this method is time-consuming and can introduce spatial and temporal error and is often constrained by the availability of external trackable features in the surrounding environment. Motion tracking software, widely used in the visual effects industry to track camera positions, has been adopted by some collision reconstructionists for determining vehicle speed from video. This study examines the
Perera, NishanGriffiths, HarrisonPrentice, Greg
Videos from cameras onboard a moving vehicle are increasingly available to collision reconstructionists. The goal of this study was to evaluate the accuracy of speeds, decelerations, and brake onset times calculated from onboard dash cameras (“dashcams”) using a match-moving technique. We equipped a single test vehicle with 5 commercially available dashcams, a 5th wheel, and a brake pedal switch to synchronize the cameras and 5th wheel. The 5th wheel data served as the reference for the vehicle kinematics. We conducted 9 tests involving a constant-speed approach (mean ± standard deviation = 57.6 ± 2.0 km/h) followed by hard braking (0.989 g ± 0.021 g). For each camera and brake test, we extracted the video and calculated the camera’s position in each frame using SynthEyes, a 3D motion tracking and video analysis program. Scale and location for the analyses were based on a 3D laser scan of the test site. From each camera’s position data, we calculated its speed before braking and its
Flynn, ThomasAhrens, MatthewYoung, ColeSiegmund, Gunter P.
Photogrammetry is a commonly used type of analysis in accident reconstruction. It allows the location of physical evidence, as shown in photographs and video, and the position and orientation of vehicles, other road users, and objects to be quantified. Lens distortion is an important consideration when using photogrammetry. Failure to account for lens distortion can result in inaccurate spatial measurements, particularly when elements of interest are located toward the edges and corners of images. Depending on whether the camera properties are known or unknown, various methods for removing lens distortion are commonly used in photogrammetric analysis. However, many of these methods assume that lens distortion is the result of a spherical lens or, more rarely, is solely due to distortion caused by other known lens types and has not been altered algorithmically by the camera. Today, several cameras on the market algorithmically alter images before saving them. These camera systems use
Pittman, KathleenMockensturm, EricBuckman, TaylorWhite, Kirsten
Vehicle-to-Infrastructure (V2I) cooperation has emerged as a fundamental technology to overcome the limitations of the individual ego-vehicle perception. Onboard perception is limited by the lack of information for understanding the environment, the lack of anticipation, the drop of performance due to occlusions and the physical limitations of embedded sensors. The perception of V2I in a cooperative manner improves the perception range of the ego vehicle by receiving information from the infrastructure that has another point of view, mounted with sensors, such as camera and LiDAR. This technical paper presents a perception pipeline developed for the infrastructure based on images with multiple viewpoints. It is designed to be scalable and has five main components: the image acquisition for the modification of camera settings and to get the pixel data, the object detection for fast and accurate detection of four wheels, two wheels and pedestrians, the data fusion module for robust
Picard, QuentinMorice, MaloFadili, MaryemPechberti, Steve
In this study, we introduce RGB2BEV-Net, an end-to-end pipeline that extends traditional BEV segmentation models by utilizing raw RGB images with Bird’s Eye View (BEV) generation. While previous work primarily focused on pre-segmented images to generate corresponding BEV maps, our approach expands this by collecting RGB images alongside their affiliated segmentation masks and BEV representations. This enables direct input of RGB camera sensors into the pipeline, reflecting real-world autonomous driving scenarios where RGB cameras are commonly used as sensors, rather than relying on pre-segmented images. Our model processes four RGB images through a segmentation layer before converting them into a segmented BEV, implemented in the PyTorch framework after being adapted from an original implementation that utilized a different framework. This adaptation was necessary to improve compatibility and ensure better integration of the entire system within autonomous vehicle applications. We
Hossain, SabirLin, Xianke
Shadow positions can be useful in determining the time of day that a photograph was taken and determining the position, size, and orientation of an object casting a shadow in a scene. Astronomical equations can predict the location of the sun relative to the earth, and therefore the position of shadows cast by objects, based on the location’s latitude and longitude as well as the date and time. 3D computer software includes these calculations as a part of their built-in sun systems. In this paper, the authors examine the sun system in the 3D modeling software 3ds Max to determine its accuracy for use in accident reconstruction. A parking lot was scanned using a FARO LiDAR scanner to create a point cloud of the environment. A camera was then set up on a tripod at the environment, and photographs were taken at various times throughout the day from the same location. This environment was 3D modeled in 3ds Max based on the point cloud, and the sun system in 3ds Max was configured using the
Barreiro, EvanErickson, MichaelSmith, ConnorCarter, NealHashemian, Alireza
Tesla Model 3 and Model Y vehicles come equipped with a standard dashcam feature with the ability to record video in multiple directions. Front, side, and rear views were readily available via direct USB download. Additional types of front and side views were indirectly available via privacy requests with Tesla. Prior research neither fully explored the four most readily available camera views across multiple vehicles nor field camera calibration techniques particularly useful for future software and hardware changes. Moving GPS instrumented vehicles were captured traveling approximately 7.2 kph to 20.4 kph across the front, side, and rear views available via direct USB download. Reverse project photogrammetry projects and video timing data successfully measured vehicle speeds with an average error of 2.45% across 25 tests. Previously researched front and rear camera calibration parameters were reaffirmed despite software changes, and additional parameters for the side cameras
Jorgensen, MichaelSwinford, ScottImada, KevinFarhat, Ali
Camera matching photogrammetry is widely used in the field of accident reconstruction for mapping accident scenes, modeling vehicle damage from post collision photographs, analyzing sight lines, and video tracking. A critical aspect of camera matching photogrammetry is determining the focal length and Field of View (FOV) of the photograph being analyzed. The intent of this research is to analyze the accuracy of the metadata reported focal length and FOV. The FOV from photographs captured by over 20 different cameras of various makes, models, sensor sizes, and focal lengths will be measured using a controlled and repeatable testing methodology. The difference in measured FOV versus reported FOV will be presented and analyzed. This research will provide analysts with a dataset showing the possible error in metadata reported FOV. Analysts should consider the metadata reported FOV as a starting point for photogrammetric analysis and understand that the FOV calculated from the image
Smith, Connor A.Erickson, MichaelHashemian, Alireza
This paper introduces a method to solve the instantaneous speed and acceleration of a vehicle from one or more sources of video evidence by using optimization to determine the best fit speed profile that tracks the measured path of a vehicle through a scene. Mathematical optimization is the process of seeking the variables that drive an objective function to some optimal value, usually a minimum, subject to constraints on the variables. In the video analysis problem, the analyst is seeking a speed profile that tracks measured vehicle positions over time. Measured positions and observations in the video constrain the vehicle’s motion and can be used to determine the vehicle’s instantaneous speed and acceleration. The variables are the vehicle’s initial speed and an unknown number of periods of approximately constant acceleration. Optimization can be used to determine the speed profile that minimizes the total error between the vehicle’s calculated distance traveled at each measured
Snyder, SeanCallahan, MichaelWilhelm, ChristopherJohnk, ChrisLowi, AlvinBretting, Gerald
Video analysis plays a major role in many forensic fields. Many articles, publications, and presentations have covered the importance and difficulty in properly establishing frame timing. In many cases, the analyst is given video files that do not contain native metadata. In other cases, the files contain video recordings of the surveillance playback monitor which eliminates all original metadata from the video recording. These “video of video” recordings prevent an analyst from determining frame timing using metadata from the original file. However, within many of these video files, timestamp information is visually imprinted onto each frame. Analyses that rely on timing of events captured in video may benefit from these imprinted timestamps, but for forensic purposes, it is important to establish the accuracy and reliability of these timestamps. The purpose of this research is to examine the accuracy of these timestamps and to establish if they can be used to determine the timing
Molnar, BenjaminTerpstra, TobyVoitel, Tilo
Dash cameras (dashcams) can provide collision reconstructionists with quantifiable vehicle position and speed estimates. These estimates are achieved by tracking 2D video features with camera-tracking software to solve for the time history of camera position, and speed can then be calculated from the position-time history. Not all scenes have the same geometric features in quality or abundance. In this study, we compared the vehicle position and derived-speed estimates from dashcam video for different numbers and spatial distributions of tracked features that mimicked the continuum between barren environments and feature-rich environments. We used video from a dashcam mounted in a vehicle undergoing straight-line emergency braking. The surrounding environment had abundant trackable features on both sides of the road, including road markings, streetlights, signs, trees, and buildings. We first created a reference solution using SynthEyes, a 3D camera- and object-tracking program, and
Young, ColeAhrens, MatthewFlynn, ThomasSiegmund, Gunter P.
This study investigates the ignitability of hydrogen in an optical heavy-duty SI engine. While the ignition energy of hydrogen is exceptionally low, the high load and lean mixtures used in heavy-duty hydrogen engines lead to a high gas density, resulting in a much higher breakdown voltage than in light-duty SI engines. Spark plug wear is a concern, so there is a need to minimise the spark energy while maintaining combustion stability, even at challenging conditions for ignition. This work consists of a two-stage experimental study performed in an optical engine. In the first part, we mapped the combustion stability and frequency of misfires with two different ignition systems: a DC inductive discharge ignition system, and a closed-loop controlled capacitive AC system. The equivalence ratio and dwell time were varied for the inductive system while the capacitive system instead varied spark duration and spark current in addition to equivalence ratio. A key finding was that spark energy
Hallstadius, PeterSaha, AnupamSridhara, AravindAndersson, Öivind
The current leading experimental platform for engine visualization research is the optical engine, which features transparent window components classified into two types: partially visible windows and fully visible windows. Due to structural limitations, fully visible windows cannot be employed under certain complex or extreme operating conditions, leading to the acquisition of only local in-cylinder combustion images and resulting in information loss. This study introduces a method for reconstructing in-cylinder combustion images from local images using deep learning techniques. The experiments were conducted using an optical engine specifically designed for spark-ignition combustion modes, capturing in-cylinder flame images under various conditions with high-speed cameras. The primary focus was on reconstructing the flame edge, with in-cylinder combustion images categorized into three types: images where the flame edge is fully within the partially visible window, partly within the
Wang, MianhengZhang, YixiaoDu, HaoyuXiao, MaMao, JianshuFang, Yuwen
Deliberate modifications to infrastructure can significantly enhance machine vision recognition of road sections designed for Vulnerable Road Users, such as green bike lanes. This study evaluates how green bike lanes, compared to unpainted lanes, enhance machine vision recognition and vulnerable road users safety by keeping vehicles at a safe distance and preventing encroachment into designated bike lanes. Conducted at the American Center for Mobility, this study utilizes a vehicle equipped with a front-facing camera to assess green bike lane recognition capabilities across various environmental conditions including dry daytime, dry nighttime, rain, fog, and snow. Data collection involved gathering a comprehensive dataset under diverse conditions and generating masks for lane markings to perform comparative analysis for training Advanced Driver Assistance Systems. Quality measurement and statistical analysis are used to evaluate the effectiveness of machine vision recognition using
Ponnuru, Venkata Naga RithikaDas, SushantaGrant, JosephNaber, JeffreyBahramgiri, Mojtaba
This study outlines a camera-based perspective transformation method for measuring driver direct visibility, which produces 360-degree view maps of the nearest visible ground points. This method is ideal for field data collection due to its portability and minimal space requirements. Compared with ground truth assessments using a physical grid, this method was found to have a high level of accuracy, with all points in the vehicle front varying less than 0.30 m and varying less than 0.6 m for the A- and B-pillars. Points out of the rear window varied up to 2.4 m and were highly sensitive to differences in the chosen pixel due to their greater distance from the camera. Repeatability through trials of multiple measurements per vehicle and reproducibility through measures from multiple data collectors produced highly similar results, with the greatest variations ranging from 0.19 to 1.38 m. Additionally, three different camera lenses were evaluated, resulting in comparable results within
Mueller, BeckyBragg, HadenBird, Teddy
This paper presents advanced intelligent monitoring methods aimed at enhancing the quality and durability of asphalt pavement construction. The study focuses on two critical tasks: foreign object detection and the uniform application of tack coat oil. For object recognition, the YOLOv5 algorithm is employed, which provides real-time detection capabilities essential for construction environments where timely decisions are crucial. A meticulously annotated dataset comprising 4,108 images, created with the LabelImg tool, ensures the accurate detection of foreign objects such as leaves and cigarette butts. By utilizing pre-trained weights during model training, the research achieved significant improvements in key performance metrics, including precision and recall rates. In addition to object detection, the study explores color space analysis through the HSV (Hue, Saturation, Value) model to effectively differentiate between coated and uncoated pavement areas following the application of
Hu, YufanFan, JianweiTang, FanlongMa, Tao
Vehicle localization in enclosed environments, such as indoor parking lots, tunnels, and confined areas, presents significant challenges and has garnered considerable research interest. This paper proposes a localization technique based on an onboard binocular camera system, utilizing binocular ranging and spatial intersection algorithms to achieve active localization. The method involves pre-deploying reference points with known coordinates within the experimental space, using binocular ranging to measure the distance between the camera and the reference points, and applying the spatial intersection algorithm to calculate the camera’s center coordinates, thereby completing the localization process. Experimental results demonstrate that the proposed algorithm achieves sub-meter level localization accuracy. Localization accuracy is significantly influenced by the calibration precision of the binocular camera and the number of reference points. Higher calibration precision and a greater
Feifei, LiHaoping, QiYi, Wei
The modern-day vehicle’s driverless or driver-assisted systems are developed by sensing the surroundings using a combination of camera, lidar, and other related sensors by forming an accurate perception of the driving environment. Machine learning algorithms help in forming perception and perform planning and control of the vehicle. The control of the vehicle which reflects safety depends on the accurate understanding of the surroundings by the trained machine learning models by subdividing a camera image fed into multiple segments or objects. The semantic segmentation system comes with the objective of assigning predefined class labels such as tree, road, and the like to each pixel of an image. Any security attacks on pixel classification nodes of the segmentation systems based on deep learning result in the failure of the driver assistance or autonomous vehicle safety functionalities due to a falsely formed perception. The security compromisations on the pixel classification head of
Prashanth, K.Y.Rohitha , U.M.
Human-wildlife conflicts pose significant challenges to both conservation efforts and community well-being. As these conflicts escalate globally, innovative technologies become imperative for effective and humane management strategies. This paper presents an integrated autonomous drone solution designed to mitigate human-wildlife conflicts by leveraging technologies in drone surveillance and artificial intelligence. The proposed system consists of stationary IR cameras that are setup within the conflict prone areas, which utilizes machine learning to identify the presence of wild animals and to send the corresponding location to a drone docking station. An autonomous drone equipped with high-resolution IR cameras and sensors is deployed from the docking station to the provided location. The drone camera utilizes object detection technology to scan the specified zone to detect the animal and emit animal repelling ultrasonic sound from a device integrated to the drone to achieve non
Sadanandan, VaishnavSadique, AnwarGeorge, Angeo PradeepVinod, VishalRaveendran, Darshan Unni
Seoul National University College of Engineering announced that researchers from the Department of Electrical and Computer Engineering’s Optical Engineering and Quantum Electronics Laboratory have developed an optical design technology that dramatically reduces the volume of cameras with a folded lens system utilizing “metasurfaces,” a next-generation nano-optical device. By arranging metasurfaces on the glass substrate so that light can be reflected and moved around in the glass substrate in a folded manner, the researchers have realized a lens system with a thickness of 0.7 mm, which is much thinner than existing refractive lens systems. The research, which was supported by the Samsung Future Technology Development Program and the Institute of Information & Communications Technology Planning & Evaluation (IITP), was published on October 30 in the journal Science Advances. Traditional cameras are designed to stack multiple glass lenses to refract light when capturing images. While
A team led by University of Maryland computer scientists invented a camera mechanism that improves how robots see and react to the world around them. Inspired by how the human eye works, their innovative camera system mimics the tiny involuntary movements used by the eye to maintain clear and stable vision over time. The team’s prototyping and testing of the camera — called the Artificial Microsaccade-Enhanced Event Camera (AMI-EV) — was detailed in a paper published in the journal Science Robotics in May 2024.
Sometimes, we try to capture a QR code with a good digital camera on a smartphone, but the reading eventually fails. This usually happens when the QR code itself is of poor image quality, or if it has been printed on surfaces that are not flat — deformed or with irregularities of unknown pattern — such as the wrapping of a courier package or a tray of prepared food. Now, a team from the University of Barcelona (UB) and the Universitat Oberta de Catalunya (UOC) has designed a methodology that facilitates the recognition of QR codes in these physical environments, where reading is more complicated.
The flow structure and unsteadiness of shock wave–boundary layer interaction (SWBLI) has been studied using rainbow schlieren deflectometry (RSD), ensemble averaging, fast Fourier transform (FFT), and snapshot proper orthogonal decomposition (POD) techniques. Shockwaves were generated in a test section by subjecting a Mach = 3.1 free-stream flow to a 12° isosceles triangular prism. The RSD pictures captured with a high-speed camera at 5000 frames/s rate were used to determine the transverse ray deflections at each pixel of the pictures. The interaction region structure is described statistically with the ensemble average and root mean square deflections. The FFT technique was used to determine the frequency content of the flow field. Results indicate that dominant frequencies were in the range of 400 Hz–900 Hz. The Strouhal numbers calculated using the RSD data were in the range of 0.025–0.07. The snapshot POD technique was employed to analyze flow structures and their associated
Datta, NarendraOlcmen, SemihKolhe, Pankaj
Cooperative perception has attracted wide attention given its capability to leverage shared information across connected automated vehicles (CAVs) and smart infrastructure to address the occlusion and sensing range limitation issues. To date, existing research is mainly focused on prototyping cooperative perception using only one type of sensor such as LiDAR and camera. In such cases, the performance of cooperative perception is constrained by individual sensor limitations. To exploit the multi-modality of sensors to further improve distant object detection accuracy, in this paper, we propose a unified multi-modal multi-agent cooperative perception framework that integrates camera and LiDAR data to enhance perception performance in intelligent transportation systems. By leveraging the complementary strengths of LiDAR and camera sensors, our framework utilizes the geometry information from LiDAR and the semantic information from cameras to achieve an accurate cooperative perception
Meng, ZonglinXia, XinZheng, ZhaoliangGao, LetianLiu, WeiZhu, JiaqiMa, Jiaqi
Cameras are crucial sensors in intelligent driving systems. Due to the optical windows of these cameras generally being exposed, they are highly susceptible to contaminant from external dust, mud, and other contaminants. These contaminants can degrade the vehicle’s perception capabilities, posing safety risks. Therefore, research on the identification and automatic cleaning of optical window surface contamination for automotive cameras is essential. This paper constructs a dataset of contaminated images of automotive cameras using a method based on shooting and image fusion. By introducing the SE attention mechanism and replacing the YOLOv8 backbone network with FasterNet, this paper proposed the SEFaster-YOLOv8 model. Experimental results show that the SEFaster-YOLOv8 model reduces the parameter count by 37.6% compared to the original YOLOv8 model. The mAP@0.5 and mAP@0.5:0.95 reach 95.7% and 66.9%, respectively, representing improvements of 1.8% and 1.1% over the original YOLOv8
Ran, LujiaHu, ZongjieLu, XiangxiangWu, Zhijun
Recently, four-dimensional (4D) radar has shown unique advantages in the field of odometry estimation due to its low cost, all-weather use, and dynamic and static recognition. These features complement the performance of monocular cameras, which provide rich information but are easily affected by lighting. However, the construction of deep radar visual odometry faces the following challenges: (1) the 4D radar point cloud is very sparse; (2) due to the penetration ability of 4D radar, it will produce mismatches with pixels when projected onto the image plane. In order to enrich the point cloud information and improve the accuracy of modal correspondence, this paper proposes a low-cost fusion odometry method based on 4D radar and pseudo-LiDAR, 4DRPLO-Net. This method proposes a new framework that uses 4D radar points and pseudo-LiDAR points generated by images to construct odometry, bridging the gap between 4D radar and images in three-dimensional (3D) space. Specifically, the pseudo
Huang, MinqingLu, ShouyiZhuo, Guirong
This project presents the development of an advanced Autonomous Mobile Robot (AMR) designed to autonomously lift and maneuver four-wheel drive vehicles into parking spaces without human intervention. By leveraging cutting-edge camera and sensor technologies, the AMR integrates LIDAR for precise distance measurements and obstacle detection, high-resolution cameras for capturing detailed images of the parking environment, and object recognition algorithms for accurately identifying and selecting available parking spaces. These integrated technologies enable the AMR to navigate complex parking lots, optimize space utilization, and provide seamless automated parking. The AMR autonomously detects free parking spaces, lifts the vehicle, and parks it with high precision, making the entire parking process autonomous and highly efficient. This project pushes the boundaries of autonomous vehicle technology, aiming to contribute significantly to smarter and more efficient urban mobility systems.
Atheef, M. SyedSundar, K. ShamKumar, P. P. PremKarthika, J.
Object detection (OD) is one of the most important aspects in Autonomous Driving (AD) application. This depends on the strategic sensor’s selection and placement of sensors around the vehicle. The sensors should be selected based on various constraints such as range, use-case, and cost limitation. This paper introduces a systematic approach for identifying the optimal practices for selecting sensors in AD object detection, offering guidance for those looking to expand their expertise in this field and select the most suitable sensors accordingly. In general, object detection typically involves utilizing RADAR, LiDAR, and cameras. RADAR excels in accurately measuring longitudinal distances over both long and short ranges, but its accuracy in lateral distances is limited. LiDAR is known for its ability to provide accurate range data, but it struggles to identify objects in various weather conditions. On the other hand, camera-based systems offer superior recognition capabilities but lack
Maktedar, AsrarulhaqChatterjee, Mayurika
Researchers led by Professor Young Min Song from the Gwangju Institute of Science and Technology (GIST) have unveiled a vision system inspired by feline eyes to enhance object detection in various lighting conditions. Featuring a unique shape and reflective surface, the system reduces glare in bright environments and boosts sensitivity in low-light scenarios. By filtering unnecessary details, this technology significantly improves the performance of single-lens cameras, representing a notable advancement in robotic vision capabilities.
There are certain situations when landing an Advanced Air Mobility (AAM) aircraft is required to be performed without assistance from GPS data. For example, AAM aircraft flying in an urban environment with tall buildings and narrow canyons may affect the ability of the AAM aircraft to effectively use GPS to access a landing area. Incorporating a vision-based navigation method, NASA Ames has developed a novel Alternative Position, Navigation, and Timing (APNT) solution for AAM aircraft in environments where GPS is not available.
Object detection is one of the core tasks in autonomous driving perception systems. Most perception algorithms commonly use cameras and LiDAR sensors, but the robustness is insufficient in harsh environments such as heavy rain and fog. Moreover, velocity of objects is crucial for identifying motion states. The next generation of 4D millimeter-wave radar retains traditional radar advantages in robustness and speed measurement, while also providing height information, higher resolution and density. 4D radar has great potential in the field of 3D object detection. However, existing methods overlook the need for specific feature extraction modules for 4D millimeter-wave radar, which can lead to potential information loss. In this study, we propose RadarPillarDet, a novel approach for extracting features from 4D radar to achieve high-quality object detection. Specifically, our method introduces a dual-stream encoder (DSE) module, which combines traditional multilayer perceptron and
Yang, LongZheng, LianqingMo, JingyueBai, JieZhu, XichanMa, Zhixiong
In non-cooperative environments, unmanned aerial vehicles (UAVs) have to land without artificial markers, which is a key step towards achieving full autonomy. However, the existing vision-based schemes have the common problems of poor robustness and generalization, and the LiDAR-based schemes have the disadvantages of low resolution, high power consumption and high weight. In this paper, we propose an UAV landing system equipped with a binocular camera to preform 3D reconstruction and select the safe landing zone. The whole system only consists of a stereo camera, and the innovation of the solution is fusing the stereo matching algorithm and monocular depth estimation(MDE) model to get a robust prediction on the metric depth. The whole landing system consists of a stereo matching module, a monocular depth estimation (MDE) module, a depth fusion module, and a safe landing zone selection module. The stereo matching module uses Semi-Global Matching (SGM) algorithm to calculate the
Zhou, YiBiaoZhang, BiHui
With the rapid advancement in unmanned aerial vehicle (UAV) technology, the demand for stable and high-precision electro-optical (EO) pods, such as cameras, lidar sensors, and infrared imaging systems, has significantly increased. However, the inherent vibrations generated by the UAV’s propulsion system and aerodynamic disturbances pose significant challenges to the stability and accuracy of these payloads. To address this issue, this paper presents a study on the application of high-static low-dynamic stiffness (HSLDS) vibration isolation devices in EO payloads mounted on UAVs. The HSLDS system is designed to effectively isolate low-frequency and high-amplitude vibrations while maintaining high static stiffness, ensuring both stability during hovering and precise pointing capabilities. A nonlinear dynamic system model with two degrees of freedom is formulated for an EO pod supported by HSLDS isolators at both ends. The model’s natural frequencies are determined, and approximate
Tian, YishenGuo, GaofengWang, GuangzhaoWei, WanBao, LingcongDong, GuanLi, Liujie
Items per page:
1 – 50 of 570