Projects

Here, we would like to take the opportunity to showcase some of our reference projects:


IoT connectivity, Cloud Platform, Function Development

NISAR and e:fs join forces:
Combining new ideas from the startup ecosystem with extensive experience in series development and validation we strive to change the way how automotive software is developed today.

– Test car for data recording, remote connectivity
– Cloud platform
– IoT connectivity for car-pc to cloud platform


Traffic Scenario Recorder based on Ring-Buffer

NISAR and e:fs join forces:
Combining new ideas from the startup ecosystem with extensive experience in series development and validation we strive to change the way how automotive software is developed today.

-Sparse Scenario Recording
-Do not record hundreds of hours and km, just record what is relevant
-Store vehicle sensor and bus data in ring buffer
-Store ring buffer content based on triggers, i.e. sharp braking


ECU remote monitoring & flashing on production line

Monitoring & flashing ECU already on the production line

Software for ECUs is increasing rapidly in size. And so is the flashing time needed at End-of-line flashing stations.

With our IoT connector the ECU is already flashed while on the line, saving time at EOL.


XDP Showcase with TogetherOS

Member of the Horizon Ecosystem

Code Generation for Horizon TogetherOS from Software Architecture i.e. for a camera fusion system.

Built into NISAR Architect Design Studio


XDP Showcase with AUTOSAR

Phase 1:
Integrating Nisar ADS with ETAS AUTOSAR Toolchain, enabling the generation of AUTOSAR code from abstract architectures

Phase 2:
Enabling the bridge from pre-development in ROS to AUTOSAR via the abstract architecture


Multi Radar Sensor Fusion

Creation of a central 360° environmental model from several corner radars in ROS

Two variants:

1.Free-Space model using occupancy grid maps

2.Geometrical obstacle representation using bounding polygons


Surround View and Parking Assistant

Creation of a central 360° environmental model from several cameras

Three steps:
1. Bird-eye view for the driver by merging 6 cameras
2. Detecting of free and occupied parking spaces
3. Dubbin / Reed based trajectory planner

Simulation and real vehicle


Radar-Camera-Fusion & Object-Tracking

Car equipped with radar and cameras, each with their own object tracking
• Fusion of the tracks
• Increased accuracy
• Increased robustness
• Tracking beyond the limits of the sensors’ fields of view
• Tracking of lanes and assignment of objects
• Planning the navigable space taking into account the objects’ own movement and prediction


AI-based Cyber Security

AI-based design wizards to enable integrating Cyber Security already in early design phases while at the same time easing the task for Cyber Security Engineers

Published at AI.BAY conference, 2023, with TH Ingolstadt CARISSMA, D.Bayerl


Our test vehicle in Shanghai

Test Vehicle Platform
– MG4 Mulan

Sensor Setup
– 7 Cameras – 2-8M Pixels
– 4 Fisheye cameras
– 3 LRR-Radars

Software Components integrated
– BEV Model for Perception
– HD Map Fusion
– Radar + Camera Object Fusion
– NoA Navigation on Autopilot
– Middleware: TROS, ROS and Autosar

Current Hardware Platforms
– HR J5 x 2 256 TOPs + Semidrive X9U
– HR J5 x 1 + NXP SR32G
– NVIDIA Orin + TC397
– HR J3 x 1 3 TOPs + Aurix


…additional impressions…