You are required to read and agree to the below before accessing a full-text version of an article in the IDE article repository.
The full-text document you are about to access is subject to national and international copyright laws. In most cases (but not necessarily all) the consequence is that personal use is allowed given that the copyright owner is duly acknowledged and respected. All other use (typically) require an explicit permission (often in writing) by the copyright owner.
For the reports in this repository we specifically note that
By accepting I agree to acknowledge and respect the rights of the copyright owner of the document I am about to access.
If you are in doubt, feel free to contact email@example.com
Hybrid systems, and related formalisms, have been successfully used to model Cyber-Physical Systems. However, mathematical models are always a simplification of the system they are meant to describe, and one must be aware of this mismatch, when using these models to analyze a system. In safety analysis it is acceptable to use over-approximations of the system behavior, indeed they are the bread and butter of counterexample guided abstraction refinement (CEGAR). We propose a notion of system behavior robust wrt arbitrary small over-approximations, and argue that it is particularly appropriate for safety analysis.