Universität Stuttgart

Permanent URI for this communityhttps://elib.uni-stuttgart.de/handle/11682/1

Browse

Search Results

Now showing 1 - 10 of 240
  • Thumbnail Image
    ItemOpen Access
    Anonymisierung von Daten : von der Literatur zum Automobilbereich
    (2023) Herkommer, Jan
    Die Datenanonymisierung im Automobilbereich gewinnt immer mehr an Bedeutung. Jedoch gibt es kaum Literatur und Ansätze, die sich mit der Anonymisierung von Automobildaten beschäftigen. In dieser Arbeit werden deshalb mit Hilfe einer strukturierten Literaturrecherche die aktuell verbreitetsten Verfahren und Anwendungsbereiche erörtert und die wichtigsten Erkenntnisse der Recherche zusammengefasst. So werden bei den analysierten Paper der Anwendungsbereich, die Methodik sowie der zu anonymisierende Datentyp ermittelt. DesWeiteren werden die Metriken zum Vergleich von unterschiedlichen Ansätzen betrachtet. Mit Hilfe dieser Erkenntnisse wird im Anschluss auf die Anonymisierung von Fahrzeugdaten anhand verschiedener Anwendungsfälle eingegangen und Herausforderungen und Lösungsansätze skizziert. Zuletzt wird beispielhaft ein Ansatz zur Anonymisierung von Routen implementiert, um mit Hilfe eines GPS-Sensors aufgezeichnete Fahrzeugrouten zu anonymisieren. Dabei werden zusätzliche Probleme wie der Umgang mit Messungenauigkeiten und Messfehlern sowie die tatsächlichen Auswirkungen von reduzierter Datennutzbarkeit verdeutlicht.
  • Thumbnail Image
    ItemOpen Access
    Development of an Euler-Lagrangian framework for point-particle tracking to enable efficient multiscale simulations of complex flows
    (2023) Kschidock, Helena
    In this work, we implement, test, and validate an Euler-Lagrangian point-particle tracking framework for the commercial aerodynamics and aeroacoustics simulation tool ultraFluidX, which is based on the Lattice Boltzmann Method and optimized for GPUs. Our framework successfully simulates one-way and two-way coupled particle-laden flows based on drag forces and gravitation. Trilinear interpolation is used for determining the fluid's macroscopic properties at the particle position. Object and domain boundary conditions are implemented using a planar surface approximation. The whole particle framework is run within three dedicated GPU kernels, and data is only copied back to the CPU upon output. We show validation for the velocity interpolation, gravitational acceleration, back-coupling forces and boundary conditions, and test runtimes and memory requirements. We also propose the next steps required to make the particle framework ready for use in engineering applications.
  • Thumbnail Image
    ItemOpen Access
    Concepts and methods for the design, configuration and selection of machine learning solutions in manufacturing
    (2021) Villanueva Zacarias, Alejandro Gabriel; Mitschang, Bernhard (Prof. Dr.-Ing. habil.)
    The application of Machine Learning (ML) techniques and methods is common practice in manufacturing companies. They assign teams to the development of ML solutions to support individual use cases. This dissertation refers as ML solution to the set of software components and learning algorithms to deliver a predictive capability based on available use case data, their (hyper) paremeters and technical settings. Currently, development teams face four challenges that complicate the development of ML solutions. First, they lack a formal approach to specify ML solutions that can trace the impact of individual solution components on domain-specific requirements. Second, they lack an approach to document the configurations chosen to build an ML solution, therefore ensuring the reproducibility of the performance obtained. Third, they lack an approach to recommend and select ML solutions that is intuitive for non ML experts. Fourth, they lack a comprehensive sequence of steps that ensures both best practices and the consideration of technical and domain-specific aspects during the development process. Overall, the inability to address these challenges leads to longer development times and higher development costs, as well as less suitable ML solutions that are more difficult to understand and to reuse. This dissertation presents concepts to address these challenges. They are Axiomatic Design for Machine Learning (AD4ML), the ML solution profiling framework and AssistML. AD4ML is a concept for the structured and agile specification of ML solutions. AD4ML establishes clear relationships between domain-specific requirements and concrete software components. AD4ML specifications can thus be validated regarding domain expert requirements before implementation. The ML solution profiling framework employs metadata to document important characteristics of data, technical configurations, and parameter values of software components as well as multiple performance metrics. These metadata constitute the foundations for the reproducibility of ML solutions. AssistML recommends ML solutions for new use cases. AssistML searches among documented ML solutions those that better fulfill the performance preferences of the new use case. The selected solutions are then presented to decision-makers in an intuitive way. Each of these concepts was evaluated and implemented. Combined, these concepts offer development teams a technology-agnostic approach to build ML solutions. The use of these concepts brings multiple benefits, i. e., shorter development times, more efficient development projects, and betterinformed decisions about the development and selection of ML solutions.
  • Thumbnail Image
    ItemOpen Access
    Adaptive robust scheduling in wireless Time-Sensitive Networks (TSN)
    (2024) Egger, Simon
    The correct operation of upper-layer services is unattainable in wireless Time-Sensitive Networks (TSN) if the schedule cannot provide formal reliability guarantees to each stream. Still, current TSN scheduling literature leaves reliability, let alone provable reliability, either poorly quantified or entirely unaddressed. This work aims to remedy this shortcoming by designing an adaptive mechanism to compute robust schedules. For static wireless channels, robust schedules enforce the streams' reliability requirements by allocating sufficiently large wireless transmission intervals and by isolating omission faults. While robustness against omission faults is conventionally achieved by strictly isolating each transmission, we show that controlled interleaving of wireless streams is crucial for finding eligible schedules. We adapt the Disjunctive Graph Model (DGM) from job-shop scheduling to design TSN-DGM as a metaheuristic scheduler that can schedule up to one hundred wireless streams with fifty cross-traffic streams in under five minutes. In comparison, we demonstrate that strict transmission isolation already prohibits scheduling a few wireless streams. For dynamic wireless channels, we introduce shuffle graphs as a linear-time adaptation strategy that converts reliability surpluses from improving wireless links into slack and reliability impairments from degrading wireless links into tardiness. While TSN-DGM is able to improve the adapted schedule considerably within ten seconds of reactive rescheduling, we justify that the reliability contracts between upper-layer services and the infrastructure provider should specify a worst-case channel degradation beyond which no punctuality guarantees can be made.
  • Thumbnail Image
    ItemOpen Access
    Improving usability of gaze and voice based text entry systems
    (2023) Sengupta, Korok; Staab, Steffen (Prof. Dr.)
  • Thumbnail Image
    ItemOpen Access
    Neural Networks on Microsoft HoloLens 2
    (2021) Lazar, Léon
    The goal of the present Bachelor thesis is to enable comparing different approaches of integrating Neural Networks in HoloLens 2 applications in a quantitative and qualitative manner by defining highly diagnostic criteria. Moreover, multiple different approaches to accomplish the integration are proposed, implemented and evaluated using the aforementioned criteria. Finally, the work gives an expressive overview of all working approaches. The basic requirements are that Neural Networks trained by TensorFlow/Keras can be used and executed directly on the HoloLens 2 without requiring an internet connection. Furthermore, the Neural Networks have to be integrable in Mixed/Augmented Reality applications. In total four approaches are proposed: TensorFlow.js, Unity Barracuda, TensorFlow.NET, and Windows Machine Learning which is an already existing approach. For each working approach a benchmarking application is developed which runs a common reference model on a test datatset to measure inference time and accuracy. Moreover, a small proof of concept application is developed in order to show that the approach also works with real Augmented Reality applications. The application uses a MobileNetV2 model to classify image frames coming from the webcam and displays the results to the user. All the feasible approaches are evaluated using the aforementioned evaluation criteria which include ease of implementation, performance, accuracy, compatibility with Machine Learning frameworks and pre-trained models, and integrability with 3D frameworks. The Barracuda, TensorFlow.js and WinML approaches turned out to be feasible. Barracuda, which only can be integrated in Unity applications, is the most performant framework since it can make use of GPU inference. After that follows TensorFlow.js which can be integrated in JavaScript Augmented Reality frameworks such as A-Frame. Windows ML can currently only use CPU inference on the HoloLens 2 and is therefore the slowest one. It can be integrated in Unity projects with some difficulties as well as plain Win32 and UWP apps. Barracuda and Windows Machine Learning are also integrated in a biomechanical visualization application based on Unity for performing simulations. The results of this thesis make the different approaches for integrating Neural Networks on the HoloLens 2 comparable. Now an informed decision which approach is the best for a specific application can be made. Furthermore, the work shows that the use of Barracuda or TensorFlow.js on the HoloLens 2 is feasible and superior compared to the existing WinML approach.
  • Thumbnail Image
    ItemOpen Access
    Models for data-efficient reinforcement learning on real-world applications
    (2021) Dörr, Andreas; Toussaint, Marc (Prof. Dr.)
    Large-scale deep Reinforcement Learning is strongly contributing to many recently published success stories of Artificial Intelligence. These techniques enabled computer systems to autonomously learn and master challenging problems, such as playing the game of Go or complex strategy games such as Star-Craft on human levels or above. Naturally, the question arises which problems could be addressed with these Reinforcement Learning technologies in industrial applications. So far, machine learning technologies based on (semi-)supervised learning create the most visible impact in industrial applications. For example, image, video or text understanding are primarily dominated by models trained and derived autonomously from large-scale data sets with modern (deep) machine learning methods. Reinforcement Learning, on the opposite side, however, deals with temporal decision-making problems and is much less commonly found in the industrial context. In these problems, current decisions and actions inevitably influence the outcome and success of a process much further down the road. This work strives to address some of the core problems, which prevent the effective use of Reinforcement Learning in industrial settings. Autonomous learning of new skills is always guided by existing priors that allow for generalization from previous experience. In some scenarios, non-existing or uninformative prior knowledge can be mitigated by vast amounts of experience for a particular task at hand. Typical industrial processes are, however, operated in very restricted, tightly calibrated operating points. Exploring the space of possible actions or changes to the process naively on the search for improved performance tends to be costly or even prohibitively dangerous. Therefore, one reoccurring subject throughout this work is the emergence of priors and model structures that allow for efficient use of all available experience data. A promising direction is Model-Based Reinforcement Learning, which is explored in the first part of this work. This part derives an automatic tuning method for one of themostcommonindustrial control architectures, the PID controller. By leveraging all available data about the system’s behavior in learning a system dynamics model, the derived method can efficiently tune these controllers from scratch. Although we can easily incorporate all data into dynamics models, real systems expose additional problems to the dynamics modeling and learning task. Characteristics such as non-Gaussian noise, latent states, feedback control or non-i.i.d. data regularly prevent using off-the-shelf modeling tools. Therefore, the second part of this work is concerned with the derivation of modeling solutions that are particularly suited for the reinforcement learning problem. Despite the predominant focus on model-based reinforcement learning as a promising, data-efficient learning tool, this work’s final part revisits model assumptions in a separate branch of reinforcement learning algorithms. Again, generalization and, therefore, efficient learning in model-based methods is primarily driven by the incorporated model assumptions (e.g., smooth dynamics), which real, discontinuous processes might heavily violate. To this end, a model-free reinforcement learning is presented that carefully reintroduces prior model structure to facilitate efficient learning without the need for strong dynamic model priors. The methods and solutions proposed in this work are grounded in the challenges experienced when operating with real-world hardware systems. With applications on a humanoid upper-body robot or an autonomous model race car, the proposed methods are demonstrated to successfully model and master their complex behavior.
  • Thumbnail Image
    ItemOpen Access
    Hochperformante Auflösung kleiner Referenzen in verteilten Systemen
    (2023) Waimer, Joel
    Bereits mit dem Aufkommen erster Filesharing-Systeme wurde die Entwicklung effektiver Verfahren zum Auffinden von mittels global eindeutiger Referenzen bezeichneter Datenobjekte in Peer-to-Peer-Systemen intensiv diskutiert und vorangetrieben, mit dem Ergebnis zahlreicher konkret ausgearbeiteter Lösungsansätze. Unbeachtet geblieben ist dabei jedoch der, für Filesharing-Systeme nicht lohnenswerte, für kleinere verteilte Datenspeichersysteme aber durchaus vorteilhafte Einsatz kleiner Referenzen, in der, aufgrund neuer Fortschritte im Bereich der Speicherdichte, damit einhergehenden dünnen Besetzung dieser kleinen Adressräume, durch welche allerdings die den Verfahren zueigenen Garantien bezüglich der benötigten Anzahl an Schritten zur Auflösung einer Referenz innerhalb des Systems stark verzerrt werden und sich die je Auflösung nötige Laufzeit vergrößert. Diese Arbeit beleuchtet zunächst die Grundlagen der mit der Auflösung von Datenreferenzen in verteilten Speichersystemen einhergehenden Problematiken, beschreibt die beiden Verfahren Chord [SMK+01] und Koorde [KK03] und misst anschließend deren Leistungsfähigkeit in dünn besetzten Adressräumen, unter der Verwendung kleiner Referenzen; mit den Messungen kann schließlich die Vermutung eines negativen Einflusses der dünnen Besetztheit des Adressraums auf die benötigte Laufzeit je Auflösung bestätigt werden. Eingegangen wurde hierbei auch auf mögliche Gegenmaßnahmen zur Verbesserung der Leistungsfähigkeit der beiden Verfahren, wobei hier die Verbindung der beiden untersuchten Verfahren mit einer Abwandlung des beim Distance-Halving-Netzwerk [NW03] eingesetzten Initialisierungsverfahrens zu einer nahezu gleichmäßigen Aufteilung des Adressraumes auf die einzelnen Knoten hier großes Potential besitzt, da so einer Entartung der Pfadlänge je Auflösung entgegengewirkt werden kann; zudem zeigten sich in den Messungen stark ungünstige Auswirkungen einer naiven, iterativen Implementierung des Chord-Verfahrens gegenüber Koorde.
  • Thumbnail Image
    ItemOpen Access
    Efficient exploratory clustering analyses in large-scale exploration processes
    (2021) Fritz, Manuel; Behringer, Michael; Tschechlov, Dennis; Schwarz, Holger
    Clustering is a fundamental primitive in manifold applications. In order to achieve valuable results in exploratory clustering analyses, parameters of the clustering algorithm have to be set appropriately, which is a tremendous pitfall. We observe multiple challenges for large-scale exploration processes. On the one hand, they require specific methods to efficiently explore large parameter search spaces. On the other hand, they often exhibit large runtimes, in particular when large datasets are analyzed using clustering algorithms with super-polynomial runtimes, which repeatedly need to be executed within exploratory clustering analyses. We address these challenges as follows: First, we present LOG-Means and show that it provides estimates for the number of clusters in sublinear time regarding the defined search space, i.e., provably requiring less executions of a clustering algorithm than existing methods. Second, we demonstrate how to exploit fundamental characteristics of exploratory clustering analyses in order to significantly accelerate the (repetitive) execution of clustering algorithms on large datasets. Third, we show how these challenges can be tackled at the same time. To the best of our knowledge, this is the first work which simultaneously addresses the above-mentioned challenges. In our comprehensive evaluation, we unveil that our proposed methods significantly outperform state-of-the-art methods, thus especially supporting novice analysts for exploratory clustering analyses in large-scale exploration processes.
  • Thumbnail Image
    ItemOpen Access
    Data-integrated methods for performance improvement of massively parallel coupled simulations
    (2022) Totounferoush, Amin; Schulte, Miriam (Prof. Dr.)
    This thesis presents data-integrated methods to improve the computational performance of partitioned multi-physics simulations, particularly on highly parallel systems. Partitioned methods allow using available single-physic solvers and well-validated numerical methods for multi-physics simulations by decomposing the domain into smaller sub-domains. Each sub-domain is solved by a separate solver and an external library is incorporated to couple the solvers. This significantly reduces the software development cost and enhances flexibility, while it introduces new challenges that must be addressed carefully. These challenges include but are not limited to, efficient data communication between sub-domains, data mapping between not-matching meshes, inter-solver load balancing, and equation coupling. In the current work, inter-solver communication is improved by introducing a two-level communication initialization scheme to the coupling library preCICE. The new method significantly speed-ups the initialization and removes memory bottlenecks of the previous implementation. In addition, a data-driven inter-solver load balancing method is developed to efficiently distribute available computational resources between coupled single-physic solvers. This method employs both regressions and deep neural networks (DNN) for modeling the performance of the solvers and derives and solves an optimization problem to distribute the available CPU and GPU cores among solvers. To accelerate the equation coupling between strongly coupled solvers, a hybrid framework is developed that integrates DNNs and classical solvers. The DNN computes a solution estimation for each time step which is used by classical solvers as a first guess to compute the final solution. To preserve DNN's efficiency during the simulation, a dynamic re-training strategy is introduced that updates the DNN's weights on-the-fly. The cheap but accurate solution estimation by the DNN surrogate solver significantly reduces the number of subsequent classical iterations necessary for solution convergence. Finally, a highly scalable simulation environment is introduced for fluid-structure interaction problems. The environment consists of highly parallel numerical solvers and an efficient and scalable coupling library. This framework is able to efficiently exploit both CPU-only and hybrid CPU-GPU machines. Numerical performance investigations using a complex test case demonstrate a very high parallel efficiency on a large number of CPUs and a significant speed-up due to the GPU acceleration.