In order to identify the modal parameters of civil structures it is vital to distinguish the defective data from that of appropriate and accurate data. The defects in data may be due to various reasons like defects in the data collection, malfunctioning of sensors, etc. For this purpose Exploratory Data Analysis (EDA) was engaged toenvisage the distribution of sensor’s data and to detect the malfunctioning with in the sensors. Then outlier analysis was performed to remove those data points which may disrupt the accurate data analysis. Then Data Driven Stochastic Sub-space Identification (DATA-SSI) was engaged to perform the modal parameter identification. In the end to validate the accuracy of the proposed method stabilization diagrams were plotted. Sutong Bridge, one of the largest span cable stayed bridge was used as a case study and the suggested technique was employed. The results obtained after employing the above mentioned techniques are very valuable, accurate and effective.
A common observation of everyday life reveals the growing importance of data science methods, which are increasingly more and more important part of the mainstream of knowledge generation process. Digital technologies and their potential for data collection and data processing have initiated the birth of the fourth paradigm of science, based on Big Data. Key to these transformations is datafication and data mining that allow the discovery of knowledge from contaminated data. The main purpose of the considerations presented here is to describe the phenomena that make up these processes and indicate their possible epistemological consequences. It has been assumed that increasing datafication tendencies may result in the formation of a data- centric perception of all aspects of reality, making data and the methods of their processing a kind of higher instance shaping human thinking about the world. This research is theoretical in nature. Such issues as the process of datafication and data science have been analyzed with a focus on the areas that raise doubts about the validity of this form of cognition.
Decision-making processes, including the ones related to ill-structured problems, are of considerable significance in the area of construction projects. Computer-aided inference under such conditions requires the employment of specific methods and tools (non-algorithmic ones), the best recognized and successfully used in practice represented by expert systems. The knowledge indispensable for such systems to perform inference is most frequently acquired directly from experts (through a dialogue: a domain expert - a knowledge engineer) and from various source documents. Little is known, however, about the possibility of automating knowledge acquisition in this area and as a result, in practice it is scarcely ever used. lt has to be noted that in numerous areas of management more and more attention is paid to the issue of acquiring knowledge from available data. What is known and successfully employed in the practice of aiding the decision-making is the different methods and tools. The paper attempts to select methods for knowledge discovery in data and presents possible ways of representing the acquired knowledge as well as sample tools (including programming ones), allowing for the use of this knowledge in the area under consideration.
In the paper the phenomenon of big data is presented. I pay my special attention to the relation of this phenomenon to research work in experimental sciences. I search for answers to two questions. First, do the research methods proposed within the paradigm big data can be applied in experimental sciences? Second, does applying the research methods subject to the big data paradigm lead, in consequence, to a new understanding of science?
We talk to Roman Topór-Mądry, MD, chairman of the PAS Committee on Public health, and Tomasz Zdrojewski, MD, from the Jagiellonian University’s Public Health Institute, coauthors of the first Report on Diabetes in Poland, about counting the number of diabetics and data-gathering techniques.
The problem of poor quality of traffic accident data assembled in national databases has been addressed in European project InDeV. Vulnerable road users (pedestrians, cyclists, motorcyclists and moped riders) are especially affected by underreporting of accidents and misreporting of injury severity. Analyses of data from the European CARE database shows differences between countries in accident number trends as well as in fatality and injury rates which are difficult to explain. A survey of InDeV project partners from 7 EU countries helped to identify differences in their countries in accident and injury definitions as well as in reporting and data checking procedures. Measures to improve the quality of accident data are proposed such as including pedestrian falls in accident statistics, precisely defining minimum injury and combining police accident records with hospital data.
Power big data contains a lot of information related to equipment fault. The analysis and processing of power big data can realize fault diagnosis. This study mainly analyzed the application of association rules in power big data processing. Firstly, the association rules and the Apriori algorithm were introduced. Then, aiming at the shortage of the Apriori algorithm, an IM-Apriori algorithm was designed, and a simulation experiment was carried out. The results showed that the IM-Apriori algorithm had a significant advantage over the Apriori algorithm in the running time. When the number of transactions was 100 000, the running of the IM-Apriori algorithm was 38.42% faster than that of the Apriori algorithm. The IM-Apriori algorithm was little affected by the value of supportmin. Compared with the Extreme Learning Machine (ELM), the IM-Apriori algorithm had better accuracy. The experimental results show the effectiveness of the IM-Apriori algorithm in fault diagnosis, and it can be further promoted and applied in power grid equipment.
The paper indicates the significance of the problem of foundry processes parameters stability supervision and assessment. The parameters, which can be effectively tracked and analysed using dedicated computer systems for data acquisition and exploration (Acquisition and Data Mining systems, A&D systems) were pointed out. The state of research and methods of solving production problems with the help of computational intelligence systems (Computational Intelligence, CI) were characterised. The research part shows capabilities of an original A&DM system in the aspect of selected analyses of recorded data for cast defects (effect) forecast on the example of a chosen iron foundry. Implementation tests and analyses were performed based on selected assortments for grey and nodular cast iron grades (castings with 50 kg maximum weight, casting on automatic moulding lines for disposable green sand moulds). Validation tests results, applied methods and algorithms (the original system’s operation in real production conditions) confirmed the effectiveness of the assumptions and application of the methods described. Usability, as well as benefits of using A&DM systems in foundries are measurable and lead to stabilisation of production conditions in particular sections included in the area of use of these systems, and as a result to improvement of casting quality and reduction of defect number.
A data warehouse (DW) is a large centralized database that stores data integrated from multiple, usually heterogeneous external
data sources (EDSs). DW content is processed by so called On-Line Analytical Processing applications, that analyze business trends, discover anomalies and hidden dependencies between data. These applications are part of decision support systems. EDSs constantly change their content and often change their structures. These changes have to be propagated into a DW, causing its evolution. The propagation of content changes is implemented by means of materialized views. Whereas the propagation of structural changes is mainly based on temporal extensions and schema evolution, that limits the application of these techniques. Our approach to handling the evolution of a DW is based on schema and data versioning. This mechanism is the core of, so called, a multiversion data warehouse. A multiversion DW is composed of the set of its versions. A single DWversion is in turn composed of a schema version and the set of data described by this schema version. Every DW version stores a DW state which is valid within a certain time period. In this paper we present: (1) a formal model of a multiversion data warehouse, (2) the set of operators with their formal semantics that support a DW evolution, (3) the impact analysis of the operators on DW data and user analytical queries. The presented formal model was a basis for implementing a multiversion DW prototype system.