This book guides animal ecologists, biologists and wildlife and data managers through a step-by-step procedure to build their own advanced software platforms to manage and process wildlife tracking data. This unique, problem-solving-oriented guide focuses on how to extract the most from GPS animal tracking data, while preventing error propagation and optimizing analysis performance. Based on the open source PostgreSQL/PostGIS spatial database, the software platform will allow researchers and managers to integrate and harmonize GPS tracking data together with animal characteristics, environmental data sets, including remote sensing image time series, and other bio-logged data, such as acceleration data. Moreover, the book shows how the powerful R statistical environment can be integrated into the software platform, either connecting the database with R, or embedding the same tools in the database through the PostgreSQL extension Pl/R. The client/server architecture allows users to remotely connect a number of software applications that can be used as a database front end, including GIS software and WebGIS. Each chapter offers a real-world data management and processing problem that is discussed in its biological context; solutions are proposed and exemplified through ad hoc SQL code, progressively exploring the potential of spatial database functions applied to the respective wildlife tracking case. Finally, wildlife tracking management issues are discussed in the increasingly widespread framework of collaborative science and data sharing. GPS animal telemetry data from a real study, freely available online, are used to demonstrate the proposed examples. This book is also suitable for undergraduate and graduate students, if accompanied by the basics of databases.
Mining Very Large Databases with Parallel Processing addresses the problem of large-scale data mining. It is an interdisciplinary text, describing advances in the integration of three computer science areas, namely `intelligent' (machine learning-based) data mining techniques, relational databases and parallel processing. The basic idea is to use concepts and techniques of the latter two areas - particularly parallel processing - to speed up and scale up data mining algorithms.
This new book aims to provide both beginners and experts with a completely algorithmic approach to data analysis and conceptual modeling, database design, implementation, and tuning, starting from vague and incomplete customer requests and ending with IBM DB/2, Oracle, MySQL, MS SQL Server, or Access based software applications. A rich panoply of solutions to actual useful data sub-universes (e.g. business, university, public and home library, geography, history, etc.) is provided, constituting a powerful library of examples.
Four data models are presented and used: the graphical Entity-Relationship, the mathematical EMDM, the physical Relational, and the logical deterministic deductive Datalog ones. For each one of them, best practice rules, algorithms, and the theory beneath are clearly separated. Four case studies, from a simple public library example, to a complex geographical study are fully presented, on all needed levels.
Several dozens of real-life exercises are proposed, out of which at least one per chapter is completely solved. Both major historical and up-to-date references are provided for each of the four data models considered.
The book provides a library of useful solutions to real-life problems and provides valuable knowledge on data analysis and modeling, database design, implementation, and fine tuning.
Website Hosting Articles
Website Hosting Books