Workshops
Five workshops are accepted. Accepted workshop papers will be published as Springer's LNCS series with the main conference proceedings. The registration for the workshops will also be integrated into the main DASFAA 2012 conference (single registration for both workshops and conference).- The Second International Workshop on Flash-based Database Systems (FlashDB 2012)
- FlashDB 2012 Homage Page: http://idke.ruc.edu.cn/FlashDB_2012
- Information Technologies for the Maritime and Logistics (ITEMS'2012)
- ITEMS 2012 Homage Page: http://www.maritimestudies.nus.edu.sg/ITEMS2012/ITEMS2012.html
- The Third International Workshop for Social Networks and Social Web Mining (SNSM 2012)
- SNSM 2012 Homage Page:
http://e-research.csm.vu.edu.au/files/snsm2012/index.html
- SNSM 2012 Homage Page:
- The Second International Workshop on Spatial Information
Modeling, Management and Mining (SIM3-2012)
- SIM3 2012 Homage Page:
http://www.ucalgary.ca/sim3/
- SIM3 2012 Homage Page:
- Data Quality in Data Integration System (DQDI’2012)
- DQDI’2012 Homage Page:
http://itee.uq.edu.au/~dke-dq/DQIS2012/
Recently, new storage media such as flash memory and phase change memory have been developed very quickly, which bring big challenges to the architecture of computer systems as well as the design of system software. In particular, NAND flash (either SLC- or MLC-based) in the form of solid state disks (SSDs) has been an alternative to traditional magnetic disks, both in the home-user environment and in the enterprise computing environment, due to its shock-resistance, low power consumption, non-volatility, and high I/O speed. The special features of flash memory and other new storage media impose new challenges to traditional data management technologies. As a result, traditional database architectures and algorithms designed for magnetic-disk-based storage fail to utilize new storage media efficiently. Meanwhile, the new characteristics of modern storage media, such as not-in-place update and asymmetric read/write/erase latencies of flash memory, also bring great challenges in optimizing database performance, by using new query processing algorithms, indexes, buffer management schemes, and new transaction processing protocols. Consequently, how to exploit the characteristics of flash memory and other new storage media has become an important topic of database systems research. In order to make use of the characteristics of flash memory and other new storage media, the data management community needs to rethink the algorithms and technical issues in magnetic-disk-oriented database systems and gets them adapted to the advances in the underlying storage infrastructure.
The aim of this one-day workshop is to bring together researchers who are interested in optimizing database performance on flash memory or other new storage media by designing new data management techniques and tools. We welcome papers that address fundamental research issues in this challenging area, with emphasizes on new algorithms of flash-based data management. We also encourage papers to report on system level research related to data management on flash-based solid state disks or other new storage devices. The topics of interest include, but are not limited to, the following:
- Design of new storage devices, e.g., flash-based solid state disks
- Database architectures for new storage devices
- Green-computing and energy-efficient techniques for new storage devices
- Cost models and query optimization for new storage devices
- Index structures optimized for new storage devices
- Buffer management policies for new storage devices
- Multi-level storage systems
- Performance analysis of database workloads on new storage devices
- Transaction processing in database systems on new storage devices
- Data management issues for hybrid storage architecture
- New workloads and benchmark tools for algorithms on new storage devices
According to the International Maritime Organization more than 90% of the global trade, whether it is oil and gas, bulk or containerized cargo, is carried by sea. To this volume leisure, passenger and military shipping must be added to account for the traffic. The stakeholders in the maritime industry are numerous. Information Technology for the Maritime and Logistics (ITEMS) is an international symposium that aims at creating a forum for researchers and professionals to present and discuss the latest research results and developments of information technologies for the maritime and logistics. The workshop's scope includes all important aspects of information technology, including simulation of maritime systems (vessels interactions, trajectories, domains); acquisition, processing, and management of maritime data; robotic aspects of port management; web technologies for various aspects of maritime operations; artificial intelligence, decision support systems and data analytics for various maritime partners.
We solicit papers that address all important aspects of information technologies for the maritime industry. Topics of interest include but are not limited to:
- Navigation and Surveillance Systems such as Electronic Chart Display and Information System (ECDIS), Long Range Identification and Tracking (LRIT), and Automatic Identification Systems (ASI)
- Wireless Sensor Networks (WSN) and Radio-Frequency Identification (RFID) for asset tracking and infrastructure maintenance
- Management of maritime information
- Decision Support Systems for maritime operations
- Ship movement and traffic simulation
- Geo-nautical information systems
- Maritime standards and security
- Robotics and port management
- Ship design methods & tools
- Maritime data analytics
- Artificial Intelligence
- Shipping networks
- Maritime logistics
- Web technologies
Today the emergence of web-based communities and hosted services such as social networking sites, wikis and folksonomies, brings in tremendous freedom of Web autonomy and facilitate collaboration and sharing between users. And along with the interactions between users and computers, social media are rapidly becoming an important part of our digital experience, ranging from digital textual information to rich multimedia formats. These aspects and characteristics form the core of the second generation of Web.
A prominent challenge lies in modeling and mining this vast volume of data to extract, represent and exploit meaningful knowledge, and to leverage structures and dynamics of emerging social networks residing in the social media. Social networks and social media mining combines data mining with social computing as a promising direction and offers unique opportunities for developing novel algorithms and tools ranging from text and content mining to link mining and community detection and so on.
This workshop will bring together the academia, researchers and industrial practitioners from computer science, information systems, statistics, sociology, behavior science and organization science discipline, and provide a forum for disseminating and exchanging recent advances in the field of social networks and social media, from the perspectives of data management and mining.
We solicit original and high quality submissions addressing all aspects of this field. The topics of interest include, but are not limited to the following:
- Computational models for social media
- Query languages for social networks and social media
- Influence, trust, and privacy
- Collaborative filtering and content ranking using social media
- Data management in collaborative open applications
- Social reputation and recommendation systems, trust management
- Search in social networks and social media
- Interoperability among social applications and social media
- Techniques for social-network analysis/mining and for the analysis of social-media phenomena
- Adversarial blogging and counter measures
- Link analysis and network structure discovery
- Community detection and evolution
- Blog search and retrieval
- Group interaction, collaboration, and recommendation
- Knowledge discovery (collective wisdom, trend analysis, and topic detection)
- Social aspects of Blogosphere
Nowadays, spatial data exists pervasively in various information systems and applications. The unprecedented amount of spatial data that has been amassed and that is being produced in an increasing speed, via various facilities such as sensors, GPS receivers, smart phones and remote sensing, calls for extensive, deep and sustaining research on spatial information modeling, management and mining. In the past decade, we witnessed increasing research interests in these areas from database, data mining and geographic information systems (GIS) communities.
Following the success of SIM3-2011, SIM3-2012 workshop sticks to the tradition that brings together researchers, developers, users, and practitioner carrying out research and development in spatial information modeling, management and mining, and fosters discussions in all aspects of these research areas. The workshop will provide a forum for original research contributions and practical experiences of spatial information modeling, management and mining and will highlight future trends in these topics.
Therefore, we will invite submissions that address theoretical, technical and practical issues of spatial information modeling, management, and mining. Suggested topics include, but are not limited to:
Spatial information modeling
- Spatial data modeling
- Conceptual and logical models for spatial data
- Spatial services modeling
- Spatial information grid
- Ontologies for GIS applications
- Schema mapping and evolution
- Spatial semantic issues
Spatial information management
- Spatial query languages and interfaces
- Spatial query processing and optimization
- Spatial search engines
- Spatial data management in cloud
- Spatial information integration
- Spatial information interoperability and standards
- Spatial metadata management
- Spatial information visualization
- Location-based services
Spatial information mining
- Spatial correlation analysis
- Spatial clustering
- Spatial regression and classification
- Spatial trend detection
- Spatial outlier detection
- Spatial data warehousing
Data integration has been a subject of intense research and development for over three decades. Basically the goal of a data integration system is to provide a uniform interface to a multitude of data sources. Difficulties in overcoming the schematic, syntactic and semantic differences of data from multiple autonomous and heterogeneous sources are well recognized, and have resulted in a multi-billion dollar data integration market and growing. With the phenomenal increase in the scale and disparity of data, the problems associated with data integration have increased dramatically.
A fundamental aspect of user satisfaction from an integration system is the quality of data. Industry reports indicate that expensive data integration initiatives stemming from migrations, mergers, legacy upgrades etc, succeed in achieving a common technology platform, but are rejected by the user communities due to the presence (or exposure) of poor data quality. Poor data quality is known to compromise the credibility and efficiency of commercial as well as public endeavours. Several developments from industry as well as academia have contributed significantly towards addressing the problem.
These typically include analysts and practitioners who have contributed to the design of strategies and methodologies for data governance; solution architects including software vendors who have contributed towards appropriate system architectures that promote data integration; and data experts who have contributed to data quality problems such as duplicate detection, identification of outliers, consistency checking and many more through the use of computational techniques. The attainment of true data quality lies at the convergence of the three aspects, namely organizational, architectural and computational.
The workshop will provide a forum to bring together diverse researchers and make a consolidated contribution to new and extended methods to address the challenges of data quality in data integration system. Topics covered by the workshop include at least the following:
- Data integration, linkage and fusion
- Entity resolution, duplicate detection, and consistency checking
- Data profiling and measurement
- Use of data mining for data quality assessment
- Methods for data transformation, reconciliation, consolidation
- Algorithms for data cleansing
- Data quality and cleansing in information extraction
- Dealing with uncertain or noisy data (e.g., sensor data)
- Data lineage and provenance
- Models, frameworks, methodologies and metrics for data quality
- Metadata management for data quality
- Application specific data quality, case studies, experience reports
- User/social perceptive on data quality and cleansing
- Data quality and cleansing for complex data (e.g. documents, semi-structured data, XMLs, multimedia data, graphs, biosequences, etc.)