- Dr. Divesh Srivastava, AT&T Research, USA
- Dr. Sang Kyun Cha, Seoul National University & SAP Labs Korea, South Korea
Enabling Real Time Data Analysis
Network-based services have become a ubiquitous part of our lives, to the point where individuals and businesses have often come to critically rely on them. Building and maintaining such reliable, high performance network and service infrastructures requires the ability to rapidly investigate and resolve complex service and performance impacting issues. To achieve this, it is important to collect, correlate and analyze massive amounts of data from a diverse collection of data sources in real time.
We have designed and implemented a variety of data systems at AT&T Labs-Research to build highly scalable databases that support real time data collection, correlation and analysis, including (a) the Daytona data management system, (b) the DataDepot data warehousing system, (c) the GS tool data stream management system, and (d) the Bistro data feed manager. Together, these data systems have enabled the creation and maintenance of a data warehouse and data analysis infrastructure for troubleshooting complex issues in the network. We describe these data systems and their key research contributions in this talk.
Divesh Srivastava is the head of the Database Research Department at AT&T Labs-Research. He received his Ph.D. from the University of Wisconsin, Madison, and his B.Tech from the Indian Institute of Technology, Bombay. He is on the board of trustees of the VLDB Endowment and an associate editor of the ACM Transactions on Database Systems. He has served as the program committee co-chair of many conferences, including VLDB 2007. He has been invited to present keynotes at many conferences, including VLDB 2010. His research interests and publications span a wide variety of topics in data management.
A New Paradigm of Thinking and Architecture for Real-Time Information Processing at Fingertips
Today's enterprise-scale information systems comprise of complex vertical tiers of database, application, web, and mobile servers. Horizontal tiers of OLTP and OLAP systems add further complexity to enterprise information management. Historically, such introduction of vertical and horizontal tiers was inevitable to address the complexity and performance problems in the course of building up enterprise applications by divide and conquer. However, these tiers have accumulated so much redundancy and overhead over time, making the overall system difficult and expensive to maintain.
Over past decades, we have persistently observed exponential growth of hardware power following the well-known Moore's law. A commodity server can now have hundreds of cores and terabytes of memory, which were not conceivable other than in supercomputers several years ago, at a fraction of cost. This trend is likely to continue at least several years, and at least ten times of further increase of hardware processing power is expected in the near future.
The dramatic hardware advance has brought us to an inflection point that we can eliminate these complex tiers to streamline information delivery to the new generation of end users demanding real-time decision making at fingertips any time anywhere. SAP HANA platform was designed with this rethinking of tiers in enterprise-scale information systems, leveraging the hardware advance and SAP's knowledge of enterprise applications. It enables running OLTP, OLAP, and text processing in a single run-time environment in a scalable way. The foundation of SAP HANA platform is a massively parallel distributed integrated in-memory row and column database system. This talk presents a new paradigm of thinking and architecture underlying SAP HANA platform.
Dr. Sang Kyun Cha has been a professor at Seoul National University in School of Electrical Engineering and Computer Science since February 1992.
He received his Ph.D. in database systems at Stanford University in 1991 with Gio Wiederhold. Before joining the university, he experienced working at various global IT companies: Computer Science Center of Texas Instruments, Inc. (in 1985), IBM Palo Alto Scientific Center (as a consultant in 1990-1991) and HP Laboratories, Palo Alto (an in-memory personal data management project, from 1991 to 1992).
From August 2001 to August 2002, he visited the database research group in Computer Science Department of Stanford University and was responsible for organizing CS545 InfoSeminar course.
In 2000, he founded Transact In Memory, Inc. (TIM) in Korea to fund the next-generation In-Memory Database engine development and in 2002 moved its headquarter to Silicon Valley. In November 2005, TIM was acquired by SAP and its Korean entity was transformed to SAP Labs Korea. By SAP's request, he continues to lead SAP Labs Korea and contributes to SAP's HANA In-Memory DB development.
In March 2012, he was elected as an external director on the board of KT (Korea Telecom).