Research theme

Big sciences such as astronomy and meteorology, AI technology including deep learning, and blockchain technology that forms the basis of virtual currency are prosperous. Since these technologies are all based on large-scale data, a basic technology to manage them is necessary. This basic technology is called a data system. There are various data systems, such as traditional relational databases and file systems, graphs and time series databases tailored to specific purposes, and many products are scattered. Although it is such a diverse data system, the only elemental technologies are actually query processing and transaction processing. The query process is a process that interprets the content of the user asking the system for "Please give me such data" and returns the processing result. An example of a query language is SQL. Transaction processing is processing performed when money is transferred to a bank or when data is stored in a computer, and has a characteristic called ACID. If you master these two technologies from the basics, you will be able to understand any data system related to big science, AI infrastructure, blockchain, etc.

Telescope data astronomy

The Subaru Telescope used by astronomers who collaborate with us produces as much as 300GB of data every night. Data processing called astronomical pipeline is necessary to obtain beautiful images by processing this large-scale data. This requires faster processing called transactions, which we are working on. Our transaction processing technology is tens to tens of thousands of times the performance of commercial systems. this is Joint research with Kavli IPMU, Institute of Statistical Mathematics, and NTT CS Laboratories. Thank you for your support from JST CREST.


Real-time data kernel

A real-time database is required to realize a variety of real-time artificial intelligence that realizes a distribution system that fluctuates from time to time, a gravitational wave candidate celestial body that suddenly appears, and the immediate environment necessary for evacuation after the earthquake disaster. This system requires an HTAP (Hybrid Transactional & Analytical Processing) architecture that tightly combines transaction processing (OLTP) and data analysis (OLAP). This is a joint research with Nautilus Technologies, NEC, Pasco, Makoto Onizuka (Osaka Univ.), And Professor Keiji Ishikawa (Nagoya Univ.). We have received support from NEDO.


Block chain

To realize a gateless ticket gate and a cashierless convenience store, a huge amount of billing processing is required. Because this billing process requires high reliability, a distributed agreement technology that synchronizes data on multiple machines is required. High-speed communication mechanism (Remote Designing technology using Direct Memory Access). This project is a joint research with a financial institution.


Non-volatile memory

Non-volatile memory such as Intel Optane Persistency Memory (memory that does not lose its contents even when the power is turned off) has come out. Non-volatile memory is a new technology, and it is a device with an enormous number of unknowns about what applications it can support and how it should be designed. In joint research with a company that researches and develops non-volatile memory, we design a new memory architecture and a new data infrastructure that uses it.


Integration of artificial intelligence and data infrastructure

We are researching the creation of a database system that provides various machine learning technologies, the index mechanism using deep learning, and the creation of query optimizers using deep learning. Supported by joint research with a company and Scientific Grant-in-Aid for Scientific Research B .


Large-scale data integration

After a huge amount of data has been collected, there is a growing need to know what it means and how it relates to other data. Research on data profiling is underway to solve this problem. This study is focused on petabyte-scale data held by a company.