Hadoop Integration

SAP Sybase IQ includes a UDF API that you can use to build MapReduce components, which can be used for Hadoop integration. The SAP Sybase solutions store has examples of Hadoop integration.

The MapReduce programming model is designed for massively parallel distributed computing. The MapReduce programming model consists of two main stages:

Apache Hadoop is a MapReduce implementation. Hadoop is a Java software framework that automates scheduling of map and reduce jobs. 

SAP Sybase IQ supports Hadoop-like parallel scheduling using Table Parameterized Functions (TPFs), a class of external user-defined functions. TPFs accept arbitrary rowsets of table-valued input parameters, and can be parallized in a distributed server environment. You can specify partitioning and ordering requirements on the TPF input. As a developer, you can use TPFs to exploit the MapReduce paradigm from within the database server, using SQL.

For TPF fundamentals, see the User-Defined Functions guide.