This project is widely inspired (fork) from https://github.com/brock-acryl/dremio-hivejdbc-connector and is a customization to connect Dremio on a HiveServer2 instance behind a Knox Gateway.
It is highly recommended to connect to Dremio using the native Hive Metastore connector, as this will provide the best performance. This Connector will allow you to use a JDBC connection to Hive.
The Advanced Relational Pushdown (ARP) Framework allows for the creation of Dremio plugins for any data source which has a JDBC driver and accepts SQL as a query language. More details on this here : https://docs.dremio.com/current/developer/arp-connector/
There are two files that are necessary for creation of an ARP-based plugin: the storage plugin configuration, which is code, and the plugin ARP file, which is a YAML (https://yaml.org/) file.
The storage plugin configuration file tells Dremio what the name of the plugin should be, what connection options should be displayed in the source UI, what the name of the ARP file is, which JDBC driver to use and how to make a connection to the JDBC driver.
The ARP YAML file is what is used to modify the SQL queries that are sent to the JDBC driver, allowing you to specify support for different data types and functions, as well as rewrite them if tweaks need to be made for your specific data source.
The ARP file is broken down into several sections:
metadata
- This section outlines some high level metadata about the plugin.
syntax
- This section allows for specifying some general syntax items like the identifier quote character.
data_types
- This section outlines which data types are supported by the plugin, their names as they appear in the JDBC driver, and how they map to Dremio types.
relational_algebra - This section is divided up into a number of other subsections:
- aggregation
- Specify what aggregate functions, such as SUM, MAX, etc, are supported and what signatures they have. You can also specify a rewrite to alter the SQL for how this is issued.
- except/project/join/sort/union/union_all/values
- These sections indicate if the specific operation is supported or not.
- expressions
- This section outlines general operations that are supported. The main sections are:
- operators
- Outlines which scalar functions, such as SIN, SUBSTR, LOWER, etc, are supported, along with the signatures of these functions which are supported. Finally, you can also specify a rewrite to alter the SQL for how this is issued.
- variable_length_operators
- The same as operators, but allows specification of functions which may have a variable number of arguments, such as AND and OR.
If an operation or function is not specified in the ARP file, then Dremio will handle the operation itself. Any operations which are indicated as supported but need to be stacked on operations which are not will not be pushed down to the SQL query.
- In root directory with the pom.xml file run
mvn clean install - Take the resulting .jar file in the target folder and put it in the \dremio\jars folder in Dremio
- Take the Hive JDBC driver from (https://downloads.cloudera.com/connectors/ClouderaHiveJDBC-2.6.25.1033.zip) and put in in the \dremio\jars\3rdparty folder
- Restart Dremio
Once the connector has been added, log in to Dremio. Inside Dremio, click on the plus sign to add a new data source, and select HiveJDBC from the list of available sources. In the 'New HiveJDBC Source' window enter the connection string, user, password to connect to and click Save (jdbc:hive2://:).
- Optimize pushdown