BigData@Polito Access Instructions

User Account

To access the BigData@Polito Cluster you need a user account. Request the credentials.

For support, accounts or technical issues, write to supporto@smartdata.polito.it.

Jupyter Interface – Spawning a Server

The primary interface for the cluster is via JupyterLab:  https://jupyter.polito.it/

After login, spawn a server among the images at your disposition.

Coming soon: With the new BigData Cluster, more powerful images will become available for all users. More powerful images will have a limited lifetime (i.e., the biggest images will be culled everyday). Less powerful images can be used for longer interval instead (e.g., up to a week).

Jupyter Interface – Main Interface

In the main jupyter interface you can start PySpark notebooks, running your master in the notebook server (local) or in the big data cluster (yarn). You can open your terminal or interact with your home folder in the left panel. Other notebooks are also available, such as R and Octave. 

Accessing BigData@Polito Hadoop Services

<< The preferable interface to interact with the cluster is the JupyterHub listed above >>

You can also access the cluster via the Access Gateway bigdatalab.polito.it. Connecting to

$ ssh your_username@bigdatalab.polito.it

You can interact with the HDFS file system, copy your data to/from the cluster, and submit your jobs. You can find here a list of available commands. Note that limited resources are available in the gateway and strict quotas are enforced. If you need to perform heavy analysis in the gateway, fire a large image via JupyterHub.

Web Interfaces to the Hadoop Services

For more flexible access to the system, you can use one of the web interfaces for the services. The preferred means to interact with  the services is via the Hue service: https://bigdatalab.polito.it/

Again, use your credentials for signing in, then you can interact with the services (e.g., HDFS).

Advanced Use Cases

 Other Hadoop interfaces are available only from Polito wired network: 
HDFS https://ma1-bigdata.polito.it:9871/https://mb1-bigdata.polito.it:9871/
YARN Job History https://ma1-bigdata.polito.it:19890/jobhistory/
YARN Resource Manager https://ma1-bigdata.polito.it:8090/cluster/

https://mb1-bigdata.polito.it:8090/cluster/

For accessing the YARN Job History, the YARN Resource Manager and the HDFS web interfaces, you will need to get a kerberos ticket in your PC and your browser must support HTTP SPNEGO, as described below.

Security Information

HTTPS and TLS connections to the private BigData@Polito servers (available only from Polito network) are signed by our BigData Lab Root CA. You can either add in your browsers security exceptions to all the servers involved, or you can decide to trust the BigData Lab Root CA installing, in order, its certificates on your system.

  1. BigData Lab Root CA certificate
  2. BigData Lab Intermediated CA certificate

Using your own Hadoop Clients

For complex development or special needs, e.g. large local disk space, we suggest you to deploy your own local Hadoop system and configure to act as a client for the BigData@Polito cluster.

The software currently used in on BigData@Polito is the Cloudera CDH distribution. Consult the installation requirements and instructions for any information: you will need to perform an “unmanaged deployment” using the provided packages. If the current Cloudera version is more recent than the one on the cluster, follow the instructions about “Installing an Earlier Release”.

You still need valid BigData@Polito credentials, the correct Kerberos configuration, and you must verify that AES-256 Encryption is enabled for your Java JDK. More information here.

After installing the software, you need to configure it to interact the BigData@Polito cluster, downloading the following Client Configuration files, and deploying them in the respective configuration directories:

Please notice that we cannot help you directly with the deployment of your personal Hadoop system, beside giving you the general information about the required software versions and providing you the Client configuration files: for any other problem you must contact your IT support.

Kerberos Client Configuration

Edit your Kerberos client configuration file, usually located in /etc/krb5.conf, to add the BigData@Polito servers:

[libdefaults]
 default_realm = BIGDATA.POLITO.IT
[realms]
BIGDATA.POLITO.IT = {
   kdc = ma1-bigdata.polito.it
   kdc = mb1-bigdata.polito.it
   admin_server = ma1-bigdata.polito.it
}

Verify you can obtain your Kerberos ticket issuing the command

$ kinit <your_username>

If the system is correctly configured you will receive no error message. You can verify the status of your ticket with the command

$ klist

If you cannot obtain the ticket, check that the Kerberos KDCs can be reached from your terminal, or contact your IT support.

Web Browser Configuration

To use the Web Interfaces for the BigData Hadoop Services you must configure your browser to utilize the Simple and Protected GSS-API Negotiation (SPNEGO) mechanism.

HTTP SPNEGO is enabled by default on Safari and Google Chrome. Instructions on how to enable SPNEGO in Firefox and other browsers are available at the links below. The HTTP SPNEGO whitelist must include the “.polito.it” domain.