Skip to content

Latest commit

 

History

History

zeppelin

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 

Zeppelin Chart

Zeppelin is a web based notebook for interactive data analytics with Spark, SQL and Scala.

Chart Details

Installing the Chart

To install the chart:

$ helm install banzaicloud-stable/zeppelin

Configuration

The following tables lists the configurable parameters of the Zeppelin Sever chart and their default values, in case you want to preserve your Spark application logs on S3 or Azure storage.

Parameter Required Description Example
username no Admin username, by default is admin
password no Salted password of admin user, by default is zeppelin You can salt your own password using shiro cli tool java -jar ~/dev/tools/shiro-tools-hasher-1.3.2-cli.jar -p
userCredentialSecretName no Credentials above are set in a K8s secret. Instead of specifying username & password directly you can provide the name of this K8s secret containing these fields
logService.host yes if you want to send logs to Syslog Host address of Syslog service 10.44.0.12
logService.zeppelinLogPort yes if you want to send logs to Syslog UDP port for Zeppelin logs 512
logService.sparkLogPort yes if you want to send logs to Syslog UDP port for Spark Driver and Executor logs 512
logService.applicationLogPort yes if you want to send logs to Syslog UDP port for Application logs 512
logService.applicationLoggerName no Name of log4j logger for Application logs by default: application
logService.zeppelinLogLevel no log4j log level for Zeppelin logs by default: DEBUG
logService.zeppelinLogPattern no log4j log pattern for Zeppelin logs by default: "%5p [%d] ({%t} %F[%M]:%L) - %m%n"
logService.sparkLogLevel no log4j log level for Spark logs by default: INFO
logService.sparkLogPattern no log4j log pattern for Spark logs by default: "[%p] %c:%L - %m%n"
logService.applicationLogLevel no log4j log level for Application logs by default: INFO
logService.applicationLogPattern no log4j log pattern for Application logs by default: "[%p] %c:%L - %m%n"
sparkSubmitOptions.eventLogDirectory yes the URL to the directory for event logs s3a://yourBucketName
wasb://your_blob_container_name@you_storage_account_name.blob.core.windows.net
gs://yourBucketName
notebookStorage.type no storage type for notebooks s3
azure
gs
by default no storage is configured
notebookStorage.path no storage path for notebooks bucket name in case of S3 / GS, file share name for Azure
azureStorageAccountName only in case of using Azure Storage Name of your Azure storage account see Notes
azureStorageAccessKey only in case of using Azure Storage Access key for your Azure storage account see Notes

Notes

  • in case of using S3 and Google Storage, we don't pass credentials and access keys we're using IAM roles and policies on Amazon and Service Account based access on Google Cloud
  • in case of Azure the storage account name would be the dns prefix it's created (e.g. mystorage.blob.core.windows.net - the name would be mystorage), and you can you either the primary or secondary keys