SCALABLE COMPUTE INFRASTRUCUTRE

Building systems from scratch; creating and managing virtual machines (QEMU-KVM), also Docker containers; creating HDFS clusters, installing Spark and other Hadoop ecosystem components.

Managing Petabyte distributed fault-tolerant file system, Apache Hadoop HDFS, via both Ambari Fileviewer and CLI.

Scalable statistical programming [15 years], in SAS-HP, SparkR/MLLib, and Tensorflow/MXNet.

Scalable general programming in PySpark [4 years] for Spark, Hadoop cluster/HDFS, Data Lake.

Scalable information-retrieval programming [12 years], in Java, Solr and Lucene (NoSQL).

SQL programming [15+ years], in DB2, MySQL/MariaDB, as well as Hive and SparkQL on Hadoop cluster.

Fluent in Linux (CentOS/RedHat, Fedora, Ubuntu) [14 years], and Windows/DOS/PowerShell [20+ years] platforms.