SCALABLE COMPUTE INFRASTRUCUTRE
Building systems from scratch; creating and managing virtual machines (QEMU-KVM), also Docker containers; creating HDFS clusters, installing Spark and other Hadoop ecosystem components.
Managing Petabyte distributed fault-tolerant file system, Apache Hadoop HDFS, via both Ambari Fileviewer and CLI.
Scalable statistical programming [15 years], in SAS-HP, SparkR/MLLib, and Tensorflow/MXNet.
Scalable general programming in PySpark [4 years] for Spark, Hadoop cluster/HDFS, Data Lake.
Scalable information-retrieval programming [12 years], in Java, Solr and Lucene (NoSQL).
SQL programming [15+ years], in DB2, MySQL/MariaDB, as well as Hive and SparkQL on Hadoop cluster.
Fluent in Linux (CentOS/RedHat, Fedora, Ubuntu) [14 years], and Windows/DOS/PowerShell [20+ years] platforms.