fbpx Essam | DevReady

Let’s Get Started

Essam
HireEssam

Essam

Parlin, New Jersey --:--:--

Essam is an innovative, experienced, results-driven Sr. Architect with 30+ years of experience, combining hands-on, architectural, and team leadership talents to design and develop high-volume data products and microservices architecture. Proven expertise in roles where the organization’s strategy is to create breakthrough solutions to make the best use of the data using AI and analytics. Passionate about building superior data-driven products using Structured and Unstructured Data Integration, Knowledge Representation; Semantic Web; Linked Data, Graph Analytics; NLP; and Machine Learning. Leverage technical, business, and analytical skills to align enterprise technology strategies and processes with strategic business plans through business process re-engineering, and innovative architectural solutions.  Skilled at conveying complex technical solutions to individuals and groups from professionals to C-level executives.

Hire Essam
Skills
Years
1
2
3
4
5
6
7
8
9
10+
Airflow
DevOps
Scala
Hadoop
ERP
Pytorch
S3
ElasticSearch
Oracle
Dremio
AWS
Data Lakes
SQL
Kubernetes
MongoDB
PowerBuilder
Kafka
Talend
OWL
R2RML
GraphQL
Enterprise Edition
Spark
Twill
WebLogic
API Gateway
Silk
Python
Jupyter
Java
NoSQL
Apache
SQL Server
R
DB/400 Servers
MAPICS
IBM
Oracle PL/QL
J2EE
Globus
Developer Personality

Independent

Collaborative

Trailblazer

Conservative

Generalist

Specialist

Planner

Doer

Idealist

Pragmatist

Abstraction

Control

100
50
0
50
100
Feature Experience

AWS

Big Data

AI & Machine Learning

Text Analytics & NLP

MODERATE
EXTENSIVE
EXPERT
Cultural Experience

Agile - Cross-functional Teams

Marketing / Retail

Healthcare / Pharmaceuticals

Education

MODERATE
EXTENSIVE
EXPERT
Portfolio

Memorial Sloan Kettering

Architect - Lead Data Engineer

Work Experience : 2020 - present

Architecting and implementing of MSK MIND computational oncology AI platform to integrate multidimensional datasets and enable advanced analytics including machine learning and artificial intelligence, using cutting edge big data and AI technologies, to predict patient outcomes and ultimately be translated to improved patient care:
• Implemented radiology and pathology data ingestion pipeline and integrated the data into a virtual data lake.
• Implemented ML pipeline by doing data mining for clinical data radiology reports to predict metastatic events using CNN and word embedding.
Technologies: Spark, AirFlow, MLOps (guild), distributed computing (Ray/Dask), Delta Lake, CI/CD (CircleCI, Flux CD), ML/AI, HPC, Python, PyTorch, Data Virtualization (Dremio).

More

Elsevier

Solutions Architect / Life Sciences Solutions

Work Experience : 2015-2020

Elsevier is an information and analytics company, and one of the world’s major providers of scientific, technical, and medical information.
“Find all compounds share multiple targets with drug compounds and”, “Study the most significant pathways that are associated with a given adverse drug reaction”: These are the kind of natural language questions, and more, that could be answered with this patented Semantic Data Integration, Search & AI Platform for Life Sciences – Entellect. The platform ingests structured and unstructured data streams, extracts knowledge from the content using NLP and Machine Learning, semantically integrates the data, and creates data products for consumption system antic search engine, AI platform, and graph analytics engine.

Role included:

  • Patent co-authoring;
  • Worked with the Product Owner to define the product vision, strategy, business case, and roadmap;
  • Platform architecture and design based on Microservices Architecture (Kubernetes/AWS EKS) and AWS Serverless, more specifically:
    • Designed structured data ingestion pipeline (Technology: Scala, S3, AWS Lambda, Kafka Streams, R2RML)
    • Designed unstructured content ingestion pipelines (Technology: Java, S3, AWS Lambda, Kafka Streams, Spark);
    • Designed platform consumers pipelines, and consumption data models (Technology: AWS Glue, AWS Lambda, AWS Step Functions, AWS Elasticsearch, AWS Neptune, EMR);
    • Designed Semantic Search as a Service to enable cognitive search on Elasticsearch (Technology: Java, Elasticsearch)
    • Designed Life Sciences Integrated Ontology Model to integrate Life Sciences
      datasets based on cheminf, CHEBI, BAO, and UniProt ontologies;
    • Designed Data Governance Platform (data flow, policies, and procedures) to manage
      vocabularies, taxonomies, ontologies, and sources metadata (Technology: TopBraid, AWS
      Neptune);
    • Designed platform security, Authentication, and Authorization Framework (Technology: Auth0, Ambassador API Gateway, ITSIO Service Mesh);
    • Designed AI platform (Technology: Jupyter Notebook, Jupyterhub. Spark, HDFS)
    • Worked with the Data Scientist to extract life science concepts from the content (Technology: Python, scikit-learn, LDA, SVM)
  • Technology selection and PoC, and leading a global team for MVP implementation;
  • Rearchitected Reaxys and Pathway Studio products to onboard onto the platform by migrating Oracle relational databases into KDB NoSQL database;
  • Worked with engineering squads on platform implementation;
  • Worked with the Enterprise Data Architects for alignment with the Enterprise Data Model;
  • Worked with the Enterprise Security Architects to enable SSO and to ensure compliance with security
    standards;
  • Created AWS Infrastructure Topology and worked with the infrastructure and DevOps team for deployment;
  • Worked with the product owners to gather data analytics use cases and requirements;
  • Worked with Data Scientists to create implementation notebooks, train and deploy ML models.
More

Pearson

Sr. Architect

Work Experience : 2005-2015

Pearson is the world’s leading learning company, providing a range of education products and services to institutions, governments, and individual learners. My role was researching new data products, leading the development of PoC/Prototypes, and in some instances Beta releases, and developing of product roadmap and strategy.
• Integrated nine different Product and Customer Masters (PMI/CDI) by architecting a Data Hub based on Semantic Data Integration, Graph modeling, and ontology mapping approaches. Used OWL Ontology, Talend ETL, Oracle DB, Oracle ESB, and Java.
• Architected a data-driven approach for content development, publishing, and analytics to improve learning efficacy and outcomes by implementing a “Data Lake” with an underline “Education Graph” using Big Data and Graph/Linked Data paradigms and technologies. The Data Lake enabled global users to discover, search, share, reuse, manage and collaborate throughout the content life-cycle, with near real-time Data & Content Analytics. Used Apache Kafka for real-time distributed messaging and data streaming to build the data pipelines, Hadoop/Pig for graph ETL and building, and OntoText GraphDB as RDF triple/graph store. Deployed the Data Lake in three global regions, and used Apache Yarn and Twill for distributed application management.
• Developed a Predictive Model to identify Risks and opportunities in the UK EdExcel Exam business to predict schools likely to switch providers. Ingested and aggregated 5 years of GSCE grade data sets into Hbase using Pig. Fetched data into R Language in-memory datasets using rhbase. Trained binary classifier (stayed vs. switched) using caret package in R. Initial data analysis shows 67% precision and 32% recall.
• Increased content share and reuse by architecting Chaski Enterprise Search & Retrieval platform. Used ElasticSearch, MongoDB, and Java to build content sources connectors.
• Implemented Recommendation Engine using Similarity Search and Content Classification for the Learning Object Discovery Engine content recommendation and curriculum standards mapping. Used Silk for link discovery, ElasticSearch Search, Apache Jena, Python, and Apache Jena TDB Graph Store.
• Increased content discoverability by 75% by architecting text analytics and NLP pipeline based on Linked Data and Semantic Web principles to enable automatic content semantic enrichment throughout the content development life-cycle. Used DBpedia Spotlight, UIMA, Apache Jena, and Java.

More

First Genetic Trust

Sr. Technical Manager

Categories

Work Experience : 2001-2005

First Genetic Trust is a provider of secure electronic clinical data collection (EDC), Pharmacogenomics clinical trial management, and secure genetic data banking (data warehouse) products and services. My primary responsibility included: gathering internal and external requirements, architecting the product solution, integrating with client’s and partner’s applications, Leading onsite/offshore development teams, coordinating
across functional teams, delivery of development artifacts, and managing client expectations. Some of my accomplishments are:
Product Management:
• Worked with senior management to define product development strategy and roadmap.
• Managed the entire product life cycle to successfully deliver three enTRUST releases and four minor releases.
• Defined and documented functional specifications and software architecture for the company’s product line, and created reference implementations.
• Designed the pheno-genotypic mapping graph/ontology model.
• Evangelized the product by developing technical white papers and presentations.
• Enabled FDA 21 CFR Part 11 compliance across the product line.
Integration Solutions
• Business Cases: (1) Enroll patients, capture clinical data, extract genetic data, and transform and load to the genetic bank. Integration Scenarios: Architected SOA implementation using process integration, data integration, EAI, and Role-Based Access portals (patient/physician/researcher). (2) Integrate patients’ phenotypic and genetic data into a Genetic Safe. (3) Integrate the Generic Safe with the NCI Center for Bioinformatics Cancer Grid.
• Integration Scenarios: Data integration, data federation, and composite application.
• Technical Environment: WebLogic, Oracle, J2EE, LDAP, JAAS. Globus, DQP, Oracle, J2EE, data encryption technology, OWL Ontology, and Search.

More

marchFIRST / Whittman-Hart

Principal Consultant

Work Experience : 1998-2001

Participated in pre-sales/post-sales presentations and prototyping, defined data warehouse, and web-based architectural solutions for clients, create software architecture and design documents, worked with cross-functional teams and lead development teams towards successful implementation.
Selected list of projects:
• HealthMarket, inc, CT (www.HealthMarket.com B2C Online Health Plans and Services)
• ARMADA Hess, Woodbridge, NJ (www.Hesstoytruck.com B2C Online Marketing)
• Jones of New York, PA (Sales Tracking Data Warehouse)
• Avon Products, NY (International Product Distribution Data Warehouse)

Skills

More

BIDM

Technical Lead

Categories

Work Experience : 1993-1998

Worked with clients, and managed onsite and offshore project teams for Container Shipping application development. Enforced standards, code-reuse, code reviews, and unit/integration/system/acceptance testing.
Technical environment: Client/Server (PowerBuilder), IBM Mid-range AS/400, Oracle, PL/SQL.
Selected list of projects:
• Atlantic Container Line, S. Plainfield, NJ (Container Shipping Application)
• United Arab Shipping, Cranford, NJ/Kuwait (Container Shipping Application)

More

Prior Experience

Software Engineer / Systems Engineer

Work Experience : 1987-1993

From 1987 to 1993 worked as Software Engineer and Systems Engineer on consulting engagements executing projects (MAPICS/ERP implementations, Distributed DB/400 servers, Oracle Forms application) at multinational firms in Egypt (General Motors Egypt), Germany (IBM), and the United States.

More

Hire Essam