Custom Operator (don't know anything)

No comments
Controller Runtime: The Kubernetes controller-runtime Project is a set of go libraries for building Controllers Operator SDK: The Operator SDK is a framework that uses the controller-runtime library to make writing operators easier 

https://sdk.operatorframework.io/docs/building-operators/ansible/quickstart/
https://sdk.operatorframework.io/docs/building-operators/ansible/tutorial/
https://sdk.operatorframework.io/docs/building-operators/ansible/reference/dependent-watches/
https://docs.openshift.com/container-platform/4.7/operators/operator_sdk/ansible/osdk-ansible-inside-operator.html


https://itnext.io/a-practical-kubernetes-operator-using-ansible-an-example-d3a9d3674d5b
https://two-oes.medium.com/building-custom-ansible-based-operator-for-openshift-4-ec681fa0466d




Example:

operator-sdk init --domain operator.redhatgov.io --plugins ansible

operator-sdk create api --group workshops --version v1alpha1 --kind Workshop --generate-role


It will create some directory structure(defaults file will be available)



Creating CRD:

kubectl create -f config/crd/bases/workshops.operator.redhatgov.io_workshops.yaml


Run custom Operator:
command: make deploy


create custom resource now
-----

apiVersion: workshops.operator.redhatgov.io/v1alpha1

#workshops.operator.redhatgov.io/v1alpha1

kind: Workshop

metadata:

  name: example-workshop

spec:

  # Add fields here

  cr_my_replicas: 1



kubectl create -f config/crd/bases/workshops_v1_workshop_cr.yaml

you can check logs of custom - operator

kubectl logs -f workshop-controller-manager-9f6ff675b-rbcqw -n workshop-system



In RBAC role.yaml
I updated below

### added by prakash ###

  - apiGroups:

      - "*"

    resources:

      - "*"

    verbs:

      - "*"






No comments :

Post a Comment

GCP Data Life Cycle

No comments

 Data lifecycle


Mainly data life cycle has 4 steps:

1. Ingest - ( 
to pull in the raw data )
2. Store - ( tstore in a format that is durable and can be easily accessed)
3. Process and analyze - (data is transformed from raw form into actionable information)
4. Explore and Visualize - (to convert the results of the analysis into a format that is easy to draw insights from)




Ingest
1) ingesting app data
2) ingesting Streaming data
3) ingesting batch data





















Store



Cloud Storage:

backing up and archiving 
storage and delivery of content
* cloud storage can be accessed by dataflow for transformation and loading into other systems such as Bigtable or BigQuery.
* For Hadoop and Spark jobs, data from Cloud Storage can be natively accessed by using Dataproc.
* BigQuery natively supports importing CSV, JSON, and Avro files from a specified Cloud Storage bucket.


Cloud Storage for Firebase:
good fit for storing and retrieving assets such as images, audio, video, and other user-generated content in mobile and web apps.



Cloud SQL:

fully managed, cloud-native RDBMS that offers both MySQL and PostgreSQL engines with built-in support for replication.
offers built-in backup and restoration, high availability, and read replicas.
* Cloud SQL supports RDBMS workloads up to 30 TB for both MySQL and PostgreSQL
* Data stored in Cloud SQL is encrypted both in transit and at rest
* For OLTP Cloud SQL is appropriate
For OLAP workloads, consider BigQuery
*  If your workload requires dynamic schemas, consider Datastore.
*  You can use
Dataflow or Dataproc to create ETL jobs that pull data from Cloud SQL and insert it into other storage systems.



Bigtable: Managed wide-column NoSQL
managed, high-performance NoSQL database service designed for terabyte- to petabyte-scale workloads

provides consistent, low-latency, and high-throughput storage for large-scale NoSQL data
Bigtable is built for real-time app serving workloads, as well as large-scale analytical workloads.
* Use case:
 1) 
Real-time app data
 2) 
Stream processing (pub/sub => dataflow(transform) => BigTable)
 3) 
IoT time series data (sensor/streamed data => Bigtable (time series schema))
 4) AdTech workloads (can be used to store and track ad impressions which can be used by dataproc and dataflow for processing and analysing)
5) data ingestion (cloud storage => dataflow/dataproc => Bigtable)
6) Analytical Workloads (Bigtable=> dataflow (complex aggrregation) => dataproc((
Dataproc can be used to execute Hadoop or Spark processing and machine-learning tasks.))
7) Apache HBase replacement 

note: 
While Bigtable is considered an OLTP system, it doesn't support multi-row transactions, SQL queries or joins. For those use cases, consider either Cloud SQL or Datastore.


fully managed relational database service for mission-critical OLTP apps
Like relational databases, Spanner supports schemas, ACID transactions, and SQL queries
Spanner also performs automatic sharding while serving data with single-digit millisecond latencies
uses cases for Spanner:
   1) financial services (
strong consistency across read/write operations without scarificing HA)
   2) Ad tech (low-latency querying without compromising scale or availability.)
   3) Retail and Global Supply Chain(Spanner offers automatic, global, synchronous replication with low latency, which means that data is always consistent and highly available.)


* NoSQL database that stores JSON data
JSON data can be synchronized in real time to connected clients across different platforms
to build a real-time experience serving millions of users without compromising responsiveness.
* Use Cases:
  1) chat ans social (
Store and retrieve images, audio, video, and other user-generated content.)
  2) Mobile Games (Keep track of game progress and statistics across devices and device platforms)


=============================================================
(Process and analyze)
to derive business value and insights from data, you must transform and analyze it.

 


No comments :

Post a Comment

GCP Helicopter Racing League

No comments

Helicopter Racing League

Overview

  • Regional league
  • Offers a paid service to stream the races all over the world with live telemetry and predictions throughout each race.

Solution Concept

  •  migrate their existing service to a new platform 
  • to expand their use of managed AI and ML services to facilitate race predictions.
  • they want to move the serving of their content, both real-time and recorded, closer to their users.

Existing Technical Environment

  •  public cloud-first company
  • core of their mission-critical applications runs on their current public cloud provider.
  • Video recording and editing is performed at the race tracks,
  •  the content is encoded and transcoded, where needed, in the cloud.
  • Enterprise-grade connectivity and local compute is provided by truck-mounted mobile data centers.
  • Their race prediction services are hosted exclusively on their existing public cloud provider.[CloudML/Tensorflow/MLWorkflow]
  • Existing content is stored in an object storage service on their existing public cloud provider.[Google Storage bucket]
  • Video encoding and transcoding is performed on VMs created for each job.
  • Race predictions are performed using TensorFlow running on VMs in the current public cloud provider.[CloudML/Tensorflow]

Business Requirement

HRL’s owners want to expand their predictive capabilities and reduce latency for their viewers in emerging markets. Their requirements are:

  • Support ability to expose the predictive models to partners.[Cloud Endpoint]
  • Increase predictive capabilities during and before races:(Race result/Mechanical Failure/Crowd sentiment) [Data Ingestion/Data storage/ Processing]
  • Increase telemetry and create additional insights.
  • Measure fan engagement with new predictions.
  • Enhance global availability and quality of the broadcasts.
  • Increase the number of concurrent viewers.
  • Minimize operational complexity.
  • Ensure compliance with regulations.
  • Create a merchandising revenue stream .[Cloud Endpoint]

Technical Requirements   

  • Maintain or increase prediction throughput and accuracy.
  • Reduce viewer latency.
  • Increase transcoding performance.
  • Create real-time analytics of viewer consumption patterns and engagement.
  • Create a data mart to enable processing of large volumes of race data

Executive Statement

  •  enhanced video streams [AutoML Video Intelligence/Video Intelligence API]
  • to include predictions of events within the race (e.g., overtaking).
  •  Our current platform allows us to predict race outcomes but lacks the facility to support real-time predictions during races and the capacity to process season-long results.

 

=======================Our Analysis==========================

  • Streaming ==>> Data Processing
  • Predictions ==>> Machine Learning

Enterprise grade connectivity ==>>   


Holistic security: Enterprise-grade necessitates a holistic approach towards security, across products, processes, and applications.
Integration: Enterprise-grade expects tools and technologies to add to and extend existing tools, such that end users don’t have to face any disruption in their routine work.
Productivity: Enterprise-grade requires technologies to be attuned to the idea of “more work in less time,” such that users don’t feel inclined to use consumer-grade alternatives for more convenience.
Support: vendors need to support not only the internal stakeholder but also make sure the technology gels well into the larger distributed ecosystem.
Granular control: Enterprise-grade technology must offer companies deep control over policies for accessibility of content, and the ability to manage user environments suitable for specific user groups




Reference Link:
[AutoML Video Intelligence/Video Intelligence API]

[Cloud Endpoint]

[Data Ingestion/Data storage/ Processing]







services category




No comments :

Post a Comment