Abhijna D V
Member since 2022
Diamond League
48005 points
Member since 2022
Cloud technology can bring great value to an organization, and combining the power of cloud technology with data has the potential to unlock even more value and create new customer experiences. “Exploring Data Transformation with Google Cloud” explores the value data can bring to an organization and ways Google Cloud can make data useful and accessible. Part of the Cloud Digital Leader learning path, this course aims to help individuals grow in their role and build the future of their business.
This course aims to upskill Google Cloud partners to perform specific tasks of migrating data from Microsoft SQL Server to CloudSQL using the built-in replication capabilities of SQL Server. Sample data will be used during the migration. Learners will complete several labs that focus on the process of transferring schema, data, and related processes to corresponding Google Cloud products. One or more challenge labs will test the learner's understanding of the topics.
Processing streaming data is becoming increasingly popular as streaming enables businesses to get real-time metrics on business operations. This course covers how to build streaming data pipelines on Google Cloud. Pub/Sub is described for handling incoming streaming data. The course also covers how to apply aggregations and transformations to streaming data using Dataflow, and how to store processed records to BigQuery or Bigtable for analysis. Learners get hands-on experience building streaming data pipeline components on Google Cloud by using QwikLabs.
Data pipelines typically fall under one of the Extract and Load (EL), Extract, Load and Transform (ELT) or Extract, Transform and Load (ETL) paradigms. This course describes which paradigm should be used and when for batch data. Furthermore, this course covers several technologies on Google Cloud for data transformation including BigQuery, executing Spark on Dataproc, pipeline graphs in Cloud Data Fusion and serverless data processing with Dataflow. Learners get hands-on experience building data pipeline components on Google Cloud using Qwiklabs.
Migration from Oracle to Cloud Spanner using HarbourBridge. This course describes an example scenario that uses sample data during the migration. This process includes using HarbourBridge for Assessment, Schema Conversion, Schema Transformation, Data Migration, and supporting tools for data validation.
Migration from MySQL to Cloud Spanner using Dataflow that includes sample mock data and all necessary steps with initial assessment to validation including taking care of migrating users and grants.
This workload aims to upskill Google Cloud partners to perform specific tasks associated with priority workloads. Learners will perform the tasks for migrating data from AWS Redshift to BigQuery using BigQuery Data Transfer Service, which includes sample mock data. Learners will complete a challenge lab that focuses on the process of transferring both schema and data from a Redshift data warehouse to BigQuery.
This workload aims to upskill Google Cloud partners to perform specific tasks associated with priority workloads. Learners will perform the tasks of migrating data from Snowflake to BigQuery. Sample data will be used during the migration. Learners will complete several labs that focus on the process of transferring schema, data and related processes to corresponding Google Cloud products.There will be one or more challenge labs that will test the learners' understanding of the topics. "This learning path aims to upskill Google Cloud partners to perform specific tasks associated with priority workloads. Learners will perform the tasks of migrating data from Snowflake to BigQuery.
In this course, we see what the common challenges faced by data analysts are and how to solve them with the big data tools on Google Cloud. You’ll pick up some SQL along the way and become very familiar with using BigQuery and Dataprep to analyze and transform your datasets. This is the first course of the From Data to Insights with Google Cloud series. After completing this course, enroll in the Creating New BigQuery Datasets and Visualizing Insights course.
This course discusses the key elements of Google's Data Warehouse solution portfolio and strategy.
This course continues to explore the implementation of data load and transformation pipelines for a BigQuery Data Warehouse using Dataflow.
This course explores how to implement a streaming analytics solution using Pub/Sub.
This course explores how to implement a streaming analytics solution using Dataflow and BigQuery.
This course explores the Geographic Information Systems (GIS), GIS Visualization, and machine learning enhancements to BigQuery.
This course explores how to leverage Looker to create data experiences and gain insights with modern business intelligence (BI) and reporting.
Welcome to Intro to Data Lakes, where we discuss how to create a scalable and secure data lake on Google Cloud that allows enterprises to ingest, store, process, and analyze any type or volume of full fidelity data.
Welcome to Migrate Workflows, where we discuss how to migrate Spark and Hadoop tasks and workflows to Google Cloud.
Welcome to Data Governance, where we discuss how to implement data governance on Google Cloud.
This workload aims to upskill Google Cloud partners to perform specific tasks associated with priority workloads. Learners will perform the tasks of Migration from Teradata to BigQuery using the Data Transfer Service and the Teradata TPT Export Utility. Sample Data will be used during both methods. Learners will complete a challenge lab that focuses on the process of transferring both schema, data and SQL from a Teradata data warehouse to BigQuery.
In this course, you explore the four components that make up the BigQuery Migration Service. They are Migration Assessment, SQL Translation, Data Transfer Service, and Data Validation. You will use each of these tools to perform a migration using to BigQuery.
This course covers BigQuery fundamentals for professionals who are familiar with SQL-based cloud data warehouses in Snowflake and want to begin working in BigQuery. Through interactive lecture content and hands-on labs, you learn how to provision resources, create and share data assets, ingest data, and optimize query performance in BigQuery. Drawing upon your knowledge of Snowflake, you also learn about similarities and differences between Snowflake and BigQuery to help you get started with data warehouses in BigQuery. After this course, you can continue your BigQuery journey by completing the skill badge quest titled Build and Optimize Data Warehouses with BigQuery.
In this course, you will receive technical training for Enterprise Data Warehouses solutions using BigQuery based on the best practices developed internally by Google’s technical sales and services organizations. The course will also provide guidance and training on key technical challenges that can arise when migrating existing Enterprise Data Warehouses and ETL pipelines to Google Cloud. You will get hands-on experience with real migration tasks, such as data migration, schema optimization, and SQL Query conversion and optimization. The course will also cover key aspects of ETL pipeline migration to Dataproc as well as using Pub/Sub, Dataflow, and Cloud Data Fusion, giving you hands-on experience using all of these tools for Data Warehouse ETL pipelines.
This course identifies best practices for migrating data warehouses to BigQuery and the key skills required to perform successful migration.
Perform a migration from Oracle to BigQuery using SQL Translation and DataFlow using Sample Data. Learners will complete a quiz that focuses on the process of transferring both schema and data from an Oracle enterprise data warehouse to BigQuery.
This workload aims to upskill Google Cloud partners to perform specific tasks associated with priority workloads. Learners will perform the tasks of migrating data from five products hosted on Cloudera or Hortonworks to corresponding Google Cloud services and hosted products. The migration solutions addressed will be: HDFS data to Google Cloud Dataproc and Cloud Storage Hive data to Cloud Dataproc and the Cloud Dataproc Metastore Hive data to Google Cloud BigQuery Impala data to Google Cloud BigQuery HBase to Google Cloud Bigtable Sample data will be used during all five migrations. Learners will complete several labs that focus on the process of transferring schema, data and related processes to corresponding Google Cloud products.There will be one or more challenge labs that will test the learners understanding of the topics.
This course covers BigQuery fundamentals for professionals who are familiar with SQL-based cloud data warehouses in Redshift and want to begin working in BigQuery. Through interactive lecture content and hands-on labs, you learn how to provision resources, create and share data assets, ingest data, and optimize query performance in BigQuery. Drawing upon your knowledge of Redshift, you also learn about similarities and differences between Redshift and BigQuery to help you get started with data warehouses in BigQuery. After this course, you can continue your BigQuery journey by completing the skill badge quest titled Build and Optimize Data Warehouses with BigQuery.
This course introduces participants to MLOps tools and best practices for deploying, evaluating, monitoring and operating production ML systems on Google Cloud. MLOps is a discipline focused on the deployment, testing, monitoring, and automation of ML systems in production. Machine Learning Engineering professionals use tools for continuous improvement and evaluation of deployed models. They work with (or can be) Data Scientists, who develop models, to enable velocity and rigor in deploying the best performing models.
This course teaches you how to create an image captioning model by using deep learning. You learn about the different components of an image captioning model, such as the encoder and decoder, and how to train and evaluate your model. By the end of this course, you will be able to create your own image captioning models and use them to generate captions for images
Welcome to Design in BigQuery, where we map Enterprise Data Warehouse concepts and components to BigQuery and Google data services with a focus on schema design.
This course covers BigQuery fundamentals for professionals who are familiar with SQL-based cloud data warehouses in Teradata and want to begin working in BigQuery. Through interactive lecture content and hands-on labs, you learn how to provision resources, create and share data assets, ingest data, and optimize query performance in BigQuery. Drawing upon your knowledge of Teradata, you also learn about similarities and differences between Teradata and BigQuery to help you get started with data warehouses in BigQuery. After this course, you can continue your BigQuery journey by completing the skill badge quest titled Build and Optimize Data Warehouses with BigQuery.
This course provides partners the skills required to scope, design and deploy Document AI solutions for enterprise customers utilizing use-cases from both the procurement and lending arenas.
Incorporating machine learning into data pipelines increases the ability to extract insights from data. This course covers ways machine learning can be included in data pipelines on Google Cloud. For little to no customization, this course covers AutoML. For more tailored machine learning capabilities, this course introduces Notebooks and BigQuery machine learning (BigQuery ML). Also, this course covers how to productionalize machine learning solutions by using Vertex AI.
As the use of enterprise Artificial Intelligence and Machine Learning continues to grow, so too does the importance of building it responsibly. A challenge for many is that talking about responsible AI can be easier than putting it into practice. If you’re interested in learning how to operationalize responsible AI in your organization, this course is for you. In this course, you will learn how Google Cloud does this today, together with best practices and lessons learned, to serve as a framework for you to build your own responsible AI approach.
Welcome to "Virtual Agent Development in Dialogflow CX for Citizen Devs", the second course in the "Customer Experiences with Contact Center AI" series. In this course, learn how to develop customer conversational solutions using Contact Center Artificial Intelligence (CCAI). In this course, you'll be introduced to adding voice (telephony) as a communication channel to your virtual agent conversations using Dialogflow CX.
Welcome to "Virtual Agent Development in Dialogflow CX for Software Devs", the third course in the "Customer Experiences with Contact Center AI" series. In this course, learn how to develop more customized customer conversational solutions using Contact Center Artificial Intelligence (CCAI). In this course, you'll be introduced to more advanced and customized handling for virtual agent conversations that need to look up and convey dynamic data, and methods available to you for testing your virtual agent and logs which can be useful for understanding issues that arise. This is an intermediate course, intended for learners with the following type of role: Software developers: Codes computer software in a programming language (e.g., C++, Python, Javascript) and often using an SDK/API.
Welcome to "CCAI Operations and Implementation", the fourth course in the "Customer Experiences with Contact Center AI" series. In this course, learn some best practices for integrating conversational solutions with your existing contact center software, establishing a framework for human agent assistance, and implementing solutions securely and at scale. In this course, you'll be introduced to Agent Assist and the technology it uses so you can delight your customers with the efficiencies and accuracy of services provided when customers require human agents, connectivity protocols, APIs, and platforms which you can use to create an integration between your virtual agent and the services already established for your business, Dialogflow's Environment Management tool for deployment of different versions of your virtual agent for various purposes, compliance measures and regulations you should be aware of when bringing your virtual agent to production, and you'll be given tips from virtua…
Welcome to "Virtual Agent Development in Dialogflow ES for Citizen Devs", the second course in the "Customer Experiences with Contact Center AI" series. In this course, learn how to develop customer conversational solutions using Contact Center Artificial Intelligence (CCAI). You will use Dialogflow ES to create virtual agents and test them using the Dialogflow ES simulator. This course also provides best practices on developing virtual agents. You will also be introduced to adding voice (telephony) as a communication channel to your virtual agent conversations. Through a combination of presentations, demos, and hands-on labs, participants learn how to create virtual agents. This is an intermediate course, intended for learners with the following types of roles: Conversational designers: Designs the user experience of a virtual assistant. Translates the brand's business requirements into natural dialog flows. Citizen developers: Creates new business applications fo…
Welcome to "CCAI Virtual Agent Development in Dialogflow ES for Software Developers", the third course in the "Customer Experiences with Contact Center AI" series. In this course, learn to use additional features of Dialogflow ES for your virtual agent, create a Firestore instance to store customer data, and implement cloud functions that access the data. With the ability to read and write customer data, learner’s virtual agents are conversationally dynamic and able to defer contact center volume from human agents. You'll be introduced to methods for testing your virtual agent and logs which can be useful for understanding issues that arise. Lastly, learn about connectivity protocols, APIs, and platforms for integrating your virtual agent with services already established for your business.
Welcome to "CCAI Conversational Design Fundamentals", the first course in the "Customer Experiences with Contact Center AI" series. In this course, learn how to design customer conversational solutions using Contact Center Artificial Intelligence (CCAI). You will be introduced to CCAI and its three pillars (Dialogflow, Agent Assist, and Insights), and the concepts behind conversational experiences and how the study of them influences the design of your virtual agent. After taking this course you will be prepared to take your virtual agent design to the next level of intelligent conversation.
This workload aims to upskill Google Cloud partners to perform specific tasks for modernization using LookML on BigQuery. A proof-of-concept will take learners through the process of creating LookML visualizations on BigQuery. During this course, learners will be guided specifically on how to write Looker modeling language, also known as LookML and create semantic data models, and learn how LookML constructs SQL queries against BigQuery. At a high level, this course will focus on basic LookML to create and access BigQuery objects, and optimize BigQuery objects with LookML.
This course covers BigQuery fundamentals for professionals who are familiar with SQL-based cloud data warehouses in Oracle and want to begin working in BigQuery. Through interactive lecture content and hands-on labs, you learn how to provision resources, create and share data assets, ingest data, and optimize query performance in BigQuery. Drawing upon your knowledge of Oracle, you also learn about similarities and differences between Oracle and BigQuery to help you get started with data warehouses in BigQuery. After this course, you can continue your BigQuery journey by completing the skill badge quest titled Build and Optimize Data Warehouses with BigQuery.
This learning experience guides you through the process of utilizing various data sources and multiple Google Cloud products (including BigQuery and Google Sheets using Connected Sheets) to analyze, visualize, and interpret data to answer specific questions and share insights with key decision makers.
This course continues to explore the implementation of data load and transformation pipelines for a BigQuery Data Warehouse using Cloud Data Fusion.
This course explores the implementation of data load and transformation pipelines for a BigQuery Data Warehouse using Dataproc.
Welcome to Optimize in BigQuery, where we map Enterprise Data Warehouse concepts and components to BigQuery and Google data services with a focus on optimization.
The two key components of any data pipeline are data lakes and warehouses. This course highlights use-cases for each type of storage and dives into the available data lake and warehouse solutions on Google Cloud in technical detail. Also, this course describes the role of a data engineer, the benefits of a successful data pipeline to business operations, and examines why data engineering should be done in a cloud environment. This is the first course of the Data Engineering on Google Cloud series. After completing this course, enroll in the Building Batch Data Pipelines on Google Cloud course.
In this course, you learn how to do the kind of data exploration and analysis in Looker that would formerly be done primarily by SQL developers or analysts. Upon completion of this course, you will be able to leverage Looker's modern analytics platform to find and explore relevant content in your organization’s Looker instance, ask questions of your data, create new metrics as needed, and build and share visualizations and dashboards to facilitate data-driven decision making.
This course introduces the Google Cloud big data and machine learning products and services that support the data-to-AI lifecycle. It explores the processes, challenges, and benefits of building a big data pipeline and machine learning models with Vertex AI on Google Cloud.