As tecnologias de nuvem podem agregar muito valor a uma organização e, ao combinar esse poder com dados, o potencial de crescer e criar novas experiências para os clientes é ainda maior. O curso "Como é feita a transformação de dados com o Google Cloud" mostra como os dados agregam valor às organizações e como o Google Cloud torna esses dados eficientes e acessíveis. Este curso, que faz parte do programa de aprendizado do Líder digital do Cloud, se destina às pessoas que querem crescer na profissão e construir o futuro da empresa.
This course aims to upskill Google Cloud partners to perform specific tasks of migrating data from Microsoft SQL Server to CloudSQL using the built-in replication capabilities of SQL Server. Sample data will be used during the migration. Learners will complete several labs that focus on the process of transferring schema, data, and related processes to corresponding Google Cloud products. One or more challenge labs will test the learner's understanding of the topics.
O processamento de dados de streaming é cada vez mais usado pelas empresas para gerar métricas sobre as operações comerciais em tempo real. Neste curso, você vai aprender a criar pipelines de dados de streaming no Google Cloud. O Pub/Sub é apresentado como a ferramenta para gerenciar dados de streaming de entrada. No curso, também abordamos a aplicação de agregações e transformações a dados de streaming usando o Dataflow, além de formas de armazenar registros processados no BigQuery ou no Bigtable para análise. Os participantes vão ganhar experiência prática na criação de componentes de pipelines de dados de streaming no Google Cloud usando o Qwiklabs.
Os pipelines de dados geralmente se encaixam em um desses três paradigmas: extração e carregamento (EL), extração, carregamento e transformação (ELT) ou extração, transformação e carregamento (ETL). Este curso descreve qual paradigma deve ser usado em determinadas situações e quando isso ocorre com dados em lote. Além disso, vamos falar sobre várias tecnologias no Google Cloud para transformação de dados, incluindo o BigQuery, a execução do Spark no Dataproc, gráficos de pipeline no Cloud Data Fusion e processamento de dados sem servidor com o Dataflow. Os participantes vão ganhar experiência prática na criação de componentes de pipelines de dados no Google Cloud usando o Qwiklabs.
Migration from Oracle to Cloud Spanner using HarbourBridge. This course describes an example scenario that uses sample data during the migration. This process includes using HarbourBridge for Assessment, Schema Conversion, Schema Transformation, Data Migration, and supporting tools for data validation.
Migration from MySQL to Cloud Spanner using Dataflow that includes sample mock data and all necessary steps with initial assessment to validation including taking care of migrating users and grants.
This workload aims to upskill Google Cloud partners to perform specific tasks associated with priority workloads. Learners will perform the tasks for migrating data from AWS Redshift to BigQuery using BigQuery Data Transfer Service, which includes sample mock data. Learners will complete a challenge lab that focuses on the process of transferring both schema and data from a Redshift data warehouse to BigQuery.
This workload aims to upskill Google Cloud partners to perform specific tasks associated with priority workloads. Learners will perform the tasks of migrating data from Snowflake to BigQuery. Sample data will be used during the migration. Learners will complete several labs that focus on the process of transferring schema, data and related processes to corresponding Google Cloud products.There will be one or more challenge labs that will test the learners' understanding of the topics. "This learning path aims to upskill Google Cloud partners to perform specific tasks associated with priority workloads. Learners will perform the tasks of migrating data from Snowflake to BigQuery.
Neste curso, conhecemos os desafios mais comuns enfrentados pelos analistas de dados e como resolvê-los com as ferramentas de big data no Google Cloud. Ao longo do caminho, você vai aprender um pouco de SQL e se familiarizar com o uso do BigQuery e do Dataprep para analisar e transformar seus conjuntos de dados. Este é o primeiro curso da série From Data to Insights with Google Cloud. Depois de concluir este curso, inscreva-se no curso Creating New BigQuery Datasets and Visualizing Insights.
This course discusses the key elements of Google's Data Warehouse solution portfolio and strategy.
This course continues to explore the implementation of data load and transformation pipelines for a BigQuery Data Warehouse using Dataflow.
This course explores how to implement a streaming analytics solution using Pub/Sub.
This course explores how to implement a streaming analytics solution using Dataflow and BigQuery.
This course explores the Geographic Information Systems (GIS), GIS Visualization, and machine learning enhancements to BigQuery.
This course explores how to leverage Looker to create data experiences and gain insights with modern business intelligence (BI) and reporting.
Welcome to Intro to Data Lakes, where we discuss how to create a scalable and secure data lake on Google Cloud that allows enterprises to ingest, store, process, and analyze any type or volume of full fidelity data.
Welcome to Migrate Workflows, where we discuss how to migrate Spark and Hadoop tasks and workflows to Google Cloud.
Welcome to Data Governance, where we discuss how to implement data governance on Google Cloud.
This workload aims to upskill Google Cloud partners to perform specific tasks associated with priority workloads. Learners will perform the tasks of Migration from Teradata to BigQuery using the Data Transfer Service and the Teradata TPT Export Utility. Sample Data will be used during both methods. Learners will complete a challenge lab that focuses on the process of transferring both schema, data and SQL from a Teradata data warehouse to BigQuery.
In this course, you explore the four components that make up the BigQuery Migration Service. They are Migration Assessment, SQL Translation, Data Transfer Service, and Data Validation. You will use each of these tools to perform a migration using to BigQuery.
This course covers BigQuery fundamentals for professionals who are familiar with SQL-based cloud data warehouses in Snowflake and want to begin working in BigQuery. Through interactive lecture content and hands-on labs, you learn how to provision resources, create and share data assets, ingest data, and optimize query performance in BigQuery. Drawing upon your knowledge of Snowflake, you also learn about similarities and differences between Snowflake and BigQuery to help you get started with data warehouses in BigQuery. After this course, you can continue your BigQuery journey by completing the skill badge quest titled Build and Optimize Data Warehouses with BigQuery.
In this course, you will receive technical training for Enterprise Data Warehouses solutions using BigQuery based on the best practices developed internally by Google’s technical sales and services organizations. The course will also provide guidance and training on key technical challenges that can arise when migrating existing Enterprise Data Warehouses and ETL pipelines to Google Cloud. You will get hands-on experience with real migration tasks, such as data migration, schema optimization, and SQL Query conversion and optimization. The course will also cover key aspects of ETL pipeline migration to Dataproc as well as using Pub/Sub, Dataflow, and Cloud Data Fusion, giving you hands-on experience using all of these tools for Data Warehouse ETL pipelines.
This course identifies best practices for migrating data warehouses to BigQuery and the key skills required to perform successful migration.
Perform a migration from Oracle to BigQuery using SQL Translation and DataFlow using Sample Data. Learners will complete a quiz that focuses on the process of transferring both schema and data from an Oracle enterprise data warehouse to BigQuery.
This workload aims to upskill Google Cloud partners to perform specific tasks associated with priority workloads. Learners will perform the tasks of migrating data from five products hosted on Cloudera or Hortonworks to corresponding Google Cloud services and hosted products. The migration solutions addressed will be: HDFS data to Google Cloud Dataproc and Cloud Storage Hive data to Cloud Dataproc and the Cloud Dataproc Metastore Hive data to Google Cloud BigQuery Impala data to Google Cloud BigQuery HBase to Google Cloud Bigtable Sample data will be used during all five migrations. Learners will complete several labs that focus on the process of transferring schema, data and related processes to corresponding Google Cloud products.There will be one or more challenge labs that will test the learners understanding of the topics.
This course covers BigQuery fundamentals for professionals who are familiar with SQL-based cloud data warehouses in Redshift and want to begin working in BigQuery. Through interactive lecture content and hands-on labs, you learn how to provision resources, create and share data assets, ingest data, and optimize query performance in BigQuery. Drawing upon your knowledge of Redshift, you also learn about similarities and differences between Redshift and BigQuery to help you get started with data warehouses in BigQuery. After this course, you can continue your BigQuery journey by completing the skill badge quest titled Build and Optimize Data Warehouses with BigQuery.
Neste curso, os participantes vão conhecer as ferramentas de MLOps e as práticas recomendadas para a implantação, a avaliação, o monitoramento e a operação de sistemas de ML de produção no Google Cloud. MLOps é uma disciplina com foco na implantação, no teste, no monitoramento e na automação de sistemas de ML em produção. Profissionais de engenharia de machine learning usam ferramentas para fazer melhorias contínuas e avaliações de modelos implantados. São profissionais que trabalham com ciências de dados e desenvolvem modelos para garantir a velocidade e o rigor na implantação de modelos com melhor desempenho.
Neste curso, ensinamos a criar um modelo de legenda para imagens usando aprendizado profundo. Você vai aprender sobre os diferentes componentes de um modelo de legenda para imagens, como o codificador e decodificador, e de que forma treinar e avaliar seu modelo. Ao final deste curso, você será capaz de criar e usar seus próprios modelos de legenda para imagens.
Welcome to Design in BigQuery, where we map Enterprise Data Warehouse concepts and components to BigQuery and Google data services with a focus on schema design.
This course covers BigQuery fundamentals for professionals who are familiar with SQL-based cloud data warehouses in Teradata and want to begin working in BigQuery. Through interactive lecture content and hands-on labs, you learn how to provision resources, create and share data assets, ingest data, and optimize query performance in BigQuery. Drawing upon your knowledge of Teradata, you also learn about similarities and differences between Teradata and BigQuery to help you get started with data warehouses in BigQuery. After this course, you can continue your BigQuery journey by completing the skill badge quest titled Build and Optimize Data Warehouses with BigQuery.
This course provides partners the skills required to scope, design and deploy Document AI solutions for enterprise customers utilizing use-cases from both the procurement and lending arenas.
A incorporação de machine learning em pipelines de dados aumenta a capacidade de extrair insights dessas informações. Neste curso, mostramos as várias formas de incluir essa tecnologia em pipelines de dados do Google Cloud. Para casos de pouca ou nenhuma personalização, vamos falar sobre o AutoML. Para usar recursos de machine learning mais personalizados, vamos apresentar os Notebooks e o machine learning do BigQuery (BigQuery ML). No curso, você também vai aprender sobre a produção de soluções de machine learning usando a Vertex AI.
Quanto maior é o uso da inteligência artificial empresarial e do machine learning, mais importante é desenvolvê-los de maneira responsável. Para muitos, falar sobre a IA responsável pode ser mais fácil, mas colocá-la em prática é um desafio. Se você tem interesse em aprender a operacionalizar a IA responsável na sua organização, este curso é para você. Nele, você vai aprender como o Google Cloud faz isso hoje, além de analisar práticas recomendadas e lições aprendidas, a fim de criar uma base para elaborar sua própria abordagem de IA responsável.
Welcome to "Virtual Agent Development in Dialogflow CX for Citizen Devs", the second course in the "Customer Experiences with Contact Center AI" series. In this course, learn how to develop customer conversational solutions using Contact Center Artificial Intelligence (CCAI). In this course, you'll be introduced to adding voice (telephony) as a communication channel to your virtual agent conversations using Dialogflow CX.
Welcome to "Virtual Agent Development in Dialogflow CX for Software Devs", the third course in the "Customer Experiences with Contact Center AI" series. In this course, learn how to develop more customized customer conversational solutions using Contact Center Artificial Intelligence (CCAI). In this course, you'll be introduced to more advanced and customized handling for virtual agent conversations that need to look up and convey dynamic data, and methods available to you for testing your virtual agent and logs which can be useful for understanding issues that arise. This is an intermediate course, intended for learners with the following type of role: Software developers: Codes computer software in a programming language (e.g., C++, Python, Javascript) and often using an SDK/API.
Welcome to "CCAI Operations and Implementation", the fourth course in the "Customer Experiences with Contact Center AI" series. In this course, learn some best practices for integrating conversational solutions with your existing contact center software, establishing a framework for human agent assistance, and implementing solutions securely and at scale. In this course, you'll be introduced to Agent Assist and the technology it uses so you can delight your customers with the efficiencies and accuracy of services provided when customers require human agents, connectivity protocols, APIs, and platforms which you can use to create an integration between your virtual agent and the services already established for your business, Dialogflow's Environment Management tool for deployment of different versions of your virtual agent for various purposes, compliance measures and regulations you should be aware of when bringing your virtual agent to production, and you'll be given tips from virtua…
Welcome to "Virtual Agent Development in Dialogflow ES for Citizen Devs", the second course in the "Customer Experiences with Contact Center AI" series. In this course, learn how to develop customer conversational solutions using Contact Center Artificial Intelligence (CCAI). You will use Dialogflow ES to create virtual agents and test them using the Dialogflow ES simulator. This course also provides best practices on developing virtual agents. You will also be introduced to adding voice (telephony) as a communication channel to your virtual agent conversations. Through a combination of presentations, demos, and hands-on labs, participants learn how to create virtual agents. This is an intermediate course, intended for learners with the following types of roles: Conversational designers: Designs the user experience of a virtual assistant. Translates the brand's business requirements into natural dialog flows. Citizen developers: Creates new business applications fo…
Welcome to "CCAI Virtual Agent Development in Dialogflow ES for Software Developers", the third course in the "Customer Experiences with Contact Center AI" series. In this course, learn to use additional features of Dialogflow ES for your virtual agent, create a Firestore instance to store customer data, and implement cloud functions that access the data. With the ability to read and write customer data, learner’s virtual agents are conversationally dynamic and able to defer contact center volume from human agents. You'll be introduced to methods for testing your virtual agent and logs which can be useful for understanding issues that arise. Lastly, learn about connectivity protocols, APIs, and platforms for integrating your virtual agent with services already established for your business.
Welcome to "CCAI Conversational Design Fundamentals", the first course in the "Customer Experiences with Contact Center AI" series. In this course, learn how to design customer conversational solutions using Contact Center Artificial Intelligence (CCAI). You will be introduced to CCAI and its three pillars (Dialogflow, Agent Assist, and Insights), and the concepts behind conversational experiences and how the study of them influences the design of your virtual agent. After taking this course you will be prepared to take your virtual agent design to the next level of intelligent conversation.
This workload aims to upskill Google Cloud partners to perform specific tasks for modernization using LookML on BigQuery. A proof-of-concept will take learners through the process of creating LookML visualizations on BigQuery. During this course, learners will be guided specifically on how to write Looker modeling language, also known as LookML and create semantic data models, and learn how LookML constructs SQL queries against BigQuery. At a high level, this course will focus on basic LookML to create and access BigQuery objects, and optimize BigQuery objects with LookML.
This course covers BigQuery fundamentals for professionals who are familiar with SQL-based cloud data warehouses in Oracle and want to begin working in BigQuery. Through interactive lecture content and hands-on labs, you learn how to provision resources, create and share data assets, ingest data, and optimize query performance in BigQuery. Drawing upon your knowledge of Oracle, you also learn about similarities and differences between Oracle and BigQuery to help you get started with data warehouses in BigQuery. After this course, you can continue your BigQuery journey by completing the skill badge quest titled Build and Optimize Data Warehouses with BigQuery.
This learning experience guides you through the process of utilizing various data sources and multiple Google Cloud products (including BigQuery and Google Sheets using Connected Sheets) to analyze, visualize, and interpret data to answer specific questions and share insights with key decision makers.
This course continues to explore the implementation of data load and transformation pipelines for a BigQuery Data Warehouse using Cloud Data Fusion.
This course explores the implementation of data load and transformation pipelines for a BigQuery Data Warehouse using Dataproc.
Welcome to Optimize in BigQuery, where we map Enterprise Data Warehouse concepts and components to BigQuery and Google data services with a focus on optimization.
Os dois principais componentes de um pipeline de dados são data lakes e warehouses. Neste curso, destacamos os casos de uso para cada tipo de armazenamento e as soluções de data lake e warehouse disponíveis no Google Cloud de forma detalhada e técnica. Além disso, também descrevemos o papel de um engenheiro de dados, os benefícios de um pipeline de dados funcional para operações comerciais e analisamos por que a engenharia de dados deve ser feita em um ambiente de nuvem. Este é o primeiro curso da série "Engenharia de dados no Google Cloud". Após a conclusão, recomendamos que você comece o curso "Como criar pipelines de dados em lote no Google Cloud".
In this course, you learn how to do the kind of data exploration and analysis in Looker that would formerly be done primarily by SQL developers or analysts. Upon completion of this course, you will be able to leverage Looker's modern analytics platform to find and explore relevant content in your organization’s Looker instance, ask questions of your data, create new metrics as needed, and build and share visualizations and dashboards to facilitate data-driven decision making.
Este curso apresenta os produtos e serviços de Big Data e machine learning do Google Cloud que auxiliam no ciclo de vida de dados para IA. Ele explica os processos, os desafios e os benefícios de criar um pipeline de Big Data e modelos de machine learning com a Vertex AI no Google Cloud.