加入 登录

Chinwong Tangchringchai

成为会员时间:2023

白银联赛

2600 积分
Generative AI Fundamentals - 简体中文 Earned Jun 14, 2023 EDT
负责任的 AI 简介 Earned Jun 14, 2023 EDT
大型语言模型简介 Earned Jun 13, 2023 EDT
生成式 AI 简介 Earned Jun 12, 2023 EDT
Preparing for your Professional Data Engineer Journey Earned Feb 28, 2023 EST
在 Google Cloud 上为机器学习 API 准备数据 Earned Feb 23, 2023 EST
利用 BigQuery ML 构建预测模型时的数据工程处理 Earned Feb 22, 2023 EST
Serverless Data Processing with Dataflow: Operations Earned Feb 22, 2023 EST
Serverless Data Processing with Dataflow: Develop Pipelines Earned Feb 20, 2023 EST
Serverless Data Processing with Dataflow: Foundations Earned Feb 13, 2023 EST
Smart Analytics, Machine Learning, and AI on Google Cloud Earned Feb 13, 2023 EST
Building Resilient Streaming Analytics Systems on Google Cloud Earned Feb 3, 2023 EST
Building Batch Data Pipelines on Google Cloud Earned Feb 1, 2023 EST
在 Compute Engine 上实现负载均衡 Earned Jan 25, 2023 EST
Modernizing Data Lakes and Data Warehouses with Google Cloud Earned Jan 24, 2023 EST
Google Cloud Big Data and Machine Learning Fundamentals Earned Jan 20, 2023 EST
[DEPRECATED] Data Engineering Earned Jan 16, 2023 EST

完成 Introduction to Generative AI、Introduction to Large Language Models 和 Introduction to Responsible AI 三门课程,赢取技能徽章。通过最终测验,即表明您理解了生成式 AI 的基本概念。 技能徽章是由 Google Cloud 颁发的数字徽章,旨在认可您对 Google Cloud 产品与服务的了解程度。公开您的个人资料并将技能徽章添加到您的社交媒体个人资料中,以此来分享您获得的成就。

了解详情

这是一节入门级微课程,旨在解释什么是负责任的 AI、它的重要性,以及 Google 如何在自己的产品中实现负责任的 AI。此外,本课程还介绍了 Google 的 7 个 AI 开发原则。

了解详情

这是一节入门级微学习课程,探讨什么是大型语言模型 (LLM)、适合的应用场景以及如何使用提示调整来提升 LLM 性能,还介绍了可以帮助您开发自己的 Gen AI 应用的各种 Google 工具。

了解详情

这是一节入门级微课程,旨在解释什么是生成式 AI、它的用途以及与传统机器学习方法的区别。该课程还介绍了可以帮助您开发自己的生成式 AI 应用的各种 Google 工具。

了解详情

This course helps learners create a study plan for the PDE (Professional Data Engineer) certification exam. Learners explore the breadth and scope of the domains covered in the exam. Learners assess their exam readiness and create their individual study plan.

了解详情

完成入门级技能徽章课程在 Google Cloud 上为机器学习 API 准备数据,展示以下技能: 使用 Dataprep by Trifacta 清理数据、在 Dataflow 中运行数据流水线、在 Dataproc 中创建集群和运行 Apache Spark 作业,以及调用机器学习 API,包括 Cloud Natural Language API、Google Cloud Speech-to-Text API 和 Video Intelligence API。 技能徽章是由 Google Cloud 颁发的专属数字徽章,旨在认可您在 Google Cloud 产品与服务方面的熟练度; 您需要在交互式实操环境中参加考核,证明自己运用所学知识的能力后才能获得。完成此技能徽章课程和作为最终评估的实验室挑战赛, 获得技能徽章,在您的人际圈中炫出自己的技能。

了解详情

完成中级技能徽章课程利用 BigQuery ML 构建预测模型时的数据工程处理, 展示自己在以下方面的技能:利用 Dataprep by Trifacta 构建 BigQuery 数据转换流水线; 利用 Cloud Storage、Dataflow 和 BigQuery 构建提取、转换和加载 (ETL) 工作流; 以及利用 BigQuery ML 构建机器学习模型。 技能徽章是由 Google Cloud 颁发的专属数字徽章,旨在认可 您对 Google Cloud 产品与服务的熟练度;您需要在 交互式实操环境中参加考核,证明自己运用所学知识的能力后才能获得此徽章。完成技能徽章课程和 作为最终评估的实验室挑战赛,即可获得数字徽章, 在您的人际圈中炫出自己的技能。

了解详情

In the last installment of the Dataflow course series, we will introduce the components of the Dataflow operational model. We will examine tools and techniques for troubleshooting and optimizing pipeline performance. We will then review testing, deployment, and reliability best practices for Dataflow pipelines. We will conclude with a review of Templates, which makes it easy to scale Dataflow pipelines to organizations with hundreds of users. These lessons will help ensure that your data platform is stable and resilient to unanticipated circumstances.

了解详情

In this second installment of the Dataflow course series, we are going to be diving deeper on developing pipelines using the Beam SDK. We start with a review of Apache Beam concepts. Next, we discuss processing streaming data using windows, watermarks and triggers. We then cover options for sources and sinks in your pipelines, schemas to express your structured data, and how to do stateful transformations using State and Timer APIs. We move onto reviewing best practices that help maximize your pipeline performance. Towards the end of the course, we introduce SQL and Dataframes to represent your business logic in Beam and how to iteratively develop pipelines using Beam notebooks.

了解详情

This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.

了解详情

Incorporating machine learning into data pipelines increases the ability to extract insights from data. This course covers ways machine learning can be included in data pipelines on Google Cloud. For little to no customization, this course covers AutoML. For more tailored machine learning capabilities, this course introduces Notebooks and BigQuery machine learning (BigQuery ML). Also, this course covers how to productionalize machine learning solutions by using Vertex AI.

了解详情

Processing streaming data is becoming increasingly popular as streaming enables businesses to get real-time metrics on business operations. This course covers how to build streaming data pipelines on Google Cloud. Pub/Sub is described for handling incoming streaming data. The course also covers how to apply aggregations and transformations to streaming data using Dataflow, and how to store processed records to BigQuery or Bigtable for analysis. Learners get hands-on experience building streaming data pipeline components on Google Cloud by using QwikLabs.

了解详情

Data pipelines typically fall under one of the Extract and Load (EL), Extract, Load and Transform (ELT) or Extract, Transform and Load (ETL) paradigms. This course describes which paradigm should be used and when for batch data. Furthermore, this course covers several technologies on Google Cloud for data transformation including BigQuery, executing Spark on Dataproc, pipeline graphs in Cloud Data Fusion and serverless data processing with Dataflow. Learners get hands-on experience building data pipeline components on Google Cloud using Qwiklabs.

了解详情

完成入门级在 Compute Engine 上实现负载均衡技能徽章课程,展示自己在以下方面的技能: 编写 gcloud 命令和使用 Cloud Shell,在 Compute Engine 中创建和部署虚拟机, 以及配置网络和 HTTP 负载均衡器。 技能徽章是由 Google Cloud 颁发的专属数字徽章, 旨在认可您在 Google Cloud 产品与服务方面的熟练度; 该课程会检验您在交互式实操环境中运用所学知识的 能力。完成此技能徽章课程和作为最终评估的实验室挑战赛, 即可获得技能徽章,并在您的圈子中秀一秀。

了解详情

The two key components of any data pipeline are data lakes and warehouses. This course highlights use-cases for each type of storage and dives into the available data lake and warehouse solutions on Google Cloud in technical detail. Also, this course describes the role of a data engineer, the benefits of a successful data pipeline to business operations, and examines why data engineering should be done in a cloud environment. This is the first course of the Data Engineering on Google Cloud series. After completing this course, enroll in the Building Batch Data Pipelines on Google Cloud course.

了解详情

This course introduces the Google Cloud big data and machine learning products and services that support the data-to-AI lifecycle. It explores the processes, challenges, and benefits of building a big data pipeline and machine learning models with Vertex AI on Google Cloud.

了解详情

This advanced-level quest is unique amongst the other catalog offerings. The labs have been curated to give IT professionals hands-on practice with topics and services that appear in the Google Cloud Certified Professional Data Engineer Certification. From Big Query, to Dataprep, to Cloud Composer, this quest is composed of specific labs that will put your Google Cloud data engineering knowledge to the test. Be aware that while practice with these labs will increase your skills and abilities, you will need other preparation, too. The exam is quite challenging and external studying, experience, and/or background in cloud data engineering is recommended. Looking for a hands on challenge lab to demonstrate your skills and validate your knowledge? On completing this quest, enroll in and finish the additional challenge lab at the end of the Engineer Data in the Google Cloud to receive an exclusive Google Cloud digital badge.

了解详情