Modeling Data Warehouses using Apache Hive

4.2
15件の評価
提供:
Coursera Project Network
このガイド付きプロジェクトでは、次のことを行います。

Design a data warehouse

Build a data warehouse using Apache Hive over apache Hadoop

Query a data warehouse for specific records

Clock1 hour 40 minutes
Beginner初級
Cloudダウンロード不要
Video分割画面ビデオ
Comment Dots英語
Laptopデスクトップのみ

By the end of this project, you will create your own data warehouse by using a star schema modeling technique that holds baseball raw data gathered from multiple baseball leagues in the States. During the course of this project, you will be using Apache Hadoop and Apache Hive as the main tools to build your very own Data Warehouse using HQL as the querying language, which is very similar to SQL. The learner will be introduced to the star schema modeling technique, database creation and handling, data insertion, and data storage using two different methods in Apache Hadoop. Data Warehouses are viewed as the core of any application, as any further step that requires dealing with the data stored such as Data Analytics, building a recommendation system, or applying any sort of Machine Learning will inevitably point out to the start point, which is the data warehouse modeling. This guided project is for students who are aiming to start their careers in the broad field of Data Engineering and the specific field of Data Warehouses Modeling. It will also be of interest to people who are keen on exploring Big Data platforms such as Apache Hadoop and Apache Hive. Note: This project works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.

あなたが開発するスキル

Data Warehouse DevelopmentDatabase (DB) DesignData WarehousingDatabase Management Systems

ステップバイステップで学習します

ワークエリアを使用した分割画面で再生するビデオでは、講師がこれらの手順を説明します。

  1. By the end of Task 1,you will be able to identify the star schema format in our project description.

  2. By the end of Task 2, you will be able to identify the attributes and their data types.

  3. By the end of Task 3, you will be able to create a database and tables in the database using Apache Hadoop and Apache Hive.

  4. By the end of Task 4, you will be able to insert data into tables manually.

  5. By the end of Task 5, you will be able to amend tables by adding primary and foreign keys constraints.

  6. By the end of Task 6,   you will be able to apply useful queries to your data warehouse to fetch specific data records.

ガイド付きプロジェクトの仕組み

ワークスペースは、ブラウザに完全にロードされたクラウドデスクトップですので、ダウンロードは不要です

分割画面のビデオで、講師が手順ごとにガイドします

レビュー

MODELING DATA WAREHOUSES USING APACHE HIVE からの人気レビュー

すべてのレビューを見る

よくある質問

よくある質問

さらに質問がある場合は、受講者向けヘルプセンターにアクセスしてください。