bigquery

BigQuery import and processing pipelines

  • 所有者: HTTPArchive/bigquery
  • 平台:
  • 許可證:
  • 分類:
  • 主題:
  • 喜歡:
    0
      比較:

Github星跟蹤圖

HTTP Archive + BigQuery data import

Note: you don't need to import this data yourself, the BigQuery dataset is public! Getting started.

However, if you do want your own private copy of the dataset... The following import and sync scripts will help you import the HTTP Archive dataset into BigQuery and keep it up to date.

$> sh sync.sh Jun_15_2013
$> sh sync.sh mobile_Jun_15_2013

That's all there is to it. The sync script handles all the necessary processing:

  • Archives are fetched from archive.org (and cached locally)
  • Archived CSV is transformed to BigQuery compatible escaping
    • You will need +pigz+ installed for parallel compression
  • Request files are split into <1GB compressed CSV's
  • Resulting pages and request data is synced to a Google Storage bucket
  • BigQuery import is kicked off for each of compressed archives on Google Storage

After the upload is complete, a copy of the latest tables can be made with:

$> bq.py cp runs.2013_06_15_pages runs.latest_pages
$> bq.py cp runs.2013_06_15_pages_mobile runs.latest_pages_mobile
$> bq.py cp runs.2013_06_15_requests runs.latest_requests
$> bq.py cp runs.2013_06_15_requests_mobiel runs.latest_requests_mobile

(MIT License) - Copyright (c) 2013 Ilya Grigorik

主要指標

概覽
名稱與所有者HTTPArchive/bigquery
主編程語言Jupyter Notebook
編程語言Ruby (語言數: 6)
平台
許可證
所有者活动
創建於2013-06-04 00:58:31
推送於2025-03-02 12:54:00
最后一次提交2025-03-02 12:53:58
發布數0
用户参与
星數67
關注者數13
派生數23
提交數502
已啟用問題?
問題數42
打開的問題數2
拉請求數149
打開的拉請求數0
關閉的拉請求數14
项目设置
已啟用Wiki?
已存檔?
是復刻?
已鎖定?
是鏡像?
是私有?