<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question How to use Google Analytics API to import data to datawarehouse built on hadoop in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/How-to-use-Google-Analytics-API-to-import-data-to/m-p/153635#M116093</link>
    <description>&lt;P&gt;How to to use Google Analytics API to import data in hadoop file system? Is there a tool that can interact directly with the API and import data to the warehouse or Do I need to write programs for it? Could someone please point me to an example that shows import of data from Google Analytics to HDFS, more precisely to hive ?&lt;/P&gt;</description>
    <pubDate>Fri, 16 Sep 2022 10:20:28 GMT</pubDate>
    <dc:creator>simran_k</dc:creator>
    <dc:date>2022-09-16T10:20:28Z</dc:date>
    <item>
      <title>How to use Google Analytics API to import data to datawarehouse built on hadoop</title>
      <link>https://community.cloudera.com/t5/Support-Questions/How-to-use-Google-Analytics-API-to-import-data-to/m-p/153635#M116093</link>
      <description>&lt;P&gt;How to to use Google Analytics API to import data in hadoop file system? Is there a tool that can interact directly with the API and import data to the warehouse or Do I need to write programs for it? Could someone please point me to an example that shows import of data from Google Analytics to HDFS, more precisely to hive ?&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 10:20:28 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/How-to-use-Google-Analytics-API-to-import-data-to/m-p/153635#M116093</guid>
      <dc:creator>simran_k</dc:creator>
      <dc:date>2022-09-16T10:20:28Z</dc:date>
    </item>
    <item>
      <title>Re: How to use Google Analytics API to import data to datawarehouse built on hadoop</title>
      <link>https://community.cloudera.com/t5/Support-Questions/How-to-use-Google-Analytics-API-to-import-data-to/m-p/153636#M116094</link>
      <description>&lt;P&gt;Sounds like &lt;A href="https://community.hortonworks.com/questions/33961/how-to-import-data-return-by-google-analytic-s-api.html" target="_blank"&gt;https://community.hortonworks.com/questions/33961/how-to-import-data-return-by-google-analytic-s-api.html&lt;/A&gt; was a repost of this earlier question.  I provided my (more generic) answer over there, but maybe someone has a more specific response tied directly to Google Analytics and Hadoop.  Good luck!&lt;/P&gt;</description>
      <pubDate>Wed, 18 May 2016 21:03:38 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/How-to-use-Google-Analytics-API-to-import-data-to/m-p/153636#M116094</guid>
      <dc:creator>LesterMartin</dc:creator>
      <dc:date>2016-05-18T21:03:38Z</dc:date>
    </item>
    <item>
      <title>Re: How to use Google Analytics API to import data to datawarehouse built on hadoop</title>
      <link>https://community.cloudera.com/t5/Support-Questions/How-to-use-Google-Analytics-API-to-import-data-to/m-p/153637#M116095</link>
      <description>&lt;P&gt;We kind of built a data warehouse around the same idea that you have talked about in your article. &lt;/P&gt;&lt;P&gt;Integrating Salesforce and Google analytics as data-warehouse @infocaptor &lt;A href="http://www.infocaptor.com" target="_blank"&gt;http://www.infocaptor.com&lt;/A&gt;&lt;/P&gt;&lt;P&gt;The benefit is you can also co-relate with your financial data
When you design using GA api, you need to load the initial historical data for a certain date range. This has its own complications as you might run into segmentation issues, loss of data etc. You need to handle pagination etc. &lt;/P&gt;&lt;P&gt;Once the initial data load is complete, you then run it in incremental mode where you just bring new data only. This data gets appended to the same Data warehouse tables and does not cause duplicate with overlapping dates.
The minimum you would need to design is some kind of background daemon that runs everyday or at some frequency. &lt;/P&gt;&lt;P&gt;You will need job tables to monitor the success and failure of the extracts so that it can resume from where the error occurred. &lt;/P&gt;&lt;P&gt;Some of the other considerations &lt;/P&gt;&lt;P&gt;1. What happens if you run the extract for the same data range &lt;/P&gt;&lt;P&gt;2. What if a job fails for certain dates
It is important to set your primary keys for your DW target tables.&lt;/P&gt;&lt;P&gt;The extracted data is stored as CSV files and these can be easily pushed to Hadoop file system.&lt;/P&gt;</description>
      <pubDate>Sat, 21 Jan 2017 01:41:57 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/How-to-use-Google-Analytics-API-to-import-data-to/m-p/153637#M116095</guid>
      <dc:creator>contact</dc:creator>
      <dc:date>2017-01-21T01:41:57Z</dc:date>
    </item>
  </channel>
</rss>

