This document discusses architectural considerations for analyzing clickstream data using Hadoop. It covers choices for data storage layers like HDFS vs HBase, data formats like Avro and Parquet, partitioning strategies, and data ingestion using tools like Flume and Kafka. It also discusses processing engines like MapReduce, Spark and Impala and how they can be used to sessionize data and perform other analytics.