#1
  1. No Profile Picture
    Contributing User
    Devshed Regular (2000 - 2499 posts)

    Join Date
    Sep 2006
    Posts
    2,030
    Rep Power
    535

    How does Google Analytics determine real users?


    We all now know that Google Anal (oops, I mean Analytics) provides a reporting API to request a page view count by URL. And we also all now concede that GA is amazing.

    My question is purely academic and I have no practical use at the moment, however, I am curious how GA does it? Is it server side browser sniffing? Is there a client side JS component, and if so, does it not count non-JS client hits? Does it constantly change to keep ahead of the newest Bots? Does it use some totally different AI?
  2. #2
  3. --
    Devshed Expert (3500 - 3999 posts)

    Join Date
    Jul 2012
    Posts
    3,959
    Rep Power
    1014
    Hi,

    I'm not aware of any member who works at Google, so not sure what you expect from us. A bunch of speculations?
    The 6 worst sins of security ē How to (properly) access a MySQL database with PHP

    Why canít I use certain words like "drop" as part of my Security Question answers?
    There are certain words used by hackers to try to gain access to systems and manipulate data; therefore, the following words are restricted: "select," "delete," "update," "insert," "drop" and "null".
  4. #3
  5. No Profile Picture
    Contributing User
    Devshed Regular (2000 - 2499 posts)

    Join Date
    Sep 2006
    Posts
    2,030
    Rep Power
    535
    Originally Posted by Jacques1
    Hi,

    I'm not aware of any member who works at Google, so not sure what you expect from us. A bunch of speculations?
    Just a logical hypothesis.
  6. #4
  7. Wiser? Not exactly.
    Devshed God 1st Plane (5500 - 5999 posts)

    Join Date
    May 2001
    Location
    Bonita Springs, FL
    Posts
    5,952
    Rep Power
    4033
    My guess would be they start with some basic useragent / ip filtering of known bots and then move on to some heuristic algorithms that use data gathered across several several sites and across time to help filter things further.

    For example if you see a particular agent requesting a page on a site every couple of seconds for an extended period you could reasonably assume that is a bot crawling the site.

    If you see an agent hitting a bunch of completely unrelated sites consistently or at the same time you could make an assumption that they are likely a bot as well.

    A key thing to keep in mind is that even though GA may only show you information relevant to your site, it has data from thousands of sites at it's disposal when attempting to determine patterns or trends in browsing behaviors. All that data plus a few smart people can lead to some fairly accurate results.
    Recycle your old CD's, don't just trash them



    If I helped you out, show some love with some reputation, or tip with Bitcoins to 1N645HfYf63UbcvxajLKiSKpYHAq2Zxud
  8. #5
  9. No Profile Picture
    Contributing User
    Devshed Regular (2000 - 2499 posts)

    Join Date
    Sep 2006
    Posts
    2,030
    Rep Power
    535
    Originally Posted by kicken
    All that data plus a few smart people can lead to some fairly accurate results.
    Thanks Kicken, I appreciate your response. Seems like a very reasonable guess, and believe you are correct.

IMN logo majestic logo threadwatch logo seochat tools logo