Is there something that can generate random Internet usage to make the real sites I go to a bit obfuscated?

I’m thinking something that runs on my server, and simply visits a random website. It probably shouldn’t actually be random, and some sort of tweaking would be great. Like the ability to have it visit every news site there is. That way the ISP will have a harder time telling my political bias.

The threat model for this is below using a VPN for normal usage, although getting a dedicated VPN IP address is a project for one day.

          • sugar_in_your_tea@sh.itjust.works
            link
            fedilink
            arrow-up
            1
            ·
            4 months ago

            It helps a lot. Someone would need to go through the effort of tracking your VPN service and correlating that to you.

            OP seems to worry about their ISP, and their ISP can’t see anything at the VPN, so they’d need to go out of their way to gather that info. So as long as OP trusts the VPN, they’re probably fine from everyone but state actors and well funded private investigators.

            If you want to go beyond that, Tor is your next option. That should be effective against all but the most determined state actors.

    • Dust0741@lemmy.worldOP
      link
      fedilink
      arrow-up
      5
      arrow-down
      2
      ·
      4 months ago

      As I mentioned I have a server, and I use a VPN to connect always to it. This makes using a paid VPN a bit harder. The dedicated VPN IP should fix this issue but I haven’t looked into how difficult that’d be.

      • Imprint9816@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        5
        ·
        edit-2
        4 months ago

        Ahh i see.

        Yeah it really slims down your VPN choices as having an IP address associated with your account makes it much more identifiable. So some providers wont offer them (such as mullvad).

        It also usually costs more. The one I know offers a static IP is express VPN and ive heard Proton has plans on offering it. It looks like PIA offers it too.

        • Dust0741@lemmy.worldOP
          link
          fedilink
          arrow-up
          2
          ·
          4 months ago

          Yup. Tailscale+Mullvad isn’t a bad option, but I’d rather not depend on tailscale and a true local connection will always be better. But then you have to pay through tailscale and then more identifiable.

      • sugar_in_your_tea@sh.itjust.works
        link
        fedilink
        arrow-up
        1
        ·
        4 months ago

        So are you worried about your server’s ISP? If so, you could run your own VPN on your server so your traffic to your server would be protected.

        That wouldn’t protect outgoing traffic, which may be what you’re concerned about (i.e. if you’re using your server as a SOCKS proxy or something).

  • foreverunsure@pawb.social
    link
    fedilink
    arrow-up
    10
    ·
    4 months ago

    you can run an ArchiveTeam Warrior on your server and choose the URLs project. if i understand correctly, the Warrior will continuously visit randomly discovered websites to download their contents and upload them to a server that later feeds the data into the Internet Archive. best of both worlds - your ISP has a harder time distinguishing your real traffic from the ArchiveTeam-generated one, and your server is actively contributing to IA.

    • Dust0741@lemmy.worldOP
      link
      fedilink
      arrow-up
      3
      ·
      4 months ago

      Very similar yes. Trackmenot but for any site not just search engines. Although it may be a good option too

      • WhatAmLemmy@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        4 months ago

        If you’re on a VPN then the thousand other people using that server provide that type of obfuscation.

        What you seem to want is a web crawler that is perceived to be a real human and navigates alternate sites in real time. That’s a near-impossible cat and mouse game. Ultimately not worth the effort. All you have to do is be harder to track than the majority. The value for ad tech is in efficiently profiling the 99% — not the 1% of paranoid folk.

  • Archon of the Valley@infosec.pub
    link
    fedilink
    English
    arrow-up
    8
    ·
    4 months ago

    First off, if you’re concerned about ISPs selling your data (couldn’t exactly tell if that’s a part of your concern), switching to private DNS provider and enabling DNS over HTTPS/TLS can significantly cut down on that, since most of what ISPs sell comes from DNS requests. That being said, they can still tell what sites you visit if you don’t use a VPN/Tor, but they’re less likely to care unless you’re doing something illegal.

    In terms of your obfuscation plan, I’m not sure that’d do much; if anything, it’d make you stand out more. A bunch of random traffic, even tweaked to fit your browsing habits, probably would look suspicious on their end and it wouldn’t actually hide or disguise anything.

    So ideally, you’re just going to want to figure something out to set up some sort of VPN at some point. Switching DNS providers might be a bit of help in reducing sale of your traffic data, however. My recommendation is Quad9 but any privacy-friendly provider is fine.

  • GluWu@lemm.ee
    link
    fedilink
    arrow-up
    8
    ·
    4 months ago

    It’d be pretty quick to write a script that loads a randomly selected url from a prepopulated list at random intervals. Could probably do it in grease monkey directly in Firefox so you could use other tools in addition like adnausium and a client spoofer.

        • LWD@lemm.ee
          link
          fedilink
          arrow-up
          2
          ·
          4 months ago

          chrome:site-engagement for a slightly more accessible list

      • GluWu@lemm.ee
        link
        fedilink
        arrow-up
        2
        ·
        edit-2
        4 months ago

        Just start listing the most popular and generic sites. Then Google a topic like technology and copy whatever those sites are. I imagine you could have a pretty decent list populated in 15 minutes. You could also just ask chatgpt to create lists of the top 100 sites for “x”.

        What would write in? I might be willing to help because this interests me as well.

        • Dust0741@lemmy.worldOP
          link
          fedilink
          arrow-up
          3
          ·
          4 months ago

          That’s a good idea.

          Probably just a shell script. Someone mentioned using curl so that’d be pretty easy

          • GluWu@lemm.ee
            link
            fedilink
            arrow-up
            2
            ·
            4 months ago

            Let me know if you start working on anything. I want to try to use greasemonkey, I haven’t in years.

            • Dust0741@lemmy.worldOP
              link
              fedilink
              arrow-up
              1
              ·
              edit-2
              3 months ago

              Little curl shell script that works:

              #!/bin/bash
              
              # Random_Curl_Request.sh
              
              # CSV file containing websites
              CSV_FILE="/home/user/Documents/randomSiteVisitor/websites.csv"
              
              # Get a random line from the CSV file
              RANDOM_LINE=$(shuf -n 1 "$CSV_FILE")
              
              # Extract the website URL from the random line
              WEBSITE=$(echo $RANDOM_LINE | cut -d ',' -f 1)
              
              # Make a curl request to the random website every minute
              while true; do
                  curl $WEBSITE
                  sleep 60
              
                  # Get a new random line from the CSV file
                  RANDOM_LINE=$(shuf -n 1 "$CSV_FILE")
              
                  # Extract the website URL from the new random line
                  NEW_WEBSITE=$(echo $RANDOM_LINE | cut -d ',' -f 1)
              
                  # Update the website URL for the next iteration
                  WEBSITE=$NEW_WEBSITE
              done
              
              
  • gibson@sopuli.xyz
    link
    fedilink
    arrow-up
    6
    ·
    4 months ago

    It usually isn’t super hard to tell apart randomized junk like this from real human patterns. That is why Tor Browser for example tries its best to make everyone look the same instead of randomizing everything.

    That said, for the mere purpose of throwing off the ISPs profiling algorithms, you could make a relatively simple python program to solve this. A naive solution would just do an http GET to each site, but a better solution would mimic human web browsing:

    If you have no programming capability this will be rough. If you have at least a little you can follow tutorials and use an LLM to help you.

    The main issue with this goal is that it isn’t possible to tell how advanced your ISP’s profiling is, so you have no way to know if your solution is effective.

    Feel free to DM me if you go this route.

  • catloaf@lemm.ee
    link
    fedilink
    English
    arrow-up
    6
    ·
    4 months ago

    Just curl a bunch of sites at random times? Under https, everything in the URL except the domain is encrypted, so it’ll look roughly like a regular user requesting a page.

    • Dust0741@lemmy.worldOP
      link
      fedilink
      arrow-up
      2
      ·
      edit-2
      4 months ago

      I like this. Is there some sort of list of safe sites that exists that I could use in a script?

      Edit: something like this

  • JoeKrogan@lemmy.world
    link
    fedilink
    arrow-up
    3
    ·
    4 months ago

    You could set up a tor relay and use it too. They they will just see connections to tor.

    Or vpn and seed torrents

  • delirious_owl@discuss.online
    link
    fedilink
    arrow-up
    2
    ·
    edit-2
    4 months ago

    Create a guest WiFi network called “free community internet” with no password that is rate-limited to ~5% of your total bandwidth

    Run a tor exit node. But you should notify the police first.