• 7 Posts
  • 254 Comments
Joined 2 years ago
cake
Cake day: June 2nd, 2023

help-circle







  • Can confirm, I’ve recently got some cameras and set up Frigate and it’s been great. Not using Reolink but the ones I have work well enough. I have a TPLink that I like, and a Hilook starlight camera that I am not convinced on as it doesn’t seem to have auto-exposure adjustment. Both work well for object detection, though there’s a bit of a learning curve with frigate needing to be configured via YAML for a lot of things.

    I’ve also started playing with Frigate’s face detection but I don’t think the cameras are really positioned for it. It probably makes more sense for a front door camera getting a good view of the person.

    I’ve also got Home Assistant picking up the frigate camera streams which works well too.


  • Dave@lemmy.nztoSelfhosted@lemmy.worldImmich mobile app sync V2
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    1 month ago

    Local storage on a VPS is expensive, and I’ve never been happy with a lower powered server serving media. Personally I self-host and send a backup to Backblaze B2 for offsite (using Rclone).

    I use Borgmatic for incremental, deduplicated backups but make sure you save your encryption key somewhere you can access it if your house burns down.








  • I think you’d be right that the direct cost of running the crawler and index would not be the issue. But fighting SEO to keep your results decent is probably a cost that dwarfs the basic technical cost of running the crawler and index.

    And you’d need a technical security team on top of things as link farms aren’t your only risk, I’m sure there are countless ways to manipulate the algorithm to put your site on top that Google probably have multiple teams working on fighting it full time.

    Many of these things would likely not be a problem for a startup, though. No one is paying SEO firms big money to get into a search index no one has heard of and hardly anyone uses, so these costs probably grow exponentially over time as you become more well known.


  • I’m not disputing that you might be right, but the internet archive runs a very different service. Mainly that Google needs to continuously prune their 400 billion page index because of link rot. The Internet Archive has the opposite aim, they are preserving sites that no longer exist.

    I’m also not sure they even crawl. Do sites get added on user request? When looking at a medium popularity page, you see it only has a couple of scrapes a year.

    None of them. At least, none that I’m aware of. I just don’t think that direct expenses are the reason that there are are only two major web search tools. I also don’t think Google and bing are good examples to point at when estimating the cost of running a complete search engine.

    I would suggest direct expenses are the barrier, but perhaps crawling is not the main expense. I would be interested to know any speculations you have outside of expenses that cause a barrier?


  • That website claims they add 3-5 billion pages a month. Google is doing that in a day or three, as recency of information is very important in search. Plus that site claims 100 billion pages to Googles 400 billion. It’s still an impressive project.

    Size isn’t everything, so the real question is: what search site uses only the common crawl index and has results on par with bing or google?