Inspired by a comment on my last post.
I feel like I never have a solution that allows me to control it while also being automated to such a degree that I don’t have a huge confusing backup if I don’t do finances for days or weeks.
Inspired by a comment on my last post.
I feel like I never have a solution that allows me to control it while also being automated to such a degree that I don’t have a huge confusing backup if I don’t do finances for days or weeks.
This won’t help you, but I want to brag. I started using Quicken to track my finances at the turn of the century, back when it was all local storage. Quicken 2012 was the last iteration that used http (not https) to update stock prices. When they discontinued support, I captured the interaction and deciphered the formats. Wrote a proxy to intercept the request, look up the security info, and send back the data.
So, I self-host quicken.com. It’s saved me having to update Quicken or submit to their subscription model.
That’s really cool! How did you do that?
He mentioned it used http, so the traffic is not encrypted. You can easily monitor http traffic with wireshark.
It is pretty easy. There’s tons of tutorials and walkthroughs for doing it, but anyone familiar with UIs will be able to work it out pretty quickly I think. Maybe a friction point in using the filter query, but again there’s tons of walkthroughs and guides for using it online.
If you can’t conceptualize a packet, or sockets, or network flows, even with the help of online guides/manuals, I guess it wouldn’t be easy. In that case I’d be wondering why someone would want to use those tools in the first place though, as then they probably wouldn’t have the skills necessary to leverage the information gleaned from the tool in any useful way.
Edit - As we’re in the self-hosted community, I’d argue that anyone who is self-hosting anything would probably be able to easily install wireshark and view http requests, both individual packets and the stream as a whole.
Thank you
Super easy, as it turns out. I run my own DNS and web servers, so I pointed quicken.com at my web server to capture the request, then used curl to capture the response. Both turned out to be plain ASCII, request like
as POST data, and responses like
plus a whole slew of other optional fields for fundamentals, dividends, etc. It was a simpler time on the internet, when no one cared about leaking data and companies didn’t care if a handful of geeks reversed engineered their data structures.