So does the fact that they aren’t converging near the same point indicate there is a problem with my architecture and model design?
Software Engineer (iOS - ForeFlight) 🖥📱, student pilot ✈️, HUGE Colorado Avalanche fan 🥅, entrepreneur (rrainn, Inc.) ⭐️ https://charlie.fish
So does the fact that they aren’t converging near the same point indicate there is a problem with my architecture and model design?
Got it. I’ll try with some more values and see what that leads to.
So does that mean my learning rate might be too high and it’s overshooting the optimal solution sometimes based on those random weights?
I think what you’re referring to with iterating through algorithms and such is called hyper parameter tuning. I think there is a tool called Keras Tuner you can use for this.
However. I’m incredibly skeptical that will work in this situation because of how variable the results are between runs. I run it with the same input, same code, everything, and get wildly different results. So I think in order for that to be effective it needs to be fairly consistent between runs.
I could be totally off base here tho. (I haven’t worked with this stuff a ton yet).
Thanks so much for the reply!
The convolution size seems a little small
I changed this to 5 instead of 3, and hard to tell if that made much of an improvement. It still is pretty inconsistent between training runs.
If it doesn’t I’d look into reducing the number of filters or the dense layer. Reducing the available space can force an overfitting network to figure out more general solutions
I’ll try reducing the dense layer from 128 to 64 next.
Lastly, I bet someone else has either solved the same problem as an exercise or something similar and you could check out their network architecture to see if your solution is in the ballpark of something that works
This is a great idea. I did a quick Google search and nothing stood out to start. But I’ll dig deeper more.
It’s still super weird to me that with zero changes how variable it can be. I don’t change anything, and one run it is consistently improving for a few epochs, the next run it’s a lot less accurate to start and declines after the first epoch.
That’s attached to the instance? Do you have a screenshot maybe?
What is the error that you get?
Yes. It just will fill your feed with a bunch of things you might not care about. But admin vs non admin doesn’t matter in the context of what I said.
Your instance is the one that federates. However it starts with a user subscribing to that content. Your instance won’t federate normally without user interaction.
Normally the solution for the second part is relays. But that isn’t something Lemmy supports currently. This issue is very common with smaller instances. It isn’t as big of a deal with bigger instances since users are more likely to have subscribed to more communities that will automatically be federated to your instance. You could experiment with creating a user and subscribing to a bunch of communities so they get federated to your instance.
It’s not really any different than hosting any other service.
I was lucky to get in in the early days when posting Mastodon handles on Twitter was common so was able to easily migrate. But this is a problem with ActivityPub right now I feel like. Discovery algorithms can be awful in the timeline, but so useful for finding people/communities to follow.
Yep just saw that too after I researched it a bit more. What is strange is I don’t remember Eve Energy having a firmware update since then. Makes me wonder if they had it ready to go in previous firmware versions based on internal specs they saw? Or maybe I just forgot about a firmware update I did.
but as the Matter standard doesn’t yet support energy monitoring, users are limited to basic features like on and off and scheduling
- from this link
Granted the article is almost a year old. But I just didn’t realize that Matter now supports energy monitoring. Somehow I just missed that news.
I’m not aware of any official Ubiquiti certifications. Maybe it was a 3rd party certification? Someone else might know more than I do tho.
I know I’m not necessarily the target audience for this. But it feels too expensive. 6x the price of Cloudflare R2, almost 13x the price of Wasabi. Even iCloud storage is $0.99 for 50 GB with a 5 GB free tier. But again, I know I’m not necessarily the target audience as I have a lot of technical skills that maybe average users don’t have.
If you ever get around to building an API, and are interested in partnerships, let me know. Maybe there is a possibility for integration into [email protected] 😉.
This worked!!! However it now looks like I have to pass in 32 (batch size) comments in order to run a prediction in Core ML now? Kinda strange when I could pass in a single string to TensorFlow to run a prediction on.
Also it seems to be much slower than my Create ML model I was playing with. Went from 0.05 ms on average for the Create ML model to 0.47 ms on average for this TensorFlow model. Looks like this TensorFlow model also is running 100% on the CPU (not taking advantage of GPU or Neural Engine).
Obviously there are some major advantages to using TensorFlow (ie. I can run on a server environment, I can better control stopping training early based on that val_accuracy
metric, etc). But Create ML seems to really win in other areas like being able to pass in a simple string (and not having to worry about tokenization), not having to pass in 32 strings in a single prediction, and the performance.
Maybe I should lower my batch_size? I’ve heard there are pros and cons to lowering & increasing batch_size. Haven’t played around with it too much yet.
Am I just missing something in this analysis?
I really appreciate your help and advice!
Interesting! I’ll try this tonight and see how it goes. Really appreciate your reply tho. I’ll let you know the outcome.
Got it. Thanks for the reply! So is Keras just a dependency used in TensorFlow?
From what I’ve seen TensorFlow is still more popular. But that might be starting to change. Maybe we need to make a PyTorch community as well 🤔
I wish it worked on more webpages. But totally agree.
Ok I changed the Conv2D layer to be 10x10. I also changed the dense units to 64. Here is just a single run of that with a Confusion Matrix.
I don’t really see a bias towards non-blurred images.