tazz tone

tazztone
ยท

AI & ML interests

None yet

Recent Activity

Organizations

None yet

tazztone's activity

updated a Space about 2 months ago
upvoted an article about 2 months ago
Reacted to m-ric's post with ๐Ÿ‘€ about 2 months ago
view post
Post
3214
๐ŸŒŽ ๐“๐ก๐ž ๐Ÿ๐ข๐ซ๐ฌ๐ญ ๐ž๐ฏ๐ž๐ซ ๐…๐จ๐ฎ๐ง๐๐š๐ญ๐ข๐จ๐ง ๐ฐ๐ž๐š๐ญ๐ก๐ž๐ซ ๐ฆ๐จ๐๐ž๐ฅ: ๐๐ซ๐ข๐ญ๐ก๐ฏ๐ข ๐–๐ฑ๐‚ ๐ž๐ง๐š๐›๐ฅ๐ž๐ฌ ๐ฅ๐ข๐Ÿ๐ž-๐ฌ๐š๐ฏ๐ข๐ง๐  ๐ฐ๐ž๐š๐ญ๐ก๐ž๐ซ ๐ฉ๐ซ๐ž๐๐ข๐œ๐ญ๐ข๐จ๐ง๐ฌ

Hurricane Katrina killed hundreds of people as it made landfall on New Orleans in 2005 - many of these deaths could have been avoided if alerts had been given one day earlier. Accurate weather forecasts are really life-saving.

๐Ÿ”ฅย Now, NASA and IBM just dropped a game-changing new model: the first ever foundation model for weather! This means, it's the first time we have a generalist model not restricted to one task, but able to predict 160 weather variables!

Prithvi WxC (Prithvi, โ€œเคชเฅƒเคฅเฅเคตเฅ€โ€, is the Sanskrit name for Earth) - is a 2.3 billion parameter model, with an architecture close to previous vision transformers like Hiera.

๐Ÿ’กย But it comes with some important tweaks: under the hood, Prithvi WxC uses a clever transformer-based architecture with 25 encoder and 5 decoder blocks. It alternates between "local" and "global" attention to capture both regional and global weather patterns.

๐—ž๐—ฒ๐˜† ๐—ถ๐—ป๐˜€๐—ถ๐—ด๐—ต๐˜๐˜€:
๐Ÿ”ฎ Nails short-term forecasts - Prithvi WxC crushed it on 6-12 hour predictions, even outperforming some traditional numerical weather models
๐ŸŒ€ Tracks hurricanes like a champ - For Hurricane Ida, it predicted the landfall location within 5 km (vs 20+ km errors from other AI models), which is a huge progress!
๐Ÿ” 6x downscaling power - Can zoom in on weather data to 6x higher resolution with 4x lower error than basic methods
๐ŸŒŠ Models elusive gravity waves - Accurately simulates these crucial but hard-to-capture atmospheric oscillations

As climate change intensifies, tools like Prithvi WxC will become more and more crucial to avoid disasters!

Announcement post ๐Ÿ‘‰ https://newsroom.ibm.com/2024-09-23-ibm-and-nasa-release-open-source-ai-model-on-hugging-face-for-weather-and-climate-applications

Model on the Hub ๐Ÿ‘‰ https://huggingface.co/Prithvi-WxC

Thank you @clem for highlighting it!
New activity in ashen0209/Flux-Dev2Pro 2 months ago
replied to MonsterMMORPG's post 2 months ago
view reply

do the finetunes still make other (background) people look like the trained character, as it does with LoRAs?

Reacted to MonsterMMORPG's post with ๐Ÿ‘ 2 months ago
view post
Post
3304
Full Fine Tuning of FLUX yields way better results than LoRA training as expected, overfitting and bleeding reduced a lot

Configs and Full Experiments
Full configs and grid files shared here : https://www.patreon.com/posts/kohya-flux-fine-112099700

Details
I am still rigorously testing different hyperparameters and comparing impact of each one to find the best workflow
So far done 16 different full trainings and completing 8 more at the moment
I am using my poor overfit 15 images dataset for experimentation (4th image)
I have already proven that when I use a better dataset it becomes many times betters and generate expressions perfectly
Here example case : https://www.reddit.com/r/FluxAI/comments/1ffz9uc/tried_expressions_with_flux_lora_training_with_my/
Conclusions
When the results are analyzed, Fine Tuning is way lesser overfit and more generalized and better quality
In first 2 images, it is able to change hair color and add beard much better, means lesser overfit
In the third image, you will notice that the armor is much better, thus lesser overfit
I noticed that the environment and clothings are much lesser overfit and better quality
Disadvantages
Kohya still doesnโ€™t have FP8 training, thus 24 GB GPUs gets a huge speed drop
Moreover, 48 GB GPUs has to use Fused Back Pass optimization, thus have some speed drop
16 GB GPUs gets way more aggressive speed drop due to lack of FP8
Clip-L and T5 trainings still not supported
Speeds
Rank 1 Fast Config โ€” uses 27.5 GB VRAM, 6.28 second / it (LoRA is 4.85 second / it)
Rank 1 Slower Config โ€” uses 23.1 GB VRAM, 14.12 second / it (LoRA is 4.85 second / it)
Rank 1 Slowest Config โ€” uses 15.5 GB VRAM, 39 second / it (LoRA is 6.05 second / it)
Final Info
Saved checkpoints are FP16 and thus 23.8 GB (no Clip-L or T5 trained)
According to the Kohya, applied optimizations doesnโ€™t change quality so all configs are ranked as Rank 1 at the moment
I am still testing whether these optimizations make any impact on quality or not
  • 2 replies
ยท