@Lisa03312 Profile picture

Lisa033

@Lisa03312

Joined December 2017
Similar User
Rothschild.eth photo

@punk1685

Jaxay Shah photo

@jaxayshah

EL COBA 🦎 photo

@Miisssliz

Prophet.eth photo

@Prophet87_eth

هكر _هكر أخلاقي photo

@1903SavasBJK

MAEDA, Kazuho 前田 一歩 photo

@maesocio

Drah Cula photo

@drah_cula

tricia photo

@tdl8_

Lima Jmr photo

@LimaJmr8

ItIsAllFake photo

@FesslerThedfess

Lisa033 Reposted

A comparison of the exact same model training for all 3 approaches 1) Pure PyTorch: github.com/rasbt/pytorch-… (25 min) 2) PyTorch + Fabric: github.com/rasbt/pytorch-… (1.7 min) 3) PyTorch + Trainer: github.com/rasbt/faster-p… (2.7 min, the 1 extra min is for logging+checkpointing)


Lisa033 Reposted

Easy to miss in the @pytorch 2.0 release notes, they've added a small, but useful feature: torch.device, which previously just returned a device object, can now be used as a context manager. 0/8


Lisa033 Reposted

What an awesome week for open source and the PyTorch ecosystem with three big launches! - PyTorch 2.0 - Lightning Trainer 2.0 for PyTorch - Fabric for PyTorch! Just updated my "faster PyTorch" article to include the latest tools! 🔗 sebastianraschka.com/blog/2023/pyto…

rasbt's tweet image. What an awesome  week for open source and the PyTorch ecosystem with three big launches!

- PyTorch 2.0
- Lightning Trainer 2.0 for PyTorch
- Fabric for PyTorch!

Just updated my "faster PyTorch" article to include the latest tools!

🔗 <a style="text-decoration: none;" rel="nofollow" target="_blank" href="https://t.co/36ptnaFOzz">sebastianraschka.com/blog/2023/pyto…</a>

Lisa033 Reposted

Btw if you have an hour to spare on the weekend, I am covering how to use the Trainer to fit PyTorch models in the newly released Unit 5: lightning.ai/pages/courses/…

rasbt's tweet image. Btw if you have an hour to spare on the weekend, I am covering how to use the Trainer to fit PyTorch models in the newly released Unit 5: 
<a style="text-decoration: none;" rel="nofollow" target="_blank" href="https://t.co/MJB1pr3VeQ">lightning.ai/pages/courses/…</a>

What an awesome week for open source and the PyTorch ecosystem with three big launches! - PyTorch 2.0 - Lightning Trainer 2.0 for PyTorch - Fabric for PyTorch! Just updated my "faster PyTorch" article to include the latest tools! 🔗 sebastianraschka.com/blog/2023/pyto…

rasbt's tweet image. What an awesome  week for open source and the PyTorch ecosystem with three big launches!

- PyTorch 2.0
- Lightning Trainer 2.0 for PyTorch
- Fabric for PyTorch!

Just updated my "faster PyTorch" article to include the latest tools!

🔗 <a style="text-decoration: none;" rel="nofollow" target="_blank" href="https://t.co/36ptnaFOzz">sebastianraschka.com/blog/2023/pyto…</a>


Lisa033 Reposted

But note that vision transformers (ViTs) are not free from any inductive biases! ViTs focus more on global relationships due to the patchification & self-attention mechanism, which often leads to the perception that they act as low-pass filters, emphasizing (or recognizing)…

rasbt's tweet image. But note that vision transformers (ViTs) are not free from any inductive biases!

ViTs focus more on global relationships due to the patchification &amp; self-attention mechanism, which often leads to the perception that they act as low-pass filters, emphasizing (or recognizing)…

Similar to fully-connected networks, the ViT architecture (and transformer architecture in general) lacks the inductive bias for spatial invariance/equivariance that convolutional networks have. Consequently, ViTs require more data for pretraining to acquire useful "priors" from…

rasbt's tweet image. Similar to fully-connected networks, the ViT architecture (and transformer architecture in general) lacks the inductive bias for spatial invariance/equivariance that convolutional networks have. Consequently, ViTs require more data for pretraining to acquire useful "priors" from…


Lisa033 Reposted

Plot twist: they didn’t disclose any details because there was actually no innovation to report, just a bit more finetuning, which would have looked underwhelming given all the hype. By not sharing any details they made GPT-4 seem like a bigger innovation/deal than it really is.

GPT-4 was interesting for a hot second, I'll admit. But today is a new day: time to move on and get back to discussing research and open source.



Lisa033 Reposted

Exactly! This is a great example of how CVPR publicity restrictions very effectively prevent unfair public visibility on social media for research from prestigious institutes like Ivy League. oh wait…

I don’t mean to be disrespectful to either @fchollet or the authors of ViperGPT (its great work!), but i have to point out that our work VisProg did something similar 4 months ago but it just went unnoticed because of CVPR’s publicity restrictions. @CVPR @CVPRConf



Lisa033 Reposted

For a hot second, I was wondering how relevant weight decay still is. Instead of asking ChatGPT, I ran a simple experiment (*when a picture says more than a thousand words*)

rasbt's tweet image. For a hot second, I was wondering how relevant weight decay still is. 
Instead of asking ChatGPT, I ran a simple experiment (*when a picture says more than a thousand words*)

Lisa033 Reposted

Some unsolicited writing advice when you are trying to fit things into an 8-page paper limit / 1-page rebuttal limit: ... without any use of bias units ... ... without using bias units ... Does LayerNorm have an effect on ...? Does LayerNorm affect ...? The theorem does not…


Lisa033 Reposted

Classical Theory: garbage in garbage out Minor domain shift: gold in garbage out Diffusion models: garbage in gold out


Lisa033 Reposted

New AI research & news everywhere! A short post on my personal approach to keeping up with things. sebastianraschka.com/blog/2023/keep…


Lisa033 Reposted

🤔 Want to improve your PyTorch model's training performance without sacrificing accuracy? Learn how you can cut training time on a single GPU from 22.53 mins to 2.75 mins and maintain prediction accuracy🤯🚀 Check out this blog by @rasbt: lightning.ai/pages/communit… #PyTorch


Lisa033 Reposted

This blog on attention and the intuition by @rasbt is so well written! Bonus: It also has the code to run with. sebastianraschka.com/blog/2023/self…


Lisa033 Reposted

Reviewers: If you ask the authors to do something and they have followed through successfully, or you made a claim authors successfully refuted, then you need to be prepared to change your recommendation to positive. #ICML2023 1/3


Lisa033 Reposted

We're excited to release Lit-LLaMA🦙, a minimal, optimized rewrite of LLaMA for training and inference licensed under Apache 2.0 🎉 Check out the repo👉👉 github.com/Lightning-AI/l…

LightningAI's tweet image. We're excited to release Lit-LLaMA🦙, a minimal, optimized rewrite of LLaMA for training and inference licensed under Apache 2.0 🎉

Check out the repo👉👉 <a style="text-decoration: none;" rel="nofollow" target="_blank" href="https://t.co/Q3y56BXaAr">github.com/Lightning-AI/l…</a>

Lisa033 Reposted

I am old enough to remember people cheering AI when it was defeating the human Go champion


Lisa033 Reposted

Want to get into AI? My book is a 775-page journey from the fundamentals of machine learning to finetuning large language models. Today is the last day to catch "Machine Learning with PyTorch & Scikit-Learn book" during the Spring Sale – 25% off! 🌱📚 amazon.com/Machine-Learni…


Lisa033 Reposted

I didn't sign "the letter". Current AI poses lots of risks, but describing these systems as "ever more powerful digital minds" that no one can control is likely to make the problem even worse. What's needed: more transparency and better public discourse.


Lisa033 Reposted

Speaking of better public discourse, @MelMitchell1 has an excellent newsletter/blog on the state & problems of large language models. (It's free from attention-seeking headlines, and thus probably not as popular as it should be.) Highly recommended: aiguide.substack.com

I didn't sign "the letter". Current AI poses lots of risks, but describing these systems as "ever more powerful digital minds" that no one can control is likely to make the problem even worse. What's needed: more transparency and better public discourse.



Lisa033 Reposted

Open Source Sunday! Just released a new version of MLxtend: rasbt.github.io/mlxtend/ Featuring - a snappier ExhaustiveFeatureSelector - the H-Mine frequent pattern mining algo - multiprocessing for plot_decision_regions Thx to contributors, Fatih Sen, Nima Sarajpoor & others

rasbt's tweet image. Open Source Sunday!
Just released a new version of MLxtend: <a style="text-decoration: none;" rel="nofollow" target="_blank" href="https://t.co/7wS1gbADJ9">rasbt.github.io/mlxtend/</a>

Featuring
- a snappier ExhaustiveFeatureSelector
- the H-Mine frequent pattern mining algo
- multiprocessing for plot_decision_regions

Thx to contributors, Fatih Sen, Nima Sarajpoor &amp; others

Loading...

Something went wrong.


Something went wrong.