@strategist922 Profile picture

James Chang

@strategist922

I use data, analysis and visualization to advance understanding, make inferences, and make things better.

Joined April 2010
Similar User
@mariuskarma photo

@mariuskarma

Lone Star Analysis photo

@LoneStarAnalysi

Talk Data to Me photo

@talkdatatomee

nikosv photo

@nikosvg

jamshaidsohail photo

@jamshaidsohail5

K-State Data Science photo

@kstate_bigdata

John Letteboer photo

@jletteboer

Chatmine photo

@chatmine

Cashbook Software photo

@CashbookLtd

John Hassman photo

@JohnHassman

Data2Ag photo

@Data2Ag

Hugo Alatrista-Salas photo

@HugoAlatsal

James Chang Reposted

Wow nvidia just published a 72B model with is ~on par with llama 3.1 405B in math and coding evals and also has vision 🤯

Tweet Image 1

James Chang Reposted

50 AI Tools to Turn Hours of Work into Minutes: 1. Creative Brainstorming - Claude AI - ChatGPT 4 - Bing Chat - Perplexity - Better research 2. Image Creation & Editing - Midjourney - Bing Create - Leap AI - Astira AI - Stable Diffusion 3. Note-Taking & Summarizing -…

Tweet Image 1

James Chang Reposted

(1/7) Physics of LM, Part 2.2 with 8 results on "LLM how to learn from mistakes" now on arxiv: arxiv.org/abs/2408.16293. We explore the possibility to enable models to correct errors immediately after they are made (no multi-round prompting). Check out the slides in this thread.

Tweet Image 1

James Chang Reposted

(1/n) Training LLMs can be hindered by out-of-memory, scaling batch size, and seq length. Add one line to boost multi-GPU training throughput by 20% and reduce memory usage by 60%. Introducing Liger-Kernel: Efficient Triton Kernels for LLM Training. github.com/linkedin/Liger…

Tweet Image 1

James Chang Reposted

All right, fully open-source code, Apache license for anyone and any company to use freely: github.com/leptonai/searc… Our goal: enable creators and enterprises to build AI applications as easy as possible, like this search application. Happy Friday and have fun!

Building an AI app has never been easier. Over the weekend, we built a demo for conversational search with <500 lines of python, and it's live at search.lepton.run. Give it a shot! Code to be open sourced soon as we clean up all "# temp scaffolds" stuff. (1/x)



James Chang Reposted

What if I told you that you can simultaneously enhance an LLM's task performance and reduce its size with no additional training? We find selective low-rank reduction of matrices in a transformer can improve its performance on language understanding tasks, at times by 30% pts!🧵

Tweet Image 1

James Chang Reposted

Top 9 Spaces from all time 1. @StabilityAI Stable Diffusion 2.1 2. @huggingface Open LLM Leaderboard 3. @craiyonAI DALL-E Mini 4. @AIatMeta MusicGen 5. @flngr Comic Factory 6. @angrypenguinPNG Illusion Diffusion 7. @pharmapsychotic CLIP Interrogator 8. @Microsoft HuggingGPT 9.…

Tweet Image 1

James Chang Reposted

If you think this is a normal driving video, you'd be wrong. This entire video is generated by @wayve_ai's generative AI model, GAIA-1. The model was built to generate realistic driving scenes into the future to improve safety of self-driving cars in the real world.


James Chang Reposted

Our latest paper (arxiv.org/abs/2309.02390) provides a general theory explaining when and why grokking (aka delayed generalisation) occurs – a theory so precise that we can predict hyperparameters that lead to partial grokking, and design interventions that reverse grokking! 🧵👇

Tweet Image 1
Tweet Image 2

James Chang Reposted

📣 Today we launched an overhauled NLP course to 600 students in the online MS programs at UT Austin. 98 YouTube videos 🎥 + readings 📖 open to all! cs.utexas.edu/~gdurrett/cour… w/5 hours of new 🎥 on LLMs, RLHF, chain-of-thought, etc! Meme trailer 🎬 youtu.be/DcB6ZPReeuU 🧵


James Chang Reposted

Today I finally read about speculative decoding, and it's a brilliant idea to speed up inference of LLMs! I recommend the amazing blog by @joao_gante if you want to understand how it works (and is already implemented in 🤗): huggingface.co/blog/assisted-…

Full F16 precision 34B Code Llama at >20 t/s on M2 Ultra



James Chang Reposted

Today my Transformers-Tutorials repo hit 2,000 stars on @github! 🤩 Very greatful :) the repo contains many tutorial notebooks on inference + fine-tuning with custom data for Transformers on all kinds of data; text, images, scanned PDFs, videos ⭐ github.com/NielsRogge/Tra…


Cohere Blog: Generative AI with Cohere: Part 5 - Chaining Prompts txt.cohere.com/generative-ai-…


James Chang Reposted

Did you know? 🕷️ The venom of the wandering spider, a.k.a. the banana spider (Phoneutria nigriventer), contains a surprising secret. It's been found to induce erections in males that last longer than 2 hours 😳, earning its reputation as a natural Viagra.

Tweet Image 1

James Chang Reposted

Thrilled to see Vicuna-33B top on the AlpacaEval leaderboard! Nonetheless, it's crucial to recognize that open models are still lagging behind in some areas, such as math, coding, and extraction as per our latest MT-bench study [2, 3]. Plus, GPT-4 may occasionally misjudge,…

Tweet Image 1
Tweet Image 2

James Chang Reposted

Full details in the paper: arxiv.org/abs/2306.11644 Awesome collaboration with our (also awesome) @MSFTResearch team! Cc a few authors with an active twitter account: @EldanRonen (we follow-up on his TinyStories w. Yuanzhi Li!) @JyotiAneja @sytelus @AdilSlm @YiZhangZZZ @xinw_ai


Loading...

Something went wrong.


Something went wrong.