@SalehSoltan Profile picture

Saleh Soltan

@SalehSoltan

Principal Applied Scientist @Amazon AGI |Ph.D. @Columbia 2017 Views of my own.

Joined October 2017
Similar User
ACM SIGCOMM photo

@ACMSIGCOMM

Zeyuan Allen-Zhu | I'm NOT @ NeurIPS 2024 photo

@ZeyuanAllenZhu

Ang Li photo

@angcharlesli

Roi Reichart photo

@roireichart

AdapterHub photo

@AdapterHub

ARM Lab at University of Michigan photo

@umicharmlab

Gil Zussman photo

@gil_zussman

Center for Smart Streetscapes photo

@CS3_NSF

Genta Winata photo

@gentaiscool

Ish Jain photo

@ish_iitk

Tingjun Chen photo

@TingjunChen

Boris Hanin photo

@BorisHanin

Shreyansh Singh photo

@shreyansh_26

Mahshid Ghasemi photo

@Mahshid95521230

Jonas Gehring photo

@jnsgehring

Saleh Soltan Reposted

Amazon has launched Nova, a highly competitive family of foundation models. Nova Pro, Lite and Flash set new standards for the intelligence that can be accessed at the price and speed these models are offered at. Nova Pro, the flagship model, ranks amongst the leading frontier…

ArtificialAnlys's tweet image. Amazon has launched Nova, a highly competitive family of foundation models. Nova Pro, Lite and Flash set new standards for the intelligence that can be accessed at the price and speed these models are offered at.

Nova Pro, the flagship model, ranks amongst the leading frontier…

It takes a great team to build strong models, but it takes an exceptional team to build state-of-the-art models. I'm incredibly proud of what we've achieved as a team in bringing Nova models to our customers. To many more models to come 🍾 #nova #Amazon aws.amazon.com/ai/generative-…


Saleh Soltan Reposted

The ddl is Oct. 1st! Here is the website sites.google.com/view/neurips20…

Fanghui_SgrA's tweet image. The ddl is Oct. 1st! Here is the website <a style="text-decoration: none;" rel="nofollow" target="_blank" href="https://t.co/uZS0sR2WQ8">sites.google.com/view/neurips20…</a>

Our fine-tuning workshop@NeurIPS’24 @neurips24fitml has the following amazing speakers and panelists! Welcome to submit your work and participate in! This workshop aims to explore the frontiers of knowledge of how to efficiently handle nonlinearity!

Fanghui_SgrA's tweet image. Our fine-tuning workshop@NeurIPS’24 @neurips24fitml has the following amazing speakers and panelists! Welcome to submit your work and participate in! This workshop aims to explore the frontiers of knowledge of how to efficiently handle nonlinearity!


Saleh Soltan Reposted

Our fine-tuning workshop@NeurIPS’24 @neurips24fitml has the following amazing speakers and panelists! Welcome to submit your work and participate in! This workshop aims to explore the frontiers of knowledge of how to efficiently handle nonlinearity!

Fanghui_SgrA's tweet image. Our fine-tuning workshop@NeurIPS’24 @neurips24fitml has the following amazing speakers and panelists! Welcome to submit your work and participate in! This workshop aims to explore the frontiers of knowledge of how to efficiently handle nonlinearity!

Saleh Soltan Reposted

Happy to announce the #NeurIPS2024 Workshop on Fine-Tuning in Modern Machine Learning: Principles and Scalability! @neurips24fitml For details: sites.google.com/view/neurips20… Organize with @Grigoris_c @BeidiChen @BurkholzRebekka @SalehSoltan @AngelikiGiannou Masashi, @CevherLIONS


Can't agree more

Francois Chollet: OpenAI has set back the progress towards AGI by 5-10 years because frontier research is no longer being published and LLMs are an offramp on the path to AGI



Saleh Soltan Reposted

We should call models like Llama 3, Mixtral, etc. “open-weight models”, not “open-source models”. For a model to be open-source, the code and training data need to be public (good examples: GPT-J, OLMo, RedPajama, StarCoder, K2, etc.). Weights are like an exe file, which would be…


Amazing collaboration between @BMW and @alexa99 AI (led by my colleague @samridhishree) to use the LLMs trained by our team to help customers navigate the features in their car easier! youtube.com/watch?v=ZXCjw_…


Now I can freely say what we have been tirelessly working on in the past few months and more. Super excited about the launch!

Amazon’s assistant now has its own Alexa LLM poised to improve the brains in your smart home thanks to its extensive knowledge of smart home APIs. trib.al/5Gce4o3



Saleh Soltan Reposted

In case there was any doubt, LLaMA 2 is NOT open source as confirmed by the Open Source Initiative. As much as I appreciate Meta's openness compared to others, I find this misleading marketing a bit irritating.

Congratulations but please watch your language: The license authorizes only some commercial uses. The term Open Source has a clear, well understood meaning that excludes putting any restrictions on commercial use. See `2. Additional Commercial Terms` github.com/facebookresear…



Saleh Soltan Reposted

I'm supporting the international protest to stop a proposal for an intensive octopus farm in the Canary Islands. For my video, and much more: pacma.es/actualidad/sto…


Saleh Soltan Reposted

Tragic news out of Turkey and the surrounding region. We’re activating Amazon’s logistics and delivery network to quickly get donated relief items such as blankets, tents, heaters, food, diapers, baby food & medicine to those in need. aboutamazon.com/news/community…


We have published a blogpost summarizing some of our recent efforts on using LLMs for generating synthetic annotated data: @AmazonScience amazon.science/blog/using-lar…


Now that AlexaTM 20B is publicly released, you can use it to generate multilingual data for Semantic Parsing (e.g., mTOP). Check our paper for the prompt and results (to appear in AACL 2022): arxiv.org/abs/2210.07074

SalehSoltan's tweet image. Now that AlexaTM 20B is publicly released, you can use it to generate multilingual data for Semantic Parsing (e.g., mTOP). Check our paper for the prompt and results (to appear in AACL 2022): <a style="text-decoration: none;" rel="nofollow" target="_blank" href="https://t.co/d77E7IyVzk">arxiv.org/abs/2210.07074</a>

We just released AlexaTM 20B on Amazon SageMaker Jumpstart. You can now deploy and run inference with the AlexaTM 20B model through JumpStart APIs: aws.amazon.com/blogs/machine-…


Saleh Soltan Reposted

When people say GPT-3, do they mean the original GPT-3 or InstructGPT? And which version? It makes a huge difference, so it'd be nice to explicitly specify davinci, text-davinci-002, etc. when making a claim about GPT-3.


Saleh Soltan Reposted

🚨 New preprint! 🚨 We refine least-to-most prompting and achieve sota on CFQ (95% accuracy), outperforming previous fully supervised methods. Joint first author work with the formidable Nathanael Schärli.

Compositional Semantic Parsing with Large Language Models abs: arxiv.org/abs/2209.15003

_akhaliq's tweet image. Compositional Semantic Parsing with Large Language Models
abs: <a style="text-decoration: none;" rel="nofollow" target="_blank" href="https://t.co/1k6rGYVNUW">arxiv.org/abs/2209.15003</a>


Check out our new work on generating annotated data for NLU for completely new domains and intents by fine-tuning a large-scale seq2seq model (AlexaTM 5B). Full paper (to appear in @coling2022): arxiv.org/pdf/2209.09900…

SalehSoltan's tweet image. Check out our new work on generating annotated data for NLU for completely new domains and intents by fine-tuning a large-scale seq2seq model (AlexaTM 5B). 

Full paper (to appear in @coling2022): <a style="text-decoration: none;" rel="nofollow" target="_blank" href="https://t.co/Jtuig6ot96">arxiv.org/pdf/2209.09900…</a>

Loading...

Something went wrong.


Something went wrong.